Bug Summary

File:linux-user/mmap.c
Location:line 747, column 13
Description:Value stored to 'host_addr' is never read

Annotated Source Code

1/*
2 * mmap support for qemu
3 *
4 * Copyright (c) 2003 Fabrice Bellard
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
10 *
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, see <http://www.gnu.org/licenses/>.
18 */
19#include <stdlib.h>
20#include <stdio.h>
21#include <stdarg.h>
22#include <string.h>
23#include <unistd.h>
24#include <errno(*__errno_location ()).h>
25#include <sys/types.h>
26#include <sys/stat.h>
27#include <sys/mman.h>
28#include <linux1/mman.h>
29#include <linux1/unistd.h>
30
31#include "qemu.h"
32#include "qemu-common.h"
33
34//#define DEBUG_MMAP
35
36static pthread_mutex_t mmap_mutex = PTHREAD_MUTEX_INITIALIZER{ { 0, 0, 0, 0, 0, 0, { 0, 0 } } };
37static __thread int mmap_lock_count;
38
39void mmap_lock(void)
40{
41 if (mmap_lock_count++ == 0) {
42 pthread_mutex_lock(&mmap_mutex);
43 }
44}
45
46void mmap_unlock(void)
47{
48 if (--mmap_lock_count == 0) {
49 pthread_mutex_unlock(&mmap_mutex);
50 }
51}
52
53/* Grab lock to make sure things are in a consistent state after fork(). */
54void mmap_fork_start(void)
55{
56 if (mmap_lock_count)
57 abort();
58 pthread_mutex_lock(&mmap_mutex);
59}
60
61void mmap_fork_end(int child)
62{
63 if (child)
64 pthread_mutex_init(&mmap_mutex, NULL((void*)0));
65 else
66 pthread_mutex_unlock(&mmap_mutex);
67}
68
69/* NOTE: all the constants are the HOST ones, but addresses are target. */
70int target_mprotect(abi_ulong start, abi_ulong len, int prot)
71{
72 abi_ulong end, host_start, host_end, addr;
73 int prot1, ret;
74
75#ifdef DEBUG_MMAP
76 printf("mprotect: start=0x" TARGET_ABI_FMT_lx"%016" "l" "x"
77 "len=0x" TARGET_ABI_FMT_lx"%016" "l" "x" " prot=%c%c%c\n", start, len,
78 prot & PROT_READ0x1 ? 'r' : '-',
79 prot & PROT_WRITE0x2 ? 'w' : '-',
80 prot & PROT_EXEC0x4 ? 'x' : '-');
81#endif
82
83 if ((start & ~TARGET_PAGE_MASK~((1 << 13) - 1)) != 0)
84 return -EINVAL22;
85 len = TARGET_PAGE_ALIGN(len)(((len) + (1 << 13) - 1) & ~((1 << 13) - 1));
86 end = start + len;
87 if (end < start)
88 return -EINVAL22;
89 prot &= PROT_READ0x1 | PROT_WRITE0x2 | PROT_EXEC0x4;
90 if (len == 0)
91 return 0;
92
93 mmap_lock();
94 host_start = start & qemu_host_page_mask;
95 host_end = HOST_PAGE_ALIGN(end)(((end) + qemu_host_page_size - 1) & qemu_host_page_mask);
96 if (start > host_start) {
97 /* handle host page containing start */
98 prot1 = prot;
99 for(addr = host_start; addr < start; addr += TARGET_PAGE_SIZE(1 << 13)) {
100 prot1 |= page_get_flags(addr);
101 }
102 if (host_end == host_start + qemu_host_page_size) {
103 for(addr = end; addr < host_end; addr += TARGET_PAGE_SIZE(1 << 13)) {
104 prot1 |= page_get_flags(addr);
105 }
106 end = host_end;
107 }
108 ret = mprotect(g2h(host_start)((void *)((unsigned long)(target_ulong)(host_start) + guest_base
))
, qemu_host_page_size, prot1 & PAGE_BITS(0x0001 | 0x0002 | 0x0004));
109 if (ret != 0)
110 goto error;
111 host_start += qemu_host_page_size;
112 }
113 if (end < host_end) {
114 prot1 = prot;
115 for(addr = end; addr < host_end; addr += TARGET_PAGE_SIZE(1 << 13)) {
116 prot1 |= page_get_flags(addr);
117 }
118 ret = mprotect(g2h(host_end - qemu_host_page_size)((void *)((unsigned long)(target_ulong)(host_end - qemu_host_page_size
) + guest_base))
, qemu_host_page_size,
119 prot1 & PAGE_BITS(0x0001 | 0x0002 | 0x0004));
120 if (ret != 0)
121 goto error;
122 host_end -= qemu_host_page_size;
123 }
124
125 /* handle the pages in the middle */
126 if (host_start < host_end) {
127 ret = mprotect(g2h(host_start)((void *)((unsigned long)(target_ulong)(host_start) + guest_base
))
, host_end - host_start, prot);
128 if (ret != 0)
129 goto error;
130 }
131 page_set_flags(start, start + len, prot | PAGE_VALID0x0008);
132 mmap_unlock();
133 return 0;
134error:
135 mmap_unlock();
136 return ret;
137}
138
139/* map an incomplete host page */
140static int mmap_frag(abi_ulong real_start,
141 abi_ulong start, abi_ulong end,
142 int prot, int flags, int fd, abi_ulong offset)
143{
144 abi_ulong real_end, addr;
145 void *host_start;
146 int prot1, prot_new;
147
148 real_end = real_start + qemu_host_page_size;
149 host_start = g2h(real_start)((void *)((unsigned long)(target_ulong)(real_start) + guest_base
))
;
150
151 /* get the protection of the target pages outside the mapping */
152 prot1 = 0;
153 for(addr = real_start; addr < real_end; addr++) {
154 if (addr < start || addr >= end)
155 prot1 |= page_get_flags(addr);
156 }
157
158 if (prot1 == 0) {
159 /* no page was there, so we allocate one */
160 void *p = mmap(host_start, qemu_host_page_size, prot,
161 flags | MAP_ANONYMOUS0x20, -1, 0);
162 if (p == MAP_FAILED((void *) -1))
163 return -1;
164 prot1 = prot;
165 }
166 prot1 &= PAGE_BITS(0x0001 | 0x0002 | 0x0004);
167
168 prot_new = prot | prot1;
169 if (!(flags & MAP_ANONYMOUS0x20)) {
170 /* msync() won't work here, so we return an error if write is
171 possible while it is a shared mapping */
172 if ((flags & MAP_TYPE0x0f) == MAP_SHARED0x01 &&
173 (prot & PROT_WRITE0x2))
174 return -1;
175
176 /* adjust protection to be able to read */
177 if (!(prot1 & PROT_WRITE0x2))
178 mprotect(host_start, qemu_host_page_size, prot1 | PROT_WRITE0x2);
179
180 /* read the corresponding file data */
181 if (pread(fd, g2h(start)((void *)((unsigned long)(target_ulong)(start) + guest_base)), end - start, offset) == -1)
182 return -1;
183
184 /* put final protection */
185 if (prot_new != (prot1 | PROT_WRITE0x2))
186 mprotect(host_start, qemu_host_page_size, prot_new);
187 } else {
188 /* just update the protection */
189 if (prot_new != prot1) {
190 mprotect(host_start, qemu_host_page_size, prot_new);
191 }
192 }
193 return 0;
194}
195
196#if HOST_LONG_BITS64 == 64 && TARGET_ABI_BITS64 == 64
197# define TASK_UNMAPPED_BASE(1ul << 38) (1ul << 38)
198#elif defined(__CYGWIN__)
199/* Cygwin doesn't have a whole lot of address space. */
200# define TASK_UNMAPPED_BASE(1ul << 38) 0x18000000
201#else
202# define TASK_UNMAPPED_BASE(1ul << 38) 0x40000000
203#endif
204abi_ulong mmap_next_start = TASK_UNMAPPED_BASE(1ul << 38);
205
206unsigned long last_brk;
207
208#ifdef CONFIG_USE_GUEST_BASE1
209/* Subroutine of mmap_find_vma, used when we have pre-allocated a chunk
210 of guest address space. */
211static abi_ulong mmap_find_vma_reserved(abi_ulong start, abi_ulong size)
212{
213 abi_ulong addr;
214 abi_ulong end_addr;
215 int prot;
216 int looped = 0;
217
218 if (size > RESERVED_VAreserved_va) {
219 return (abi_ulong)-1;
220 }
221
222 size = HOST_PAGE_ALIGN(size)(((size) + qemu_host_page_size - 1) & qemu_host_page_mask
)
;
223 end_addr = start + size;
224 if (end_addr > RESERVED_VAreserved_va) {
225 end_addr = RESERVED_VAreserved_va;
226 }
227 addr = end_addr - qemu_host_page_size;
228
229 while (1) {
230 if (addr > end_addr) {
231 if (looped) {
232 return (abi_ulong)-1;
233 }
234 end_addr = RESERVED_VAreserved_va;
235 addr = end_addr - qemu_host_page_size;
236 looped = 1;
237 continue;
238 }
239 prot = page_get_flags(addr);
240 if (prot) {
241 end_addr = addr;
242 }
243 if (addr + size == end_addr) {
244 break;
245 }
246 addr -= qemu_host_page_size;
247 }
248
249 if (start == mmap_next_start) {
250 mmap_next_start = addr;
251 }
252
253 return addr;
254}
255#endif
256
257/*
258 * Find and reserve a free memory area of size 'size'. The search
259 * starts at 'start'.
260 * It must be called with mmap_lock() held.
261 * Return -1 if error.
262 */
263abi_ulong mmap_find_vma(abi_ulong start, abi_ulong size)
264{
265 void *ptr, *prev;
266 abi_ulong addr;
267 int wrapped, repeat;
268
269 /* If 'start' == 0, then a default start address is used. */
270 if (start == 0) {
271 start = mmap_next_start;
272 } else {
273 start &= qemu_host_page_mask;
274 }
275
276 size = HOST_PAGE_ALIGN(size)(((size) + qemu_host_page_size - 1) & qemu_host_page_mask
)
;
277
278#ifdef CONFIG_USE_GUEST_BASE1
279 if (RESERVED_VAreserved_va) {
280 return mmap_find_vma_reserved(start, size);
281 }
282#endif
283
284 addr = start;
285 wrapped = repeat = 0;
286 prev = 0;
287
288 for (;; prev = ptr) {
289 /*
290 * Reserve needed memory area to avoid a race.
291 * It should be discarded using:
292 * - mmap() with MAP_FIXED flag
293 * - mremap() with MREMAP_FIXED flag
294 * - shmat() with SHM_REMAP flag
295 */
296 ptr = mmap(g2h(addr)((void *)((unsigned long)(target_ulong)(addr) + guest_base)), size, PROT_NONE0x0,
297 MAP_ANONYMOUS0x20|MAP_PRIVATE0x02|MAP_NORESERVE0x4000, -1, 0);
298
299 /* ENOMEM, if host address space has no memory */
300 if (ptr == MAP_FAILED((void *) -1)) {
301 return (abi_ulong)-1;
302 }
303
304 /* Count the number of sequential returns of the same address.
305 This is used to modify the search algorithm below. */
306 repeat = (ptr == prev ? repeat + 1 : 0);
307
308 if (h2g_valid(ptr + size - 1)({ unsigned long __guest = (unsigned long)(ptr + size - 1) - guest_base
; (__guest < (1ul << 44)) && (!reserved_va ||
(__guest < reserved_va)); })
) {
309 addr = h2g(ptr)({ ((({ unsigned long __guest = (unsigned long)(ptr) - guest_base
; (__guest < (1ul << 44)) && (!reserved_va ||
(__guest < reserved_va)); })) ? (void) (0) : __assert_fail
("({ unsigned long __guest = (unsigned long)(ptr) - guest_base; (__guest < (1ul << 44)) && (!reserved_va || (__guest < reserved_va)); })"
, "/home/stefan/src/qemu/qemu.org/qemu/linux-user/mmap.c", 309
, __PRETTY_FUNCTION__)); ({ unsigned long __ret = (unsigned long
)(ptr) - guest_base; (abi_ulong)__ret; }); })
;
310
311 if ((addr & ~TARGET_PAGE_MASK~((1 << 13) - 1)) == 0) {
312 /* Success. */
313 if (start == mmap_next_start && addr >= TASK_UNMAPPED_BASE(1ul << 38)) {
314 mmap_next_start = addr + size;
315 }
316 return addr;
317 }
318
319 /* The address is not properly aligned for the target. */
320 switch (repeat) {
321 case 0:
322 /* Assume the result that the kernel gave us is the
323 first with enough free space, so start again at the
324 next higher target page. */
325 addr = TARGET_PAGE_ALIGN(addr)(((addr) + (1 << 13) - 1) & ~((1 << 13) - 1));
326 break;
327 case 1:
328 /* Sometimes the kernel decides to perform the allocation
329 at the top end of memory instead. */
330 addr &= TARGET_PAGE_MASK~((1 << 13) - 1);
331 break;
332 case 2:
333 /* Start over at low memory. */
334 addr = 0;
335 break;
336 default:
337 /* Fail. This unaligned block must the last. */
338 addr = -1;
339 break;
340 }
341 } else {
342 /* Since the result the kernel gave didn't fit, start
343 again at low memory. If any repetition, fail. */
344 addr = (repeat ? -1 : 0);
345 }
346
347 /* Unmap and try again. */
348 munmap(ptr, size);
349
350 /* ENOMEM if we checked the whole of the target address space. */
351 if (addr == (abi_ulong)-1) {
352 return (abi_ulong)-1;
353 } else if (addr == 0) {
354 if (wrapped) {
355 return (abi_ulong)-1;
356 }
357 wrapped = 1;
358 /* Don't actually use 0 when wrapping, instead indicate
359 that we'd truly like an allocation in low memory. */
360 addr = (mmap_min_addr > TARGET_PAGE_SIZE(1 << 13)
361 ? TARGET_PAGE_ALIGN(mmap_min_addr)(((mmap_min_addr) + (1 << 13) - 1) & ~((1 << 13
) - 1))
362 : TARGET_PAGE_SIZE(1 << 13));
363 } else if (wrapped && addr >= start) {
364 return (abi_ulong)-1;
365 }
366 }
367}
368
369/* NOTE: all the constants are the HOST ones */
370abi_long target_mmap(abi_ulong start, abi_ulong len, int prot,
371 int flags, int fd, abi_ulong offset)
372{
373 abi_ulong ret, end, real_start, real_end, retaddr, host_offset, host_len;
374
375 mmap_lock();
376#ifdef DEBUG_MMAP
377 {
378 printf("mmap: start=0x" TARGET_ABI_FMT_lx"%016" "l" "x"
379 " len=0x" TARGET_ABI_FMT_lx"%016" "l" "x" " prot=%c%c%c flags=",
380 start, len,
381 prot & PROT_READ0x1 ? 'r' : '-',
382 prot & PROT_WRITE0x2 ? 'w' : '-',
383 prot & PROT_EXEC0x4 ? 'x' : '-');
384 if (flags & MAP_FIXED0x10)
385 printf("MAP_FIXED ");
386 if (flags & MAP_ANONYMOUS0x20)
387 printf("MAP_ANON ");
388 switch(flags & MAP_TYPE0x0f) {
389 case MAP_PRIVATE0x02:
390 printf("MAP_PRIVATE ");
391 break;
392 case MAP_SHARED0x01:
393 printf("MAP_SHARED ");
394 break;
395 default:
396 printf("[MAP_TYPE=0x%x] ", flags & MAP_TYPE0x0f);
397 break;
398 }
399 printf("fd=%d offset=" TARGET_ABI_FMT_lx"%016" "l" "x" "\n", fd, offset);
400 }
401#endif
402
403 if (offset & ~TARGET_PAGE_MASK~((1 << 13) - 1)) {
404 errno(*__errno_location ()) = EINVAL22;
405 goto fail;
406 }
407
408 len = TARGET_PAGE_ALIGN(len)(((len) + (1 << 13) - 1) & ~((1 << 13) - 1));
409 if (len == 0)
410 goto the_end;
411 real_start = start & qemu_host_page_mask;
412 host_offset = offset & qemu_host_page_mask;
413
414 /* If the user is asking for the kernel to find a location, do that
415 before we truncate the length for mapping files below. */
416 if (!(flags & MAP_FIXED0x10)) {
417 host_len = len + offset - host_offset;
418 host_len = HOST_PAGE_ALIGN(host_len)(((host_len) + qemu_host_page_size - 1) & qemu_host_page_mask
)
;
419 start = mmap_find_vma(real_start, host_len);
420 if (start == (abi_ulong)-1) {
421 errno(*__errno_location ()) = ENOMEM12;
422 goto fail;
423 }
424 }
425
426 /* When mapping files into a memory area larger than the file, accesses
427 to pages beyond the file size will cause a SIGBUS.
428
429 For example, if mmaping a file of 100 bytes on a host with 4K pages
430 emulating a target with 8K pages, the target expects to be able to
431 access the first 8K. But the host will trap us on any access beyond
432 4K.
433
434 When emulating a target with a larger page-size than the hosts, we
435 may need to truncate file maps at EOF and add extra anonymous pages
436 up to the targets page boundary. */
437
438 if ((qemu_real_host_page_size < TARGET_PAGE_SIZE(1 << 13))
439 && !(flags & MAP_ANONYMOUS0x20)) {
440 struct stat sb;
441
442 if (fstat (fd, &sb) == -1)
443 goto fail;
444
445 /* Are we trying to create a map beyond EOF?. */
446 if (offset + len > sb.st_size) {
447 /* If so, truncate the file map at eof aligned with
448 the hosts real pagesize. Additional anonymous maps
449 will be created beyond EOF. */
450 len = (sb.st_size - offset);
451 len += qemu_real_host_page_size - 1;
452 len &= ~(qemu_real_host_page_size - 1);
453 }
454 }
455
456 if (!(flags & MAP_FIXED0x10)) {
457 unsigned long host_start;
458 void *p;
459
460 host_len = len + offset - host_offset;
461 host_len = HOST_PAGE_ALIGN(host_len)(((host_len) + qemu_host_page_size - 1) & qemu_host_page_mask
)
;
462
463 /* Note: we prefer to control the mapping address. It is
464 especially important if qemu_host_page_size >
465 qemu_real_host_page_size */
466 p = mmap(g2h(start)((void *)((unsigned long)(target_ulong)(start) + guest_base)), host_len, prot,
467 flags | MAP_FIXED0x10 | MAP_ANONYMOUS0x20, -1, 0);
468 if (p == MAP_FAILED((void *) -1))
469 goto fail;
470 /* update start so that it points to the file position at 'offset' */
471 host_start = (unsigned long)p;
472 if (!(flags & MAP_ANONYMOUS0x20)) {
473 p = mmap(g2h(start)((void *)((unsigned long)(target_ulong)(start) + guest_base)), len, prot,
474 flags | MAP_FIXED0x10, fd, host_offset);
475 if (p == MAP_FAILED((void *) -1)) {
476 munmap(g2h(start)((void *)((unsigned long)(target_ulong)(start) + guest_base)), host_len);
477 goto fail;
478 }
479 host_start += offset - host_offset;
480 }
481 start = h2g(host_start)({ ((({ unsigned long __guest = (unsigned long)(host_start) -
guest_base; (__guest < (1ul << 44)) && (!reserved_va
|| (__guest < reserved_va)); })) ? (void) (0) : __assert_fail
("({ unsigned long __guest = (unsigned long)(host_start) - guest_base; (__guest < (1ul << 44)) && (!reserved_va || (__guest < reserved_va)); })"
, "/home/stefan/src/qemu/qemu.org/qemu/linux-user/mmap.c", 481
, __PRETTY_FUNCTION__)); ({ unsigned long __ret = (unsigned long
)(host_start) - guest_base; (abi_ulong)__ret; }); })
;
482 } else {
483 if (start & ~TARGET_PAGE_MASK~((1 << 13) - 1)) {
484 errno(*__errno_location ()) = EINVAL22;
485 goto fail;
486 }
487 end = start + len;
488 real_end = HOST_PAGE_ALIGN(end)(((end) + qemu_host_page_size - 1) & qemu_host_page_mask);
489
490 /*
491 * Test if requested memory area fits target address space
492 * It can fail only on 64-bit host with 32-bit target.
493 * On any other target/host host mmap() handles this error correctly.
494 */
495 if ((unsigned long)start + len - 1 > (abi_ulong) -1) {
496 errno(*__errno_location ()) = EINVAL22;
497 goto fail;
498 }
499
500 /* worst case: we cannot map the file because the offset is not
501 aligned, so we read it */
502 if (!(flags & MAP_ANONYMOUS0x20) &&
503 (offset & ~qemu_host_page_mask) != (start & ~qemu_host_page_mask)) {
504 /* msync() won't work here, so we return an error if write is
505 possible while it is a shared mapping */
506 if ((flags & MAP_TYPE0x0f) == MAP_SHARED0x01 &&
507 (prot & PROT_WRITE0x2)) {
508 errno(*__errno_location ()) = EINVAL22;
509 goto fail;
510 }
511 retaddr = target_mmap(start, len, prot | PROT_WRITE0x2,
512 MAP_FIXED0x10 | MAP_PRIVATE0x02 | MAP_ANONYMOUS0x20,
513 -1, 0);
514 if (retaddr == -1)
515 goto fail;
516 if (pread(fd, g2h(start)((void *)((unsigned long)(target_ulong)(start) + guest_base)), len, offset) == -1)
517 goto fail;
518 if (!(prot & PROT_WRITE0x2)) {
519 ret = target_mprotect(start, len, prot);
520 if (ret != 0) {
521 start = ret;
522 goto the_end;
523 }
524 }
525 goto the_end;
526 }
527
528 /* handle the start of the mapping */
529 if (start > real_start) {
530 if (real_end == real_start + qemu_host_page_size) {
531 /* one single host page */
532 ret = mmap_frag(real_start, start, end,
533 prot, flags, fd, offset);
534 if (ret == -1)
535 goto fail;
536 goto the_end1;
537 }
538 ret = mmap_frag(real_start, start, real_start + qemu_host_page_size,
539 prot, flags, fd, offset);
540 if (ret == -1)
541 goto fail;
542 real_start += qemu_host_page_size;
543 }
544 /* handle the end of the mapping */
545 if (end < real_end) {
546 ret = mmap_frag(real_end - qemu_host_page_size,
547 real_end - qemu_host_page_size, real_end,
548 prot, flags, fd,
549 offset + real_end - qemu_host_page_size - start);
550 if (ret == -1)
551 goto fail;
552 real_end -= qemu_host_page_size;
553 }
554
555 /* map the middle (easier) */
556 if (real_start < real_end) {
557 void *p;
558 unsigned long offset1;
559 if (flags & MAP_ANONYMOUS0x20)
560 offset1 = 0;
561 else
562 offset1 = offset + real_start - start;
563 p = mmap(g2h(real_start)((void *)((unsigned long)(target_ulong)(real_start) + guest_base
))
, real_end - real_start,
564 prot, flags, fd, offset1);
565 if (p == MAP_FAILED((void *) -1))
566 goto fail;
567 }
568 }
569 the_end1:
570 page_set_flags(start, start + len, prot | PAGE_VALID0x0008);
571 the_end:
572#ifdef DEBUG_MMAP
573 printf("ret=0x" TARGET_ABI_FMT_lx"%016" "l" "x" "\n", start);
574 page_dump(stdoutstdout);
575 printf("\n");
576#endif
577 tb_invalidate_phys_range(start, start + len, 0);
578 mmap_unlock();
579 return start;
580fail:
581 mmap_unlock();
582 return -1;
583}
584
585static void mmap_reserve(abi_ulong start, abi_ulong size)
586{
587 abi_ulong real_start;
588 abi_ulong real_end;
589 abi_ulong addr;
590 abi_ulong end;
591 int prot;
592
593 real_start = start & qemu_host_page_mask;
594 real_end = HOST_PAGE_ALIGN(start + size)(((start + size) + qemu_host_page_size - 1) & qemu_host_page_mask
)
;
595 end = start + size;
596 if (start > real_start) {
597 /* handle host page containing start */
598 prot = 0;
599 for (addr = real_start; addr < start; addr += TARGET_PAGE_SIZE(1 << 13)) {
600 prot |= page_get_flags(addr);
601 }
602 if (real_end == real_start + qemu_host_page_size) {
603 for (addr = end; addr < real_end; addr += TARGET_PAGE_SIZE(1 << 13)) {
604 prot |= page_get_flags(addr);
605 }
606 end = real_end;
607 }
608 if (prot != 0)
609 real_start += qemu_host_page_size;
610 }
611 if (end < real_end) {
612 prot = 0;
613 for (addr = end; addr < real_end; addr += TARGET_PAGE_SIZE(1 << 13)) {
614 prot |= page_get_flags(addr);
615 }
616 if (prot != 0)
617 real_end -= qemu_host_page_size;
618 }
619 if (real_start != real_end) {
620 mmap(g2h(real_start)((void *)((unsigned long)(target_ulong)(real_start) + guest_base
))
, real_end - real_start, PROT_NONE0x0,
621 MAP_FIXED0x10 | MAP_ANONYMOUS0x20 | MAP_PRIVATE0x02 | MAP_NORESERVE0x4000,
622 -1, 0);
623 }
624}
625
626int target_munmap(abi_ulong start, abi_ulong len)
627{
628 abi_ulong end, real_start, real_end, addr;
629 int prot, ret;
630
631#ifdef DEBUG_MMAP
632 printf("munmap: start=0x" TARGET_ABI_FMT_lx"%016" "l" "x" " len=0x"
633 TARGET_ABI_FMT_lx"%016" "l" "x" "\n",
634 start, len);
635#endif
636 if (start & ~TARGET_PAGE_MASK~((1 << 13) - 1))
637 return -EINVAL22;
638 len = TARGET_PAGE_ALIGN(len)(((len) + (1 << 13) - 1) & ~((1 << 13) - 1));
639 if (len == 0)
640 return -EINVAL22;
641 mmap_lock();
642 end = start + len;
643 real_start = start & qemu_host_page_mask;
644 real_end = HOST_PAGE_ALIGN(end)(((end) + qemu_host_page_size - 1) & qemu_host_page_mask);
645
646 if (start > real_start) {
647 /* handle host page containing start */
648 prot = 0;
649 for(addr = real_start; addr < start; addr += TARGET_PAGE_SIZE(1 << 13)) {
650 prot |= page_get_flags(addr);
651 }
652 if (real_end == real_start + qemu_host_page_size) {
653 for(addr = end; addr < real_end; addr += TARGET_PAGE_SIZE(1 << 13)) {
654 prot |= page_get_flags(addr);
655 }
656 end = real_end;
657 }
658 if (prot != 0)
659 real_start += qemu_host_page_size;
660 }
661 if (end < real_end) {
662 prot = 0;
663 for(addr = end; addr < real_end; addr += TARGET_PAGE_SIZE(1 << 13)) {
664 prot |= page_get_flags(addr);
665 }
666 if (prot != 0)
667 real_end -= qemu_host_page_size;
668 }
669
670 ret = 0;
671 /* unmap what we can */
672 if (real_start < real_end) {
673 if (RESERVED_VAreserved_va) {
674 mmap_reserve(real_start, real_end - real_start);
675 } else {
676 ret = munmap(g2h(real_start)((void *)((unsigned long)(target_ulong)(real_start) + guest_base
))
, real_end - real_start);
677 }
678 }
679
680 if (ret == 0) {
681 page_set_flags(start, start + len, 0);
682 tb_invalidate_phys_range(start, start + len, 0);
683 }
684 mmap_unlock();
685 return ret;
686}
687
688abi_long target_mremap(abi_ulong old_addr, abi_ulong old_size,
689 abi_ulong new_size, unsigned long flags,
690 abi_ulong new_addr)
691{
692 int prot;
693 void *host_addr;
694
695 mmap_lock();
696
697 if (flags & MREMAP_FIXED2) {
698 host_addr = (void *) syscall(__NR_mremap25, g2h(old_addr)((void *)((unsigned long)(target_ulong)(old_addr) + guest_base
))
,
699 old_size, new_size,
700 flags,
701 g2h(new_addr)((void *)((unsigned long)(target_ulong)(new_addr) + guest_base
))
);
702
703 if (RESERVED_VAreserved_va && host_addr != MAP_FAILED((void *) -1)) {
704 /* If new and old addresses overlap then the above mremap will
705 already have failed with EINVAL. */
706 mmap_reserve(old_addr, old_size);
707 }
708 } else if (flags & MREMAP_MAYMOVE1) {
709 abi_ulong mmap_start;
710
711 mmap_start = mmap_find_vma(0, new_size);
712
713 if (mmap_start == -1) {
714 errno(*__errno_location ()) = ENOMEM12;
715 host_addr = MAP_FAILED((void *) -1);
716 } else {
717 host_addr = (void *) syscall(__NR_mremap25, g2h(old_addr)((void *)((unsigned long)(target_ulong)(old_addr) + guest_base
))
,
718 old_size, new_size,
719 flags | MREMAP_FIXED2,
720 g2h(mmap_start)((void *)((unsigned long)(target_ulong)(mmap_start) + guest_base
))
);
721 if ( RESERVED_VAreserved_va ) {
722 mmap_reserve(old_addr, old_size);
723 }
724 }
725 } else {
726 int prot = 0;
727 if (RESERVED_VAreserved_va && old_size < new_size) {
728 abi_ulong addr;
729 for (addr = old_addr + old_size;
730 addr < old_addr + new_size;
731 addr++) {
732 prot |= page_get_flags(addr);
733 }
734 }
735 if (prot == 0) {
736 host_addr = mremap(g2h(old_addr)((void *)((unsigned long)(target_ulong)(old_addr) + guest_base
))
, old_size, new_size, flags);
737 if (host_addr != MAP_FAILED((void *) -1) && RESERVED_VAreserved_va && old_size > new_size) {
738 mmap_reserve(old_addr + old_size, new_size - old_size);
739 }
740 } else {
741 errno(*__errno_location ()) = ENOMEM12;
742 host_addr = MAP_FAILED((void *) -1);
743 }
744 /* Check if address fits target address space */
745 if ((unsigned long)host_addr + new_size > (abi_ulong)-1) {
746 /* Revert mremap() changes */
747 host_addr = mremap(g2h(old_addr)((void *)((unsigned long)(target_ulong)(old_addr) + guest_base
))
, new_size, old_size, flags)
;
Value stored to 'host_addr' is never read
748 errno(*__errno_location ()) = ENOMEM12;
749 host_addr = MAP_FAILED((void *) -1);
750 }
751 }
752
753 if (host_addr == MAP_FAILED((void *) -1)) {
754 new_addr = -1;
755 } else {
756 new_addr = h2g(host_addr)({ ((({ unsigned long __guest = (unsigned long)(host_addr) - guest_base
; (__guest < (1ul << 44)) && (!reserved_va ||
(__guest < reserved_va)); })) ? (void) (0) : __assert_fail
("({ unsigned long __guest = (unsigned long)(host_addr) - guest_base; (__guest < (1ul << 44)) && (!reserved_va || (__guest < reserved_va)); })"
, "/home/stefan/src/qemu/qemu.org/qemu/linux-user/mmap.c", 756
, __PRETTY_FUNCTION__)); ({ unsigned long __ret = (unsigned long
)(host_addr) - guest_base; (abi_ulong)__ret; }); })
;
757 prot = page_get_flags(old_addr);
758 page_set_flags(old_addr, old_addr + old_size, 0);
759 page_set_flags(new_addr, new_addr + new_size, prot | PAGE_VALID0x0008);
760 }
761 tb_invalidate_phys_range(new_addr, new_addr + new_size, 0);
762 mmap_unlock();
763 return new_addr;
764}
765
766int target_msync(abi_ulong start, abi_ulong len, int flags)
767{
768 abi_ulong end;
769
770 if (start & ~TARGET_PAGE_MASK~((1 << 13) - 1))
771 return -EINVAL22;
772 len = TARGET_PAGE_ALIGN(len)(((len) + (1 << 13) - 1) & ~((1 << 13) - 1));
773 end = start + len;
774 if (end < start)
775 return -EINVAL22;
776 if (end == start)
777 return 0;
778
779 start &= qemu_host_page_mask;
780 return msync(g2h(start)((void *)((unsigned long)(target_ulong)(start) + guest_base)), end - start, flags);
781}