thread: add tsan annotations to QemuSpin
[qemu/ar7.git] / linux-user / syscall.c
blob97de9fb5c92a9ee01918ffe3900cd5a32325482e
1 /*
2 * Linux syscalls
4 * Copyright (c) 2003 Fabrice Bellard
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, see <http://www.gnu.org/licenses/>.
19 #define _ATFILE_SOURCE
20 #include "qemu/osdep.h"
21 #include "qemu/cutils.h"
22 #include "qemu/path.h"
23 #include "qemu/memfd.h"
24 #include "qemu/queue.h"
25 #include <elf.h>
26 #include <endian.h>
27 #include <grp.h>
28 #include <sys/ipc.h>
29 #include <sys/msg.h>
30 #include <sys/wait.h>
31 #include <sys/mount.h>
32 #include <sys/file.h>
33 #include <sys/fsuid.h>
34 #include <sys/personality.h>
35 #include <sys/prctl.h>
36 #include <sys/resource.h>
37 #include <sys/swap.h>
38 #include <linux/capability.h>
39 #include <sched.h>
40 #include <sys/timex.h>
41 #include <sys/socket.h>
42 #include <linux/sockios.h>
43 #include <sys/un.h>
44 #include <sys/uio.h>
45 #include <poll.h>
46 #include <sys/times.h>
47 #include <sys/shm.h>
48 #include <sys/sem.h>
49 #include <sys/statfs.h>
50 #include <utime.h>
51 #include <sys/sysinfo.h>
52 #include <sys/signalfd.h>
53 //#include <sys/user.h>
54 #include <netinet/ip.h>
55 #include <netinet/tcp.h>
56 #include <linux/wireless.h>
57 #include <linux/icmp.h>
58 #include <linux/icmpv6.h>
59 #include <linux/errqueue.h>
60 #include <linux/random.h>
61 #ifdef CONFIG_TIMERFD
62 #include <sys/timerfd.h>
63 #endif
64 #ifdef CONFIG_EVENTFD
65 #include <sys/eventfd.h>
66 #endif
67 #ifdef CONFIG_EPOLL
68 #include <sys/epoll.h>
69 #endif
70 #ifdef CONFIG_ATTR
71 #include "qemu/xattr.h"
72 #endif
73 #ifdef CONFIG_SENDFILE
74 #include <sys/sendfile.h>
75 #endif
76 #ifdef CONFIG_KCOV
77 #include <sys/kcov.h>
78 #endif
80 #define termios host_termios
81 #define winsize host_winsize
82 #define termio host_termio
83 #define sgttyb host_sgttyb /* same as target */
84 #define tchars host_tchars /* same as target */
85 #define ltchars host_ltchars /* same as target */
87 #include <linux/termios.h>
88 #include <linux/unistd.h>
89 #include <linux/cdrom.h>
90 #include <linux/hdreg.h>
91 #include <linux/soundcard.h>
92 #include <linux/kd.h>
93 #include <linux/mtio.h>
94 #include <linux/fs.h>
95 #include <linux/fd.h>
96 #if defined(CONFIG_FIEMAP)
97 #include <linux/fiemap.h>
98 #endif
99 #include <linux/fb.h>
100 #if defined(CONFIG_USBFS)
101 #include <linux/usbdevice_fs.h>
102 #include <linux/usb/ch9.h>
103 #endif
104 #include <linux/vt.h>
105 #include <linux/dm-ioctl.h>
106 #include <linux/reboot.h>
107 #include <linux/route.h>
108 #include <linux/filter.h>
109 #include <linux/blkpg.h>
110 #include <netpacket/packet.h>
111 #include <linux/netlink.h>
112 #include <linux/if_alg.h>
113 #include <linux/rtc.h>
114 #include <sound/asound.h>
115 #include "linux_loop.h"
116 #include "uname.h"
118 #include "qemu.h"
119 #include "qemu/guest-random.h"
120 #include "qemu/selfmap.h"
121 #include "user/syscall-trace.h"
122 #include "qapi/error.h"
123 #include "fd-trans.h"
124 #include "tcg/tcg.h"
126 #ifndef CLONE_IO
127 #define CLONE_IO 0x80000000 /* Clone io context */
128 #endif
130 /* We can't directly call the host clone syscall, because this will
131 * badly confuse libc (breaking mutexes, for example). So we must
132 * divide clone flags into:
133 * * flag combinations that look like pthread_create()
134 * * flag combinations that look like fork()
135 * * flags we can implement within QEMU itself
136 * * flags we can't support and will return an error for
138 /* For thread creation, all these flags must be present; for
139 * fork, none must be present.
141 #define CLONE_THREAD_FLAGS \
142 (CLONE_VM | CLONE_FS | CLONE_FILES | \
143 CLONE_SIGHAND | CLONE_THREAD | CLONE_SYSVSEM)
145 /* These flags are ignored:
146 * CLONE_DETACHED is now ignored by the kernel;
147 * CLONE_IO is just an optimisation hint to the I/O scheduler
149 #define CLONE_IGNORED_FLAGS \
150 (CLONE_DETACHED | CLONE_IO)
152 /* Flags for fork which we can implement within QEMU itself */
153 #define CLONE_OPTIONAL_FORK_FLAGS \
154 (CLONE_SETTLS | CLONE_PARENT_SETTID | \
155 CLONE_CHILD_CLEARTID | CLONE_CHILD_SETTID)
157 /* Flags for thread creation which we can implement within QEMU itself */
158 #define CLONE_OPTIONAL_THREAD_FLAGS \
159 (CLONE_SETTLS | CLONE_PARENT_SETTID | \
160 CLONE_CHILD_CLEARTID | CLONE_CHILD_SETTID | CLONE_PARENT)
162 #define CLONE_INVALID_FORK_FLAGS \
163 (~(CSIGNAL | CLONE_OPTIONAL_FORK_FLAGS | CLONE_IGNORED_FLAGS))
165 #define CLONE_INVALID_THREAD_FLAGS \
166 (~(CSIGNAL | CLONE_THREAD_FLAGS | CLONE_OPTIONAL_THREAD_FLAGS | \
167 CLONE_IGNORED_FLAGS))
169 /* CLONE_VFORK is special cased early in do_fork(). The other flag bits
170 * have almost all been allocated. We cannot support any of
171 * CLONE_NEWNS, CLONE_NEWCGROUP, CLONE_NEWUTS, CLONE_NEWIPC,
172 * CLONE_NEWUSER, CLONE_NEWPID, CLONE_NEWNET, CLONE_PTRACE, CLONE_UNTRACED.
173 * The checks against the invalid thread masks above will catch these.
174 * (The one remaining unallocated bit is 0x1000 which used to be CLONE_PID.)
177 /* Define DEBUG_ERESTARTSYS to force every syscall to be restarted
178 * once. This exercises the codepaths for restart.
180 //#define DEBUG_ERESTARTSYS
182 //#include <linux/msdos_fs.h>
183 #define VFAT_IOCTL_READDIR_BOTH _IOR('r', 1, struct linux_dirent [2])
184 #define VFAT_IOCTL_READDIR_SHORT _IOR('r', 2, struct linux_dirent [2])
186 #undef _syscall0
187 #undef _syscall1
188 #undef _syscall2
189 #undef _syscall3
190 #undef _syscall4
191 #undef _syscall5
192 #undef _syscall6
194 #define _syscall0(type,name) \
195 static type name (void) \
197 return syscall(__NR_##name); \
200 #define _syscall1(type,name,type1,arg1) \
201 static type name (type1 arg1) \
203 return syscall(__NR_##name, arg1); \
206 #define _syscall2(type,name,type1,arg1,type2,arg2) \
207 static type name (type1 arg1,type2 arg2) \
209 return syscall(__NR_##name, arg1, arg2); \
212 #define _syscall3(type,name,type1,arg1,type2,arg2,type3,arg3) \
213 static type name (type1 arg1,type2 arg2,type3 arg3) \
215 return syscall(__NR_##name, arg1, arg2, arg3); \
218 #define _syscall4(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4) \
219 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4) \
221 return syscall(__NR_##name, arg1, arg2, arg3, arg4); \
224 #define _syscall5(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
225 type5,arg5) \
226 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4,type5 arg5) \
228 return syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5); \
232 #define _syscall6(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
233 type5,arg5,type6,arg6) \
234 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4,type5 arg5, \
235 type6 arg6) \
237 return syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5, arg6); \
241 #define __NR_sys_uname __NR_uname
242 #define __NR_sys_getcwd1 __NR_getcwd
243 #define __NR_sys_getdents __NR_getdents
244 #define __NR_sys_getdents64 __NR_getdents64
245 #define __NR_sys_getpriority __NR_getpriority
246 #define __NR_sys_rt_sigqueueinfo __NR_rt_sigqueueinfo
247 #define __NR_sys_rt_tgsigqueueinfo __NR_rt_tgsigqueueinfo
248 #define __NR_sys_syslog __NR_syslog
249 #if defined(__NR_futex)
250 # define __NR_sys_futex __NR_futex
251 #endif
252 #if defined(__NR_futex_time64)
253 # define __NR_sys_futex_time64 __NR_futex_time64
254 #endif
255 #define __NR_sys_inotify_init __NR_inotify_init
256 #define __NR_sys_inotify_add_watch __NR_inotify_add_watch
257 #define __NR_sys_inotify_rm_watch __NR_inotify_rm_watch
258 #define __NR_sys_statx __NR_statx
260 #if defined(__alpha__) || defined(__x86_64__) || defined(__s390x__)
261 #define __NR__llseek __NR_lseek
262 #endif
264 /* Newer kernel ports have llseek() instead of _llseek() */
265 #if defined(TARGET_NR_llseek) && !defined(TARGET_NR__llseek)
266 #define TARGET_NR__llseek TARGET_NR_llseek
267 #endif
269 #define __NR_sys_gettid __NR_gettid
270 _syscall0(int, sys_gettid)
272 /* For the 64-bit guest on 32-bit host case we must emulate
273 * getdents using getdents64, because otherwise the host
274 * might hand us back more dirent records than we can fit
275 * into the guest buffer after structure format conversion.
276 * Otherwise we emulate getdents with getdents if the host has it.
278 #if defined(__NR_getdents) && HOST_LONG_BITS >= TARGET_ABI_BITS
279 #define EMULATE_GETDENTS_WITH_GETDENTS
280 #endif
282 #if defined(TARGET_NR_getdents) && defined(EMULATE_GETDENTS_WITH_GETDENTS)
283 _syscall3(int, sys_getdents, uint, fd, struct linux_dirent *, dirp, uint, count);
284 #endif
285 #if (defined(TARGET_NR_getdents) && \
286 !defined(EMULATE_GETDENTS_WITH_GETDENTS)) || \
287 (defined(TARGET_NR_getdents64) && defined(__NR_getdents64))
288 _syscall3(int, sys_getdents64, uint, fd, struct linux_dirent64 *, dirp, uint, count);
289 #endif
290 #if defined(TARGET_NR__llseek) && defined(__NR_llseek)
291 _syscall5(int, _llseek, uint, fd, ulong, hi, ulong, lo,
292 loff_t *, res, uint, wh);
293 #endif
294 _syscall3(int, sys_rt_sigqueueinfo, pid_t, pid, int, sig, siginfo_t *, uinfo)
295 _syscall4(int, sys_rt_tgsigqueueinfo, pid_t, pid, pid_t, tid, int, sig,
296 siginfo_t *, uinfo)
297 _syscall3(int,sys_syslog,int,type,char*,bufp,int,len)
298 #ifdef __NR_exit_group
299 _syscall1(int,exit_group,int,error_code)
300 #endif
301 #if defined(TARGET_NR_set_tid_address) && defined(__NR_set_tid_address)
302 _syscall1(int,set_tid_address,int *,tidptr)
303 #endif
304 #if defined(__NR_futex)
305 _syscall6(int,sys_futex,int *,uaddr,int,op,int,val,
306 const struct timespec *,timeout,int *,uaddr2,int,val3)
307 #endif
308 #if defined(__NR_futex_time64)
309 _syscall6(int,sys_futex_time64,int *,uaddr,int,op,int,val,
310 const struct timespec *,timeout,int *,uaddr2,int,val3)
311 #endif
312 #define __NR_sys_sched_getaffinity __NR_sched_getaffinity
313 _syscall3(int, sys_sched_getaffinity, pid_t, pid, unsigned int, len,
314 unsigned long *, user_mask_ptr);
315 #define __NR_sys_sched_setaffinity __NR_sched_setaffinity
316 _syscall3(int, sys_sched_setaffinity, pid_t, pid, unsigned int, len,
317 unsigned long *, user_mask_ptr);
318 #define __NR_sys_getcpu __NR_getcpu
319 _syscall3(int, sys_getcpu, unsigned *, cpu, unsigned *, node, void *, tcache);
320 _syscall4(int, reboot, int, magic1, int, magic2, unsigned int, cmd,
321 void *, arg);
322 _syscall2(int, capget, struct __user_cap_header_struct *, header,
323 struct __user_cap_data_struct *, data);
324 _syscall2(int, capset, struct __user_cap_header_struct *, header,
325 struct __user_cap_data_struct *, data);
326 #if defined(TARGET_NR_ioprio_get) && defined(__NR_ioprio_get)
327 _syscall2(int, ioprio_get, int, which, int, who)
328 #endif
329 #if defined(TARGET_NR_ioprio_set) && defined(__NR_ioprio_set)
330 _syscall3(int, ioprio_set, int, which, int, who, int, ioprio)
331 #endif
332 #if defined(TARGET_NR_getrandom) && defined(__NR_getrandom)
333 _syscall3(int, getrandom, void *, buf, size_t, buflen, unsigned int, flags)
334 #endif
336 #if defined(TARGET_NR_kcmp) && defined(__NR_kcmp)
337 _syscall5(int, kcmp, pid_t, pid1, pid_t, pid2, int, type,
338 unsigned long, idx1, unsigned long, idx2)
339 #endif
342 * It is assumed that struct statx is architecture independent.
344 #if defined(TARGET_NR_statx) && defined(__NR_statx)
345 _syscall5(int, sys_statx, int, dirfd, const char *, pathname, int, flags,
346 unsigned int, mask, struct target_statx *, statxbuf)
347 #endif
348 #if defined(TARGET_NR_membarrier) && defined(__NR_membarrier)
349 _syscall2(int, membarrier, int, cmd, int, flags)
350 #endif
352 static bitmask_transtbl fcntl_flags_tbl[] = {
353 { TARGET_O_ACCMODE, TARGET_O_WRONLY, O_ACCMODE, O_WRONLY, },
354 { TARGET_O_ACCMODE, TARGET_O_RDWR, O_ACCMODE, O_RDWR, },
355 { TARGET_O_CREAT, TARGET_O_CREAT, O_CREAT, O_CREAT, },
356 { TARGET_O_EXCL, TARGET_O_EXCL, O_EXCL, O_EXCL, },
357 { TARGET_O_NOCTTY, TARGET_O_NOCTTY, O_NOCTTY, O_NOCTTY, },
358 { TARGET_O_TRUNC, TARGET_O_TRUNC, O_TRUNC, O_TRUNC, },
359 { TARGET_O_APPEND, TARGET_O_APPEND, O_APPEND, O_APPEND, },
360 { TARGET_O_NONBLOCK, TARGET_O_NONBLOCK, O_NONBLOCK, O_NONBLOCK, },
361 { TARGET_O_SYNC, TARGET_O_DSYNC, O_SYNC, O_DSYNC, },
362 { TARGET_O_SYNC, TARGET_O_SYNC, O_SYNC, O_SYNC, },
363 { TARGET_FASYNC, TARGET_FASYNC, FASYNC, FASYNC, },
364 { TARGET_O_DIRECTORY, TARGET_O_DIRECTORY, O_DIRECTORY, O_DIRECTORY, },
365 { TARGET_O_NOFOLLOW, TARGET_O_NOFOLLOW, O_NOFOLLOW, O_NOFOLLOW, },
366 #if defined(O_DIRECT)
367 { TARGET_O_DIRECT, TARGET_O_DIRECT, O_DIRECT, O_DIRECT, },
368 #endif
369 #if defined(O_NOATIME)
370 { TARGET_O_NOATIME, TARGET_O_NOATIME, O_NOATIME, O_NOATIME },
371 #endif
372 #if defined(O_CLOEXEC)
373 { TARGET_O_CLOEXEC, TARGET_O_CLOEXEC, O_CLOEXEC, O_CLOEXEC },
374 #endif
375 #if defined(O_PATH)
376 { TARGET_O_PATH, TARGET_O_PATH, O_PATH, O_PATH },
377 #endif
378 #if defined(O_TMPFILE)
379 { TARGET_O_TMPFILE, TARGET_O_TMPFILE, O_TMPFILE, O_TMPFILE },
380 #endif
381 /* Don't terminate the list prematurely on 64-bit host+guest. */
382 #if TARGET_O_LARGEFILE != 0 || O_LARGEFILE != 0
383 { TARGET_O_LARGEFILE, TARGET_O_LARGEFILE, O_LARGEFILE, O_LARGEFILE, },
384 #endif
385 { 0, 0, 0, 0 }
388 static int sys_getcwd1(char *buf, size_t size)
390 if (getcwd(buf, size) == NULL) {
391 /* getcwd() sets errno */
392 return (-1);
394 return strlen(buf)+1;
397 #ifdef TARGET_NR_utimensat
398 #if defined(__NR_utimensat)
399 #define __NR_sys_utimensat __NR_utimensat
400 _syscall4(int,sys_utimensat,int,dirfd,const char *,pathname,
401 const struct timespec *,tsp,int,flags)
402 #else
403 static int sys_utimensat(int dirfd, const char *pathname,
404 const struct timespec times[2], int flags)
406 errno = ENOSYS;
407 return -1;
409 #endif
410 #endif /* TARGET_NR_utimensat */
412 #ifdef TARGET_NR_renameat2
413 #if defined(__NR_renameat2)
414 #define __NR_sys_renameat2 __NR_renameat2
415 _syscall5(int, sys_renameat2, int, oldfd, const char *, old, int, newfd,
416 const char *, new, unsigned int, flags)
417 #else
418 static int sys_renameat2(int oldfd, const char *old,
419 int newfd, const char *new, int flags)
421 if (flags == 0) {
422 return renameat(oldfd, old, newfd, new);
424 errno = ENOSYS;
425 return -1;
427 #endif
428 #endif /* TARGET_NR_renameat2 */
430 #ifdef CONFIG_INOTIFY
431 #include <sys/inotify.h>
433 #if defined(TARGET_NR_inotify_init) && defined(__NR_inotify_init)
434 static int sys_inotify_init(void)
436 return (inotify_init());
438 #endif
439 #if defined(TARGET_NR_inotify_add_watch) && defined(__NR_inotify_add_watch)
440 static int sys_inotify_add_watch(int fd,const char *pathname, int32_t mask)
442 return (inotify_add_watch(fd, pathname, mask));
444 #endif
445 #if defined(TARGET_NR_inotify_rm_watch) && defined(__NR_inotify_rm_watch)
446 static int sys_inotify_rm_watch(int fd, int32_t wd)
448 return (inotify_rm_watch(fd, wd));
450 #endif
451 #ifdef CONFIG_INOTIFY1
452 #if defined(TARGET_NR_inotify_init1) && defined(__NR_inotify_init1)
453 static int sys_inotify_init1(int flags)
455 return (inotify_init1(flags));
457 #endif
458 #endif
459 #else
460 /* Userspace can usually survive runtime without inotify */
461 #undef TARGET_NR_inotify_init
462 #undef TARGET_NR_inotify_init1
463 #undef TARGET_NR_inotify_add_watch
464 #undef TARGET_NR_inotify_rm_watch
465 #endif /* CONFIG_INOTIFY */
467 #if defined(TARGET_NR_prlimit64)
468 #ifndef __NR_prlimit64
469 # define __NR_prlimit64 -1
470 #endif
471 #define __NR_sys_prlimit64 __NR_prlimit64
472 /* The glibc rlimit structure may not be that used by the underlying syscall */
473 struct host_rlimit64 {
474 uint64_t rlim_cur;
475 uint64_t rlim_max;
477 _syscall4(int, sys_prlimit64, pid_t, pid, int, resource,
478 const struct host_rlimit64 *, new_limit,
479 struct host_rlimit64 *, old_limit)
480 #endif
483 #if defined(TARGET_NR_timer_create)
484 /* Maxiumum of 32 active POSIX timers allowed at any one time. */
485 static timer_t g_posix_timers[32] = { 0, } ;
487 static inline int next_free_host_timer(void)
489 int k ;
490 /* FIXME: Does finding the next free slot require a lock? */
491 for (k = 0; k < ARRAY_SIZE(g_posix_timers); k++) {
492 if (g_posix_timers[k] == 0) {
493 g_posix_timers[k] = (timer_t) 1;
494 return k;
497 return -1;
499 #endif
501 /* ARM EABI and MIPS expect 64bit types aligned even on pairs or registers */
502 #ifdef TARGET_ARM
503 static inline int regpairs_aligned(void *cpu_env, int num)
505 return ((((CPUARMState *)cpu_env)->eabi) == 1) ;
507 #elif defined(TARGET_MIPS) && (TARGET_ABI_BITS == 32)
508 static inline int regpairs_aligned(void *cpu_env, int num) { return 1; }
509 #elif defined(TARGET_PPC) && !defined(TARGET_PPC64)
510 /* SysV AVI for PPC32 expects 64bit parameters to be passed on odd/even pairs
511 * of registers which translates to the same as ARM/MIPS, because we start with
512 * r3 as arg1 */
513 static inline int regpairs_aligned(void *cpu_env, int num) { return 1; }
514 #elif defined(TARGET_SH4)
515 /* SH4 doesn't align register pairs, except for p{read,write}64 */
516 static inline int regpairs_aligned(void *cpu_env, int num)
518 switch (num) {
519 case TARGET_NR_pread64:
520 case TARGET_NR_pwrite64:
521 return 1;
523 default:
524 return 0;
527 #elif defined(TARGET_XTENSA)
528 static inline int regpairs_aligned(void *cpu_env, int num) { return 1; }
529 #else
530 static inline int regpairs_aligned(void *cpu_env, int num) { return 0; }
531 #endif
533 #define ERRNO_TABLE_SIZE 1200
535 /* target_to_host_errno_table[] is initialized from
536 * host_to_target_errno_table[] in syscall_init(). */
537 static uint16_t target_to_host_errno_table[ERRNO_TABLE_SIZE] = {
541 * This list is the union of errno values overridden in asm-<arch>/errno.h
542 * minus the errnos that are not actually generic to all archs.
544 static uint16_t host_to_target_errno_table[ERRNO_TABLE_SIZE] = {
545 [EAGAIN] = TARGET_EAGAIN,
546 [EIDRM] = TARGET_EIDRM,
547 [ECHRNG] = TARGET_ECHRNG,
548 [EL2NSYNC] = TARGET_EL2NSYNC,
549 [EL3HLT] = TARGET_EL3HLT,
550 [EL3RST] = TARGET_EL3RST,
551 [ELNRNG] = TARGET_ELNRNG,
552 [EUNATCH] = TARGET_EUNATCH,
553 [ENOCSI] = TARGET_ENOCSI,
554 [EL2HLT] = TARGET_EL2HLT,
555 [EDEADLK] = TARGET_EDEADLK,
556 [ENOLCK] = TARGET_ENOLCK,
557 [EBADE] = TARGET_EBADE,
558 [EBADR] = TARGET_EBADR,
559 [EXFULL] = TARGET_EXFULL,
560 [ENOANO] = TARGET_ENOANO,
561 [EBADRQC] = TARGET_EBADRQC,
562 [EBADSLT] = TARGET_EBADSLT,
563 [EBFONT] = TARGET_EBFONT,
564 [ENOSTR] = TARGET_ENOSTR,
565 [ENODATA] = TARGET_ENODATA,
566 [ETIME] = TARGET_ETIME,
567 [ENOSR] = TARGET_ENOSR,
568 [ENONET] = TARGET_ENONET,
569 [ENOPKG] = TARGET_ENOPKG,
570 [EREMOTE] = TARGET_EREMOTE,
571 [ENOLINK] = TARGET_ENOLINK,
572 [EADV] = TARGET_EADV,
573 [ESRMNT] = TARGET_ESRMNT,
574 [ECOMM] = TARGET_ECOMM,
575 [EPROTO] = TARGET_EPROTO,
576 [EDOTDOT] = TARGET_EDOTDOT,
577 [EMULTIHOP] = TARGET_EMULTIHOP,
578 [EBADMSG] = TARGET_EBADMSG,
579 [ENAMETOOLONG] = TARGET_ENAMETOOLONG,
580 [EOVERFLOW] = TARGET_EOVERFLOW,
581 [ENOTUNIQ] = TARGET_ENOTUNIQ,
582 [EBADFD] = TARGET_EBADFD,
583 [EREMCHG] = TARGET_EREMCHG,
584 [ELIBACC] = TARGET_ELIBACC,
585 [ELIBBAD] = TARGET_ELIBBAD,
586 [ELIBSCN] = TARGET_ELIBSCN,
587 [ELIBMAX] = TARGET_ELIBMAX,
588 [ELIBEXEC] = TARGET_ELIBEXEC,
589 [EILSEQ] = TARGET_EILSEQ,
590 [ENOSYS] = TARGET_ENOSYS,
591 [ELOOP] = TARGET_ELOOP,
592 [ERESTART] = TARGET_ERESTART,
593 [ESTRPIPE] = TARGET_ESTRPIPE,
594 [ENOTEMPTY] = TARGET_ENOTEMPTY,
595 [EUSERS] = TARGET_EUSERS,
596 [ENOTSOCK] = TARGET_ENOTSOCK,
597 [EDESTADDRREQ] = TARGET_EDESTADDRREQ,
598 [EMSGSIZE] = TARGET_EMSGSIZE,
599 [EPROTOTYPE] = TARGET_EPROTOTYPE,
600 [ENOPROTOOPT] = TARGET_ENOPROTOOPT,
601 [EPROTONOSUPPORT] = TARGET_EPROTONOSUPPORT,
602 [ESOCKTNOSUPPORT] = TARGET_ESOCKTNOSUPPORT,
603 [EOPNOTSUPP] = TARGET_EOPNOTSUPP,
604 [EPFNOSUPPORT] = TARGET_EPFNOSUPPORT,
605 [EAFNOSUPPORT] = TARGET_EAFNOSUPPORT,
606 [EADDRINUSE] = TARGET_EADDRINUSE,
607 [EADDRNOTAVAIL] = TARGET_EADDRNOTAVAIL,
608 [ENETDOWN] = TARGET_ENETDOWN,
609 [ENETUNREACH] = TARGET_ENETUNREACH,
610 [ENETRESET] = TARGET_ENETRESET,
611 [ECONNABORTED] = TARGET_ECONNABORTED,
612 [ECONNRESET] = TARGET_ECONNRESET,
613 [ENOBUFS] = TARGET_ENOBUFS,
614 [EISCONN] = TARGET_EISCONN,
615 [ENOTCONN] = TARGET_ENOTCONN,
616 [EUCLEAN] = TARGET_EUCLEAN,
617 [ENOTNAM] = TARGET_ENOTNAM,
618 [ENAVAIL] = TARGET_ENAVAIL,
619 [EISNAM] = TARGET_EISNAM,
620 [EREMOTEIO] = TARGET_EREMOTEIO,
621 [EDQUOT] = TARGET_EDQUOT,
622 [ESHUTDOWN] = TARGET_ESHUTDOWN,
623 [ETOOMANYREFS] = TARGET_ETOOMANYREFS,
624 [ETIMEDOUT] = TARGET_ETIMEDOUT,
625 [ECONNREFUSED] = TARGET_ECONNREFUSED,
626 [EHOSTDOWN] = TARGET_EHOSTDOWN,
627 [EHOSTUNREACH] = TARGET_EHOSTUNREACH,
628 [EALREADY] = TARGET_EALREADY,
629 [EINPROGRESS] = TARGET_EINPROGRESS,
630 [ESTALE] = TARGET_ESTALE,
631 [ECANCELED] = TARGET_ECANCELED,
632 [ENOMEDIUM] = TARGET_ENOMEDIUM,
633 [EMEDIUMTYPE] = TARGET_EMEDIUMTYPE,
634 #ifdef ENOKEY
635 [ENOKEY] = TARGET_ENOKEY,
636 #endif
637 #ifdef EKEYEXPIRED
638 [EKEYEXPIRED] = TARGET_EKEYEXPIRED,
639 #endif
640 #ifdef EKEYREVOKED
641 [EKEYREVOKED] = TARGET_EKEYREVOKED,
642 #endif
643 #ifdef EKEYREJECTED
644 [EKEYREJECTED] = TARGET_EKEYREJECTED,
645 #endif
646 #ifdef EOWNERDEAD
647 [EOWNERDEAD] = TARGET_EOWNERDEAD,
648 #endif
649 #ifdef ENOTRECOVERABLE
650 [ENOTRECOVERABLE] = TARGET_ENOTRECOVERABLE,
651 #endif
652 #ifdef ENOMSG
653 [ENOMSG] = TARGET_ENOMSG,
654 #endif
655 #ifdef ERKFILL
656 [ERFKILL] = TARGET_ERFKILL,
657 #endif
658 #ifdef EHWPOISON
659 [EHWPOISON] = TARGET_EHWPOISON,
660 #endif
663 static inline int host_to_target_errno(int err)
665 if (err >= 0 && err < ERRNO_TABLE_SIZE &&
666 host_to_target_errno_table[err]) {
667 return host_to_target_errno_table[err];
669 return err;
672 static inline int target_to_host_errno(int err)
674 if (err >= 0 && err < ERRNO_TABLE_SIZE &&
675 target_to_host_errno_table[err]) {
676 return target_to_host_errno_table[err];
678 return err;
681 static inline abi_long get_errno(abi_long ret)
683 if (ret == -1)
684 return -host_to_target_errno(errno);
685 else
686 return ret;
689 const char *target_strerror(int err)
691 if (err == TARGET_ERESTARTSYS) {
692 return "To be restarted";
694 if (err == TARGET_QEMU_ESIGRETURN) {
695 return "Successful exit from sigreturn";
698 if ((err >= ERRNO_TABLE_SIZE) || (err < 0)) {
699 return NULL;
701 return strerror(target_to_host_errno(err));
704 #define safe_syscall0(type, name) \
705 static type safe_##name(void) \
707 return safe_syscall(__NR_##name); \
710 #define safe_syscall1(type, name, type1, arg1) \
711 static type safe_##name(type1 arg1) \
713 return safe_syscall(__NR_##name, arg1); \
716 #define safe_syscall2(type, name, type1, arg1, type2, arg2) \
717 static type safe_##name(type1 arg1, type2 arg2) \
719 return safe_syscall(__NR_##name, arg1, arg2); \
722 #define safe_syscall3(type, name, type1, arg1, type2, arg2, type3, arg3) \
723 static type safe_##name(type1 arg1, type2 arg2, type3 arg3) \
725 return safe_syscall(__NR_##name, arg1, arg2, arg3); \
728 #define safe_syscall4(type, name, type1, arg1, type2, arg2, type3, arg3, \
729 type4, arg4) \
730 static type safe_##name(type1 arg1, type2 arg2, type3 arg3, type4 arg4) \
732 return safe_syscall(__NR_##name, arg1, arg2, arg3, arg4); \
735 #define safe_syscall5(type, name, type1, arg1, type2, arg2, type3, arg3, \
736 type4, arg4, type5, arg5) \
737 static type safe_##name(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
738 type5 arg5) \
740 return safe_syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5); \
743 #define safe_syscall6(type, name, type1, arg1, type2, arg2, type3, arg3, \
744 type4, arg4, type5, arg5, type6, arg6) \
745 static type safe_##name(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
746 type5 arg5, type6 arg6) \
748 return safe_syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5, arg6); \
751 safe_syscall3(ssize_t, read, int, fd, void *, buff, size_t, count)
752 safe_syscall3(ssize_t, write, int, fd, const void *, buff, size_t, count)
753 safe_syscall4(int, openat, int, dirfd, const char *, pathname, \
754 int, flags, mode_t, mode)
755 #if defined(TARGET_NR_wait4) || defined(TARGET_NR_waitpid)
756 safe_syscall4(pid_t, wait4, pid_t, pid, int *, status, int, options, \
757 struct rusage *, rusage)
758 #endif
759 safe_syscall5(int, waitid, idtype_t, idtype, id_t, id, siginfo_t *, infop, \
760 int, options, struct rusage *, rusage)
761 safe_syscall3(int, execve, const char *, filename, char **, argv, char **, envp)
762 #if defined(TARGET_NR_select) || defined(TARGET_NR__newselect) || \
763 defined(TARGET_NR_pselect6)
764 safe_syscall6(int, pselect6, int, nfds, fd_set *, readfds, fd_set *, writefds, \
765 fd_set *, exceptfds, struct timespec *, timeout, void *, sig)
766 #endif
767 #if defined(TARGET_NR_ppoll) || defined(TARGET_NR_poll)
768 safe_syscall5(int, ppoll, struct pollfd *, ufds, unsigned int, nfds,
769 struct timespec *, tsp, const sigset_t *, sigmask,
770 size_t, sigsetsize)
771 #endif
772 safe_syscall6(int, epoll_pwait, int, epfd, struct epoll_event *, events,
773 int, maxevents, int, timeout, const sigset_t *, sigmask,
774 size_t, sigsetsize)
775 #if defined(__NR_futex)
776 safe_syscall6(int,futex,int *,uaddr,int,op,int,val, \
777 const struct timespec *,timeout,int *,uaddr2,int,val3)
778 #endif
779 #if defined(__NR_futex_time64)
780 safe_syscall6(int,futex_time64,int *,uaddr,int,op,int,val, \
781 const struct timespec *,timeout,int *,uaddr2,int,val3)
782 #endif
783 safe_syscall2(int, rt_sigsuspend, sigset_t *, newset, size_t, sigsetsize)
784 safe_syscall2(int, kill, pid_t, pid, int, sig)
785 safe_syscall2(int, tkill, int, tid, int, sig)
786 safe_syscall3(int, tgkill, int, tgid, int, pid, int, sig)
787 safe_syscall3(ssize_t, readv, int, fd, const struct iovec *, iov, int, iovcnt)
788 safe_syscall3(ssize_t, writev, int, fd, const struct iovec *, iov, int, iovcnt)
789 safe_syscall5(ssize_t, preadv, int, fd, const struct iovec *, iov, int, iovcnt,
790 unsigned long, pos_l, unsigned long, pos_h)
791 safe_syscall5(ssize_t, pwritev, int, fd, const struct iovec *, iov, int, iovcnt,
792 unsigned long, pos_l, unsigned long, pos_h)
793 safe_syscall3(int, connect, int, fd, const struct sockaddr *, addr,
794 socklen_t, addrlen)
795 safe_syscall6(ssize_t, sendto, int, fd, const void *, buf, size_t, len,
796 int, flags, const struct sockaddr *, addr, socklen_t, addrlen)
797 safe_syscall6(ssize_t, recvfrom, int, fd, void *, buf, size_t, len,
798 int, flags, struct sockaddr *, addr, socklen_t *, addrlen)
799 safe_syscall3(ssize_t, sendmsg, int, fd, const struct msghdr *, msg, int, flags)
800 safe_syscall3(ssize_t, recvmsg, int, fd, struct msghdr *, msg, int, flags)
801 safe_syscall2(int, flock, int, fd, int, operation)
802 #ifdef TARGET_NR_rt_sigtimedwait
803 safe_syscall4(int, rt_sigtimedwait, const sigset_t *, these, siginfo_t *, uinfo,
804 const struct timespec *, uts, size_t, sigsetsize)
805 #endif
806 safe_syscall4(int, accept4, int, fd, struct sockaddr *, addr, socklen_t *, len,
807 int, flags)
808 #if defined(TARGET_NR_nanosleep)
809 safe_syscall2(int, nanosleep, const struct timespec *, req,
810 struct timespec *, rem)
811 #endif
812 #ifdef TARGET_NR_clock_nanosleep
813 safe_syscall4(int, clock_nanosleep, const clockid_t, clock, int, flags,
814 const struct timespec *, req, struct timespec *, rem)
815 #endif
816 #ifdef __NR_ipc
817 safe_syscall6(int, ipc, int, call, long, first, long, second, long, third,
818 void *, ptr, long, fifth)
819 #endif
820 #ifdef __NR_msgsnd
821 safe_syscall4(int, msgsnd, int, msgid, const void *, msgp, size_t, sz,
822 int, flags)
823 #endif
824 #ifdef __NR_msgrcv
825 safe_syscall5(int, msgrcv, int, msgid, void *, msgp, size_t, sz,
826 long, msgtype, int, flags)
827 #endif
828 #ifdef __NR_semtimedop
829 safe_syscall4(int, semtimedop, int, semid, struct sembuf *, tsops,
830 unsigned, nsops, const struct timespec *, timeout)
831 #endif
832 #ifdef TARGET_NR_mq_timedsend
833 safe_syscall5(int, mq_timedsend, int, mqdes, const char *, msg_ptr,
834 size_t, len, unsigned, prio, const struct timespec *, timeout)
835 #endif
836 #ifdef TARGET_NR_mq_timedreceive
837 safe_syscall5(int, mq_timedreceive, int, mqdes, char *, msg_ptr,
838 size_t, len, unsigned *, prio, const struct timespec *, timeout)
839 #endif
840 /* We do ioctl like this rather than via safe_syscall3 to preserve the
841 * "third argument might be integer or pointer or not present" behaviour of
842 * the libc function.
844 #define safe_ioctl(...) safe_syscall(__NR_ioctl, __VA_ARGS__)
845 /* Similarly for fcntl. Note that callers must always:
846 * pass the F_GETLK64 etc constants rather than the unsuffixed F_GETLK
847 * use the flock64 struct rather than unsuffixed flock
848 * This will then work and use a 64-bit offset for both 32-bit and 64-bit hosts.
850 #ifdef __NR_fcntl64
851 #define safe_fcntl(...) safe_syscall(__NR_fcntl64, __VA_ARGS__)
852 #else
853 #define safe_fcntl(...) safe_syscall(__NR_fcntl, __VA_ARGS__)
854 #endif
856 static inline int host_to_target_sock_type(int host_type)
858 int target_type;
860 switch (host_type & 0xf /* SOCK_TYPE_MASK */) {
861 case SOCK_DGRAM:
862 target_type = TARGET_SOCK_DGRAM;
863 break;
864 case SOCK_STREAM:
865 target_type = TARGET_SOCK_STREAM;
866 break;
867 default:
868 target_type = host_type & 0xf /* SOCK_TYPE_MASK */;
869 break;
872 #if defined(SOCK_CLOEXEC)
873 if (host_type & SOCK_CLOEXEC) {
874 target_type |= TARGET_SOCK_CLOEXEC;
876 #endif
878 #if defined(SOCK_NONBLOCK)
879 if (host_type & SOCK_NONBLOCK) {
880 target_type |= TARGET_SOCK_NONBLOCK;
882 #endif
884 return target_type;
887 static abi_ulong target_brk;
888 static abi_ulong target_original_brk;
889 static abi_ulong brk_page;
891 void target_set_brk(abi_ulong new_brk)
893 target_original_brk = target_brk = HOST_PAGE_ALIGN(new_brk);
894 brk_page = HOST_PAGE_ALIGN(target_brk);
897 //#define DEBUGF_BRK(message, args...) do { fprintf(stderr, (message), ## args); } while (0)
898 #define DEBUGF_BRK(message, args...)
900 /* do_brk() must return target values and target errnos. */
901 abi_long do_brk(abi_ulong new_brk)
903 abi_long mapped_addr;
904 abi_ulong new_alloc_size;
906 DEBUGF_BRK("do_brk(" TARGET_ABI_FMT_lx ") -> ", new_brk);
908 if (!new_brk) {
909 DEBUGF_BRK(TARGET_ABI_FMT_lx " (!new_brk)\n", target_brk);
910 return target_brk;
912 if (new_brk < target_original_brk) {
913 DEBUGF_BRK(TARGET_ABI_FMT_lx " (new_brk < target_original_brk)\n",
914 target_brk);
915 return target_brk;
918 /* If the new brk is less than the highest page reserved to the
919 * target heap allocation, set it and we're almost done... */
920 if (new_brk <= brk_page) {
921 /* Heap contents are initialized to zero, as for anonymous
922 * mapped pages. */
923 if (new_brk > target_brk) {
924 memset(g2h(target_brk), 0, new_brk - target_brk);
926 target_brk = new_brk;
927 DEBUGF_BRK(TARGET_ABI_FMT_lx " (new_brk <= brk_page)\n", target_brk);
928 return target_brk;
931 /* We need to allocate more memory after the brk... Note that
932 * we don't use MAP_FIXED because that will map over the top of
933 * any existing mapping (like the one with the host libc or qemu
934 * itself); instead we treat "mapped but at wrong address" as
935 * a failure and unmap again.
937 new_alloc_size = HOST_PAGE_ALIGN(new_brk - brk_page);
938 mapped_addr = get_errno(target_mmap(brk_page, new_alloc_size,
939 PROT_READ|PROT_WRITE,
940 MAP_ANON|MAP_PRIVATE, 0, 0));
942 if (mapped_addr == brk_page) {
943 /* Heap contents are initialized to zero, as for anonymous
944 * mapped pages. Technically the new pages are already
945 * initialized to zero since they *are* anonymous mapped
946 * pages, however we have to take care with the contents that
947 * come from the remaining part of the previous page: it may
948 * contains garbage data due to a previous heap usage (grown
949 * then shrunken). */
950 memset(g2h(target_brk), 0, brk_page - target_brk);
952 target_brk = new_brk;
953 brk_page = HOST_PAGE_ALIGN(target_brk);
954 DEBUGF_BRK(TARGET_ABI_FMT_lx " (mapped_addr == brk_page)\n",
955 target_brk);
956 return target_brk;
957 } else if (mapped_addr != -1) {
958 /* Mapped but at wrong address, meaning there wasn't actually
959 * enough space for this brk.
961 target_munmap(mapped_addr, new_alloc_size);
962 mapped_addr = -1;
963 DEBUGF_BRK(TARGET_ABI_FMT_lx " (mapped_addr != -1)\n", target_brk);
965 else {
966 DEBUGF_BRK(TARGET_ABI_FMT_lx " (otherwise)\n", target_brk);
969 #if defined(TARGET_ALPHA)
970 /* We (partially) emulate OSF/1 on Alpha, which requires we
971 return a proper errno, not an unchanged brk value. */
972 return -TARGET_ENOMEM;
973 #endif
974 /* For everything else, return the previous break. */
975 return target_brk;
978 #if defined(TARGET_NR_select) || defined(TARGET_NR__newselect) || \
979 defined(TARGET_NR_pselect6)
980 static inline abi_long copy_from_user_fdset(fd_set *fds,
981 abi_ulong target_fds_addr,
982 int n)
984 int i, nw, j, k;
985 abi_ulong b, *target_fds;
987 nw = DIV_ROUND_UP(n, TARGET_ABI_BITS);
988 if (!(target_fds = lock_user(VERIFY_READ,
989 target_fds_addr,
990 sizeof(abi_ulong) * nw,
991 1)))
992 return -TARGET_EFAULT;
994 FD_ZERO(fds);
995 k = 0;
996 for (i = 0; i < nw; i++) {
997 /* grab the abi_ulong */
998 __get_user(b, &target_fds[i]);
999 for (j = 0; j < TARGET_ABI_BITS; j++) {
1000 /* check the bit inside the abi_ulong */
1001 if ((b >> j) & 1)
1002 FD_SET(k, fds);
1003 k++;
1007 unlock_user(target_fds, target_fds_addr, 0);
1009 return 0;
1012 static inline abi_ulong copy_from_user_fdset_ptr(fd_set *fds, fd_set **fds_ptr,
1013 abi_ulong target_fds_addr,
1014 int n)
1016 if (target_fds_addr) {
1017 if (copy_from_user_fdset(fds, target_fds_addr, n))
1018 return -TARGET_EFAULT;
1019 *fds_ptr = fds;
1020 } else {
1021 *fds_ptr = NULL;
1023 return 0;
1026 static inline abi_long copy_to_user_fdset(abi_ulong target_fds_addr,
1027 const fd_set *fds,
1028 int n)
1030 int i, nw, j, k;
1031 abi_long v;
1032 abi_ulong *target_fds;
1034 nw = DIV_ROUND_UP(n, TARGET_ABI_BITS);
1035 if (!(target_fds = lock_user(VERIFY_WRITE,
1036 target_fds_addr,
1037 sizeof(abi_ulong) * nw,
1038 0)))
1039 return -TARGET_EFAULT;
1041 k = 0;
1042 for (i = 0; i < nw; i++) {
1043 v = 0;
1044 for (j = 0; j < TARGET_ABI_BITS; j++) {
1045 v |= ((abi_ulong)(FD_ISSET(k, fds) != 0) << j);
1046 k++;
1048 __put_user(v, &target_fds[i]);
1051 unlock_user(target_fds, target_fds_addr, sizeof(abi_ulong) * nw);
1053 return 0;
1055 #endif
1057 #if defined(__alpha__)
1058 #define HOST_HZ 1024
1059 #else
1060 #define HOST_HZ 100
1061 #endif
1063 static inline abi_long host_to_target_clock_t(long ticks)
1065 #if HOST_HZ == TARGET_HZ
1066 return ticks;
1067 #else
1068 return ((int64_t)ticks * TARGET_HZ) / HOST_HZ;
1069 #endif
1072 static inline abi_long host_to_target_rusage(abi_ulong target_addr,
1073 const struct rusage *rusage)
1075 struct target_rusage *target_rusage;
1077 if (!lock_user_struct(VERIFY_WRITE, target_rusage, target_addr, 0))
1078 return -TARGET_EFAULT;
1079 target_rusage->ru_utime.tv_sec = tswapal(rusage->ru_utime.tv_sec);
1080 target_rusage->ru_utime.tv_usec = tswapal(rusage->ru_utime.tv_usec);
1081 target_rusage->ru_stime.tv_sec = tswapal(rusage->ru_stime.tv_sec);
1082 target_rusage->ru_stime.tv_usec = tswapal(rusage->ru_stime.tv_usec);
1083 target_rusage->ru_maxrss = tswapal(rusage->ru_maxrss);
1084 target_rusage->ru_ixrss = tswapal(rusage->ru_ixrss);
1085 target_rusage->ru_idrss = tswapal(rusage->ru_idrss);
1086 target_rusage->ru_isrss = tswapal(rusage->ru_isrss);
1087 target_rusage->ru_minflt = tswapal(rusage->ru_minflt);
1088 target_rusage->ru_majflt = tswapal(rusage->ru_majflt);
1089 target_rusage->ru_nswap = tswapal(rusage->ru_nswap);
1090 target_rusage->ru_inblock = tswapal(rusage->ru_inblock);
1091 target_rusage->ru_oublock = tswapal(rusage->ru_oublock);
1092 target_rusage->ru_msgsnd = tswapal(rusage->ru_msgsnd);
1093 target_rusage->ru_msgrcv = tswapal(rusage->ru_msgrcv);
1094 target_rusage->ru_nsignals = tswapal(rusage->ru_nsignals);
1095 target_rusage->ru_nvcsw = tswapal(rusage->ru_nvcsw);
1096 target_rusage->ru_nivcsw = tswapal(rusage->ru_nivcsw);
1097 unlock_user_struct(target_rusage, target_addr, 1);
1099 return 0;
1102 #ifdef TARGET_NR_setrlimit
1103 static inline rlim_t target_to_host_rlim(abi_ulong target_rlim)
1105 abi_ulong target_rlim_swap;
1106 rlim_t result;
1108 target_rlim_swap = tswapal(target_rlim);
1109 if (target_rlim_swap == TARGET_RLIM_INFINITY)
1110 return RLIM_INFINITY;
1112 result = target_rlim_swap;
1113 if (target_rlim_swap != (rlim_t)result)
1114 return RLIM_INFINITY;
1116 return result;
1118 #endif
1120 #if defined(TARGET_NR_getrlimit) || defined(TARGET_NR_ugetrlimit)
1121 static inline abi_ulong host_to_target_rlim(rlim_t rlim)
1123 abi_ulong target_rlim_swap;
1124 abi_ulong result;
1126 if (rlim == RLIM_INFINITY || rlim != (abi_long)rlim)
1127 target_rlim_swap = TARGET_RLIM_INFINITY;
1128 else
1129 target_rlim_swap = rlim;
1130 result = tswapal(target_rlim_swap);
1132 return result;
1134 #endif
1136 static inline int target_to_host_resource(int code)
1138 switch (code) {
1139 case TARGET_RLIMIT_AS:
1140 return RLIMIT_AS;
1141 case TARGET_RLIMIT_CORE:
1142 return RLIMIT_CORE;
1143 case TARGET_RLIMIT_CPU:
1144 return RLIMIT_CPU;
1145 case TARGET_RLIMIT_DATA:
1146 return RLIMIT_DATA;
1147 case TARGET_RLIMIT_FSIZE:
1148 return RLIMIT_FSIZE;
1149 case TARGET_RLIMIT_LOCKS:
1150 return RLIMIT_LOCKS;
1151 case TARGET_RLIMIT_MEMLOCK:
1152 return RLIMIT_MEMLOCK;
1153 case TARGET_RLIMIT_MSGQUEUE:
1154 return RLIMIT_MSGQUEUE;
1155 case TARGET_RLIMIT_NICE:
1156 return RLIMIT_NICE;
1157 case TARGET_RLIMIT_NOFILE:
1158 return RLIMIT_NOFILE;
1159 case TARGET_RLIMIT_NPROC:
1160 return RLIMIT_NPROC;
1161 case TARGET_RLIMIT_RSS:
1162 return RLIMIT_RSS;
1163 case TARGET_RLIMIT_RTPRIO:
1164 return RLIMIT_RTPRIO;
1165 case TARGET_RLIMIT_SIGPENDING:
1166 return RLIMIT_SIGPENDING;
1167 case TARGET_RLIMIT_STACK:
1168 return RLIMIT_STACK;
1169 default:
1170 return code;
1174 static inline abi_long copy_from_user_timeval(struct timeval *tv,
1175 abi_ulong target_tv_addr)
1177 struct target_timeval *target_tv;
1179 if (!lock_user_struct(VERIFY_READ, target_tv, target_tv_addr, 1)) {
1180 return -TARGET_EFAULT;
1183 __get_user(tv->tv_sec, &target_tv->tv_sec);
1184 __get_user(tv->tv_usec, &target_tv->tv_usec);
1186 unlock_user_struct(target_tv, target_tv_addr, 0);
1188 return 0;
1191 static inline abi_long copy_to_user_timeval(abi_ulong target_tv_addr,
1192 const struct timeval *tv)
1194 struct target_timeval *target_tv;
1196 if (!lock_user_struct(VERIFY_WRITE, target_tv, target_tv_addr, 0)) {
1197 return -TARGET_EFAULT;
1200 __put_user(tv->tv_sec, &target_tv->tv_sec);
1201 __put_user(tv->tv_usec, &target_tv->tv_usec);
1203 unlock_user_struct(target_tv, target_tv_addr, 1);
1205 return 0;
1208 static inline abi_long copy_to_user_timeval64(abi_ulong target_tv_addr,
1209 const struct timeval *tv)
1211 struct target__kernel_sock_timeval *target_tv;
1213 if (!lock_user_struct(VERIFY_WRITE, target_tv, target_tv_addr, 0)) {
1214 return -TARGET_EFAULT;
1217 __put_user(tv->tv_sec, &target_tv->tv_sec);
1218 __put_user(tv->tv_usec, &target_tv->tv_usec);
1220 unlock_user_struct(target_tv, target_tv_addr, 1);
1222 return 0;
1225 #if defined(TARGET_NR_futex) || \
1226 defined(TARGET_NR_rt_sigtimedwait) || \
1227 defined(TARGET_NR_pselect6) || defined(TARGET_NR_pselect6) || \
1228 defined(TARGET_NR_nanosleep) || defined(TARGET_NR_clock_settime) || \
1229 defined(TARGET_NR_utimensat) || defined(TARGET_NR_mq_timedsend) || \
1230 defined(TARGET_NR_mq_timedreceive)
1231 static inline abi_long target_to_host_timespec(struct timespec *host_ts,
1232 abi_ulong target_addr)
1234 struct target_timespec *target_ts;
1236 if (!lock_user_struct(VERIFY_READ, target_ts, target_addr, 1)) {
1237 return -TARGET_EFAULT;
1239 __get_user(host_ts->tv_sec, &target_ts->tv_sec);
1240 __get_user(host_ts->tv_nsec, &target_ts->tv_nsec);
1241 unlock_user_struct(target_ts, target_addr, 0);
1242 return 0;
1244 #endif
1246 #if defined(TARGET_NR_clock_settime64) || defined(TARGET_NR_futex_time64)
1247 static inline abi_long target_to_host_timespec64(struct timespec *host_ts,
1248 abi_ulong target_addr)
1250 struct target__kernel_timespec *target_ts;
1252 if (!lock_user_struct(VERIFY_READ, target_ts, target_addr, 1)) {
1253 return -TARGET_EFAULT;
1255 __get_user(host_ts->tv_sec, &target_ts->tv_sec);
1256 __get_user(host_ts->tv_nsec, &target_ts->tv_nsec);
1257 unlock_user_struct(target_ts, target_addr, 0);
1258 return 0;
1260 #endif
1262 static inline abi_long host_to_target_timespec(abi_ulong target_addr,
1263 struct timespec *host_ts)
1265 struct target_timespec *target_ts;
1267 if (!lock_user_struct(VERIFY_WRITE, target_ts, target_addr, 0)) {
1268 return -TARGET_EFAULT;
1270 __put_user(host_ts->tv_sec, &target_ts->tv_sec);
1271 __put_user(host_ts->tv_nsec, &target_ts->tv_nsec);
1272 unlock_user_struct(target_ts, target_addr, 1);
1273 return 0;
1276 static inline abi_long host_to_target_timespec64(abi_ulong target_addr,
1277 struct timespec *host_ts)
1279 struct target__kernel_timespec *target_ts;
1281 if (!lock_user_struct(VERIFY_WRITE, target_ts, target_addr, 0)) {
1282 return -TARGET_EFAULT;
1284 __put_user(host_ts->tv_sec, &target_ts->tv_sec);
1285 __put_user(host_ts->tv_nsec, &target_ts->tv_nsec);
1286 unlock_user_struct(target_ts, target_addr, 1);
1287 return 0;
1290 #if defined(TARGET_NR_gettimeofday)
1291 static inline abi_long copy_to_user_timezone(abi_ulong target_tz_addr,
1292 struct timezone *tz)
1294 struct target_timezone *target_tz;
1296 if (!lock_user_struct(VERIFY_WRITE, target_tz, target_tz_addr, 1)) {
1297 return -TARGET_EFAULT;
1300 __put_user(tz->tz_minuteswest, &target_tz->tz_minuteswest);
1301 __put_user(tz->tz_dsttime, &target_tz->tz_dsttime);
1303 unlock_user_struct(target_tz, target_tz_addr, 1);
1305 return 0;
1307 #endif
1309 #if defined(TARGET_NR_settimeofday)
1310 static inline abi_long copy_from_user_timezone(struct timezone *tz,
1311 abi_ulong target_tz_addr)
1313 struct target_timezone *target_tz;
1315 if (!lock_user_struct(VERIFY_READ, target_tz, target_tz_addr, 1)) {
1316 return -TARGET_EFAULT;
1319 __get_user(tz->tz_minuteswest, &target_tz->tz_minuteswest);
1320 __get_user(tz->tz_dsttime, &target_tz->tz_dsttime);
1322 unlock_user_struct(target_tz, target_tz_addr, 0);
1324 return 0;
1326 #endif
1328 #if defined(TARGET_NR_mq_open) && defined(__NR_mq_open)
1329 #include <mqueue.h>
1331 static inline abi_long copy_from_user_mq_attr(struct mq_attr *attr,
1332 abi_ulong target_mq_attr_addr)
1334 struct target_mq_attr *target_mq_attr;
1336 if (!lock_user_struct(VERIFY_READ, target_mq_attr,
1337 target_mq_attr_addr, 1))
1338 return -TARGET_EFAULT;
1340 __get_user(attr->mq_flags, &target_mq_attr->mq_flags);
1341 __get_user(attr->mq_maxmsg, &target_mq_attr->mq_maxmsg);
1342 __get_user(attr->mq_msgsize, &target_mq_attr->mq_msgsize);
1343 __get_user(attr->mq_curmsgs, &target_mq_attr->mq_curmsgs);
1345 unlock_user_struct(target_mq_attr, target_mq_attr_addr, 0);
1347 return 0;
1350 static inline abi_long copy_to_user_mq_attr(abi_ulong target_mq_attr_addr,
1351 const struct mq_attr *attr)
1353 struct target_mq_attr *target_mq_attr;
1355 if (!lock_user_struct(VERIFY_WRITE, target_mq_attr,
1356 target_mq_attr_addr, 0))
1357 return -TARGET_EFAULT;
1359 __put_user(attr->mq_flags, &target_mq_attr->mq_flags);
1360 __put_user(attr->mq_maxmsg, &target_mq_attr->mq_maxmsg);
1361 __put_user(attr->mq_msgsize, &target_mq_attr->mq_msgsize);
1362 __put_user(attr->mq_curmsgs, &target_mq_attr->mq_curmsgs);
1364 unlock_user_struct(target_mq_attr, target_mq_attr_addr, 1);
1366 return 0;
1368 #endif
1370 #if defined(TARGET_NR_select) || defined(TARGET_NR__newselect)
1371 /* do_select() must return target values and target errnos. */
1372 static abi_long do_select(int n,
1373 abi_ulong rfd_addr, abi_ulong wfd_addr,
1374 abi_ulong efd_addr, abi_ulong target_tv_addr)
1376 fd_set rfds, wfds, efds;
1377 fd_set *rfds_ptr, *wfds_ptr, *efds_ptr;
1378 struct timeval tv;
1379 struct timespec ts, *ts_ptr;
1380 abi_long ret;
1382 ret = copy_from_user_fdset_ptr(&rfds, &rfds_ptr, rfd_addr, n);
1383 if (ret) {
1384 return ret;
1386 ret = copy_from_user_fdset_ptr(&wfds, &wfds_ptr, wfd_addr, n);
1387 if (ret) {
1388 return ret;
1390 ret = copy_from_user_fdset_ptr(&efds, &efds_ptr, efd_addr, n);
1391 if (ret) {
1392 return ret;
1395 if (target_tv_addr) {
1396 if (copy_from_user_timeval(&tv, target_tv_addr))
1397 return -TARGET_EFAULT;
1398 ts.tv_sec = tv.tv_sec;
1399 ts.tv_nsec = tv.tv_usec * 1000;
1400 ts_ptr = &ts;
1401 } else {
1402 ts_ptr = NULL;
1405 ret = get_errno(safe_pselect6(n, rfds_ptr, wfds_ptr, efds_ptr,
1406 ts_ptr, NULL));
1408 if (!is_error(ret)) {
1409 if (rfd_addr && copy_to_user_fdset(rfd_addr, &rfds, n))
1410 return -TARGET_EFAULT;
1411 if (wfd_addr && copy_to_user_fdset(wfd_addr, &wfds, n))
1412 return -TARGET_EFAULT;
1413 if (efd_addr && copy_to_user_fdset(efd_addr, &efds, n))
1414 return -TARGET_EFAULT;
1416 if (target_tv_addr) {
1417 tv.tv_sec = ts.tv_sec;
1418 tv.tv_usec = ts.tv_nsec / 1000;
1419 if (copy_to_user_timeval(target_tv_addr, &tv)) {
1420 return -TARGET_EFAULT;
1425 return ret;
1428 #if defined(TARGET_WANT_OLD_SYS_SELECT)
1429 static abi_long do_old_select(abi_ulong arg1)
1431 struct target_sel_arg_struct *sel;
1432 abi_ulong inp, outp, exp, tvp;
1433 long nsel;
1435 if (!lock_user_struct(VERIFY_READ, sel, arg1, 1)) {
1436 return -TARGET_EFAULT;
1439 nsel = tswapal(sel->n);
1440 inp = tswapal(sel->inp);
1441 outp = tswapal(sel->outp);
1442 exp = tswapal(sel->exp);
1443 tvp = tswapal(sel->tvp);
1445 unlock_user_struct(sel, arg1, 0);
1447 return do_select(nsel, inp, outp, exp, tvp);
1449 #endif
1450 #endif
1452 static abi_long do_pipe2(int host_pipe[], int flags)
1454 #ifdef CONFIG_PIPE2
1455 return pipe2(host_pipe, flags);
1456 #else
1457 return -ENOSYS;
1458 #endif
1461 static abi_long do_pipe(void *cpu_env, abi_ulong pipedes,
1462 int flags, int is_pipe2)
1464 int host_pipe[2];
1465 abi_long ret;
1466 ret = flags ? do_pipe2(host_pipe, flags) : pipe(host_pipe);
1468 if (is_error(ret))
1469 return get_errno(ret);
1471 /* Several targets have special calling conventions for the original
1472 pipe syscall, but didn't replicate this into the pipe2 syscall. */
1473 if (!is_pipe2) {
1474 #if defined(TARGET_ALPHA)
1475 ((CPUAlphaState *)cpu_env)->ir[IR_A4] = host_pipe[1];
1476 return host_pipe[0];
1477 #elif defined(TARGET_MIPS)
1478 ((CPUMIPSState*)cpu_env)->active_tc.gpr[3] = host_pipe[1];
1479 return host_pipe[0];
1480 #elif defined(TARGET_SH4)
1481 ((CPUSH4State*)cpu_env)->gregs[1] = host_pipe[1];
1482 return host_pipe[0];
1483 #elif defined(TARGET_SPARC)
1484 ((CPUSPARCState*)cpu_env)->regwptr[1] = host_pipe[1];
1485 return host_pipe[0];
1486 #endif
1489 if (put_user_s32(host_pipe[0], pipedes)
1490 || put_user_s32(host_pipe[1], pipedes + sizeof(host_pipe[0])))
1491 return -TARGET_EFAULT;
1492 return get_errno(ret);
1495 static inline abi_long target_to_host_ip_mreq(struct ip_mreqn *mreqn,
1496 abi_ulong target_addr,
1497 socklen_t len)
1499 struct target_ip_mreqn *target_smreqn;
1501 target_smreqn = lock_user(VERIFY_READ, target_addr, len, 1);
1502 if (!target_smreqn)
1503 return -TARGET_EFAULT;
1504 mreqn->imr_multiaddr.s_addr = target_smreqn->imr_multiaddr.s_addr;
1505 mreqn->imr_address.s_addr = target_smreqn->imr_address.s_addr;
1506 if (len == sizeof(struct target_ip_mreqn))
1507 mreqn->imr_ifindex = tswapal(target_smreqn->imr_ifindex);
1508 unlock_user(target_smreqn, target_addr, 0);
1510 return 0;
1513 static inline abi_long target_to_host_sockaddr(int fd, struct sockaddr *addr,
1514 abi_ulong target_addr,
1515 socklen_t len)
1517 const socklen_t unix_maxlen = sizeof (struct sockaddr_un);
1518 sa_family_t sa_family;
1519 struct target_sockaddr *target_saddr;
1521 if (fd_trans_target_to_host_addr(fd)) {
1522 return fd_trans_target_to_host_addr(fd)(addr, target_addr, len);
1525 target_saddr = lock_user(VERIFY_READ, target_addr, len, 1);
1526 if (!target_saddr)
1527 return -TARGET_EFAULT;
1529 sa_family = tswap16(target_saddr->sa_family);
1531 /* Oops. The caller might send a incomplete sun_path; sun_path
1532 * must be terminated by \0 (see the manual page), but
1533 * unfortunately it is quite common to specify sockaddr_un
1534 * length as "strlen(x->sun_path)" while it should be
1535 * "strlen(...) + 1". We'll fix that here if needed.
1536 * Linux kernel has a similar feature.
1539 if (sa_family == AF_UNIX) {
1540 if (len < unix_maxlen && len > 0) {
1541 char *cp = (char*)target_saddr;
1543 if ( cp[len-1] && !cp[len] )
1544 len++;
1546 if (len > unix_maxlen)
1547 len = unix_maxlen;
1550 memcpy(addr, target_saddr, len);
1551 addr->sa_family = sa_family;
1552 if (sa_family == AF_NETLINK) {
1553 struct sockaddr_nl *nladdr;
1555 nladdr = (struct sockaddr_nl *)addr;
1556 nladdr->nl_pid = tswap32(nladdr->nl_pid);
1557 nladdr->nl_groups = tswap32(nladdr->nl_groups);
1558 } else if (sa_family == AF_PACKET) {
1559 struct target_sockaddr_ll *lladdr;
1561 lladdr = (struct target_sockaddr_ll *)addr;
1562 lladdr->sll_ifindex = tswap32(lladdr->sll_ifindex);
1563 lladdr->sll_hatype = tswap16(lladdr->sll_hatype);
1565 unlock_user(target_saddr, target_addr, 0);
1567 return 0;
1570 static inline abi_long host_to_target_sockaddr(abi_ulong target_addr,
1571 struct sockaddr *addr,
1572 socklen_t len)
1574 struct target_sockaddr *target_saddr;
1576 if (len == 0) {
1577 return 0;
1579 assert(addr);
1581 target_saddr = lock_user(VERIFY_WRITE, target_addr, len, 0);
1582 if (!target_saddr)
1583 return -TARGET_EFAULT;
1584 memcpy(target_saddr, addr, len);
1585 if (len >= offsetof(struct target_sockaddr, sa_family) +
1586 sizeof(target_saddr->sa_family)) {
1587 target_saddr->sa_family = tswap16(addr->sa_family);
1589 if (addr->sa_family == AF_NETLINK &&
1590 len >= sizeof(struct target_sockaddr_nl)) {
1591 struct target_sockaddr_nl *target_nl =
1592 (struct target_sockaddr_nl *)target_saddr;
1593 target_nl->nl_pid = tswap32(target_nl->nl_pid);
1594 target_nl->nl_groups = tswap32(target_nl->nl_groups);
1595 } else if (addr->sa_family == AF_PACKET) {
1596 struct sockaddr_ll *target_ll = (struct sockaddr_ll *)target_saddr;
1597 target_ll->sll_ifindex = tswap32(target_ll->sll_ifindex);
1598 target_ll->sll_hatype = tswap16(target_ll->sll_hatype);
1599 } else if (addr->sa_family == AF_INET6 &&
1600 len >= sizeof(struct target_sockaddr_in6)) {
1601 struct target_sockaddr_in6 *target_in6 =
1602 (struct target_sockaddr_in6 *)target_saddr;
1603 target_in6->sin6_scope_id = tswap16(target_in6->sin6_scope_id);
1605 unlock_user(target_saddr, target_addr, len);
1607 return 0;
1610 static inline abi_long target_to_host_cmsg(struct msghdr *msgh,
1611 struct target_msghdr *target_msgh)
1613 struct cmsghdr *cmsg = CMSG_FIRSTHDR(msgh);
1614 abi_long msg_controllen;
1615 abi_ulong target_cmsg_addr;
1616 struct target_cmsghdr *target_cmsg, *target_cmsg_start;
1617 socklen_t space = 0;
1619 msg_controllen = tswapal(target_msgh->msg_controllen);
1620 if (msg_controllen < sizeof (struct target_cmsghdr))
1621 goto the_end;
1622 target_cmsg_addr = tswapal(target_msgh->msg_control);
1623 target_cmsg = lock_user(VERIFY_READ, target_cmsg_addr, msg_controllen, 1);
1624 target_cmsg_start = target_cmsg;
1625 if (!target_cmsg)
1626 return -TARGET_EFAULT;
1628 while (cmsg && target_cmsg) {
1629 void *data = CMSG_DATA(cmsg);
1630 void *target_data = TARGET_CMSG_DATA(target_cmsg);
1632 int len = tswapal(target_cmsg->cmsg_len)
1633 - sizeof(struct target_cmsghdr);
1635 space += CMSG_SPACE(len);
1636 if (space > msgh->msg_controllen) {
1637 space -= CMSG_SPACE(len);
1638 /* This is a QEMU bug, since we allocated the payload
1639 * area ourselves (unlike overflow in host-to-target
1640 * conversion, which is just the guest giving us a buffer
1641 * that's too small). It can't happen for the payload types
1642 * we currently support; if it becomes an issue in future
1643 * we would need to improve our allocation strategy to
1644 * something more intelligent than "twice the size of the
1645 * target buffer we're reading from".
1647 qemu_log_mask(LOG_UNIMP,
1648 ("Unsupported ancillary data %d/%d: "
1649 "unhandled msg size\n"),
1650 tswap32(target_cmsg->cmsg_level),
1651 tswap32(target_cmsg->cmsg_type));
1652 break;
1655 if (tswap32(target_cmsg->cmsg_level) == TARGET_SOL_SOCKET) {
1656 cmsg->cmsg_level = SOL_SOCKET;
1657 } else {
1658 cmsg->cmsg_level = tswap32(target_cmsg->cmsg_level);
1660 cmsg->cmsg_type = tswap32(target_cmsg->cmsg_type);
1661 cmsg->cmsg_len = CMSG_LEN(len);
1663 if (cmsg->cmsg_level == SOL_SOCKET && cmsg->cmsg_type == SCM_RIGHTS) {
1664 int *fd = (int *)data;
1665 int *target_fd = (int *)target_data;
1666 int i, numfds = len / sizeof(int);
1668 for (i = 0; i < numfds; i++) {
1669 __get_user(fd[i], target_fd + i);
1671 } else if (cmsg->cmsg_level == SOL_SOCKET
1672 && cmsg->cmsg_type == SCM_CREDENTIALS) {
1673 struct ucred *cred = (struct ucred *)data;
1674 struct target_ucred *target_cred =
1675 (struct target_ucred *)target_data;
1677 __get_user(cred->pid, &target_cred->pid);
1678 __get_user(cred->uid, &target_cred->uid);
1679 __get_user(cred->gid, &target_cred->gid);
1680 } else {
1681 qemu_log_mask(LOG_UNIMP, "Unsupported ancillary data: %d/%d\n",
1682 cmsg->cmsg_level, cmsg->cmsg_type);
1683 memcpy(data, target_data, len);
1686 cmsg = CMSG_NXTHDR(msgh, cmsg);
1687 target_cmsg = TARGET_CMSG_NXTHDR(target_msgh, target_cmsg,
1688 target_cmsg_start);
1690 unlock_user(target_cmsg, target_cmsg_addr, 0);
1691 the_end:
1692 msgh->msg_controllen = space;
1693 return 0;
1696 static inline abi_long host_to_target_cmsg(struct target_msghdr *target_msgh,
1697 struct msghdr *msgh)
1699 struct cmsghdr *cmsg = CMSG_FIRSTHDR(msgh);
1700 abi_long msg_controllen;
1701 abi_ulong target_cmsg_addr;
1702 struct target_cmsghdr *target_cmsg, *target_cmsg_start;
1703 socklen_t space = 0;
1705 msg_controllen = tswapal(target_msgh->msg_controllen);
1706 if (msg_controllen < sizeof (struct target_cmsghdr))
1707 goto the_end;
1708 target_cmsg_addr = tswapal(target_msgh->msg_control);
1709 target_cmsg = lock_user(VERIFY_WRITE, target_cmsg_addr, msg_controllen, 0);
1710 target_cmsg_start = target_cmsg;
1711 if (!target_cmsg)
1712 return -TARGET_EFAULT;
1714 while (cmsg && target_cmsg) {
1715 void *data = CMSG_DATA(cmsg);
1716 void *target_data = TARGET_CMSG_DATA(target_cmsg);
1718 int len = cmsg->cmsg_len - sizeof(struct cmsghdr);
1719 int tgt_len, tgt_space;
1721 /* We never copy a half-header but may copy half-data;
1722 * this is Linux's behaviour in put_cmsg(). Note that
1723 * truncation here is a guest problem (which we report
1724 * to the guest via the CTRUNC bit), unlike truncation
1725 * in target_to_host_cmsg, which is a QEMU bug.
1727 if (msg_controllen < sizeof(struct target_cmsghdr)) {
1728 target_msgh->msg_flags |= tswap32(MSG_CTRUNC);
1729 break;
1732 if (cmsg->cmsg_level == SOL_SOCKET) {
1733 target_cmsg->cmsg_level = tswap32(TARGET_SOL_SOCKET);
1734 } else {
1735 target_cmsg->cmsg_level = tswap32(cmsg->cmsg_level);
1737 target_cmsg->cmsg_type = tswap32(cmsg->cmsg_type);
1739 /* Payload types which need a different size of payload on
1740 * the target must adjust tgt_len here.
1742 tgt_len = len;
1743 switch (cmsg->cmsg_level) {
1744 case SOL_SOCKET:
1745 switch (cmsg->cmsg_type) {
1746 case SO_TIMESTAMP:
1747 tgt_len = sizeof(struct target_timeval);
1748 break;
1749 default:
1750 break;
1752 break;
1753 default:
1754 break;
1757 if (msg_controllen < TARGET_CMSG_LEN(tgt_len)) {
1758 target_msgh->msg_flags |= tswap32(MSG_CTRUNC);
1759 tgt_len = msg_controllen - sizeof(struct target_cmsghdr);
1762 /* We must now copy-and-convert len bytes of payload
1763 * into tgt_len bytes of destination space. Bear in mind
1764 * that in both source and destination we may be dealing
1765 * with a truncated value!
1767 switch (cmsg->cmsg_level) {
1768 case SOL_SOCKET:
1769 switch (cmsg->cmsg_type) {
1770 case SCM_RIGHTS:
1772 int *fd = (int *)data;
1773 int *target_fd = (int *)target_data;
1774 int i, numfds = tgt_len / sizeof(int);
1776 for (i = 0; i < numfds; i++) {
1777 __put_user(fd[i], target_fd + i);
1779 break;
1781 case SO_TIMESTAMP:
1783 struct timeval *tv = (struct timeval *)data;
1784 struct target_timeval *target_tv =
1785 (struct target_timeval *)target_data;
1787 if (len != sizeof(struct timeval) ||
1788 tgt_len != sizeof(struct target_timeval)) {
1789 goto unimplemented;
1792 /* copy struct timeval to target */
1793 __put_user(tv->tv_sec, &target_tv->tv_sec);
1794 __put_user(tv->tv_usec, &target_tv->tv_usec);
1795 break;
1797 case SCM_CREDENTIALS:
1799 struct ucred *cred = (struct ucred *)data;
1800 struct target_ucred *target_cred =
1801 (struct target_ucred *)target_data;
1803 __put_user(cred->pid, &target_cred->pid);
1804 __put_user(cred->uid, &target_cred->uid);
1805 __put_user(cred->gid, &target_cred->gid);
1806 break;
1808 default:
1809 goto unimplemented;
1811 break;
1813 case SOL_IP:
1814 switch (cmsg->cmsg_type) {
1815 case IP_TTL:
1817 uint32_t *v = (uint32_t *)data;
1818 uint32_t *t_int = (uint32_t *)target_data;
1820 if (len != sizeof(uint32_t) ||
1821 tgt_len != sizeof(uint32_t)) {
1822 goto unimplemented;
1824 __put_user(*v, t_int);
1825 break;
1827 case IP_RECVERR:
1829 struct errhdr_t {
1830 struct sock_extended_err ee;
1831 struct sockaddr_in offender;
1833 struct errhdr_t *errh = (struct errhdr_t *)data;
1834 struct errhdr_t *target_errh =
1835 (struct errhdr_t *)target_data;
1837 if (len != sizeof(struct errhdr_t) ||
1838 tgt_len != sizeof(struct errhdr_t)) {
1839 goto unimplemented;
1841 __put_user(errh->ee.ee_errno, &target_errh->ee.ee_errno);
1842 __put_user(errh->ee.ee_origin, &target_errh->ee.ee_origin);
1843 __put_user(errh->ee.ee_type, &target_errh->ee.ee_type);
1844 __put_user(errh->ee.ee_code, &target_errh->ee.ee_code);
1845 __put_user(errh->ee.ee_pad, &target_errh->ee.ee_pad);
1846 __put_user(errh->ee.ee_info, &target_errh->ee.ee_info);
1847 __put_user(errh->ee.ee_data, &target_errh->ee.ee_data);
1848 host_to_target_sockaddr((unsigned long) &target_errh->offender,
1849 (void *) &errh->offender, sizeof(errh->offender));
1850 break;
1852 default:
1853 goto unimplemented;
1855 break;
1857 case SOL_IPV6:
1858 switch (cmsg->cmsg_type) {
1859 case IPV6_HOPLIMIT:
1861 uint32_t *v = (uint32_t *)data;
1862 uint32_t *t_int = (uint32_t *)target_data;
1864 if (len != sizeof(uint32_t) ||
1865 tgt_len != sizeof(uint32_t)) {
1866 goto unimplemented;
1868 __put_user(*v, t_int);
1869 break;
1871 case IPV6_RECVERR:
1873 struct errhdr6_t {
1874 struct sock_extended_err ee;
1875 struct sockaddr_in6 offender;
1877 struct errhdr6_t *errh = (struct errhdr6_t *)data;
1878 struct errhdr6_t *target_errh =
1879 (struct errhdr6_t *)target_data;
1881 if (len != sizeof(struct errhdr6_t) ||
1882 tgt_len != sizeof(struct errhdr6_t)) {
1883 goto unimplemented;
1885 __put_user(errh->ee.ee_errno, &target_errh->ee.ee_errno);
1886 __put_user(errh->ee.ee_origin, &target_errh->ee.ee_origin);
1887 __put_user(errh->ee.ee_type, &target_errh->ee.ee_type);
1888 __put_user(errh->ee.ee_code, &target_errh->ee.ee_code);
1889 __put_user(errh->ee.ee_pad, &target_errh->ee.ee_pad);
1890 __put_user(errh->ee.ee_info, &target_errh->ee.ee_info);
1891 __put_user(errh->ee.ee_data, &target_errh->ee.ee_data);
1892 host_to_target_sockaddr((unsigned long) &target_errh->offender,
1893 (void *) &errh->offender, sizeof(errh->offender));
1894 break;
1896 default:
1897 goto unimplemented;
1899 break;
1901 default:
1902 unimplemented:
1903 qemu_log_mask(LOG_UNIMP, "Unsupported ancillary data: %d/%d\n",
1904 cmsg->cmsg_level, cmsg->cmsg_type);
1905 memcpy(target_data, data, MIN(len, tgt_len));
1906 if (tgt_len > len) {
1907 memset(target_data + len, 0, tgt_len - len);
1911 target_cmsg->cmsg_len = tswapal(TARGET_CMSG_LEN(tgt_len));
1912 tgt_space = TARGET_CMSG_SPACE(tgt_len);
1913 if (msg_controllen < tgt_space) {
1914 tgt_space = msg_controllen;
1916 msg_controllen -= tgt_space;
1917 space += tgt_space;
1918 cmsg = CMSG_NXTHDR(msgh, cmsg);
1919 target_cmsg = TARGET_CMSG_NXTHDR(target_msgh, target_cmsg,
1920 target_cmsg_start);
1922 unlock_user(target_cmsg, target_cmsg_addr, space);
1923 the_end:
1924 target_msgh->msg_controllen = tswapal(space);
1925 return 0;
1928 /* do_setsockopt() Must return target values and target errnos. */
1929 static abi_long do_setsockopt(int sockfd, int level, int optname,
1930 abi_ulong optval_addr, socklen_t optlen)
1932 abi_long ret;
1933 int val;
1934 struct ip_mreqn *ip_mreq;
1935 struct ip_mreq_source *ip_mreq_source;
1937 switch(level) {
1938 case SOL_TCP:
1939 /* TCP options all take an 'int' value. */
1940 if (optlen < sizeof(uint32_t))
1941 return -TARGET_EINVAL;
1943 if (get_user_u32(val, optval_addr))
1944 return -TARGET_EFAULT;
1945 ret = get_errno(setsockopt(sockfd, level, optname, &val, sizeof(val)));
1946 break;
1947 case SOL_IP:
1948 switch(optname) {
1949 case IP_TOS:
1950 case IP_TTL:
1951 case IP_HDRINCL:
1952 case IP_ROUTER_ALERT:
1953 case IP_RECVOPTS:
1954 case IP_RETOPTS:
1955 case IP_PKTINFO:
1956 case IP_MTU_DISCOVER:
1957 case IP_RECVERR:
1958 case IP_RECVTTL:
1959 case IP_RECVTOS:
1960 #ifdef IP_FREEBIND
1961 case IP_FREEBIND:
1962 #endif
1963 case IP_MULTICAST_TTL:
1964 case IP_MULTICAST_LOOP:
1965 val = 0;
1966 if (optlen >= sizeof(uint32_t)) {
1967 if (get_user_u32(val, optval_addr))
1968 return -TARGET_EFAULT;
1969 } else if (optlen >= 1) {
1970 if (get_user_u8(val, optval_addr))
1971 return -TARGET_EFAULT;
1973 ret = get_errno(setsockopt(sockfd, level, optname, &val, sizeof(val)));
1974 break;
1975 case IP_ADD_MEMBERSHIP:
1976 case IP_DROP_MEMBERSHIP:
1977 if (optlen < sizeof (struct target_ip_mreq) ||
1978 optlen > sizeof (struct target_ip_mreqn))
1979 return -TARGET_EINVAL;
1981 ip_mreq = (struct ip_mreqn *) alloca(optlen);
1982 target_to_host_ip_mreq(ip_mreq, optval_addr, optlen);
1983 ret = get_errno(setsockopt(sockfd, level, optname, ip_mreq, optlen));
1984 break;
1986 case IP_BLOCK_SOURCE:
1987 case IP_UNBLOCK_SOURCE:
1988 case IP_ADD_SOURCE_MEMBERSHIP:
1989 case IP_DROP_SOURCE_MEMBERSHIP:
1990 if (optlen != sizeof (struct target_ip_mreq_source))
1991 return -TARGET_EINVAL;
1993 ip_mreq_source = lock_user(VERIFY_READ, optval_addr, optlen, 1);
1994 ret = get_errno(setsockopt(sockfd, level, optname, ip_mreq_source, optlen));
1995 unlock_user (ip_mreq_source, optval_addr, 0);
1996 break;
1998 default:
1999 goto unimplemented;
2001 break;
2002 case SOL_IPV6:
2003 switch (optname) {
2004 case IPV6_MTU_DISCOVER:
2005 case IPV6_MTU:
2006 case IPV6_V6ONLY:
2007 case IPV6_RECVPKTINFO:
2008 case IPV6_UNICAST_HOPS:
2009 case IPV6_MULTICAST_HOPS:
2010 case IPV6_MULTICAST_LOOP:
2011 case IPV6_RECVERR:
2012 case IPV6_RECVHOPLIMIT:
2013 case IPV6_2292HOPLIMIT:
2014 case IPV6_CHECKSUM:
2015 case IPV6_ADDRFORM:
2016 case IPV6_2292PKTINFO:
2017 case IPV6_RECVTCLASS:
2018 case IPV6_RECVRTHDR:
2019 case IPV6_2292RTHDR:
2020 case IPV6_RECVHOPOPTS:
2021 case IPV6_2292HOPOPTS:
2022 case IPV6_RECVDSTOPTS:
2023 case IPV6_2292DSTOPTS:
2024 case IPV6_TCLASS:
2025 #ifdef IPV6_RECVPATHMTU
2026 case IPV6_RECVPATHMTU:
2027 #endif
2028 #ifdef IPV6_TRANSPARENT
2029 case IPV6_TRANSPARENT:
2030 #endif
2031 #ifdef IPV6_FREEBIND
2032 case IPV6_FREEBIND:
2033 #endif
2034 #ifdef IPV6_RECVORIGDSTADDR
2035 case IPV6_RECVORIGDSTADDR:
2036 #endif
2037 val = 0;
2038 if (optlen < sizeof(uint32_t)) {
2039 return -TARGET_EINVAL;
2041 if (get_user_u32(val, optval_addr)) {
2042 return -TARGET_EFAULT;
2044 ret = get_errno(setsockopt(sockfd, level, optname,
2045 &val, sizeof(val)));
2046 break;
2047 case IPV6_PKTINFO:
2049 struct in6_pktinfo pki;
2051 if (optlen < sizeof(pki)) {
2052 return -TARGET_EINVAL;
2055 if (copy_from_user(&pki, optval_addr, sizeof(pki))) {
2056 return -TARGET_EFAULT;
2059 pki.ipi6_ifindex = tswap32(pki.ipi6_ifindex);
2061 ret = get_errno(setsockopt(sockfd, level, optname,
2062 &pki, sizeof(pki)));
2063 break;
2065 case IPV6_ADD_MEMBERSHIP:
2066 case IPV6_DROP_MEMBERSHIP:
2068 struct ipv6_mreq ipv6mreq;
2070 if (optlen < sizeof(ipv6mreq)) {
2071 return -TARGET_EINVAL;
2074 if (copy_from_user(&ipv6mreq, optval_addr, sizeof(ipv6mreq))) {
2075 return -TARGET_EFAULT;
2078 ipv6mreq.ipv6mr_interface = tswap32(ipv6mreq.ipv6mr_interface);
2080 ret = get_errno(setsockopt(sockfd, level, optname,
2081 &ipv6mreq, sizeof(ipv6mreq)));
2082 break;
2084 default:
2085 goto unimplemented;
2087 break;
2088 case SOL_ICMPV6:
2089 switch (optname) {
2090 case ICMPV6_FILTER:
2092 struct icmp6_filter icmp6f;
2094 if (optlen > sizeof(icmp6f)) {
2095 optlen = sizeof(icmp6f);
2098 if (copy_from_user(&icmp6f, optval_addr, optlen)) {
2099 return -TARGET_EFAULT;
2102 for (val = 0; val < 8; val++) {
2103 icmp6f.data[val] = tswap32(icmp6f.data[val]);
2106 ret = get_errno(setsockopt(sockfd, level, optname,
2107 &icmp6f, optlen));
2108 break;
2110 default:
2111 goto unimplemented;
2113 break;
2114 case SOL_RAW:
2115 switch (optname) {
2116 case ICMP_FILTER:
2117 case IPV6_CHECKSUM:
2118 /* those take an u32 value */
2119 if (optlen < sizeof(uint32_t)) {
2120 return -TARGET_EINVAL;
2123 if (get_user_u32(val, optval_addr)) {
2124 return -TARGET_EFAULT;
2126 ret = get_errno(setsockopt(sockfd, level, optname,
2127 &val, sizeof(val)));
2128 break;
2130 default:
2131 goto unimplemented;
2133 break;
2134 #if defined(SOL_ALG) && defined(ALG_SET_KEY) && defined(ALG_SET_AEAD_AUTHSIZE)
2135 case SOL_ALG:
2136 switch (optname) {
2137 case ALG_SET_KEY:
2139 char *alg_key = g_malloc(optlen);
2141 if (!alg_key) {
2142 return -TARGET_ENOMEM;
2144 if (copy_from_user(alg_key, optval_addr, optlen)) {
2145 g_free(alg_key);
2146 return -TARGET_EFAULT;
2148 ret = get_errno(setsockopt(sockfd, level, optname,
2149 alg_key, optlen));
2150 g_free(alg_key);
2151 break;
2153 case ALG_SET_AEAD_AUTHSIZE:
2155 ret = get_errno(setsockopt(sockfd, level, optname,
2156 NULL, optlen));
2157 break;
2159 default:
2160 goto unimplemented;
2162 break;
2163 #endif
2164 case TARGET_SOL_SOCKET:
2165 switch (optname) {
2166 case TARGET_SO_RCVTIMEO:
2168 struct timeval tv;
2170 optname = SO_RCVTIMEO;
2172 set_timeout:
2173 if (optlen != sizeof(struct target_timeval)) {
2174 return -TARGET_EINVAL;
2177 if (copy_from_user_timeval(&tv, optval_addr)) {
2178 return -TARGET_EFAULT;
2181 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname,
2182 &tv, sizeof(tv)));
2183 return ret;
2185 case TARGET_SO_SNDTIMEO:
2186 optname = SO_SNDTIMEO;
2187 goto set_timeout;
2188 case TARGET_SO_ATTACH_FILTER:
2190 struct target_sock_fprog *tfprog;
2191 struct target_sock_filter *tfilter;
2192 struct sock_fprog fprog;
2193 struct sock_filter *filter;
2194 int i;
2196 if (optlen != sizeof(*tfprog)) {
2197 return -TARGET_EINVAL;
2199 if (!lock_user_struct(VERIFY_READ, tfprog, optval_addr, 0)) {
2200 return -TARGET_EFAULT;
2202 if (!lock_user_struct(VERIFY_READ, tfilter,
2203 tswapal(tfprog->filter), 0)) {
2204 unlock_user_struct(tfprog, optval_addr, 1);
2205 return -TARGET_EFAULT;
2208 fprog.len = tswap16(tfprog->len);
2209 filter = g_try_new(struct sock_filter, fprog.len);
2210 if (filter == NULL) {
2211 unlock_user_struct(tfilter, tfprog->filter, 1);
2212 unlock_user_struct(tfprog, optval_addr, 1);
2213 return -TARGET_ENOMEM;
2215 for (i = 0; i < fprog.len; i++) {
2216 filter[i].code = tswap16(tfilter[i].code);
2217 filter[i].jt = tfilter[i].jt;
2218 filter[i].jf = tfilter[i].jf;
2219 filter[i].k = tswap32(tfilter[i].k);
2221 fprog.filter = filter;
2223 ret = get_errno(setsockopt(sockfd, SOL_SOCKET,
2224 SO_ATTACH_FILTER, &fprog, sizeof(fprog)));
2225 g_free(filter);
2227 unlock_user_struct(tfilter, tfprog->filter, 1);
2228 unlock_user_struct(tfprog, optval_addr, 1);
2229 return ret;
2231 case TARGET_SO_BINDTODEVICE:
2233 char *dev_ifname, *addr_ifname;
2235 if (optlen > IFNAMSIZ - 1) {
2236 optlen = IFNAMSIZ - 1;
2238 dev_ifname = lock_user(VERIFY_READ, optval_addr, optlen, 1);
2239 if (!dev_ifname) {
2240 return -TARGET_EFAULT;
2242 optname = SO_BINDTODEVICE;
2243 addr_ifname = alloca(IFNAMSIZ);
2244 memcpy(addr_ifname, dev_ifname, optlen);
2245 addr_ifname[optlen] = 0;
2246 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname,
2247 addr_ifname, optlen));
2248 unlock_user (dev_ifname, optval_addr, 0);
2249 return ret;
2251 case TARGET_SO_LINGER:
2253 struct linger lg;
2254 struct target_linger *tlg;
2256 if (optlen != sizeof(struct target_linger)) {
2257 return -TARGET_EINVAL;
2259 if (!lock_user_struct(VERIFY_READ, tlg, optval_addr, 1)) {
2260 return -TARGET_EFAULT;
2262 __get_user(lg.l_onoff, &tlg->l_onoff);
2263 __get_user(lg.l_linger, &tlg->l_linger);
2264 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, SO_LINGER,
2265 &lg, sizeof(lg)));
2266 unlock_user_struct(tlg, optval_addr, 0);
2267 return ret;
2269 /* Options with 'int' argument. */
2270 case TARGET_SO_DEBUG:
2271 optname = SO_DEBUG;
2272 break;
2273 case TARGET_SO_REUSEADDR:
2274 optname = SO_REUSEADDR;
2275 break;
2276 #ifdef SO_REUSEPORT
2277 case TARGET_SO_REUSEPORT:
2278 optname = SO_REUSEPORT;
2279 break;
2280 #endif
2281 case TARGET_SO_TYPE:
2282 optname = SO_TYPE;
2283 break;
2284 case TARGET_SO_ERROR:
2285 optname = SO_ERROR;
2286 break;
2287 case TARGET_SO_DONTROUTE:
2288 optname = SO_DONTROUTE;
2289 break;
2290 case TARGET_SO_BROADCAST:
2291 optname = SO_BROADCAST;
2292 break;
2293 case TARGET_SO_SNDBUF:
2294 optname = SO_SNDBUF;
2295 break;
2296 case TARGET_SO_SNDBUFFORCE:
2297 optname = SO_SNDBUFFORCE;
2298 break;
2299 case TARGET_SO_RCVBUF:
2300 optname = SO_RCVBUF;
2301 break;
2302 case TARGET_SO_RCVBUFFORCE:
2303 optname = SO_RCVBUFFORCE;
2304 break;
2305 case TARGET_SO_KEEPALIVE:
2306 optname = SO_KEEPALIVE;
2307 break;
2308 case TARGET_SO_OOBINLINE:
2309 optname = SO_OOBINLINE;
2310 break;
2311 case TARGET_SO_NO_CHECK:
2312 optname = SO_NO_CHECK;
2313 break;
2314 case TARGET_SO_PRIORITY:
2315 optname = SO_PRIORITY;
2316 break;
2317 #ifdef SO_BSDCOMPAT
2318 case TARGET_SO_BSDCOMPAT:
2319 optname = SO_BSDCOMPAT;
2320 break;
2321 #endif
2322 case TARGET_SO_PASSCRED:
2323 optname = SO_PASSCRED;
2324 break;
2325 case TARGET_SO_PASSSEC:
2326 optname = SO_PASSSEC;
2327 break;
2328 case TARGET_SO_TIMESTAMP:
2329 optname = SO_TIMESTAMP;
2330 break;
2331 case TARGET_SO_RCVLOWAT:
2332 optname = SO_RCVLOWAT;
2333 break;
2334 default:
2335 goto unimplemented;
2337 if (optlen < sizeof(uint32_t))
2338 return -TARGET_EINVAL;
2340 if (get_user_u32(val, optval_addr))
2341 return -TARGET_EFAULT;
2342 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname, &val, sizeof(val)));
2343 break;
2344 #ifdef SOL_NETLINK
2345 case SOL_NETLINK:
2346 switch (optname) {
2347 case NETLINK_PKTINFO:
2348 case NETLINK_ADD_MEMBERSHIP:
2349 case NETLINK_DROP_MEMBERSHIP:
2350 case NETLINK_BROADCAST_ERROR:
2351 case NETLINK_NO_ENOBUFS:
2352 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0)
2353 case NETLINK_LISTEN_ALL_NSID:
2354 case NETLINK_CAP_ACK:
2355 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0) */
2356 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0)
2357 case NETLINK_EXT_ACK:
2358 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0) */
2359 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 20, 0)
2360 case NETLINK_GET_STRICT_CHK:
2361 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0) */
2362 break;
2363 default:
2364 goto unimplemented;
2366 val = 0;
2367 if (optlen < sizeof(uint32_t)) {
2368 return -TARGET_EINVAL;
2370 if (get_user_u32(val, optval_addr)) {
2371 return -TARGET_EFAULT;
2373 ret = get_errno(setsockopt(sockfd, SOL_NETLINK, optname, &val,
2374 sizeof(val)));
2375 break;
2376 #endif /* SOL_NETLINK */
2377 default:
2378 unimplemented:
2379 qemu_log_mask(LOG_UNIMP, "Unsupported setsockopt level=%d optname=%d\n",
2380 level, optname);
2381 ret = -TARGET_ENOPROTOOPT;
2383 return ret;
2386 /* do_getsockopt() Must return target values and target errnos. */
2387 static abi_long do_getsockopt(int sockfd, int level, int optname,
2388 abi_ulong optval_addr, abi_ulong optlen)
2390 abi_long ret;
2391 int len, val;
2392 socklen_t lv;
2394 switch(level) {
2395 case TARGET_SOL_SOCKET:
2396 level = SOL_SOCKET;
2397 switch (optname) {
2398 /* These don't just return a single integer */
2399 case TARGET_SO_PEERNAME:
2400 goto unimplemented;
2401 case TARGET_SO_RCVTIMEO: {
2402 struct timeval tv;
2403 socklen_t tvlen;
2405 optname = SO_RCVTIMEO;
2407 get_timeout:
2408 if (get_user_u32(len, optlen)) {
2409 return -TARGET_EFAULT;
2411 if (len < 0) {
2412 return -TARGET_EINVAL;
2415 tvlen = sizeof(tv);
2416 ret = get_errno(getsockopt(sockfd, level, optname,
2417 &tv, &tvlen));
2418 if (ret < 0) {
2419 return ret;
2421 if (len > sizeof(struct target_timeval)) {
2422 len = sizeof(struct target_timeval);
2424 if (copy_to_user_timeval(optval_addr, &tv)) {
2425 return -TARGET_EFAULT;
2427 if (put_user_u32(len, optlen)) {
2428 return -TARGET_EFAULT;
2430 break;
2432 case TARGET_SO_SNDTIMEO:
2433 optname = SO_SNDTIMEO;
2434 goto get_timeout;
2435 case TARGET_SO_PEERCRED: {
2436 struct ucred cr;
2437 socklen_t crlen;
2438 struct target_ucred *tcr;
2440 if (get_user_u32(len, optlen)) {
2441 return -TARGET_EFAULT;
2443 if (len < 0) {
2444 return -TARGET_EINVAL;
2447 crlen = sizeof(cr);
2448 ret = get_errno(getsockopt(sockfd, level, SO_PEERCRED,
2449 &cr, &crlen));
2450 if (ret < 0) {
2451 return ret;
2453 if (len > crlen) {
2454 len = crlen;
2456 if (!lock_user_struct(VERIFY_WRITE, tcr, optval_addr, 0)) {
2457 return -TARGET_EFAULT;
2459 __put_user(cr.pid, &tcr->pid);
2460 __put_user(cr.uid, &tcr->uid);
2461 __put_user(cr.gid, &tcr->gid);
2462 unlock_user_struct(tcr, optval_addr, 1);
2463 if (put_user_u32(len, optlen)) {
2464 return -TARGET_EFAULT;
2466 break;
2468 case TARGET_SO_PEERSEC: {
2469 char *name;
2471 if (get_user_u32(len, optlen)) {
2472 return -TARGET_EFAULT;
2474 if (len < 0) {
2475 return -TARGET_EINVAL;
2477 name = lock_user(VERIFY_WRITE, optval_addr, len, 0);
2478 if (!name) {
2479 return -TARGET_EFAULT;
2481 lv = len;
2482 ret = get_errno(getsockopt(sockfd, level, SO_PEERSEC,
2483 name, &lv));
2484 if (put_user_u32(lv, optlen)) {
2485 ret = -TARGET_EFAULT;
2487 unlock_user(name, optval_addr, lv);
2488 break;
2490 case TARGET_SO_LINGER:
2492 struct linger lg;
2493 socklen_t lglen;
2494 struct target_linger *tlg;
2496 if (get_user_u32(len, optlen)) {
2497 return -TARGET_EFAULT;
2499 if (len < 0) {
2500 return -TARGET_EINVAL;
2503 lglen = sizeof(lg);
2504 ret = get_errno(getsockopt(sockfd, level, SO_LINGER,
2505 &lg, &lglen));
2506 if (ret < 0) {
2507 return ret;
2509 if (len > lglen) {
2510 len = lglen;
2512 if (!lock_user_struct(VERIFY_WRITE, tlg, optval_addr, 0)) {
2513 return -TARGET_EFAULT;
2515 __put_user(lg.l_onoff, &tlg->l_onoff);
2516 __put_user(lg.l_linger, &tlg->l_linger);
2517 unlock_user_struct(tlg, optval_addr, 1);
2518 if (put_user_u32(len, optlen)) {
2519 return -TARGET_EFAULT;
2521 break;
2523 /* Options with 'int' argument. */
2524 case TARGET_SO_DEBUG:
2525 optname = SO_DEBUG;
2526 goto int_case;
2527 case TARGET_SO_REUSEADDR:
2528 optname = SO_REUSEADDR;
2529 goto int_case;
2530 #ifdef SO_REUSEPORT
2531 case TARGET_SO_REUSEPORT:
2532 optname = SO_REUSEPORT;
2533 goto int_case;
2534 #endif
2535 case TARGET_SO_TYPE:
2536 optname = SO_TYPE;
2537 goto int_case;
2538 case TARGET_SO_ERROR:
2539 optname = SO_ERROR;
2540 goto int_case;
2541 case TARGET_SO_DONTROUTE:
2542 optname = SO_DONTROUTE;
2543 goto int_case;
2544 case TARGET_SO_BROADCAST:
2545 optname = SO_BROADCAST;
2546 goto int_case;
2547 case TARGET_SO_SNDBUF:
2548 optname = SO_SNDBUF;
2549 goto int_case;
2550 case TARGET_SO_RCVBUF:
2551 optname = SO_RCVBUF;
2552 goto int_case;
2553 case TARGET_SO_KEEPALIVE:
2554 optname = SO_KEEPALIVE;
2555 goto int_case;
2556 case TARGET_SO_OOBINLINE:
2557 optname = SO_OOBINLINE;
2558 goto int_case;
2559 case TARGET_SO_NO_CHECK:
2560 optname = SO_NO_CHECK;
2561 goto int_case;
2562 case TARGET_SO_PRIORITY:
2563 optname = SO_PRIORITY;
2564 goto int_case;
2565 #ifdef SO_BSDCOMPAT
2566 case TARGET_SO_BSDCOMPAT:
2567 optname = SO_BSDCOMPAT;
2568 goto int_case;
2569 #endif
2570 case TARGET_SO_PASSCRED:
2571 optname = SO_PASSCRED;
2572 goto int_case;
2573 case TARGET_SO_TIMESTAMP:
2574 optname = SO_TIMESTAMP;
2575 goto int_case;
2576 case TARGET_SO_RCVLOWAT:
2577 optname = SO_RCVLOWAT;
2578 goto int_case;
2579 case TARGET_SO_ACCEPTCONN:
2580 optname = SO_ACCEPTCONN;
2581 goto int_case;
2582 default:
2583 goto int_case;
2585 break;
2586 case SOL_TCP:
2587 /* TCP options all take an 'int' value. */
2588 int_case:
2589 if (get_user_u32(len, optlen))
2590 return -TARGET_EFAULT;
2591 if (len < 0)
2592 return -TARGET_EINVAL;
2593 lv = sizeof(lv);
2594 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv));
2595 if (ret < 0)
2596 return ret;
2597 if (optname == SO_TYPE) {
2598 val = host_to_target_sock_type(val);
2600 if (len > lv)
2601 len = lv;
2602 if (len == 4) {
2603 if (put_user_u32(val, optval_addr))
2604 return -TARGET_EFAULT;
2605 } else {
2606 if (put_user_u8(val, optval_addr))
2607 return -TARGET_EFAULT;
2609 if (put_user_u32(len, optlen))
2610 return -TARGET_EFAULT;
2611 break;
2612 case SOL_IP:
2613 switch(optname) {
2614 case IP_TOS:
2615 case IP_TTL:
2616 case IP_HDRINCL:
2617 case IP_ROUTER_ALERT:
2618 case IP_RECVOPTS:
2619 case IP_RETOPTS:
2620 case IP_PKTINFO:
2621 case IP_MTU_DISCOVER:
2622 case IP_RECVERR:
2623 case IP_RECVTOS:
2624 #ifdef IP_FREEBIND
2625 case IP_FREEBIND:
2626 #endif
2627 case IP_MULTICAST_TTL:
2628 case IP_MULTICAST_LOOP:
2629 if (get_user_u32(len, optlen))
2630 return -TARGET_EFAULT;
2631 if (len < 0)
2632 return -TARGET_EINVAL;
2633 lv = sizeof(lv);
2634 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv));
2635 if (ret < 0)
2636 return ret;
2637 if (len < sizeof(int) && len > 0 && val >= 0 && val < 255) {
2638 len = 1;
2639 if (put_user_u32(len, optlen)
2640 || put_user_u8(val, optval_addr))
2641 return -TARGET_EFAULT;
2642 } else {
2643 if (len > sizeof(int))
2644 len = sizeof(int);
2645 if (put_user_u32(len, optlen)
2646 || put_user_u32(val, optval_addr))
2647 return -TARGET_EFAULT;
2649 break;
2650 default:
2651 ret = -TARGET_ENOPROTOOPT;
2652 break;
2654 break;
2655 case SOL_IPV6:
2656 switch (optname) {
2657 case IPV6_MTU_DISCOVER:
2658 case IPV6_MTU:
2659 case IPV6_V6ONLY:
2660 case IPV6_RECVPKTINFO:
2661 case IPV6_UNICAST_HOPS:
2662 case IPV6_MULTICAST_HOPS:
2663 case IPV6_MULTICAST_LOOP:
2664 case IPV6_RECVERR:
2665 case IPV6_RECVHOPLIMIT:
2666 case IPV6_2292HOPLIMIT:
2667 case IPV6_CHECKSUM:
2668 case IPV6_ADDRFORM:
2669 case IPV6_2292PKTINFO:
2670 case IPV6_RECVTCLASS:
2671 case IPV6_RECVRTHDR:
2672 case IPV6_2292RTHDR:
2673 case IPV6_RECVHOPOPTS:
2674 case IPV6_2292HOPOPTS:
2675 case IPV6_RECVDSTOPTS:
2676 case IPV6_2292DSTOPTS:
2677 case IPV6_TCLASS:
2678 #ifdef IPV6_RECVPATHMTU
2679 case IPV6_RECVPATHMTU:
2680 #endif
2681 #ifdef IPV6_TRANSPARENT
2682 case IPV6_TRANSPARENT:
2683 #endif
2684 #ifdef IPV6_FREEBIND
2685 case IPV6_FREEBIND:
2686 #endif
2687 #ifdef IPV6_RECVORIGDSTADDR
2688 case IPV6_RECVORIGDSTADDR:
2689 #endif
2690 if (get_user_u32(len, optlen))
2691 return -TARGET_EFAULT;
2692 if (len < 0)
2693 return -TARGET_EINVAL;
2694 lv = sizeof(lv);
2695 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv));
2696 if (ret < 0)
2697 return ret;
2698 if (len < sizeof(int) && len > 0 && val >= 0 && val < 255) {
2699 len = 1;
2700 if (put_user_u32(len, optlen)
2701 || put_user_u8(val, optval_addr))
2702 return -TARGET_EFAULT;
2703 } else {
2704 if (len > sizeof(int))
2705 len = sizeof(int);
2706 if (put_user_u32(len, optlen)
2707 || put_user_u32(val, optval_addr))
2708 return -TARGET_EFAULT;
2710 break;
2711 default:
2712 ret = -TARGET_ENOPROTOOPT;
2713 break;
2715 break;
2716 #ifdef SOL_NETLINK
2717 case SOL_NETLINK:
2718 switch (optname) {
2719 case NETLINK_PKTINFO:
2720 case NETLINK_BROADCAST_ERROR:
2721 case NETLINK_NO_ENOBUFS:
2722 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0)
2723 case NETLINK_LISTEN_ALL_NSID:
2724 case NETLINK_CAP_ACK:
2725 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0) */
2726 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0)
2727 case NETLINK_EXT_ACK:
2728 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0) */
2729 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 20, 0)
2730 case NETLINK_GET_STRICT_CHK:
2731 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0) */
2732 if (get_user_u32(len, optlen)) {
2733 return -TARGET_EFAULT;
2735 if (len != sizeof(val)) {
2736 return -TARGET_EINVAL;
2738 lv = len;
2739 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv));
2740 if (ret < 0) {
2741 return ret;
2743 if (put_user_u32(lv, optlen)
2744 || put_user_u32(val, optval_addr)) {
2745 return -TARGET_EFAULT;
2747 break;
2748 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0)
2749 case NETLINK_LIST_MEMBERSHIPS:
2751 uint32_t *results;
2752 int i;
2753 if (get_user_u32(len, optlen)) {
2754 return -TARGET_EFAULT;
2756 if (len < 0) {
2757 return -TARGET_EINVAL;
2759 results = lock_user(VERIFY_WRITE, optval_addr, len, 1);
2760 if (!results) {
2761 return -TARGET_EFAULT;
2763 lv = len;
2764 ret = get_errno(getsockopt(sockfd, level, optname, results, &lv));
2765 if (ret < 0) {
2766 unlock_user(results, optval_addr, 0);
2767 return ret;
2769 /* swap host endianess to target endianess. */
2770 for (i = 0; i < (len / sizeof(uint32_t)); i++) {
2771 results[i] = tswap32(results[i]);
2773 if (put_user_u32(lv, optlen)) {
2774 return -TARGET_EFAULT;
2776 unlock_user(results, optval_addr, 0);
2777 break;
2779 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0) */
2780 default:
2781 goto unimplemented;
2783 break;
2784 #endif /* SOL_NETLINK */
2785 default:
2786 unimplemented:
2787 qemu_log_mask(LOG_UNIMP,
2788 "getsockopt level=%d optname=%d not yet supported\n",
2789 level, optname);
2790 ret = -TARGET_EOPNOTSUPP;
2791 break;
2793 return ret;
2796 /* Convert target low/high pair representing file offset into the host
2797 * low/high pair. This function doesn't handle offsets bigger than 64 bits
2798 * as the kernel doesn't handle them either.
2800 static void target_to_host_low_high(abi_ulong tlow,
2801 abi_ulong thigh,
2802 unsigned long *hlow,
2803 unsigned long *hhigh)
2805 uint64_t off = tlow |
2806 ((unsigned long long)thigh << TARGET_LONG_BITS / 2) <<
2807 TARGET_LONG_BITS / 2;
2809 *hlow = off;
2810 *hhigh = (off >> HOST_LONG_BITS / 2) >> HOST_LONG_BITS / 2;
2813 static struct iovec *lock_iovec(int type, abi_ulong target_addr,
2814 abi_ulong count, int copy)
2816 struct target_iovec *target_vec;
2817 struct iovec *vec;
2818 abi_ulong total_len, max_len;
2819 int i;
2820 int err = 0;
2821 bool bad_address = false;
2823 if (count == 0) {
2824 errno = 0;
2825 return NULL;
2827 if (count > IOV_MAX) {
2828 errno = EINVAL;
2829 return NULL;
2832 vec = g_try_new0(struct iovec, count);
2833 if (vec == NULL) {
2834 errno = ENOMEM;
2835 return NULL;
2838 target_vec = lock_user(VERIFY_READ, target_addr,
2839 count * sizeof(struct target_iovec), 1);
2840 if (target_vec == NULL) {
2841 err = EFAULT;
2842 goto fail2;
2845 /* ??? If host page size > target page size, this will result in a
2846 value larger than what we can actually support. */
2847 max_len = 0x7fffffff & TARGET_PAGE_MASK;
2848 total_len = 0;
2850 for (i = 0; i < count; i++) {
2851 abi_ulong base = tswapal(target_vec[i].iov_base);
2852 abi_long len = tswapal(target_vec[i].iov_len);
2854 if (len < 0) {
2855 err = EINVAL;
2856 goto fail;
2857 } else if (len == 0) {
2858 /* Zero length pointer is ignored. */
2859 vec[i].iov_base = 0;
2860 } else {
2861 vec[i].iov_base = lock_user(type, base, len, copy);
2862 /* If the first buffer pointer is bad, this is a fault. But
2863 * subsequent bad buffers will result in a partial write; this
2864 * is realized by filling the vector with null pointers and
2865 * zero lengths. */
2866 if (!vec[i].iov_base) {
2867 if (i == 0) {
2868 err = EFAULT;
2869 goto fail;
2870 } else {
2871 bad_address = true;
2874 if (bad_address) {
2875 len = 0;
2877 if (len > max_len - total_len) {
2878 len = max_len - total_len;
2881 vec[i].iov_len = len;
2882 total_len += len;
2885 unlock_user(target_vec, target_addr, 0);
2886 return vec;
2888 fail:
2889 while (--i >= 0) {
2890 if (tswapal(target_vec[i].iov_len) > 0) {
2891 unlock_user(vec[i].iov_base, tswapal(target_vec[i].iov_base), 0);
2894 unlock_user(target_vec, target_addr, 0);
2895 fail2:
2896 g_free(vec);
2897 errno = err;
2898 return NULL;
2901 static void unlock_iovec(struct iovec *vec, abi_ulong target_addr,
2902 abi_ulong count, int copy)
2904 struct target_iovec *target_vec;
2905 int i;
2907 target_vec = lock_user(VERIFY_READ, target_addr,
2908 count * sizeof(struct target_iovec), 1);
2909 if (target_vec) {
2910 for (i = 0; i < count; i++) {
2911 abi_ulong base = tswapal(target_vec[i].iov_base);
2912 abi_long len = tswapal(target_vec[i].iov_len);
2913 if (len < 0) {
2914 break;
2916 unlock_user(vec[i].iov_base, base, copy ? vec[i].iov_len : 0);
2918 unlock_user(target_vec, target_addr, 0);
2921 g_free(vec);
2924 static inline int target_to_host_sock_type(int *type)
2926 int host_type = 0;
2927 int target_type = *type;
2929 switch (target_type & TARGET_SOCK_TYPE_MASK) {
2930 case TARGET_SOCK_DGRAM:
2931 host_type = SOCK_DGRAM;
2932 break;
2933 case TARGET_SOCK_STREAM:
2934 host_type = SOCK_STREAM;
2935 break;
2936 default:
2937 host_type = target_type & TARGET_SOCK_TYPE_MASK;
2938 break;
2940 if (target_type & TARGET_SOCK_CLOEXEC) {
2941 #if defined(SOCK_CLOEXEC)
2942 host_type |= SOCK_CLOEXEC;
2943 #else
2944 return -TARGET_EINVAL;
2945 #endif
2947 if (target_type & TARGET_SOCK_NONBLOCK) {
2948 #if defined(SOCK_NONBLOCK)
2949 host_type |= SOCK_NONBLOCK;
2950 #elif !defined(O_NONBLOCK)
2951 return -TARGET_EINVAL;
2952 #endif
2954 *type = host_type;
2955 return 0;
2958 /* Try to emulate socket type flags after socket creation. */
2959 static int sock_flags_fixup(int fd, int target_type)
2961 #if !defined(SOCK_NONBLOCK) && defined(O_NONBLOCK)
2962 if (target_type & TARGET_SOCK_NONBLOCK) {
2963 int flags = fcntl(fd, F_GETFL);
2964 if (fcntl(fd, F_SETFL, O_NONBLOCK | flags) == -1) {
2965 close(fd);
2966 return -TARGET_EINVAL;
2969 #endif
2970 return fd;
2973 /* do_socket() Must return target values and target errnos. */
2974 static abi_long do_socket(int domain, int type, int protocol)
2976 int target_type = type;
2977 int ret;
2979 ret = target_to_host_sock_type(&type);
2980 if (ret) {
2981 return ret;
2984 if (domain == PF_NETLINK && !(
2985 #ifdef CONFIG_RTNETLINK
2986 protocol == NETLINK_ROUTE ||
2987 #endif
2988 protocol == NETLINK_KOBJECT_UEVENT ||
2989 protocol == NETLINK_AUDIT)) {
2990 return -TARGET_EPFNOSUPPORT;
2993 if (domain == AF_PACKET ||
2994 (domain == AF_INET && type == SOCK_PACKET)) {
2995 protocol = tswap16(protocol);
2998 ret = get_errno(socket(domain, type, protocol));
2999 if (ret >= 0) {
3000 ret = sock_flags_fixup(ret, target_type);
3001 if (type == SOCK_PACKET) {
3002 /* Manage an obsolete case :
3003 * if socket type is SOCK_PACKET, bind by name
3005 fd_trans_register(ret, &target_packet_trans);
3006 } else if (domain == PF_NETLINK) {
3007 switch (protocol) {
3008 #ifdef CONFIG_RTNETLINK
3009 case NETLINK_ROUTE:
3010 fd_trans_register(ret, &target_netlink_route_trans);
3011 break;
3012 #endif
3013 case NETLINK_KOBJECT_UEVENT:
3014 /* nothing to do: messages are strings */
3015 break;
3016 case NETLINK_AUDIT:
3017 fd_trans_register(ret, &target_netlink_audit_trans);
3018 break;
3019 default:
3020 g_assert_not_reached();
3024 return ret;
3027 /* do_bind() Must return target values and target errnos. */
3028 static abi_long do_bind(int sockfd, abi_ulong target_addr,
3029 socklen_t addrlen)
3031 void *addr;
3032 abi_long ret;
3034 if ((int)addrlen < 0) {
3035 return -TARGET_EINVAL;
3038 addr = alloca(addrlen+1);
3040 ret = target_to_host_sockaddr(sockfd, addr, target_addr, addrlen);
3041 if (ret)
3042 return ret;
3044 return get_errno(bind(sockfd, addr, addrlen));
3047 /* do_connect() Must return target values and target errnos. */
3048 static abi_long do_connect(int sockfd, abi_ulong target_addr,
3049 socklen_t addrlen)
3051 void *addr;
3052 abi_long ret;
3054 if ((int)addrlen < 0) {
3055 return -TARGET_EINVAL;
3058 addr = alloca(addrlen+1);
3060 ret = target_to_host_sockaddr(sockfd, addr, target_addr, addrlen);
3061 if (ret)
3062 return ret;
3064 return get_errno(safe_connect(sockfd, addr, addrlen));
3067 /* do_sendrecvmsg_locked() Must return target values and target errnos. */
3068 static abi_long do_sendrecvmsg_locked(int fd, struct target_msghdr *msgp,
3069 int flags, int send)
3071 abi_long ret, len;
3072 struct msghdr msg;
3073 abi_ulong count;
3074 struct iovec *vec;
3075 abi_ulong target_vec;
3077 if (msgp->msg_name) {
3078 msg.msg_namelen = tswap32(msgp->msg_namelen);
3079 msg.msg_name = alloca(msg.msg_namelen+1);
3080 ret = target_to_host_sockaddr(fd, msg.msg_name,
3081 tswapal(msgp->msg_name),
3082 msg.msg_namelen);
3083 if (ret == -TARGET_EFAULT) {
3084 /* For connected sockets msg_name and msg_namelen must
3085 * be ignored, so returning EFAULT immediately is wrong.
3086 * Instead, pass a bad msg_name to the host kernel, and
3087 * let it decide whether to return EFAULT or not.
3089 msg.msg_name = (void *)-1;
3090 } else if (ret) {
3091 goto out2;
3093 } else {
3094 msg.msg_name = NULL;
3095 msg.msg_namelen = 0;
3097 msg.msg_controllen = 2 * tswapal(msgp->msg_controllen);
3098 msg.msg_control = alloca(msg.msg_controllen);
3099 memset(msg.msg_control, 0, msg.msg_controllen);
3101 msg.msg_flags = tswap32(msgp->msg_flags);
3103 count = tswapal(msgp->msg_iovlen);
3104 target_vec = tswapal(msgp->msg_iov);
3106 if (count > IOV_MAX) {
3107 /* sendrcvmsg returns a different errno for this condition than
3108 * readv/writev, so we must catch it here before lock_iovec() does.
3110 ret = -TARGET_EMSGSIZE;
3111 goto out2;
3114 vec = lock_iovec(send ? VERIFY_READ : VERIFY_WRITE,
3115 target_vec, count, send);
3116 if (vec == NULL) {
3117 ret = -host_to_target_errno(errno);
3118 goto out2;
3120 msg.msg_iovlen = count;
3121 msg.msg_iov = vec;
3123 if (send) {
3124 if (fd_trans_target_to_host_data(fd)) {
3125 void *host_msg;
3127 host_msg = g_malloc(msg.msg_iov->iov_len);
3128 memcpy(host_msg, msg.msg_iov->iov_base, msg.msg_iov->iov_len);
3129 ret = fd_trans_target_to_host_data(fd)(host_msg,
3130 msg.msg_iov->iov_len);
3131 if (ret >= 0) {
3132 msg.msg_iov->iov_base = host_msg;
3133 ret = get_errno(safe_sendmsg(fd, &msg, flags));
3135 g_free(host_msg);
3136 } else {
3137 ret = target_to_host_cmsg(&msg, msgp);
3138 if (ret == 0) {
3139 ret = get_errno(safe_sendmsg(fd, &msg, flags));
3142 } else {
3143 ret = get_errno(safe_recvmsg(fd, &msg, flags));
3144 if (!is_error(ret)) {
3145 len = ret;
3146 if (fd_trans_host_to_target_data(fd)) {
3147 ret = fd_trans_host_to_target_data(fd)(msg.msg_iov->iov_base,
3148 MIN(msg.msg_iov->iov_len, len));
3149 } else {
3150 ret = host_to_target_cmsg(msgp, &msg);
3152 if (!is_error(ret)) {
3153 msgp->msg_namelen = tswap32(msg.msg_namelen);
3154 msgp->msg_flags = tswap32(msg.msg_flags);
3155 if (msg.msg_name != NULL && msg.msg_name != (void *)-1) {
3156 ret = host_to_target_sockaddr(tswapal(msgp->msg_name),
3157 msg.msg_name, msg.msg_namelen);
3158 if (ret) {
3159 goto out;
3163 ret = len;
3168 out:
3169 unlock_iovec(vec, target_vec, count, !send);
3170 out2:
3171 return ret;
3174 static abi_long do_sendrecvmsg(int fd, abi_ulong target_msg,
3175 int flags, int send)
3177 abi_long ret;
3178 struct target_msghdr *msgp;
3180 if (!lock_user_struct(send ? VERIFY_READ : VERIFY_WRITE,
3181 msgp,
3182 target_msg,
3183 send ? 1 : 0)) {
3184 return -TARGET_EFAULT;
3186 ret = do_sendrecvmsg_locked(fd, msgp, flags, send);
3187 unlock_user_struct(msgp, target_msg, send ? 0 : 1);
3188 return ret;
3191 /* We don't rely on the C library to have sendmmsg/recvmmsg support,
3192 * so it might not have this *mmsg-specific flag either.
3194 #ifndef MSG_WAITFORONE
3195 #define MSG_WAITFORONE 0x10000
3196 #endif
3198 static abi_long do_sendrecvmmsg(int fd, abi_ulong target_msgvec,
3199 unsigned int vlen, unsigned int flags,
3200 int send)
3202 struct target_mmsghdr *mmsgp;
3203 abi_long ret = 0;
3204 int i;
3206 if (vlen > UIO_MAXIOV) {
3207 vlen = UIO_MAXIOV;
3210 mmsgp = lock_user(VERIFY_WRITE, target_msgvec, sizeof(*mmsgp) * vlen, 1);
3211 if (!mmsgp) {
3212 return -TARGET_EFAULT;
3215 for (i = 0; i < vlen; i++) {
3216 ret = do_sendrecvmsg_locked(fd, &mmsgp[i].msg_hdr, flags, send);
3217 if (is_error(ret)) {
3218 break;
3220 mmsgp[i].msg_len = tswap32(ret);
3221 /* MSG_WAITFORONE turns on MSG_DONTWAIT after one packet */
3222 if (flags & MSG_WAITFORONE) {
3223 flags |= MSG_DONTWAIT;
3227 unlock_user(mmsgp, target_msgvec, sizeof(*mmsgp) * i);
3229 /* Return number of datagrams sent if we sent any at all;
3230 * otherwise return the error.
3232 if (i) {
3233 return i;
3235 return ret;
3238 /* do_accept4() Must return target values and target errnos. */
3239 static abi_long do_accept4(int fd, abi_ulong target_addr,
3240 abi_ulong target_addrlen_addr, int flags)
3242 socklen_t addrlen, ret_addrlen;
3243 void *addr;
3244 abi_long ret;
3245 int host_flags;
3247 host_flags = target_to_host_bitmask(flags, fcntl_flags_tbl);
3249 if (target_addr == 0) {
3250 return get_errno(safe_accept4(fd, NULL, NULL, host_flags));
3253 /* linux returns EINVAL if addrlen pointer is invalid */
3254 if (get_user_u32(addrlen, target_addrlen_addr))
3255 return -TARGET_EINVAL;
3257 if ((int)addrlen < 0) {
3258 return -TARGET_EINVAL;
3261 if (!access_ok(VERIFY_WRITE, target_addr, addrlen))
3262 return -TARGET_EINVAL;
3264 addr = alloca(addrlen);
3266 ret_addrlen = addrlen;
3267 ret = get_errno(safe_accept4(fd, addr, &ret_addrlen, host_flags));
3268 if (!is_error(ret)) {
3269 host_to_target_sockaddr(target_addr, addr, MIN(addrlen, ret_addrlen));
3270 if (put_user_u32(ret_addrlen, target_addrlen_addr)) {
3271 ret = -TARGET_EFAULT;
3274 return ret;
3277 /* do_getpeername() Must return target values and target errnos. */
3278 static abi_long do_getpeername(int fd, abi_ulong target_addr,
3279 abi_ulong target_addrlen_addr)
3281 socklen_t addrlen, ret_addrlen;
3282 void *addr;
3283 abi_long ret;
3285 if (get_user_u32(addrlen, target_addrlen_addr))
3286 return -TARGET_EFAULT;
3288 if ((int)addrlen < 0) {
3289 return -TARGET_EINVAL;
3292 if (!access_ok(VERIFY_WRITE, target_addr, addrlen))
3293 return -TARGET_EFAULT;
3295 addr = alloca(addrlen);
3297 ret_addrlen = addrlen;
3298 ret = get_errno(getpeername(fd, addr, &ret_addrlen));
3299 if (!is_error(ret)) {
3300 host_to_target_sockaddr(target_addr, addr, MIN(addrlen, ret_addrlen));
3301 if (put_user_u32(ret_addrlen, target_addrlen_addr)) {
3302 ret = -TARGET_EFAULT;
3305 return ret;
3308 /* do_getsockname() Must return target values and target errnos. */
3309 static abi_long do_getsockname(int fd, abi_ulong target_addr,
3310 abi_ulong target_addrlen_addr)
3312 socklen_t addrlen, ret_addrlen;
3313 void *addr;
3314 abi_long ret;
3316 if (get_user_u32(addrlen, target_addrlen_addr))
3317 return -TARGET_EFAULT;
3319 if ((int)addrlen < 0) {
3320 return -TARGET_EINVAL;
3323 if (!access_ok(VERIFY_WRITE, target_addr, addrlen))
3324 return -TARGET_EFAULT;
3326 addr = alloca(addrlen);
3328 ret_addrlen = addrlen;
3329 ret = get_errno(getsockname(fd, addr, &ret_addrlen));
3330 if (!is_error(ret)) {
3331 host_to_target_sockaddr(target_addr, addr, MIN(addrlen, ret_addrlen));
3332 if (put_user_u32(ret_addrlen, target_addrlen_addr)) {
3333 ret = -TARGET_EFAULT;
3336 return ret;
3339 /* do_socketpair() Must return target values and target errnos. */
3340 static abi_long do_socketpair(int domain, int type, int protocol,
3341 abi_ulong target_tab_addr)
3343 int tab[2];
3344 abi_long ret;
3346 target_to_host_sock_type(&type);
3348 ret = get_errno(socketpair(domain, type, protocol, tab));
3349 if (!is_error(ret)) {
3350 if (put_user_s32(tab[0], target_tab_addr)
3351 || put_user_s32(tab[1], target_tab_addr + sizeof(tab[0])))
3352 ret = -TARGET_EFAULT;
3354 return ret;
3357 /* do_sendto() Must return target values and target errnos. */
3358 static abi_long do_sendto(int fd, abi_ulong msg, size_t len, int flags,
3359 abi_ulong target_addr, socklen_t addrlen)
3361 void *addr;
3362 void *host_msg;
3363 void *copy_msg = NULL;
3364 abi_long ret;
3366 if ((int)addrlen < 0) {
3367 return -TARGET_EINVAL;
3370 host_msg = lock_user(VERIFY_READ, msg, len, 1);
3371 if (!host_msg)
3372 return -TARGET_EFAULT;
3373 if (fd_trans_target_to_host_data(fd)) {
3374 copy_msg = host_msg;
3375 host_msg = g_malloc(len);
3376 memcpy(host_msg, copy_msg, len);
3377 ret = fd_trans_target_to_host_data(fd)(host_msg, len);
3378 if (ret < 0) {
3379 goto fail;
3382 if (target_addr) {
3383 addr = alloca(addrlen+1);
3384 ret = target_to_host_sockaddr(fd, addr, target_addr, addrlen);
3385 if (ret) {
3386 goto fail;
3388 ret = get_errno(safe_sendto(fd, host_msg, len, flags, addr, addrlen));
3389 } else {
3390 ret = get_errno(safe_sendto(fd, host_msg, len, flags, NULL, 0));
3392 fail:
3393 if (copy_msg) {
3394 g_free(host_msg);
3395 host_msg = copy_msg;
3397 unlock_user(host_msg, msg, 0);
3398 return ret;
3401 /* do_recvfrom() Must return target values and target errnos. */
3402 static abi_long do_recvfrom(int fd, abi_ulong msg, size_t len, int flags,
3403 abi_ulong target_addr,
3404 abi_ulong target_addrlen)
3406 socklen_t addrlen, ret_addrlen;
3407 void *addr;
3408 void *host_msg;
3409 abi_long ret;
3411 host_msg = lock_user(VERIFY_WRITE, msg, len, 0);
3412 if (!host_msg)
3413 return -TARGET_EFAULT;
3414 if (target_addr) {
3415 if (get_user_u32(addrlen, target_addrlen)) {
3416 ret = -TARGET_EFAULT;
3417 goto fail;
3419 if ((int)addrlen < 0) {
3420 ret = -TARGET_EINVAL;
3421 goto fail;
3423 addr = alloca(addrlen);
3424 ret_addrlen = addrlen;
3425 ret = get_errno(safe_recvfrom(fd, host_msg, len, flags,
3426 addr, &ret_addrlen));
3427 } else {
3428 addr = NULL; /* To keep compiler quiet. */
3429 addrlen = 0; /* To keep compiler quiet. */
3430 ret = get_errno(safe_recvfrom(fd, host_msg, len, flags, NULL, 0));
3432 if (!is_error(ret)) {
3433 if (fd_trans_host_to_target_data(fd)) {
3434 abi_long trans;
3435 trans = fd_trans_host_to_target_data(fd)(host_msg, MIN(ret, len));
3436 if (is_error(trans)) {
3437 ret = trans;
3438 goto fail;
3441 if (target_addr) {
3442 host_to_target_sockaddr(target_addr, addr,
3443 MIN(addrlen, ret_addrlen));
3444 if (put_user_u32(ret_addrlen, target_addrlen)) {
3445 ret = -TARGET_EFAULT;
3446 goto fail;
3449 unlock_user(host_msg, msg, len);
3450 } else {
3451 fail:
3452 unlock_user(host_msg, msg, 0);
3454 return ret;
3457 #ifdef TARGET_NR_socketcall
3458 /* do_socketcall() must return target values and target errnos. */
3459 static abi_long do_socketcall(int num, abi_ulong vptr)
3461 static const unsigned nargs[] = { /* number of arguments per operation */
3462 [TARGET_SYS_SOCKET] = 3, /* domain, type, protocol */
3463 [TARGET_SYS_BIND] = 3, /* fd, addr, addrlen */
3464 [TARGET_SYS_CONNECT] = 3, /* fd, addr, addrlen */
3465 [TARGET_SYS_LISTEN] = 2, /* fd, backlog */
3466 [TARGET_SYS_ACCEPT] = 3, /* fd, addr, addrlen */
3467 [TARGET_SYS_GETSOCKNAME] = 3, /* fd, addr, addrlen */
3468 [TARGET_SYS_GETPEERNAME] = 3, /* fd, addr, addrlen */
3469 [TARGET_SYS_SOCKETPAIR] = 4, /* domain, type, protocol, tab */
3470 [TARGET_SYS_SEND] = 4, /* fd, msg, len, flags */
3471 [TARGET_SYS_RECV] = 4, /* fd, msg, len, flags */
3472 [TARGET_SYS_SENDTO] = 6, /* fd, msg, len, flags, addr, addrlen */
3473 [TARGET_SYS_RECVFROM] = 6, /* fd, msg, len, flags, addr, addrlen */
3474 [TARGET_SYS_SHUTDOWN] = 2, /* fd, how */
3475 [TARGET_SYS_SETSOCKOPT] = 5, /* fd, level, optname, optval, optlen */
3476 [TARGET_SYS_GETSOCKOPT] = 5, /* fd, level, optname, optval, optlen */
3477 [TARGET_SYS_SENDMSG] = 3, /* fd, msg, flags */
3478 [TARGET_SYS_RECVMSG] = 3, /* fd, msg, flags */
3479 [TARGET_SYS_ACCEPT4] = 4, /* fd, addr, addrlen, flags */
3480 [TARGET_SYS_RECVMMSG] = 4, /* fd, msgvec, vlen, flags */
3481 [TARGET_SYS_SENDMMSG] = 4, /* fd, msgvec, vlen, flags */
3483 abi_long a[6]; /* max 6 args */
3484 unsigned i;
3486 /* check the range of the first argument num */
3487 /* (TARGET_SYS_SENDMMSG is the highest among TARGET_SYS_xxx) */
3488 if (num < 1 || num > TARGET_SYS_SENDMMSG) {
3489 return -TARGET_EINVAL;
3491 /* ensure we have space for args */
3492 if (nargs[num] > ARRAY_SIZE(a)) {
3493 return -TARGET_EINVAL;
3495 /* collect the arguments in a[] according to nargs[] */
3496 for (i = 0; i < nargs[num]; ++i) {
3497 if (get_user_ual(a[i], vptr + i * sizeof(abi_long)) != 0) {
3498 return -TARGET_EFAULT;
3501 /* now when we have the args, invoke the appropriate underlying function */
3502 switch (num) {
3503 case TARGET_SYS_SOCKET: /* domain, type, protocol */
3504 return do_socket(a[0], a[1], a[2]);
3505 case TARGET_SYS_BIND: /* sockfd, addr, addrlen */
3506 return do_bind(a[0], a[1], a[2]);
3507 case TARGET_SYS_CONNECT: /* sockfd, addr, addrlen */
3508 return do_connect(a[0], a[1], a[2]);
3509 case TARGET_SYS_LISTEN: /* sockfd, backlog */
3510 return get_errno(listen(a[0], a[1]));
3511 case TARGET_SYS_ACCEPT: /* sockfd, addr, addrlen */
3512 return do_accept4(a[0], a[1], a[2], 0);
3513 case TARGET_SYS_GETSOCKNAME: /* sockfd, addr, addrlen */
3514 return do_getsockname(a[0], a[1], a[2]);
3515 case TARGET_SYS_GETPEERNAME: /* sockfd, addr, addrlen */
3516 return do_getpeername(a[0], a[1], a[2]);
3517 case TARGET_SYS_SOCKETPAIR: /* domain, type, protocol, tab */
3518 return do_socketpair(a[0], a[1], a[2], a[3]);
3519 case TARGET_SYS_SEND: /* sockfd, msg, len, flags */
3520 return do_sendto(a[0], a[1], a[2], a[3], 0, 0);
3521 case TARGET_SYS_RECV: /* sockfd, msg, len, flags */
3522 return do_recvfrom(a[0], a[1], a[2], a[3], 0, 0);
3523 case TARGET_SYS_SENDTO: /* sockfd, msg, len, flags, addr, addrlen */
3524 return do_sendto(a[0], a[1], a[2], a[3], a[4], a[5]);
3525 case TARGET_SYS_RECVFROM: /* sockfd, msg, len, flags, addr, addrlen */
3526 return do_recvfrom(a[0], a[1], a[2], a[3], a[4], a[5]);
3527 case TARGET_SYS_SHUTDOWN: /* sockfd, how */
3528 return get_errno(shutdown(a[0], a[1]));
3529 case TARGET_SYS_SETSOCKOPT: /* sockfd, level, optname, optval, optlen */
3530 return do_setsockopt(a[0], a[1], a[2], a[3], a[4]);
3531 case TARGET_SYS_GETSOCKOPT: /* sockfd, level, optname, optval, optlen */
3532 return do_getsockopt(a[0], a[1], a[2], a[3], a[4]);
3533 case TARGET_SYS_SENDMSG: /* sockfd, msg, flags */
3534 return do_sendrecvmsg(a[0], a[1], a[2], 1);
3535 case TARGET_SYS_RECVMSG: /* sockfd, msg, flags */
3536 return do_sendrecvmsg(a[0], a[1], a[2], 0);
3537 case TARGET_SYS_ACCEPT4: /* sockfd, addr, addrlen, flags */
3538 return do_accept4(a[0], a[1], a[2], a[3]);
3539 case TARGET_SYS_RECVMMSG: /* sockfd, msgvec, vlen, flags */
3540 return do_sendrecvmmsg(a[0], a[1], a[2], a[3], 0);
3541 case TARGET_SYS_SENDMMSG: /* sockfd, msgvec, vlen, flags */
3542 return do_sendrecvmmsg(a[0], a[1], a[2], a[3], 1);
3543 default:
3544 qemu_log_mask(LOG_UNIMP, "Unsupported socketcall: %d\n", num);
3545 return -TARGET_EINVAL;
3548 #endif
3550 #define N_SHM_REGIONS 32
3552 static struct shm_region {
3553 abi_ulong start;
3554 abi_ulong size;
3555 bool in_use;
3556 } shm_regions[N_SHM_REGIONS];
3558 #ifndef TARGET_SEMID64_DS
3559 /* asm-generic version of this struct */
3560 struct target_semid64_ds
3562 struct target_ipc_perm sem_perm;
3563 abi_ulong sem_otime;
3564 #if TARGET_ABI_BITS == 32
3565 abi_ulong __unused1;
3566 #endif
3567 abi_ulong sem_ctime;
3568 #if TARGET_ABI_BITS == 32
3569 abi_ulong __unused2;
3570 #endif
3571 abi_ulong sem_nsems;
3572 abi_ulong __unused3;
3573 abi_ulong __unused4;
3575 #endif
3577 static inline abi_long target_to_host_ipc_perm(struct ipc_perm *host_ip,
3578 abi_ulong target_addr)
3580 struct target_ipc_perm *target_ip;
3581 struct target_semid64_ds *target_sd;
3583 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1))
3584 return -TARGET_EFAULT;
3585 target_ip = &(target_sd->sem_perm);
3586 host_ip->__key = tswap32(target_ip->__key);
3587 host_ip->uid = tswap32(target_ip->uid);
3588 host_ip->gid = tswap32(target_ip->gid);
3589 host_ip->cuid = tswap32(target_ip->cuid);
3590 host_ip->cgid = tswap32(target_ip->cgid);
3591 #if defined(TARGET_ALPHA) || defined(TARGET_MIPS) || defined(TARGET_PPC)
3592 host_ip->mode = tswap32(target_ip->mode);
3593 #else
3594 host_ip->mode = tswap16(target_ip->mode);
3595 #endif
3596 #if defined(TARGET_PPC)
3597 host_ip->__seq = tswap32(target_ip->__seq);
3598 #else
3599 host_ip->__seq = tswap16(target_ip->__seq);
3600 #endif
3601 unlock_user_struct(target_sd, target_addr, 0);
3602 return 0;
3605 static inline abi_long host_to_target_ipc_perm(abi_ulong target_addr,
3606 struct ipc_perm *host_ip)
3608 struct target_ipc_perm *target_ip;
3609 struct target_semid64_ds *target_sd;
3611 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0))
3612 return -TARGET_EFAULT;
3613 target_ip = &(target_sd->sem_perm);
3614 target_ip->__key = tswap32(host_ip->__key);
3615 target_ip->uid = tswap32(host_ip->uid);
3616 target_ip->gid = tswap32(host_ip->gid);
3617 target_ip->cuid = tswap32(host_ip->cuid);
3618 target_ip->cgid = tswap32(host_ip->cgid);
3619 #if defined(TARGET_ALPHA) || defined(TARGET_MIPS) || defined(TARGET_PPC)
3620 target_ip->mode = tswap32(host_ip->mode);
3621 #else
3622 target_ip->mode = tswap16(host_ip->mode);
3623 #endif
3624 #if defined(TARGET_PPC)
3625 target_ip->__seq = tswap32(host_ip->__seq);
3626 #else
3627 target_ip->__seq = tswap16(host_ip->__seq);
3628 #endif
3629 unlock_user_struct(target_sd, target_addr, 1);
3630 return 0;
3633 static inline abi_long target_to_host_semid_ds(struct semid_ds *host_sd,
3634 abi_ulong target_addr)
3636 struct target_semid64_ds *target_sd;
3638 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1))
3639 return -TARGET_EFAULT;
3640 if (target_to_host_ipc_perm(&(host_sd->sem_perm),target_addr))
3641 return -TARGET_EFAULT;
3642 host_sd->sem_nsems = tswapal(target_sd->sem_nsems);
3643 host_sd->sem_otime = tswapal(target_sd->sem_otime);
3644 host_sd->sem_ctime = tswapal(target_sd->sem_ctime);
3645 unlock_user_struct(target_sd, target_addr, 0);
3646 return 0;
3649 static inline abi_long host_to_target_semid_ds(abi_ulong target_addr,
3650 struct semid_ds *host_sd)
3652 struct target_semid64_ds *target_sd;
3654 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0))
3655 return -TARGET_EFAULT;
3656 if (host_to_target_ipc_perm(target_addr,&(host_sd->sem_perm)))
3657 return -TARGET_EFAULT;
3658 target_sd->sem_nsems = tswapal(host_sd->sem_nsems);
3659 target_sd->sem_otime = tswapal(host_sd->sem_otime);
3660 target_sd->sem_ctime = tswapal(host_sd->sem_ctime);
3661 unlock_user_struct(target_sd, target_addr, 1);
3662 return 0;
3665 struct target_seminfo {
3666 int semmap;
3667 int semmni;
3668 int semmns;
3669 int semmnu;
3670 int semmsl;
3671 int semopm;
3672 int semume;
3673 int semusz;
3674 int semvmx;
3675 int semaem;
3678 static inline abi_long host_to_target_seminfo(abi_ulong target_addr,
3679 struct seminfo *host_seminfo)
3681 struct target_seminfo *target_seminfo;
3682 if (!lock_user_struct(VERIFY_WRITE, target_seminfo, target_addr, 0))
3683 return -TARGET_EFAULT;
3684 __put_user(host_seminfo->semmap, &target_seminfo->semmap);
3685 __put_user(host_seminfo->semmni, &target_seminfo->semmni);
3686 __put_user(host_seminfo->semmns, &target_seminfo->semmns);
3687 __put_user(host_seminfo->semmnu, &target_seminfo->semmnu);
3688 __put_user(host_seminfo->semmsl, &target_seminfo->semmsl);
3689 __put_user(host_seminfo->semopm, &target_seminfo->semopm);
3690 __put_user(host_seminfo->semume, &target_seminfo->semume);
3691 __put_user(host_seminfo->semusz, &target_seminfo->semusz);
3692 __put_user(host_seminfo->semvmx, &target_seminfo->semvmx);
3693 __put_user(host_seminfo->semaem, &target_seminfo->semaem);
3694 unlock_user_struct(target_seminfo, target_addr, 1);
3695 return 0;
3698 union semun {
3699 int val;
3700 struct semid_ds *buf;
3701 unsigned short *array;
3702 struct seminfo *__buf;
3705 union target_semun {
3706 int val;
3707 abi_ulong buf;
3708 abi_ulong array;
3709 abi_ulong __buf;
3712 static inline abi_long target_to_host_semarray(int semid, unsigned short **host_array,
3713 abi_ulong target_addr)
3715 int nsems;
3716 unsigned short *array;
3717 union semun semun;
3718 struct semid_ds semid_ds;
3719 int i, ret;
3721 semun.buf = &semid_ds;
3723 ret = semctl(semid, 0, IPC_STAT, semun);
3724 if (ret == -1)
3725 return get_errno(ret);
3727 nsems = semid_ds.sem_nsems;
3729 *host_array = g_try_new(unsigned short, nsems);
3730 if (!*host_array) {
3731 return -TARGET_ENOMEM;
3733 array = lock_user(VERIFY_READ, target_addr,
3734 nsems*sizeof(unsigned short), 1);
3735 if (!array) {
3736 g_free(*host_array);
3737 return -TARGET_EFAULT;
3740 for(i=0; i<nsems; i++) {
3741 __get_user((*host_array)[i], &array[i]);
3743 unlock_user(array, target_addr, 0);
3745 return 0;
3748 static inline abi_long host_to_target_semarray(int semid, abi_ulong target_addr,
3749 unsigned short **host_array)
3751 int nsems;
3752 unsigned short *array;
3753 union semun semun;
3754 struct semid_ds semid_ds;
3755 int i, ret;
3757 semun.buf = &semid_ds;
3759 ret = semctl(semid, 0, IPC_STAT, semun);
3760 if (ret == -1)
3761 return get_errno(ret);
3763 nsems = semid_ds.sem_nsems;
3765 array = lock_user(VERIFY_WRITE, target_addr,
3766 nsems*sizeof(unsigned short), 0);
3767 if (!array)
3768 return -TARGET_EFAULT;
3770 for(i=0; i<nsems; i++) {
3771 __put_user((*host_array)[i], &array[i]);
3773 g_free(*host_array);
3774 unlock_user(array, target_addr, 1);
3776 return 0;
3779 static inline abi_long do_semctl(int semid, int semnum, int cmd,
3780 abi_ulong target_arg)
3782 union target_semun target_su = { .buf = target_arg };
3783 union semun arg;
3784 struct semid_ds dsarg;
3785 unsigned short *array = NULL;
3786 struct seminfo seminfo;
3787 abi_long ret = -TARGET_EINVAL;
3788 abi_long err;
3789 cmd &= 0xff;
3791 switch( cmd ) {
3792 case GETVAL:
3793 case SETVAL:
3794 /* In 64 bit cross-endian situations, we will erroneously pick up
3795 * the wrong half of the union for the "val" element. To rectify
3796 * this, the entire 8-byte structure is byteswapped, followed by
3797 * a swap of the 4 byte val field. In other cases, the data is
3798 * already in proper host byte order. */
3799 if (sizeof(target_su.val) != (sizeof(target_su.buf))) {
3800 target_su.buf = tswapal(target_su.buf);
3801 arg.val = tswap32(target_su.val);
3802 } else {
3803 arg.val = target_su.val;
3805 ret = get_errno(semctl(semid, semnum, cmd, arg));
3806 break;
3807 case GETALL:
3808 case SETALL:
3809 err = target_to_host_semarray(semid, &array, target_su.array);
3810 if (err)
3811 return err;
3812 arg.array = array;
3813 ret = get_errno(semctl(semid, semnum, cmd, arg));
3814 err = host_to_target_semarray(semid, target_su.array, &array);
3815 if (err)
3816 return err;
3817 break;
3818 case IPC_STAT:
3819 case IPC_SET:
3820 case SEM_STAT:
3821 err = target_to_host_semid_ds(&dsarg, target_su.buf);
3822 if (err)
3823 return err;
3824 arg.buf = &dsarg;
3825 ret = get_errno(semctl(semid, semnum, cmd, arg));
3826 err = host_to_target_semid_ds(target_su.buf, &dsarg);
3827 if (err)
3828 return err;
3829 break;
3830 case IPC_INFO:
3831 case SEM_INFO:
3832 arg.__buf = &seminfo;
3833 ret = get_errno(semctl(semid, semnum, cmd, arg));
3834 err = host_to_target_seminfo(target_su.__buf, &seminfo);
3835 if (err)
3836 return err;
3837 break;
3838 case IPC_RMID:
3839 case GETPID:
3840 case GETNCNT:
3841 case GETZCNT:
3842 ret = get_errno(semctl(semid, semnum, cmd, NULL));
3843 break;
3846 return ret;
3849 struct target_sembuf {
3850 unsigned short sem_num;
3851 short sem_op;
3852 short sem_flg;
3855 static inline abi_long target_to_host_sembuf(struct sembuf *host_sembuf,
3856 abi_ulong target_addr,
3857 unsigned nsops)
3859 struct target_sembuf *target_sembuf;
3860 int i;
3862 target_sembuf = lock_user(VERIFY_READ, target_addr,
3863 nsops*sizeof(struct target_sembuf), 1);
3864 if (!target_sembuf)
3865 return -TARGET_EFAULT;
3867 for(i=0; i<nsops; i++) {
3868 __get_user(host_sembuf[i].sem_num, &target_sembuf[i].sem_num);
3869 __get_user(host_sembuf[i].sem_op, &target_sembuf[i].sem_op);
3870 __get_user(host_sembuf[i].sem_flg, &target_sembuf[i].sem_flg);
3873 unlock_user(target_sembuf, target_addr, 0);
3875 return 0;
3878 static inline abi_long do_semop(int semid, abi_long ptr, unsigned nsops)
3880 struct sembuf sops[nsops];
3881 abi_long ret;
3883 if (target_to_host_sembuf(sops, ptr, nsops))
3884 return -TARGET_EFAULT;
3886 ret = -TARGET_ENOSYS;
3887 #ifdef __NR_semtimedop
3888 ret = get_errno(safe_semtimedop(semid, sops, nsops, NULL));
3889 #endif
3890 #ifdef __NR_ipc
3891 if (ret == -TARGET_ENOSYS) {
3892 ret = get_errno(safe_ipc(IPCOP_semtimedop, semid, nsops, 0, sops, 0));
3894 #endif
3895 return ret;
3898 struct target_msqid_ds
3900 struct target_ipc_perm msg_perm;
3901 abi_ulong msg_stime;
3902 #if TARGET_ABI_BITS == 32
3903 abi_ulong __unused1;
3904 #endif
3905 abi_ulong msg_rtime;
3906 #if TARGET_ABI_BITS == 32
3907 abi_ulong __unused2;
3908 #endif
3909 abi_ulong msg_ctime;
3910 #if TARGET_ABI_BITS == 32
3911 abi_ulong __unused3;
3912 #endif
3913 abi_ulong __msg_cbytes;
3914 abi_ulong msg_qnum;
3915 abi_ulong msg_qbytes;
3916 abi_ulong msg_lspid;
3917 abi_ulong msg_lrpid;
3918 abi_ulong __unused4;
3919 abi_ulong __unused5;
3922 static inline abi_long target_to_host_msqid_ds(struct msqid_ds *host_md,
3923 abi_ulong target_addr)
3925 struct target_msqid_ds *target_md;
3927 if (!lock_user_struct(VERIFY_READ, target_md, target_addr, 1))
3928 return -TARGET_EFAULT;
3929 if (target_to_host_ipc_perm(&(host_md->msg_perm),target_addr))
3930 return -TARGET_EFAULT;
3931 host_md->msg_stime = tswapal(target_md->msg_stime);
3932 host_md->msg_rtime = tswapal(target_md->msg_rtime);
3933 host_md->msg_ctime = tswapal(target_md->msg_ctime);
3934 host_md->__msg_cbytes = tswapal(target_md->__msg_cbytes);
3935 host_md->msg_qnum = tswapal(target_md->msg_qnum);
3936 host_md->msg_qbytes = tswapal(target_md->msg_qbytes);
3937 host_md->msg_lspid = tswapal(target_md->msg_lspid);
3938 host_md->msg_lrpid = tswapal(target_md->msg_lrpid);
3939 unlock_user_struct(target_md, target_addr, 0);
3940 return 0;
3943 static inline abi_long host_to_target_msqid_ds(abi_ulong target_addr,
3944 struct msqid_ds *host_md)
3946 struct target_msqid_ds *target_md;
3948 if (!lock_user_struct(VERIFY_WRITE, target_md, target_addr, 0))
3949 return -TARGET_EFAULT;
3950 if (host_to_target_ipc_perm(target_addr,&(host_md->msg_perm)))
3951 return -TARGET_EFAULT;
3952 target_md->msg_stime = tswapal(host_md->msg_stime);
3953 target_md->msg_rtime = tswapal(host_md->msg_rtime);
3954 target_md->msg_ctime = tswapal(host_md->msg_ctime);
3955 target_md->__msg_cbytes = tswapal(host_md->__msg_cbytes);
3956 target_md->msg_qnum = tswapal(host_md->msg_qnum);
3957 target_md->msg_qbytes = tswapal(host_md->msg_qbytes);
3958 target_md->msg_lspid = tswapal(host_md->msg_lspid);
3959 target_md->msg_lrpid = tswapal(host_md->msg_lrpid);
3960 unlock_user_struct(target_md, target_addr, 1);
3961 return 0;
3964 struct target_msginfo {
3965 int msgpool;
3966 int msgmap;
3967 int msgmax;
3968 int msgmnb;
3969 int msgmni;
3970 int msgssz;
3971 int msgtql;
3972 unsigned short int msgseg;
3975 static inline abi_long host_to_target_msginfo(abi_ulong target_addr,
3976 struct msginfo *host_msginfo)
3978 struct target_msginfo *target_msginfo;
3979 if (!lock_user_struct(VERIFY_WRITE, target_msginfo, target_addr, 0))
3980 return -TARGET_EFAULT;
3981 __put_user(host_msginfo->msgpool, &target_msginfo->msgpool);
3982 __put_user(host_msginfo->msgmap, &target_msginfo->msgmap);
3983 __put_user(host_msginfo->msgmax, &target_msginfo->msgmax);
3984 __put_user(host_msginfo->msgmnb, &target_msginfo->msgmnb);
3985 __put_user(host_msginfo->msgmni, &target_msginfo->msgmni);
3986 __put_user(host_msginfo->msgssz, &target_msginfo->msgssz);
3987 __put_user(host_msginfo->msgtql, &target_msginfo->msgtql);
3988 __put_user(host_msginfo->msgseg, &target_msginfo->msgseg);
3989 unlock_user_struct(target_msginfo, target_addr, 1);
3990 return 0;
3993 static inline abi_long do_msgctl(int msgid, int cmd, abi_long ptr)
3995 struct msqid_ds dsarg;
3996 struct msginfo msginfo;
3997 abi_long ret = -TARGET_EINVAL;
3999 cmd &= 0xff;
4001 switch (cmd) {
4002 case IPC_STAT:
4003 case IPC_SET:
4004 case MSG_STAT:
4005 if (target_to_host_msqid_ds(&dsarg,ptr))
4006 return -TARGET_EFAULT;
4007 ret = get_errno(msgctl(msgid, cmd, &dsarg));
4008 if (host_to_target_msqid_ds(ptr,&dsarg))
4009 return -TARGET_EFAULT;
4010 break;
4011 case IPC_RMID:
4012 ret = get_errno(msgctl(msgid, cmd, NULL));
4013 break;
4014 case IPC_INFO:
4015 case MSG_INFO:
4016 ret = get_errno(msgctl(msgid, cmd, (struct msqid_ds *)&msginfo));
4017 if (host_to_target_msginfo(ptr, &msginfo))
4018 return -TARGET_EFAULT;
4019 break;
4022 return ret;
4025 struct target_msgbuf {
4026 abi_long mtype;
4027 char mtext[1];
4030 static inline abi_long do_msgsnd(int msqid, abi_long msgp,
4031 ssize_t msgsz, int msgflg)
4033 struct target_msgbuf *target_mb;
4034 struct msgbuf *host_mb;
4035 abi_long ret = 0;
4037 if (msgsz < 0) {
4038 return -TARGET_EINVAL;
4041 if (!lock_user_struct(VERIFY_READ, target_mb, msgp, 0))
4042 return -TARGET_EFAULT;
4043 host_mb = g_try_malloc(msgsz + sizeof(long));
4044 if (!host_mb) {
4045 unlock_user_struct(target_mb, msgp, 0);
4046 return -TARGET_ENOMEM;
4048 host_mb->mtype = (abi_long) tswapal(target_mb->mtype);
4049 memcpy(host_mb->mtext, target_mb->mtext, msgsz);
4050 ret = -TARGET_ENOSYS;
4051 #ifdef __NR_msgsnd
4052 ret = get_errno(safe_msgsnd(msqid, host_mb, msgsz, msgflg));
4053 #endif
4054 #ifdef __NR_ipc
4055 if (ret == -TARGET_ENOSYS) {
4056 ret = get_errno(safe_ipc(IPCOP_msgsnd, msqid, msgsz, msgflg,
4057 host_mb, 0));
4059 #endif
4060 g_free(host_mb);
4061 unlock_user_struct(target_mb, msgp, 0);
4063 return ret;
4066 static inline abi_long do_msgrcv(int msqid, abi_long msgp,
4067 ssize_t msgsz, abi_long msgtyp,
4068 int msgflg)
4070 struct target_msgbuf *target_mb;
4071 char *target_mtext;
4072 struct msgbuf *host_mb;
4073 abi_long ret = 0;
4075 if (msgsz < 0) {
4076 return -TARGET_EINVAL;
4079 if (!lock_user_struct(VERIFY_WRITE, target_mb, msgp, 0))
4080 return -TARGET_EFAULT;
4082 host_mb = g_try_malloc(msgsz + sizeof(long));
4083 if (!host_mb) {
4084 ret = -TARGET_ENOMEM;
4085 goto end;
4087 ret = -TARGET_ENOSYS;
4088 #ifdef __NR_msgrcv
4089 ret = get_errno(safe_msgrcv(msqid, host_mb, msgsz, msgtyp, msgflg));
4090 #endif
4091 #ifdef __NR_ipc
4092 if (ret == -TARGET_ENOSYS) {
4093 ret = get_errno(safe_ipc(IPCOP_CALL(1, IPCOP_msgrcv), msqid, msgsz,
4094 msgflg, host_mb, msgtyp));
4096 #endif
4098 if (ret > 0) {
4099 abi_ulong target_mtext_addr = msgp + sizeof(abi_ulong);
4100 target_mtext = lock_user(VERIFY_WRITE, target_mtext_addr, ret, 0);
4101 if (!target_mtext) {
4102 ret = -TARGET_EFAULT;
4103 goto end;
4105 memcpy(target_mb->mtext, host_mb->mtext, ret);
4106 unlock_user(target_mtext, target_mtext_addr, ret);
4109 target_mb->mtype = tswapal(host_mb->mtype);
4111 end:
4112 if (target_mb)
4113 unlock_user_struct(target_mb, msgp, 1);
4114 g_free(host_mb);
4115 return ret;
4118 static inline abi_long target_to_host_shmid_ds(struct shmid_ds *host_sd,
4119 abi_ulong target_addr)
4121 struct target_shmid_ds *target_sd;
4123 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1))
4124 return -TARGET_EFAULT;
4125 if (target_to_host_ipc_perm(&(host_sd->shm_perm), target_addr))
4126 return -TARGET_EFAULT;
4127 __get_user(host_sd->shm_segsz, &target_sd->shm_segsz);
4128 __get_user(host_sd->shm_atime, &target_sd->shm_atime);
4129 __get_user(host_sd->shm_dtime, &target_sd->shm_dtime);
4130 __get_user(host_sd->shm_ctime, &target_sd->shm_ctime);
4131 __get_user(host_sd->shm_cpid, &target_sd->shm_cpid);
4132 __get_user(host_sd->shm_lpid, &target_sd->shm_lpid);
4133 __get_user(host_sd->shm_nattch, &target_sd->shm_nattch);
4134 unlock_user_struct(target_sd, target_addr, 0);
4135 return 0;
4138 static inline abi_long host_to_target_shmid_ds(abi_ulong target_addr,
4139 struct shmid_ds *host_sd)
4141 struct target_shmid_ds *target_sd;
4143 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0))
4144 return -TARGET_EFAULT;
4145 if (host_to_target_ipc_perm(target_addr, &(host_sd->shm_perm)))
4146 return -TARGET_EFAULT;
4147 __put_user(host_sd->shm_segsz, &target_sd->shm_segsz);
4148 __put_user(host_sd->shm_atime, &target_sd->shm_atime);
4149 __put_user(host_sd->shm_dtime, &target_sd->shm_dtime);
4150 __put_user(host_sd->shm_ctime, &target_sd->shm_ctime);
4151 __put_user(host_sd->shm_cpid, &target_sd->shm_cpid);
4152 __put_user(host_sd->shm_lpid, &target_sd->shm_lpid);
4153 __put_user(host_sd->shm_nattch, &target_sd->shm_nattch);
4154 unlock_user_struct(target_sd, target_addr, 1);
4155 return 0;
4158 struct target_shminfo {
4159 abi_ulong shmmax;
4160 abi_ulong shmmin;
4161 abi_ulong shmmni;
4162 abi_ulong shmseg;
4163 abi_ulong shmall;
4166 static inline abi_long host_to_target_shminfo(abi_ulong target_addr,
4167 struct shminfo *host_shminfo)
4169 struct target_shminfo *target_shminfo;
4170 if (!lock_user_struct(VERIFY_WRITE, target_shminfo, target_addr, 0))
4171 return -TARGET_EFAULT;
4172 __put_user(host_shminfo->shmmax, &target_shminfo->shmmax);
4173 __put_user(host_shminfo->shmmin, &target_shminfo->shmmin);
4174 __put_user(host_shminfo->shmmni, &target_shminfo->shmmni);
4175 __put_user(host_shminfo->shmseg, &target_shminfo->shmseg);
4176 __put_user(host_shminfo->shmall, &target_shminfo->shmall);
4177 unlock_user_struct(target_shminfo, target_addr, 1);
4178 return 0;
4181 struct target_shm_info {
4182 int used_ids;
4183 abi_ulong shm_tot;
4184 abi_ulong shm_rss;
4185 abi_ulong shm_swp;
4186 abi_ulong swap_attempts;
4187 abi_ulong swap_successes;
4190 static inline abi_long host_to_target_shm_info(abi_ulong target_addr,
4191 struct shm_info *host_shm_info)
4193 struct target_shm_info *target_shm_info;
4194 if (!lock_user_struct(VERIFY_WRITE, target_shm_info, target_addr, 0))
4195 return -TARGET_EFAULT;
4196 __put_user(host_shm_info->used_ids, &target_shm_info->used_ids);
4197 __put_user(host_shm_info->shm_tot, &target_shm_info->shm_tot);
4198 __put_user(host_shm_info->shm_rss, &target_shm_info->shm_rss);
4199 __put_user(host_shm_info->shm_swp, &target_shm_info->shm_swp);
4200 __put_user(host_shm_info->swap_attempts, &target_shm_info->swap_attempts);
4201 __put_user(host_shm_info->swap_successes, &target_shm_info->swap_successes);
4202 unlock_user_struct(target_shm_info, target_addr, 1);
4203 return 0;
4206 static inline abi_long do_shmctl(int shmid, int cmd, abi_long buf)
4208 struct shmid_ds dsarg;
4209 struct shminfo shminfo;
4210 struct shm_info shm_info;
4211 abi_long ret = -TARGET_EINVAL;
4213 cmd &= 0xff;
4215 switch(cmd) {
4216 case IPC_STAT:
4217 case IPC_SET:
4218 case SHM_STAT:
4219 if (target_to_host_shmid_ds(&dsarg, buf))
4220 return -TARGET_EFAULT;
4221 ret = get_errno(shmctl(shmid, cmd, &dsarg));
4222 if (host_to_target_shmid_ds(buf, &dsarg))
4223 return -TARGET_EFAULT;
4224 break;
4225 case IPC_INFO:
4226 ret = get_errno(shmctl(shmid, cmd, (struct shmid_ds *)&shminfo));
4227 if (host_to_target_shminfo(buf, &shminfo))
4228 return -TARGET_EFAULT;
4229 break;
4230 case SHM_INFO:
4231 ret = get_errno(shmctl(shmid, cmd, (struct shmid_ds *)&shm_info));
4232 if (host_to_target_shm_info(buf, &shm_info))
4233 return -TARGET_EFAULT;
4234 break;
4235 case IPC_RMID:
4236 case SHM_LOCK:
4237 case SHM_UNLOCK:
4238 ret = get_errno(shmctl(shmid, cmd, NULL));
4239 break;
4242 return ret;
4245 #ifndef TARGET_FORCE_SHMLBA
4246 /* For most architectures, SHMLBA is the same as the page size;
4247 * some architectures have larger values, in which case they should
4248 * define TARGET_FORCE_SHMLBA and provide a target_shmlba() function.
4249 * This corresponds to the kernel arch code defining __ARCH_FORCE_SHMLBA
4250 * and defining its own value for SHMLBA.
4252 * The kernel also permits SHMLBA to be set by the architecture to a
4253 * value larger than the page size without setting __ARCH_FORCE_SHMLBA;
4254 * this means that addresses are rounded to the large size if
4255 * SHM_RND is set but addresses not aligned to that size are not rejected
4256 * as long as they are at least page-aligned. Since the only architecture
4257 * which uses this is ia64 this code doesn't provide for that oddity.
4259 static inline abi_ulong target_shmlba(CPUArchState *cpu_env)
4261 return TARGET_PAGE_SIZE;
4263 #endif
4265 static inline abi_ulong do_shmat(CPUArchState *cpu_env,
4266 int shmid, abi_ulong shmaddr, int shmflg)
4268 abi_long raddr;
4269 void *host_raddr;
4270 struct shmid_ds shm_info;
4271 int i,ret;
4272 abi_ulong shmlba;
4274 /* find out the length of the shared memory segment */
4275 ret = get_errno(shmctl(shmid, IPC_STAT, &shm_info));
4276 if (is_error(ret)) {
4277 /* can't get length, bail out */
4278 return ret;
4281 shmlba = target_shmlba(cpu_env);
4283 if (shmaddr & (shmlba - 1)) {
4284 if (shmflg & SHM_RND) {
4285 shmaddr &= ~(shmlba - 1);
4286 } else {
4287 return -TARGET_EINVAL;
4290 if (!guest_range_valid(shmaddr, shm_info.shm_segsz)) {
4291 return -TARGET_EINVAL;
4294 mmap_lock();
4296 if (shmaddr)
4297 host_raddr = shmat(shmid, (void *)g2h(shmaddr), shmflg);
4298 else {
4299 abi_ulong mmap_start;
4301 /* In order to use the host shmat, we need to honor host SHMLBA. */
4302 mmap_start = mmap_find_vma(0, shm_info.shm_segsz, MAX(SHMLBA, shmlba));
4304 if (mmap_start == -1) {
4305 errno = ENOMEM;
4306 host_raddr = (void *)-1;
4307 } else
4308 host_raddr = shmat(shmid, g2h(mmap_start), shmflg | SHM_REMAP);
4311 if (host_raddr == (void *)-1) {
4312 mmap_unlock();
4313 return get_errno((long)host_raddr);
4315 raddr=h2g((unsigned long)host_raddr);
4317 page_set_flags(raddr, raddr + shm_info.shm_segsz,
4318 PAGE_VALID | PAGE_READ |
4319 ((shmflg & SHM_RDONLY)? 0 : PAGE_WRITE));
4321 for (i = 0; i < N_SHM_REGIONS; i++) {
4322 if (!shm_regions[i].in_use) {
4323 shm_regions[i].in_use = true;
4324 shm_regions[i].start = raddr;
4325 shm_regions[i].size = shm_info.shm_segsz;
4326 break;
4330 mmap_unlock();
4331 return raddr;
4335 static inline abi_long do_shmdt(abi_ulong shmaddr)
4337 int i;
4338 abi_long rv;
4340 mmap_lock();
4342 for (i = 0; i < N_SHM_REGIONS; ++i) {
4343 if (shm_regions[i].in_use && shm_regions[i].start == shmaddr) {
4344 shm_regions[i].in_use = false;
4345 page_set_flags(shmaddr, shmaddr + shm_regions[i].size, 0);
4346 break;
4349 rv = get_errno(shmdt(g2h(shmaddr)));
4351 mmap_unlock();
4353 return rv;
4356 #ifdef TARGET_NR_ipc
4357 /* ??? This only works with linear mappings. */
4358 /* do_ipc() must return target values and target errnos. */
4359 static abi_long do_ipc(CPUArchState *cpu_env,
4360 unsigned int call, abi_long first,
4361 abi_long second, abi_long third,
4362 abi_long ptr, abi_long fifth)
4364 int version;
4365 abi_long ret = 0;
4367 version = call >> 16;
4368 call &= 0xffff;
4370 switch (call) {
4371 case IPCOP_semop:
4372 ret = do_semop(first, ptr, second);
4373 break;
4375 case IPCOP_semget:
4376 ret = get_errno(semget(first, second, third));
4377 break;
4379 case IPCOP_semctl: {
4380 /* The semun argument to semctl is passed by value, so dereference the
4381 * ptr argument. */
4382 abi_ulong atptr;
4383 get_user_ual(atptr, ptr);
4384 ret = do_semctl(first, second, third, atptr);
4385 break;
4388 case IPCOP_msgget:
4389 ret = get_errno(msgget(first, second));
4390 break;
4392 case IPCOP_msgsnd:
4393 ret = do_msgsnd(first, ptr, second, third);
4394 break;
4396 case IPCOP_msgctl:
4397 ret = do_msgctl(first, second, ptr);
4398 break;
4400 case IPCOP_msgrcv:
4401 switch (version) {
4402 case 0:
4404 struct target_ipc_kludge {
4405 abi_long msgp;
4406 abi_long msgtyp;
4407 } *tmp;
4409 if (!lock_user_struct(VERIFY_READ, tmp, ptr, 1)) {
4410 ret = -TARGET_EFAULT;
4411 break;
4414 ret = do_msgrcv(first, tswapal(tmp->msgp), second, tswapal(tmp->msgtyp), third);
4416 unlock_user_struct(tmp, ptr, 0);
4417 break;
4419 default:
4420 ret = do_msgrcv(first, ptr, second, fifth, third);
4422 break;
4424 case IPCOP_shmat:
4425 switch (version) {
4426 default:
4428 abi_ulong raddr;
4429 raddr = do_shmat(cpu_env, first, ptr, second);
4430 if (is_error(raddr))
4431 return get_errno(raddr);
4432 if (put_user_ual(raddr, third))
4433 return -TARGET_EFAULT;
4434 break;
4436 case 1:
4437 ret = -TARGET_EINVAL;
4438 break;
4440 break;
4441 case IPCOP_shmdt:
4442 ret = do_shmdt(ptr);
4443 break;
4445 case IPCOP_shmget:
4446 /* IPC_* flag values are the same on all linux platforms */
4447 ret = get_errno(shmget(first, second, third));
4448 break;
4450 /* IPC_* and SHM_* command values are the same on all linux platforms */
4451 case IPCOP_shmctl:
4452 ret = do_shmctl(first, second, ptr);
4453 break;
4454 default:
4455 qemu_log_mask(LOG_UNIMP, "Unsupported ipc call: %d (version %d)\n",
4456 call, version);
4457 ret = -TARGET_ENOSYS;
4458 break;
4460 return ret;
4462 #endif
4464 /* kernel structure types definitions */
4466 #define STRUCT(name, ...) STRUCT_ ## name,
4467 #define STRUCT_SPECIAL(name) STRUCT_ ## name,
4468 enum {
4469 #include "syscall_types.h"
4470 STRUCT_MAX
4472 #undef STRUCT
4473 #undef STRUCT_SPECIAL
4475 #define STRUCT(name, ...) static const argtype struct_ ## name ## _def[] = { __VA_ARGS__, TYPE_NULL };
4476 #define STRUCT_SPECIAL(name)
4477 #include "syscall_types.h"
4478 #undef STRUCT
4479 #undef STRUCT_SPECIAL
4481 typedef struct IOCTLEntry IOCTLEntry;
4483 typedef abi_long do_ioctl_fn(const IOCTLEntry *ie, uint8_t *buf_temp,
4484 int fd, int cmd, abi_long arg);
4486 struct IOCTLEntry {
4487 int target_cmd;
4488 unsigned int host_cmd;
4489 const char *name;
4490 int access;
4491 do_ioctl_fn *do_ioctl;
4492 const argtype arg_type[5];
4495 #define IOC_R 0x0001
4496 #define IOC_W 0x0002
4497 #define IOC_RW (IOC_R | IOC_W)
4499 #define MAX_STRUCT_SIZE 4096
4501 #ifdef CONFIG_FIEMAP
4502 /* So fiemap access checks don't overflow on 32 bit systems.
4503 * This is very slightly smaller than the limit imposed by
4504 * the underlying kernel.
4506 #define FIEMAP_MAX_EXTENTS ((UINT_MAX - sizeof(struct fiemap)) \
4507 / sizeof(struct fiemap_extent))
4509 static abi_long do_ioctl_fs_ioc_fiemap(const IOCTLEntry *ie, uint8_t *buf_temp,
4510 int fd, int cmd, abi_long arg)
4512 /* The parameter for this ioctl is a struct fiemap followed
4513 * by an array of struct fiemap_extent whose size is set
4514 * in fiemap->fm_extent_count. The array is filled in by the
4515 * ioctl.
4517 int target_size_in, target_size_out;
4518 struct fiemap *fm;
4519 const argtype *arg_type = ie->arg_type;
4520 const argtype extent_arg_type[] = { MK_STRUCT(STRUCT_fiemap_extent) };
4521 void *argptr, *p;
4522 abi_long ret;
4523 int i, extent_size = thunk_type_size(extent_arg_type, 0);
4524 uint32_t outbufsz;
4525 int free_fm = 0;
4527 assert(arg_type[0] == TYPE_PTR);
4528 assert(ie->access == IOC_RW);
4529 arg_type++;
4530 target_size_in = thunk_type_size(arg_type, 0);
4531 argptr = lock_user(VERIFY_READ, arg, target_size_in, 1);
4532 if (!argptr) {
4533 return -TARGET_EFAULT;
4535 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
4536 unlock_user(argptr, arg, 0);
4537 fm = (struct fiemap *)buf_temp;
4538 if (fm->fm_extent_count > FIEMAP_MAX_EXTENTS) {
4539 return -TARGET_EINVAL;
4542 outbufsz = sizeof (*fm) +
4543 (sizeof(struct fiemap_extent) * fm->fm_extent_count);
4545 if (outbufsz > MAX_STRUCT_SIZE) {
4546 /* We can't fit all the extents into the fixed size buffer.
4547 * Allocate one that is large enough and use it instead.
4549 fm = g_try_malloc(outbufsz);
4550 if (!fm) {
4551 return -TARGET_ENOMEM;
4553 memcpy(fm, buf_temp, sizeof(struct fiemap));
4554 free_fm = 1;
4556 ret = get_errno(safe_ioctl(fd, ie->host_cmd, fm));
4557 if (!is_error(ret)) {
4558 target_size_out = target_size_in;
4559 /* An extent_count of 0 means we were only counting the extents
4560 * so there are no structs to copy
4562 if (fm->fm_extent_count != 0) {
4563 target_size_out += fm->fm_mapped_extents * extent_size;
4565 argptr = lock_user(VERIFY_WRITE, arg, target_size_out, 0);
4566 if (!argptr) {
4567 ret = -TARGET_EFAULT;
4568 } else {
4569 /* Convert the struct fiemap */
4570 thunk_convert(argptr, fm, arg_type, THUNK_TARGET);
4571 if (fm->fm_extent_count != 0) {
4572 p = argptr + target_size_in;
4573 /* ...and then all the struct fiemap_extents */
4574 for (i = 0; i < fm->fm_mapped_extents; i++) {
4575 thunk_convert(p, &fm->fm_extents[i], extent_arg_type,
4576 THUNK_TARGET);
4577 p += extent_size;
4580 unlock_user(argptr, arg, target_size_out);
4583 if (free_fm) {
4584 g_free(fm);
4586 return ret;
4588 #endif
4590 static abi_long do_ioctl_ifconf(const IOCTLEntry *ie, uint8_t *buf_temp,
4591 int fd, int cmd, abi_long arg)
4593 const argtype *arg_type = ie->arg_type;
4594 int target_size;
4595 void *argptr;
4596 int ret;
4597 struct ifconf *host_ifconf;
4598 uint32_t outbufsz;
4599 const argtype ifreq_arg_type[] = { MK_STRUCT(STRUCT_sockaddr_ifreq) };
4600 int target_ifreq_size;
4601 int nb_ifreq;
4602 int free_buf = 0;
4603 int i;
4604 int target_ifc_len;
4605 abi_long target_ifc_buf;
4606 int host_ifc_len;
4607 char *host_ifc_buf;
4609 assert(arg_type[0] == TYPE_PTR);
4610 assert(ie->access == IOC_RW);
4612 arg_type++;
4613 target_size = thunk_type_size(arg_type, 0);
4615 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
4616 if (!argptr)
4617 return -TARGET_EFAULT;
4618 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
4619 unlock_user(argptr, arg, 0);
4621 host_ifconf = (struct ifconf *)(unsigned long)buf_temp;
4622 target_ifc_buf = (abi_long)(unsigned long)host_ifconf->ifc_buf;
4623 target_ifreq_size = thunk_type_size(ifreq_arg_type, 0);
4625 if (target_ifc_buf != 0) {
4626 target_ifc_len = host_ifconf->ifc_len;
4627 nb_ifreq = target_ifc_len / target_ifreq_size;
4628 host_ifc_len = nb_ifreq * sizeof(struct ifreq);
4630 outbufsz = sizeof(*host_ifconf) + host_ifc_len;
4631 if (outbufsz > MAX_STRUCT_SIZE) {
4633 * We can't fit all the extents into the fixed size buffer.
4634 * Allocate one that is large enough and use it instead.
4636 host_ifconf = malloc(outbufsz);
4637 if (!host_ifconf) {
4638 return -TARGET_ENOMEM;
4640 memcpy(host_ifconf, buf_temp, sizeof(*host_ifconf));
4641 free_buf = 1;
4643 host_ifc_buf = (char *)host_ifconf + sizeof(*host_ifconf);
4645 host_ifconf->ifc_len = host_ifc_len;
4646 } else {
4647 host_ifc_buf = NULL;
4649 host_ifconf->ifc_buf = host_ifc_buf;
4651 ret = get_errno(safe_ioctl(fd, ie->host_cmd, host_ifconf));
4652 if (!is_error(ret)) {
4653 /* convert host ifc_len to target ifc_len */
4655 nb_ifreq = host_ifconf->ifc_len / sizeof(struct ifreq);
4656 target_ifc_len = nb_ifreq * target_ifreq_size;
4657 host_ifconf->ifc_len = target_ifc_len;
4659 /* restore target ifc_buf */
4661 host_ifconf->ifc_buf = (char *)(unsigned long)target_ifc_buf;
4663 /* copy struct ifconf to target user */
4665 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
4666 if (!argptr)
4667 return -TARGET_EFAULT;
4668 thunk_convert(argptr, host_ifconf, arg_type, THUNK_TARGET);
4669 unlock_user(argptr, arg, target_size);
4671 if (target_ifc_buf != 0) {
4672 /* copy ifreq[] to target user */
4673 argptr = lock_user(VERIFY_WRITE, target_ifc_buf, target_ifc_len, 0);
4674 for (i = 0; i < nb_ifreq ; i++) {
4675 thunk_convert(argptr + i * target_ifreq_size,
4676 host_ifc_buf + i * sizeof(struct ifreq),
4677 ifreq_arg_type, THUNK_TARGET);
4679 unlock_user(argptr, target_ifc_buf, target_ifc_len);
4683 if (free_buf) {
4684 free(host_ifconf);
4687 return ret;
4690 #if defined(CONFIG_USBFS)
4691 #if HOST_LONG_BITS > 64
4692 #error USBDEVFS thunks do not support >64 bit hosts yet.
4693 #endif
4694 struct live_urb {
4695 uint64_t target_urb_adr;
4696 uint64_t target_buf_adr;
4697 char *target_buf_ptr;
4698 struct usbdevfs_urb host_urb;
4701 static GHashTable *usbdevfs_urb_hashtable(void)
4703 static GHashTable *urb_hashtable;
4705 if (!urb_hashtable) {
4706 urb_hashtable = g_hash_table_new(g_int64_hash, g_int64_equal);
4708 return urb_hashtable;
4711 static void urb_hashtable_insert(struct live_urb *urb)
4713 GHashTable *urb_hashtable = usbdevfs_urb_hashtable();
4714 g_hash_table_insert(urb_hashtable, urb, urb);
4717 static struct live_urb *urb_hashtable_lookup(uint64_t target_urb_adr)
4719 GHashTable *urb_hashtable = usbdevfs_urb_hashtable();
4720 return g_hash_table_lookup(urb_hashtable, &target_urb_adr);
4723 static void urb_hashtable_remove(struct live_urb *urb)
4725 GHashTable *urb_hashtable = usbdevfs_urb_hashtable();
4726 g_hash_table_remove(urb_hashtable, urb);
4729 static abi_long
4730 do_ioctl_usbdevfs_reapurb(const IOCTLEntry *ie, uint8_t *buf_temp,
4731 int fd, int cmd, abi_long arg)
4733 const argtype usbfsurb_arg_type[] = { MK_STRUCT(STRUCT_usbdevfs_urb) };
4734 const argtype ptrvoid_arg_type[] = { TYPE_PTRVOID, 0, 0 };
4735 struct live_urb *lurb;
4736 void *argptr;
4737 uint64_t hurb;
4738 int target_size;
4739 uintptr_t target_urb_adr;
4740 abi_long ret;
4742 target_size = thunk_type_size(usbfsurb_arg_type, THUNK_TARGET);
4744 memset(buf_temp, 0, sizeof(uint64_t));
4745 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
4746 if (is_error(ret)) {
4747 return ret;
4750 memcpy(&hurb, buf_temp, sizeof(uint64_t));
4751 lurb = (void *)((uintptr_t)hurb - offsetof(struct live_urb, host_urb));
4752 if (!lurb->target_urb_adr) {
4753 return -TARGET_EFAULT;
4755 urb_hashtable_remove(lurb);
4756 unlock_user(lurb->target_buf_ptr, lurb->target_buf_adr,
4757 lurb->host_urb.buffer_length);
4758 lurb->target_buf_ptr = NULL;
4760 /* restore the guest buffer pointer */
4761 lurb->host_urb.buffer = (void *)(uintptr_t)lurb->target_buf_adr;
4763 /* update the guest urb struct */
4764 argptr = lock_user(VERIFY_WRITE, lurb->target_urb_adr, target_size, 0);
4765 if (!argptr) {
4766 g_free(lurb);
4767 return -TARGET_EFAULT;
4769 thunk_convert(argptr, &lurb->host_urb, usbfsurb_arg_type, THUNK_TARGET);
4770 unlock_user(argptr, lurb->target_urb_adr, target_size);
4772 target_size = thunk_type_size(ptrvoid_arg_type, THUNK_TARGET);
4773 /* write back the urb handle */
4774 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
4775 if (!argptr) {
4776 g_free(lurb);
4777 return -TARGET_EFAULT;
4780 /* GHashTable uses 64-bit keys but thunk_convert expects uintptr_t */
4781 target_urb_adr = lurb->target_urb_adr;
4782 thunk_convert(argptr, &target_urb_adr, ptrvoid_arg_type, THUNK_TARGET);
4783 unlock_user(argptr, arg, target_size);
4785 g_free(lurb);
4786 return ret;
4789 static abi_long
4790 do_ioctl_usbdevfs_discardurb(const IOCTLEntry *ie,
4791 uint8_t *buf_temp __attribute__((unused)),
4792 int fd, int cmd, abi_long arg)
4794 struct live_urb *lurb;
4796 /* map target address back to host URB with metadata. */
4797 lurb = urb_hashtable_lookup(arg);
4798 if (!lurb) {
4799 return -TARGET_EFAULT;
4801 return get_errno(safe_ioctl(fd, ie->host_cmd, &lurb->host_urb));
4804 static abi_long
4805 do_ioctl_usbdevfs_submiturb(const IOCTLEntry *ie, uint8_t *buf_temp,
4806 int fd, int cmd, abi_long arg)
4808 const argtype *arg_type = ie->arg_type;
4809 int target_size;
4810 abi_long ret;
4811 void *argptr;
4812 int rw_dir;
4813 struct live_urb *lurb;
4816 * each submitted URB needs to map to a unique ID for the
4817 * kernel, and that unique ID needs to be a pointer to
4818 * host memory. hence, we need to malloc for each URB.
4819 * isochronous transfers have a variable length struct.
4821 arg_type++;
4822 target_size = thunk_type_size(arg_type, THUNK_TARGET);
4824 /* construct host copy of urb and metadata */
4825 lurb = g_try_malloc0(sizeof(struct live_urb));
4826 if (!lurb) {
4827 return -TARGET_ENOMEM;
4830 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
4831 if (!argptr) {
4832 g_free(lurb);
4833 return -TARGET_EFAULT;
4835 thunk_convert(&lurb->host_urb, argptr, arg_type, THUNK_HOST);
4836 unlock_user(argptr, arg, 0);
4838 lurb->target_urb_adr = arg;
4839 lurb->target_buf_adr = (uintptr_t)lurb->host_urb.buffer;
4841 /* buffer space used depends on endpoint type so lock the entire buffer */
4842 /* control type urbs should check the buffer contents for true direction */
4843 rw_dir = lurb->host_urb.endpoint & USB_DIR_IN ? VERIFY_WRITE : VERIFY_READ;
4844 lurb->target_buf_ptr = lock_user(rw_dir, lurb->target_buf_adr,
4845 lurb->host_urb.buffer_length, 1);
4846 if (lurb->target_buf_ptr == NULL) {
4847 g_free(lurb);
4848 return -TARGET_EFAULT;
4851 /* update buffer pointer in host copy */
4852 lurb->host_urb.buffer = lurb->target_buf_ptr;
4854 ret = get_errno(safe_ioctl(fd, ie->host_cmd, &lurb->host_urb));
4855 if (is_error(ret)) {
4856 unlock_user(lurb->target_buf_ptr, lurb->target_buf_adr, 0);
4857 g_free(lurb);
4858 } else {
4859 urb_hashtable_insert(lurb);
4862 return ret;
4864 #endif /* CONFIG_USBFS */
4866 static abi_long do_ioctl_dm(const IOCTLEntry *ie, uint8_t *buf_temp, int fd,
4867 int cmd, abi_long arg)
4869 void *argptr;
4870 struct dm_ioctl *host_dm;
4871 abi_long guest_data;
4872 uint32_t guest_data_size;
4873 int target_size;
4874 const argtype *arg_type = ie->arg_type;
4875 abi_long ret;
4876 void *big_buf = NULL;
4877 char *host_data;
4879 arg_type++;
4880 target_size = thunk_type_size(arg_type, 0);
4881 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
4882 if (!argptr) {
4883 ret = -TARGET_EFAULT;
4884 goto out;
4886 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
4887 unlock_user(argptr, arg, 0);
4889 /* buf_temp is too small, so fetch things into a bigger buffer */
4890 big_buf = g_malloc0(((struct dm_ioctl*)buf_temp)->data_size * 2);
4891 memcpy(big_buf, buf_temp, target_size);
4892 buf_temp = big_buf;
4893 host_dm = big_buf;
4895 guest_data = arg + host_dm->data_start;
4896 if ((guest_data - arg) < 0) {
4897 ret = -TARGET_EINVAL;
4898 goto out;
4900 guest_data_size = host_dm->data_size - host_dm->data_start;
4901 host_data = (char*)host_dm + host_dm->data_start;
4903 argptr = lock_user(VERIFY_READ, guest_data, guest_data_size, 1);
4904 if (!argptr) {
4905 ret = -TARGET_EFAULT;
4906 goto out;
4909 switch (ie->host_cmd) {
4910 case DM_REMOVE_ALL:
4911 case DM_LIST_DEVICES:
4912 case DM_DEV_CREATE:
4913 case DM_DEV_REMOVE:
4914 case DM_DEV_SUSPEND:
4915 case DM_DEV_STATUS:
4916 case DM_DEV_WAIT:
4917 case DM_TABLE_STATUS:
4918 case DM_TABLE_CLEAR:
4919 case DM_TABLE_DEPS:
4920 case DM_LIST_VERSIONS:
4921 /* no input data */
4922 break;
4923 case DM_DEV_RENAME:
4924 case DM_DEV_SET_GEOMETRY:
4925 /* data contains only strings */
4926 memcpy(host_data, argptr, guest_data_size);
4927 break;
4928 case DM_TARGET_MSG:
4929 memcpy(host_data, argptr, guest_data_size);
4930 *(uint64_t*)host_data = tswap64(*(uint64_t*)argptr);
4931 break;
4932 case DM_TABLE_LOAD:
4934 void *gspec = argptr;
4935 void *cur_data = host_data;
4936 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_target_spec) };
4937 int spec_size = thunk_type_size(arg_type, 0);
4938 int i;
4940 for (i = 0; i < host_dm->target_count; i++) {
4941 struct dm_target_spec *spec = cur_data;
4942 uint32_t next;
4943 int slen;
4945 thunk_convert(spec, gspec, arg_type, THUNK_HOST);
4946 slen = strlen((char*)gspec + spec_size) + 1;
4947 next = spec->next;
4948 spec->next = sizeof(*spec) + slen;
4949 strcpy((char*)&spec[1], gspec + spec_size);
4950 gspec += next;
4951 cur_data += spec->next;
4953 break;
4955 default:
4956 ret = -TARGET_EINVAL;
4957 unlock_user(argptr, guest_data, 0);
4958 goto out;
4960 unlock_user(argptr, guest_data, 0);
4962 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
4963 if (!is_error(ret)) {
4964 guest_data = arg + host_dm->data_start;
4965 guest_data_size = host_dm->data_size - host_dm->data_start;
4966 argptr = lock_user(VERIFY_WRITE, guest_data, guest_data_size, 0);
4967 switch (ie->host_cmd) {
4968 case DM_REMOVE_ALL:
4969 case DM_DEV_CREATE:
4970 case DM_DEV_REMOVE:
4971 case DM_DEV_RENAME:
4972 case DM_DEV_SUSPEND:
4973 case DM_DEV_STATUS:
4974 case DM_TABLE_LOAD:
4975 case DM_TABLE_CLEAR:
4976 case DM_TARGET_MSG:
4977 case DM_DEV_SET_GEOMETRY:
4978 /* no return data */
4979 break;
4980 case DM_LIST_DEVICES:
4982 struct dm_name_list *nl = (void*)host_dm + host_dm->data_start;
4983 uint32_t remaining_data = guest_data_size;
4984 void *cur_data = argptr;
4985 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_name_list) };
4986 int nl_size = 12; /* can't use thunk_size due to alignment */
4988 while (1) {
4989 uint32_t next = nl->next;
4990 if (next) {
4991 nl->next = nl_size + (strlen(nl->name) + 1);
4993 if (remaining_data < nl->next) {
4994 host_dm->flags |= DM_BUFFER_FULL_FLAG;
4995 break;
4997 thunk_convert(cur_data, nl, arg_type, THUNK_TARGET);
4998 strcpy(cur_data + nl_size, nl->name);
4999 cur_data += nl->next;
5000 remaining_data -= nl->next;
5001 if (!next) {
5002 break;
5004 nl = (void*)nl + next;
5006 break;
5008 case DM_DEV_WAIT:
5009 case DM_TABLE_STATUS:
5011 struct dm_target_spec *spec = (void*)host_dm + host_dm->data_start;
5012 void *cur_data = argptr;
5013 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_target_spec) };
5014 int spec_size = thunk_type_size(arg_type, 0);
5015 int i;
5017 for (i = 0; i < host_dm->target_count; i++) {
5018 uint32_t next = spec->next;
5019 int slen = strlen((char*)&spec[1]) + 1;
5020 spec->next = (cur_data - argptr) + spec_size + slen;
5021 if (guest_data_size < spec->next) {
5022 host_dm->flags |= DM_BUFFER_FULL_FLAG;
5023 break;
5025 thunk_convert(cur_data, spec, arg_type, THUNK_TARGET);
5026 strcpy(cur_data + spec_size, (char*)&spec[1]);
5027 cur_data = argptr + spec->next;
5028 spec = (void*)host_dm + host_dm->data_start + next;
5030 break;
5032 case DM_TABLE_DEPS:
5034 void *hdata = (void*)host_dm + host_dm->data_start;
5035 int count = *(uint32_t*)hdata;
5036 uint64_t *hdev = hdata + 8;
5037 uint64_t *gdev = argptr + 8;
5038 int i;
5040 *(uint32_t*)argptr = tswap32(count);
5041 for (i = 0; i < count; i++) {
5042 *gdev = tswap64(*hdev);
5043 gdev++;
5044 hdev++;
5046 break;
5048 case DM_LIST_VERSIONS:
5050 struct dm_target_versions *vers = (void*)host_dm + host_dm->data_start;
5051 uint32_t remaining_data = guest_data_size;
5052 void *cur_data = argptr;
5053 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_target_versions) };
5054 int vers_size = thunk_type_size(arg_type, 0);
5056 while (1) {
5057 uint32_t next = vers->next;
5058 if (next) {
5059 vers->next = vers_size + (strlen(vers->name) + 1);
5061 if (remaining_data < vers->next) {
5062 host_dm->flags |= DM_BUFFER_FULL_FLAG;
5063 break;
5065 thunk_convert(cur_data, vers, arg_type, THUNK_TARGET);
5066 strcpy(cur_data + vers_size, vers->name);
5067 cur_data += vers->next;
5068 remaining_data -= vers->next;
5069 if (!next) {
5070 break;
5072 vers = (void*)vers + next;
5074 break;
5076 default:
5077 unlock_user(argptr, guest_data, 0);
5078 ret = -TARGET_EINVAL;
5079 goto out;
5081 unlock_user(argptr, guest_data, guest_data_size);
5083 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
5084 if (!argptr) {
5085 ret = -TARGET_EFAULT;
5086 goto out;
5088 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET);
5089 unlock_user(argptr, arg, target_size);
5091 out:
5092 g_free(big_buf);
5093 return ret;
5096 static abi_long do_ioctl_blkpg(const IOCTLEntry *ie, uint8_t *buf_temp, int fd,
5097 int cmd, abi_long arg)
5099 void *argptr;
5100 int target_size;
5101 const argtype *arg_type = ie->arg_type;
5102 const argtype part_arg_type[] = { MK_STRUCT(STRUCT_blkpg_partition) };
5103 abi_long ret;
5105 struct blkpg_ioctl_arg *host_blkpg = (void*)buf_temp;
5106 struct blkpg_partition host_part;
5108 /* Read and convert blkpg */
5109 arg_type++;
5110 target_size = thunk_type_size(arg_type, 0);
5111 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5112 if (!argptr) {
5113 ret = -TARGET_EFAULT;
5114 goto out;
5116 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
5117 unlock_user(argptr, arg, 0);
5119 switch (host_blkpg->op) {
5120 case BLKPG_ADD_PARTITION:
5121 case BLKPG_DEL_PARTITION:
5122 /* payload is struct blkpg_partition */
5123 break;
5124 default:
5125 /* Unknown opcode */
5126 ret = -TARGET_EINVAL;
5127 goto out;
5130 /* Read and convert blkpg->data */
5131 arg = (abi_long)(uintptr_t)host_blkpg->data;
5132 target_size = thunk_type_size(part_arg_type, 0);
5133 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5134 if (!argptr) {
5135 ret = -TARGET_EFAULT;
5136 goto out;
5138 thunk_convert(&host_part, argptr, part_arg_type, THUNK_HOST);
5139 unlock_user(argptr, arg, 0);
5141 /* Swizzle the data pointer to our local copy and call! */
5142 host_blkpg->data = &host_part;
5143 ret = get_errno(safe_ioctl(fd, ie->host_cmd, host_blkpg));
5145 out:
5146 return ret;
5149 static abi_long do_ioctl_rt(const IOCTLEntry *ie, uint8_t *buf_temp,
5150 int fd, int cmd, abi_long arg)
5152 const argtype *arg_type = ie->arg_type;
5153 const StructEntry *se;
5154 const argtype *field_types;
5155 const int *dst_offsets, *src_offsets;
5156 int target_size;
5157 void *argptr;
5158 abi_ulong *target_rt_dev_ptr = NULL;
5159 unsigned long *host_rt_dev_ptr = NULL;
5160 abi_long ret;
5161 int i;
5163 assert(ie->access == IOC_W);
5164 assert(*arg_type == TYPE_PTR);
5165 arg_type++;
5166 assert(*arg_type == TYPE_STRUCT);
5167 target_size = thunk_type_size(arg_type, 0);
5168 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5169 if (!argptr) {
5170 return -TARGET_EFAULT;
5172 arg_type++;
5173 assert(*arg_type == (int)STRUCT_rtentry);
5174 se = struct_entries + *arg_type++;
5175 assert(se->convert[0] == NULL);
5176 /* convert struct here to be able to catch rt_dev string */
5177 field_types = se->field_types;
5178 dst_offsets = se->field_offsets[THUNK_HOST];
5179 src_offsets = se->field_offsets[THUNK_TARGET];
5180 for (i = 0; i < se->nb_fields; i++) {
5181 if (dst_offsets[i] == offsetof(struct rtentry, rt_dev)) {
5182 assert(*field_types == TYPE_PTRVOID);
5183 target_rt_dev_ptr = (abi_ulong *)(argptr + src_offsets[i]);
5184 host_rt_dev_ptr = (unsigned long *)(buf_temp + dst_offsets[i]);
5185 if (*target_rt_dev_ptr != 0) {
5186 *host_rt_dev_ptr = (unsigned long)lock_user_string(
5187 tswapal(*target_rt_dev_ptr));
5188 if (!*host_rt_dev_ptr) {
5189 unlock_user(argptr, arg, 0);
5190 return -TARGET_EFAULT;
5192 } else {
5193 *host_rt_dev_ptr = 0;
5195 field_types++;
5196 continue;
5198 field_types = thunk_convert(buf_temp + dst_offsets[i],
5199 argptr + src_offsets[i],
5200 field_types, THUNK_HOST);
5202 unlock_user(argptr, arg, 0);
5204 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
5206 assert(host_rt_dev_ptr != NULL);
5207 assert(target_rt_dev_ptr != NULL);
5208 if (*host_rt_dev_ptr != 0) {
5209 unlock_user((void *)*host_rt_dev_ptr,
5210 *target_rt_dev_ptr, 0);
5212 return ret;
5215 static abi_long do_ioctl_kdsigaccept(const IOCTLEntry *ie, uint8_t *buf_temp,
5216 int fd, int cmd, abi_long arg)
5218 int sig = target_to_host_signal(arg);
5219 return get_errno(safe_ioctl(fd, ie->host_cmd, sig));
5222 static abi_long do_ioctl_SIOCGSTAMP(const IOCTLEntry *ie, uint8_t *buf_temp,
5223 int fd, int cmd, abi_long arg)
5225 struct timeval tv;
5226 abi_long ret;
5228 ret = get_errno(safe_ioctl(fd, SIOCGSTAMP, &tv));
5229 if (is_error(ret)) {
5230 return ret;
5233 if (cmd == (int)TARGET_SIOCGSTAMP_OLD) {
5234 if (copy_to_user_timeval(arg, &tv)) {
5235 return -TARGET_EFAULT;
5237 } else {
5238 if (copy_to_user_timeval64(arg, &tv)) {
5239 return -TARGET_EFAULT;
5243 return ret;
5246 static abi_long do_ioctl_SIOCGSTAMPNS(const IOCTLEntry *ie, uint8_t *buf_temp,
5247 int fd, int cmd, abi_long arg)
5249 struct timespec ts;
5250 abi_long ret;
5252 ret = get_errno(safe_ioctl(fd, SIOCGSTAMPNS, &ts));
5253 if (is_error(ret)) {
5254 return ret;
5257 if (cmd == (int)TARGET_SIOCGSTAMPNS_OLD) {
5258 if (host_to_target_timespec(arg, &ts)) {
5259 return -TARGET_EFAULT;
5261 } else{
5262 if (host_to_target_timespec64(arg, &ts)) {
5263 return -TARGET_EFAULT;
5267 return ret;
5270 #ifdef TIOCGPTPEER
5271 static abi_long do_ioctl_tiocgptpeer(const IOCTLEntry *ie, uint8_t *buf_temp,
5272 int fd, int cmd, abi_long arg)
5274 int flags = target_to_host_bitmask(arg, fcntl_flags_tbl);
5275 return get_errno(safe_ioctl(fd, ie->host_cmd, flags));
5277 #endif
5279 static IOCTLEntry ioctl_entries[] = {
5280 #define IOCTL(cmd, access, ...) \
5281 { TARGET_ ## cmd, cmd, #cmd, access, 0, { __VA_ARGS__ } },
5282 #define IOCTL_SPECIAL(cmd, access, dofn, ...) \
5283 { TARGET_ ## cmd, cmd, #cmd, access, dofn, { __VA_ARGS__ } },
5284 #define IOCTL_IGNORE(cmd) \
5285 { TARGET_ ## cmd, 0, #cmd },
5286 #include "ioctls.h"
5287 { 0, 0, },
5290 /* ??? Implement proper locking for ioctls. */
5291 /* do_ioctl() Must return target values and target errnos. */
5292 static abi_long do_ioctl(int fd, int cmd, abi_long arg)
5294 const IOCTLEntry *ie;
5295 const argtype *arg_type;
5296 abi_long ret;
5297 uint8_t buf_temp[MAX_STRUCT_SIZE];
5298 int target_size;
5299 void *argptr;
5301 ie = ioctl_entries;
5302 for(;;) {
5303 if (ie->target_cmd == 0) {
5304 qemu_log_mask(
5305 LOG_UNIMP, "Unsupported ioctl: cmd=0x%04lx\n", (long)cmd);
5306 return -TARGET_ENOSYS;
5308 if (ie->target_cmd == cmd)
5309 break;
5310 ie++;
5312 arg_type = ie->arg_type;
5313 if (ie->do_ioctl) {
5314 return ie->do_ioctl(ie, buf_temp, fd, cmd, arg);
5315 } else if (!ie->host_cmd) {
5316 /* Some architectures define BSD ioctls in their headers
5317 that are not implemented in Linux. */
5318 return -TARGET_ENOSYS;
5321 switch(arg_type[0]) {
5322 case TYPE_NULL:
5323 /* no argument */
5324 ret = get_errno(safe_ioctl(fd, ie->host_cmd));
5325 break;
5326 case TYPE_PTRVOID:
5327 case TYPE_INT:
5328 case TYPE_LONG:
5329 case TYPE_ULONG:
5330 ret = get_errno(safe_ioctl(fd, ie->host_cmd, arg));
5331 break;
5332 case TYPE_PTR:
5333 arg_type++;
5334 target_size = thunk_type_size(arg_type, 0);
5335 switch(ie->access) {
5336 case IOC_R:
5337 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
5338 if (!is_error(ret)) {
5339 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
5340 if (!argptr)
5341 return -TARGET_EFAULT;
5342 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET);
5343 unlock_user(argptr, arg, target_size);
5345 break;
5346 case IOC_W:
5347 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5348 if (!argptr)
5349 return -TARGET_EFAULT;
5350 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
5351 unlock_user(argptr, arg, 0);
5352 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
5353 break;
5354 default:
5355 case IOC_RW:
5356 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5357 if (!argptr)
5358 return -TARGET_EFAULT;
5359 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
5360 unlock_user(argptr, arg, 0);
5361 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
5362 if (!is_error(ret)) {
5363 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
5364 if (!argptr)
5365 return -TARGET_EFAULT;
5366 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET);
5367 unlock_user(argptr, arg, target_size);
5369 break;
5371 break;
5372 default:
5373 qemu_log_mask(LOG_UNIMP,
5374 "Unsupported ioctl type: cmd=0x%04lx type=%d\n",
5375 (long)cmd, arg_type[0]);
5376 ret = -TARGET_ENOSYS;
5377 break;
5379 return ret;
5382 static const bitmask_transtbl iflag_tbl[] = {
5383 { TARGET_IGNBRK, TARGET_IGNBRK, IGNBRK, IGNBRK },
5384 { TARGET_BRKINT, TARGET_BRKINT, BRKINT, BRKINT },
5385 { TARGET_IGNPAR, TARGET_IGNPAR, IGNPAR, IGNPAR },
5386 { TARGET_PARMRK, TARGET_PARMRK, PARMRK, PARMRK },
5387 { TARGET_INPCK, TARGET_INPCK, INPCK, INPCK },
5388 { TARGET_ISTRIP, TARGET_ISTRIP, ISTRIP, ISTRIP },
5389 { TARGET_INLCR, TARGET_INLCR, INLCR, INLCR },
5390 { TARGET_IGNCR, TARGET_IGNCR, IGNCR, IGNCR },
5391 { TARGET_ICRNL, TARGET_ICRNL, ICRNL, ICRNL },
5392 { TARGET_IUCLC, TARGET_IUCLC, IUCLC, IUCLC },
5393 { TARGET_IXON, TARGET_IXON, IXON, IXON },
5394 { TARGET_IXANY, TARGET_IXANY, IXANY, IXANY },
5395 { TARGET_IXOFF, TARGET_IXOFF, IXOFF, IXOFF },
5396 { TARGET_IMAXBEL, TARGET_IMAXBEL, IMAXBEL, IMAXBEL },
5397 { 0, 0, 0, 0 }
5400 static const bitmask_transtbl oflag_tbl[] = {
5401 { TARGET_OPOST, TARGET_OPOST, OPOST, OPOST },
5402 { TARGET_OLCUC, TARGET_OLCUC, OLCUC, OLCUC },
5403 { TARGET_ONLCR, TARGET_ONLCR, ONLCR, ONLCR },
5404 { TARGET_OCRNL, TARGET_OCRNL, OCRNL, OCRNL },
5405 { TARGET_ONOCR, TARGET_ONOCR, ONOCR, ONOCR },
5406 { TARGET_ONLRET, TARGET_ONLRET, ONLRET, ONLRET },
5407 { TARGET_OFILL, TARGET_OFILL, OFILL, OFILL },
5408 { TARGET_OFDEL, TARGET_OFDEL, OFDEL, OFDEL },
5409 { TARGET_NLDLY, TARGET_NL0, NLDLY, NL0 },
5410 { TARGET_NLDLY, TARGET_NL1, NLDLY, NL1 },
5411 { TARGET_CRDLY, TARGET_CR0, CRDLY, CR0 },
5412 { TARGET_CRDLY, TARGET_CR1, CRDLY, CR1 },
5413 { TARGET_CRDLY, TARGET_CR2, CRDLY, CR2 },
5414 { TARGET_CRDLY, TARGET_CR3, CRDLY, CR3 },
5415 { TARGET_TABDLY, TARGET_TAB0, TABDLY, TAB0 },
5416 { TARGET_TABDLY, TARGET_TAB1, TABDLY, TAB1 },
5417 { TARGET_TABDLY, TARGET_TAB2, TABDLY, TAB2 },
5418 { TARGET_TABDLY, TARGET_TAB3, TABDLY, TAB3 },
5419 { TARGET_BSDLY, TARGET_BS0, BSDLY, BS0 },
5420 { TARGET_BSDLY, TARGET_BS1, BSDLY, BS1 },
5421 { TARGET_VTDLY, TARGET_VT0, VTDLY, VT0 },
5422 { TARGET_VTDLY, TARGET_VT1, VTDLY, VT1 },
5423 { TARGET_FFDLY, TARGET_FF0, FFDLY, FF0 },
5424 { TARGET_FFDLY, TARGET_FF1, FFDLY, FF1 },
5425 { 0, 0, 0, 0 }
5428 static const bitmask_transtbl cflag_tbl[] = {
5429 { TARGET_CBAUD, TARGET_B0, CBAUD, B0 },
5430 { TARGET_CBAUD, TARGET_B50, CBAUD, B50 },
5431 { TARGET_CBAUD, TARGET_B75, CBAUD, B75 },
5432 { TARGET_CBAUD, TARGET_B110, CBAUD, B110 },
5433 { TARGET_CBAUD, TARGET_B134, CBAUD, B134 },
5434 { TARGET_CBAUD, TARGET_B150, CBAUD, B150 },
5435 { TARGET_CBAUD, TARGET_B200, CBAUD, B200 },
5436 { TARGET_CBAUD, TARGET_B300, CBAUD, B300 },
5437 { TARGET_CBAUD, TARGET_B600, CBAUD, B600 },
5438 { TARGET_CBAUD, TARGET_B1200, CBAUD, B1200 },
5439 { TARGET_CBAUD, TARGET_B1800, CBAUD, B1800 },
5440 { TARGET_CBAUD, TARGET_B2400, CBAUD, B2400 },
5441 { TARGET_CBAUD, TARGET_B4800, CBAUD, B4800 },
5442 { TARGET_CBAUD, TARGET_B9600, CBAUD, B9600 },
5443 { TARGET_CBAUD, TARGET_B19200, CBAUD, B19200 },
5444 { TARGET_CBAUD, TARGET_B38400, CBAUD, B38400 },
5445 { TARGET_CBAUD, TARGET_B57600, CBAUD, B57600 },
5446 { TARGET_CBAUD, TARGET_B115200, CBAUD, B115200 },
5447 { TARGET_CBAUD, TARGET_B230400, CBAUD, B230400 },
5448 { TARGET_CBAUD, TARGET_B460800, CBAUD, B460800 },
5449 { TARGET_CSIZE, TARGET_CS5, CSIZE, CS5 },
5450 { TARGET_CSIZE, TARGET_CS6, CSIZE, CS6 },
5451 { TARGET_CSIZE, TARGET_CS7, CSIZE, CS7 },
5452 { TARGET_CSIZE, TARGET_CS8, CSIZE, CS8 },
5453 { TARGET_CSTOPB, TARGET_CSTOPB, CSTOPB, CSTOPB },
5454 { TARGET_CREAD, TARGET_CREAD, CREAD, CREAD },
5455 { TARGET_PARENB, TARGET_PARENB, PARENB, PARENB },
5456 { TARGET_PARODD, TARGET_PARODD, PARODD, PARODD },
5457 { TARGET_HUPCL, TARGET_HUPCL, HUPCL, HUPCL },
5458 { TARGET_CLOCAL, TARGET_CLOCAL, CLOCAL, CLOCAL },
5459 { TARGET_CRTSCTS, TARGET_CRTSCTS, CRTSCTS, CRTSCTS },
5460 { 0, 0, 0, 0 }
5463 static const bitmask_transtbl lflag_tbl[] = {
5464 { TARGET_ISIG, TARGET_ISIG, ISIG, ISIG },
5465 { TARGET_ICANON, TARGET_ICANON, ICANON, ICANON },
5466 { TARGET_XCASE, TARGET_XCASE, XCASE, XCASE },
5467 { TARGET_ECHO, TARGET_ECHO, ECHO, ECHO },
5468 { TARGET_ECHOE, TARGET_ECHOE, ECHOE, ECHOE },
5469 { TARGET_ECHOK, TARGET_ECHOK, ECHOK, ECHOK },
5470 { TARGET_ECHONL, TARGET_ECHONL, ECHONL, ECHONL },
5471 { TARGET_NOFLSH, TARGET_NOFLSH, NOFLSH, NOFLSH },
5472 { TARGET_TOSTOP, TARGET_TOSTOP, TOSTOP, TOSTOP },
5473 { TARGET_ECHOCTL, TARGET_ECHOCTL, ECHOCTL, ECHOCTL },
5474 { TARGET_ECHOPRT, TARGET_ECHOPRT, ECHOPRT, ECHOPRT },
5475 { TARGET_ECHOKE, TARGET_ECHOKE, ECHOKE, ECHOKE },
5476 { TARGET_FLUSHO, TARGET_FLUSHO, FLUSHO, FLUSHO },
5477 { TARGET_PENDIN, TARGET_PENDIN, PENDIN, PENDIN },
5478 { TARGET_IEXTEN, TARGET_IEXTEN, IEXTEN, IEXTEN },
5479 { 0, 0, 0, 0 }
5482 static void target_to_host_termios (void *dst, const void *src)
5484 struct host_termios *host = dst;
5485 const struct target_termios *target = src;
5487 host->c_iflag =
5488 target_to_host_bitmask(tswap32(target->c_iflag), iflag_tbl);
5489 host->c_oflag =
5490 target_to_host_bitmask(tswap32(target->c_oflag), oflag_tbl);
5491 host->c_cflag =
5492 target_to_host_bitmask(tswap32(target->c_cflag), cflag_tbl);
5493 host->c_lflag =
5494 target_to_host_bitmask(tswap32(target->c_lflag), lflag_tbl);
5495 host->c_line = target->c_line;
5497 memset(host->c_cc, 0, sizeof(host->c_cc));
5498 host->c_cc[VINTR] = target->c_cc[TARGET_VINTR];
5499 host->c_cc[VQUIT] = target->c_cc[TARGET_VQUIT];
5500 host->c_cc[VERASE] = target->c_cc[TARGET_VERASE];
5501 host->c_cc[VKILL] = target->c_cc[TARGET_VKILL];
5502 host->c_cc[VEOF] = target->c_cc[TARGET_VEOF];
5503 host->c_cc[VTIME] = target->c_cc[TARGET_VTIME];
5504 host->c_cc[VMIN] = target->c_cc[TARGET_VMIN];
5505 host->c_cc[VSWTC] = target->c_cc[TARGET_VSWTC];
5506 host->c_cc[VSTART] = target->c_cc[TARGET_VSTART];
5507 host->c_cc[VSTOP] = target->c_cc[TARGET_VSTOP];
5508 host->c_cc[VSUSP] = target->c_cc[TARGET_VSUSP];
5509 host->c_cc[VEOL] = target->c_cc[TARGET_VEOL];
5510 host->c_cc[VREPRINT] = target->c_cc[TARGET_VREPRINT];
5511 host->c_cc[VDISCARD] = target->c_cc[TARGET_VDISCARD];
5512 host->c_cc[VWERASE] = target->c_cc[TARGET_VWERASE];
5513 host->c_cc[VLNEXT] = target->c_cc[TARGET_VLNEXT];
5514 host->c_cc[VEOL2] = target->c_cc[TARGET_VEOL2];
5517 static void host_to_target_termios (void *dst, const void *src)
5519 struct target_termios *target = dst;
5520 const struct host_termios *host = src;
5522 target->c_iflag =
5523 tswap32(host_to_target_bitmask(host->c_iflag, iflag_tbl));
5524 target->c_oflag =
5525 tswap32(host_to_target_bitmask(host->c_oflag, oflag_tbl));
5526 target->c_cflag =
5527 tswap32(host_to_target_bitmask(host->c_cflag, cflag_tbl));
5528 target->c_lflag =
5529 tswap32(host_to_target_bitmask(host->c_lflag, lflag_tbl));
5530 target->c_line = host->c_line;
5532 memset(target->c_cc, 0, sizeof(target->c_cc));
5533 target->c_cc[TARGET_VINTR] = host->c_cc[VINTR];
5534 target->c_cc[TARGET_VQUIT] = host->c_cc[VQUIT];
5535 target->c_cc[TARGET_VERASE] = host->c_cc[VERASE];
5536 target->c_cc[TARGET_VKILL] = host->c_cc[VKILL];
5537 target->c_cc[TARGET_VEOF] = host->c_cc[VEOF];
5538 target->c_cc[TARGET_VTIME] = host->c_cc[VTIME];
5539 target->c_cc[TARGET_VMIN] = host->c_cc[VMIN];
5540 target->c_cc[TARGET_VSWTC] = host->c_cc[VSWTC];
5541 target->c_cc[TARGET_VSTART] = host->c_cc[VSTART];
5542 target->c_cc[TARGET_VSTOP] = host->c_cc[VSTOP];
5543 target->c_cc[TARGET_VSUSP] = host->c_cc[VSUSP];
5544 target->c_cc[TARGET_VEOL] = host->c_cc[VEOL];
5545 target->c_cc[TARGET_VREPRINT] = host->c_cc[VREPRINT];
5546 target->c_cc[TARGET_VDISCARD] = host->c_cc[VDISCARD];
5547 target->c_cc[TARGET_VWERASE] = host->c_cc[VWERASE];
5548 target->c_cc[TARGET_VLNEXT] = host->c_cc[VLNEXT];
5549 target->c_cc[TARGET_VEOL2] = host->c_cc[VEOL2];
5552 static const StructEntry struct_termios_def = {
5553 .convert = { host_to_target_termios, target_to_host_termios },
5554 .size = { sizeof(struct target_termios), sizeof(struct host_termios) },
5555 .align = { __alignof__(struct target_termios), __alignof__(struct host_termios) },
5558 static bitmask_transtbl mmap_flags_tbl[] = {
5559 { TARGET_MAP_SHARED, TARGET_MAP_SHARED, MAP_SHARED, MAP_SHARED },
5560 { TARGET_MAP_PRIVATE, TARGET_MAP_PRIVATE, MAP_PRIVATE, MAP_PRIVATE },
5561 { TARGET_MAP_FIXED, TARGET_MAP_FIXED, MAP_FIXED, MAP_FIXED },
5562 { TARGET_MAP_ANONYMOUS, TARGET_MAP_ANONYMOUS,
5563 MAP_ANONYMOUS, MAP_ANONYMOUS },
5564 { TARGET_MAP_GROWSDOWN, TARGET_MAP_GROWSDOWN,
5565 MAP_GROWSDOWN, MAP_GROWSDOWN },
5566 { TARGET_MAP_DENYWRITE, TARGET_MAP_DENYWRITE,
5567 MAP_DENYWRITE, MAP_DENYWRITE },
5568 { TARGET_MAP_EXECUTABLE, TARGET_MAP_EXECUTABLE,
5569 MAP_EXECUTABLE, MAP_EXECUTABLE },
5570 { TARGET_MAP_LOCKED, TARGET_MAP_LOCKED, MAP_LOCKED, MAP_LOCKED },
5571 { TARGET_MAP_NORESERVE, TARGET_MAP_NORESERVE,
5572 MAP_NORESERVE, MAP_NORESERVE },
5573 { TARGET_MAP_HUGETLB, TARGET_MAP_HUGETLB, MAP_HUGETLB, MAP_HUGETLB },
5574 /* MAP_STACK had been ignored by the kernel for quite some time.
5575 Recognize it for the target insofar as we do not want to pass
5576 it through to the host. */
5577 { TARGET_MAP_STACK, TARGET_MAP_STACK, 0, 0 },
5578 { 0, 0, 0, 0 }
5582 * NOTE: TARGET_ABI32 is defined for TARGET_I386 (but not for TARGET_X86_64)
5583 * TARGET_I386 is defined if TARGET_X86_64 is defined
5585 #if defined(TARGET_I386)
5587 /* NOTE: there is really one LDT for all the threads */
5588 static uint8_t *ldt_table;
5590 static abi_long read_ldt(abi_ulong ptr, unsigned long bytecount)
5592 int size;
5593 void *p;
5595 if (!ldt_table)
5596 return 0;
5597 size = TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE;
5598 if (size > bytecount)
5599 size = bytecount;
5600 p = lock_user(VERIFY_WRITE, ptr, size, 0);
5601 if (!p)
5602 return -TARGET_EFAULT;
5603 /* ??? Should this by byteswapped? */
5604 memcpy(p, ldt_table, size);
5605 unlock_user(p, ptr, size);
5606 return size;
5609 /* XXX: add locking support */
5610 static abi_long write_ldt(CPUX86State *env,
5611 abi_ulong ptr, unsigned long bytecount, int oldmode)
5613 struct target_modify_ldt_ldt_s ldt_info;
5614 struct target_modify_ldt_ldt_s *target_ldt_info;
5615 int seg_32bit, contents, read_exec_only, limit_in_pages;
5616 int seg_not_present, useable, lm;
5617 uint32_t *lp, entry_1, entry_2;
5619 if (bytecount != sizeof(ldt_info))
5620 return -TARGET_EINVAL;
5621 if (!lock_user_struct(VERIFY_READ, target_ldt_info, ptr, 1))
5622 return -TARGET_EFAULT;
5623 ldt_info.entry_number = tswap32(target_ldt_info->entry_number);
5624 ldt_info.base_addr = tswapal(target_ldt_info->base_addr);
5625 ldt_info.limit = tswap32(target_ldt_info->limit);
5626 ldt_info.flags = tswap32(target_ldt_info->flags);
5627 unlock_user_struct(target_ldt_info, ptr, 0);
5629 if (ldt_info.entry_number >= TARGET_LDT_ENTRIES)
5630 return -TARGET_EINVAL;
5631 seg_32bit = ldt_info.flags & 1;
5632 contents = (ldt_info.flags >> 1) & 3;
5633 read_exec_only = (ldt_info.flags >> 3) & 1;
5634 limit_in_pages = (ldt_info.flags >> 4) & 1;
5635 seg_not_present = (ldt_info.flags >> 5) & 1;
5636 useable = (ldt_info.flags >> 6) & 1;
5637 #ifdef TARGET_ABI32
5638 lm = 0;
5639 #else
5640 lm = (ldt_info.flags >> 7) & 1;
5641 #endif
5642 if (contents == 3) {
5643 if (oldmode)
5644 return -TARGET_EINVAL;
5645 if (seg_not_present == 0)
5646 return -TARGET_EINVAL;
5648 /* allocate the LDT */
5649 if (!ldt_table) {
5650 env->ldt.base = target_mmap(0,
5651 TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE,
5652 PROT_READ|PROT_WRITE,
5653 MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
5654 if (env->ldt.base == -1)
5655 return -TARGET_ENOMEM;
5656 memset(g2h(env->ldt.base), 0,
5657 TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE);
5658 env->ldt.limit = 0xffff;
5659 ldt_table = g2h(env->ldt.base);
5662 /* NOTE: same code as Linux kernel */
5663 /* Allow LDTs to be cleared by the user. */
5664 if (ldt_info.base_addr == 0 && ldt_info.limit == 0) {
5665 if (oldmode ||
5666 (contents == 0 &&
5667 read_exec_only == 1 &&
5668 seg_32bit == 0 &&
5669 limit_in_pages == 0 &&
5670 seg_not_present == 1 &&
5671 useable == 0 )) {
5672 entry_1 = 0;
5673 entry_2 = 0;
5674 goto install;
5678 entry_1 = ((ldt_info.base_addr & 0x0000ffff) << 16) |
5679 (ldt_info.limit & 0x0ffff);
5680 entry_2 = (ldt_info.base_addr & 0xff000000) |
5681 ((ldt_info.base_addr & 0x00ff0000) >> 16) |
5682 (ldt_info.limit & 0xf0000) |
5683 ((read_exec_only ^ 1) << 9) |
5684 (contents << 10) |
5685 ((seg_not_present ^ 1) << 15) |
5686 (seg_32bit << 22) |
5687 (limit_in_pages << 23) |
5688 (lm << 21) |
5689 0x7000;
5690 if (!oldmode)
5691 entry_2 |= (useable << 20);
5693 /* Install the new entry ... */
5694 install:
5695 lp = (uint32_t *)(ldt_table + (ldt_info.entry_number << 3));
5696 lp[0] = tswap32(entry_1);
5697 lp[1] = tswap32(entry_2);
5698 return 0;
5701 /* specific and weird i386 syscalls */
5702 static abi_long do_modify_ldt(CPUX86State *env, int func, abi_ulong ptr,
5703 unsigned long bytecount)
5705 abi_long ret;
5707 switch (func) {
5708 case 0:
5709 ret = read_ldt(ptr, bytecount);
5710 break;
5711 case 1:
5712 ret = write_ldt(env, ptr, bytecount, 1);
5713 break;
5714 case 0x11:
5715 ret = write_ldt(env, ptr, bytecount, 0);
5716 break;
5717 default:
5718 ret = -TARGET_ENOSYS;
5719 break;
5721 return ret;
5724 #if defined(TARGET_ABI32)
5725 abi_long do_set_thread_area(CPUX86State *env, abi_ulong ptr)
5727 uint64_t *gdt_table = g2h(env->gdt.base);
5728 struct target_modify_ldt_ldt_s ldt_info;
5729 struct target_modify_ldt_ldt_s *target_ldt_info;
5730 int seg_32bit, contents, read_exec_only, limit_in_pages;
5731 int seg_not_present, useable, lm;
5732 uint32_t *lp, entry_1, entry_2;
5733 int i;
5735 lock_user_struct(VERIFY_WRITE, target_ldt_info, ptr, 1);
5736 if (!target_ldt_info)
5737 return -TARGET_EFAULT;
5738 ldt_info.entry_number = tswap32(target_ldt_info->entry_number);
5739 ldt_info.base_addr = tswapal(target_ldt_info->base_addr);
5740 ldt_info.limit = tswap32(target_ldt_info->limit);
5741 ldt_info.flags = tswap32(target_ldt_info->flags);
5742 if (ldt_info.entry_number == -1) {
5743 for (i=TARGET_GDT_ENTRY_TLS_MIN; i<=TARGET_GDT_ENTRY_TLS_MAX; i++) {
5744 if (gdt_table[i] == 0) {
5745 ldt_info.entry_number = i;
5746 target_ldt_info->entry_number = tswap32(i);
5747 break;
5751 unlock_user_struct(target_ldt_info, ptr, 1);
5753 if (ldt_info.entry_number < TARGET_GDT_ENTRY_TLS_MIN ||
5754 ldt_info.entry_number > TARGET_GDT_ENTRY_TLS_MAX)
5755 return -TARGET_EINVAL;
5756 seg_32bit = ldt_info.flags & 1;
5757 contents = (ldt_info.flags >> 1) & 3;
5758 read_exec_only = (ldt_info.flags >> 3) & 1;
5759 limit_in_pages = (ldt_info.flags >> 4) & 1;
5760 seg_not_present = (ldt_info.flags >> 5) & 1;
5761 useable = (ldt_info.flags >> 6) & 1;
5762 #ifdef TARGET_ABI32
5763 lm = 0;
5764 #else
5765 lm = (ldt_info.flags >> 7) & 1;
5766 #endif
5768 if (contents == 3) {
5769 if (seg_not_present == 0)
5770 return -TARGET_EINVAL;
5773 /* NOTE: same code as Linux kernel */
5774 /* Allow LDTs to be cleared by the user. */
5775 if (ldt_info.base_addr == 0 && ldt_info.limit == 0) {
5776 if ((contents == 0 &&
5777 read_exec_only == 1 &&
5778 seg_32bit == 0 &&
5779 limit_in_pages == 0 &&
5780 seg_not_present == 1 &&
5781 useable == 0 )) {
5782 entry_1 = 0;
5783 entry_2 = 0;
5784 goto install;
5788 entry_1 = ((ldt_info.base_addr & 0x0000ffff) << 16) |
5789 (ldt_info.limit & 0x0ffff);
5790 entry_2 = (ldt_info.base_addr & 0xff000000) |
5791 ((ldt_info.base_addr & 0x00ff0000) >> 16) |
5792 (ldt_info.limit & 0xf0000) |
5793 ((read_exec_only ^ 1) << 9) |
5794 (contents << 10) |
5795 ((seg_not_present ^ 1) << 15) |
5796 (seg_32bit << 22) |
5797 (limit_in_pages << 23) |
5798 (useable << 20) |
5799 (lm << 21) |
5800 0x7000;
5802 /* Install the new entry ... */
5803 install:
5804 lp = (uint32_t *)(gdt_table + ldt_info.entry_number);
5805 lp[0] = tswap32(entry_1);
5806 lp[1] = tswap32(entry_2);
5807 return 0;
5810 static abi_long do_get_thread_area(CPUX86State *env, abi_ulong ptr)
5812 struct target_modify_ldt_ldt_s *target_ldt_info;
5813 uint64_t *gdt_table = g2h(env->gdt.base);
5814 uint32_t base_addr, limit, flags;
5815 int seg_32bit, contents, read_exec_only, limit_in_pages, idx;
5816 int seg_not_present, useable, lm;
5817 uint32_t *lp, entry_1, entry_2;
5819 lock_user_struct(VERIFY_WRITE, target_ldt_info, ptr, 1);
5820 if (!target_ldt_info)
5821 return -TARGET_EFAULT;
5822 idx = tswap32(target_ldt_info->entry_number);
5823 if (idx < TARGET_GDT_ENTRY_TLS_MIN ||
5824 idx > TARGET_GDT_ENTRY_TLS_MAX) {
5825 unlock_user_struct(target_ldt_info, ptr, 1);
5826 return -TARGET_EINVAL;
5828 lp = (uint32_t *)(gdt_table + idx);
5829 entry_1 = tswap32(lp[0]);
5830 entry_2 = tswap32(lp[1]);
5832 read_exec_only = ((entry_2 >> 9) & 1) ^ 1;
5833 contents = (entry_2 >> 10) & 3;
5834 seg_not_present = ((entry_2 >> 15) & 1) ^ 1;
5835 seg_32bit = (entry_2 >> 22) & 1;
5836 limit_in_pages = (entry_2 >> 23) & 1;
5837 useable = (entry_2 >> 20) & 1;
5838 #ifdef TARGET_ABI32
5839 lm = 0;
5840 #else
5841 lm = (entry_2 >> 21) & 1;
5842 #endif
5843 flags = (seg_32bit << 0) | (contents << 1) |
5844 (read_exec_only << 3) | (limit_in_pages << 4) |
5845 (seg_not_present << 5) | (useable << 6) | (lm << 7);
5846 limit = (entry_1 & 0xffff) | (entry_2 & 0xf0000);
5847 base_addr = (entry_1 >> 16) |
5848 (entry_2 & 0xff000000) |
5849 ((entry_2 & 0xff) << 16);
5850 target_ldt_info->base_addr = tswapal(base_addr);
5851 target_ldt_info->limit = tswap32(limit);
5852 target_ldt_info->flags = tswap32(flags);
5853 unlock_user_struct(target_ldt_info, ptr, 1);
5854 return 0;
5857 abi_long do_arch_prctl(CPUX86State *env, int code, abi_ulong addr)
5859 return -TARGET_ENOSYS;
5861 #else
5862 abi_long do_arch_prctl(CPUX86State *env, int code, abi_ulong addr)
5864 abi_long ret = 0;
5865 abi_ulong val;
5866 int idx;
5868 switch(code) {
5869 case TARGET_ARCH_SET_GS:
5870 case TARGET_ARCH_SET_FS:
5871 if (code == TARGET_ARCH_SET_GS)
5872 idx = R_GS;
5873 else
5874 idx = R_FS;
5875 cpu_x86_load_seg(env, idx, 0);
5876 env->segs[idx].base = addr;
5877 break;
5878 case TARGET_ARCH_GET_GS:
5879 case TARGET_ARCH_GET_FS:
5880 if (code == TARGET_ARCH_GET_GS)
5881 idx = R_GS;
5882 else
5883 idx = R_FS;
5884 val = env->segs[idx].base;
5885 if (put_user(val, addr, abi_ulong))
5886 ret = -TARGET_EFAULT;
5887 break;
5888 default:
5889 ret = -TARGET_EINVAL;
5890 break;
5892 return ret;
5894 #endif /* defined(TARGET_ABI32 */
5896 #endif /* defined(TARGET_I386) */
5898 #define NEW_STACK_SIZE 0x40000
5901 static pthread_mutex_t clone_lock = PTHREAD_MUTEX_INITIALIZER;
5902 typedef struct {
5903 CPUArchState *env;
5904 pthread_mutex_t mutex;
5905 pthread_cond_t cond;
5906 pthread_t thread;
5907 uint32_t tid;
5908 abi_ulong child_tidptr;
5909 abi_ulong parent_tidptr;
5910 sigset_t sigmask;
5911 } new_thread_info;
5913 static void *clone_func(void *arg)
5915 new_thread_info *info = arg;
5916 CPUArchState *env;
5917 CPUState *cpu;
5918 TaskState *ts;
5920 rcu_register_thread();
5921 tcg_register_thread();
5922 env = info->env;
5923 cpu = env_cpu(env);
5924 thread_cpu = cpu;
5925 ts = (TaskState *)cpu->opaque;
5926 info->tid = sys_gettid();
5927 task_settid(ts);
5928 if (info->child_tidptr)
5929 put_user_u32(info->tid, info->child_tidptr);
5930 if (info->parent_tidptr)
5931 put_user_u32(info->tid, info->parent_tidptr);
5932 qemu_guest_random_seed_thread_part2(cpu->random_seed);
5933 /* Enable signals. */
5934 sigprocmask(SIG_SETMASK, &info->sigmask, NULL);
5935 /* Signal to the parent that we're ready. */
5936 pthread_mutex_lock(&info->mutex);
5937 pthread_cond_broadcast(&info->cond);
5938 pthread_mutex_unlock(&info->mutex);
5939 /* Wait until the parent has finished initializing the tls state. */
5940 pthread_mutex_lock(&clone_lock);
5941 pthread_mutex_unlock(&clone_lock);
5942 cpu_loop(env);
5943 /* never exits */
5944 return NULL;
5947 /* do_fork() Must return host values and target errnos (unlike most
5948 do_*() functions). */
5949 static int do_fork(CPUArchState *env, unsigned int flags, abi_ulong newsp,
5950 abi_ulong parent_tidptr, target_ulong newtls,
5951 abi_ulong child_tidptr)
5953 CPUState *cpu = env_cpu(env);
5954 int ret;
5955 TaskState *ts;
5956 CPUState *new_cpu;
5957 CPUArchState *new_env;
5958 sigset_t sigmask;
5960 flags &= ~CLONE_IGNORED_FLAGS;
5962 /* Emulate vfork() with fork() */
5963 if (flags & CLONE_VFORK)
5964 flags &= ~(CLONE_VFORK | CLONE_VM);
5966 if (flags & CLONE_VM) {
5967 TaskState *parent_ts = (TaskState *)cpu->opaque;
5968 new_thread_info info;
5969 pthread_attr_t attr;
5971 if (((flags & CLONE_THREAD_FLAGS) != CLONE_THREAD_FLAGS) ||
5972 (flags & CLONE_INVALID_THREAD_FLAGS)) {
5973 return -TARGET_EINVAL;
5976 ts = g_new0(TaskState, 1);
5977 init_task_state(ts);
5979 /* Grab a mutex so that thread setup appears atomic. */
5980 pthread_mutex_lock(&clone_lock);
5982 /* we create a new CPU instance. */
5983 new_env = cpu_copy(env);
5984 /* Init regs that differ from the parent. */
5985 cpu_clone_regs_child(new_env, newsp, flags);
5986 cpu_clone_regs_parent(env, flags);
5987 new_cpu = env_cpu(new_env);
5988 new_cpu->opaque = ts;
5989 ts->bprm = parent_ts->bprm;
5990 ts->info = parent_ts->info;
5991 ts->signal_mask = parent_ts->signal_mask;
5993 if (flags & CLONE_CHILD_CLEARTID) {
5994 ts->child_tidptr = child_tidptr;
5997 if (flags & CLONE_SETTLS) {
5998 cpu_set_tls (new_env, newtls);
6001 memset(&info, 0, sizeof(info));
6002 pthread_mutex_init(&info.mutex, NULL);
6003 pthread_mutex_lock(&info.mutex);
6004 pthread_cond_init(&info.cond, NULL);
6005 info.env = new_env;
6006 if (flags & CLONE_CHILD_SETTID) {
6007 info.child_tidptr = child_tidptr;
6009 if (flags & CLONE_PARENT_SETTID) {
6010 info.parent_tidptr = parent_tidptr;
6013 ret = pthread_attr_init(&attr);
6014 ret = pthread_attr_setstacksize(&attr, NEW_STACK_SIZE);
6015 ret = pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED);
6016 /* It is not safe to deliver signals until the child has finished
6017 initializing, so temporarily block all signals. */
6018 sigfillset(&sigmask);
6019 sigprocmask(SIG_BLOCK, &sigmask, &info.sigmask);
6020 cpu->random_seed = qemu_guest_random_seed_thread_part1();
6022 /* If this is our first additional thread, we need to ensure we
6023 * generate code for parallel execution and flush old translations.
6025 if (!parallel_cpus) {
6026 parallel_cpus = true;
6027 tb_flush(cpu);
6030 ret = pthread_create(&info.thread, &attr, clone_func, &info);
6031 /* TODO: Free new CPU state if thread creation failed. */
6033 sigprocmask(SIG_SETMASK, &info.sigmask, NULL);
6034 pthread_attr_destroy(&attr);
6035 if (ret == 0) {
6036 /* Wait for the child to initialize. */
6037 pthread_cond_wait(&info.cond, &info.mutex);
6038 ret = info.tid;
6039 } else {
6040 ret = -1;
6042 pthread_mutex_unlock(&info.mutex);
6043 pthread_cond_destroy(&info.cond);
6044 pthread_mutex_destroy(&info.mutex);
6045 pthread_mutex_unlock(&clone_lock);
6046 } else {
6047 /* if no CLONE_VM, we consider it is a fork */
6048 if (flags & CLONE_INVALID_FORK_FLAGS) {
6049 return -TARGET_EINVAL;
6052 /* We can't support custom termination signals */
6053 if ((flags & CSIGNAL) != TARGET_SIGCHLD) {
6054 return -TARGET_EINVAL;
6057 if (block_signals()) {
6058 return -TARGET_ERESTARTSYS;
6061 fork_start();
6062 ret = fork();
6063 if (ret == 0) {
6064 /* Child Process. */
6065 cpu_clone_regs_child(env, newsp, flags);
6066 fork_end(1);
6067 /* There is a race condition here. The parent process could
6068 theoretically read the TID in the child process before the child
6069 tid is set. This would require using either ptrace
6070 (not implemented) or having *_tidptr to point at a shared memory
6071 mapping. We can't repeat the spinlock hack used above because
6072 the child process gets its own copy of the lock. */
6073 if (flags & CLONE_CHILD_SETTID)
6074 put_user_u32(sys_gettid(), child_tidptr);
6075 if (flags & CLONE_PARENT_SETTID)
6076 put_user_u32(sys_gettid(), parent_tidptr);
6077 ts = (TaskState *)cpu->opaque;
6078 if (flags & CLONE_SETTLS)
6079 cpu_set_tls (env, newtls);
6080 if (flags & CLONE_CHILD_CLEARTID)
6081 ts->child_tidptr = child_tidptr;
6082 } else {
6083 cpu_clone_regs_parent(env, flags);
6084 fork_end(0);
6087 return ret;
6090 /* warning : doesn't handle linux specific flags... */
6091 static int target_to_host_fcntl_cmd(int cmd)
6093 int ret;
6095 switch(cmd) {
6096 case TARGET_F_DUPFD:
6097 case TARGET_F_GETFD:
6098 case TARGET_F_SETFD:
6099 case TARGET_F_GETFL:
6100 case TARGET_F_SETFL:
6101 case TARGET_F_OFD_GETLK:
6102 case TARGET_F_OFD_SETLK:
6103 case TARGET_F_OFD_SETLKW:
6104 ret = cmd;
6105 break;
6106 case TARGET_F_GETLK:
6107 ret = F_GETLK64;
6108 break;
6109 case TARGET_F_SETLK:
6110 ret = F_SETLK64;
6111 break;
6112 case TARGET_F_SETLKW:
6113 ret = F_SETLKW64;
6114 break;
6115 case TARGET_F_GETOWN:
6116 ret = F_GETOWN;
6117 break;
6118 case TARGET_F_SETOWN:
6119 ret = F_SETOWN;
6120 break;
6121 case TARGET_F_GETSIG:
6122 ret = F_GETSIG;
6123 break;
6124 case TARGET_F_SETSIG:
6125 ret = F_SETSIG;
6126 break;
6127 #if TARGET_ABI_BITS == 32
6128 case TARGET_F_GETLK64:
6129 ret = F_GETLK64;
6130 break;
6131 case TARGET_F_SETLK64:
6132 ret = F_SETLK64;
6133 break;
6134 case TARGET_F_SETLKW64:
6135 ret = F_SETLKW64;
6136 break;
6137 #endif
6138 case TARGET_F_SETLEASE:
6139 ret = F_SETLEASE;
6140 break;
6141 case TARGET_F_GETLEASE:
6142 ret = F_GETLEASE;
6143 break;
6144 #ifdef F_DUPFD_CLOEXEC
6145 case TARGET_F_DUPFD_CLOEXEC:
6146 ret = F_DUPFD_CLOEXEC;
6147 break;
6148 #endif
6149 case TARGET_F_NOTIFY:
6150 ret = F_NOTIFY;
6151 break;
6152 #ifdef F_GETOWN_EX
6153 case TARGET_F_GETOWN_EX:
6154 ret = F_GETOWN_EX;
6155 break;
6156 #endif
6157 #ifdef F_SETOWN_EX
6158 case TARGET_F_SETOWN_EX:
6159 ret = F_SETOWN_EX;
6160 break;
6161 #endif
6162 #ifdef F_SETPIPE_SZ
6163 case TARGET_F_SETPIPE_SZ:
6164 ret = F_SETPIPE_SZ;
6165 break;
6166 case TARGET_F_GETPIPE_SZ:
6167 ret = F_GETPIPE_SZ;
6168 break;
6169 #endif
6170 default:
6171 ret = -TARGET_EINVAL;
6172 break;
6175 #if defined(__powerpc64__)
6176 /* On PPC64, glibc headers has the F_*LK* defined to 12, 13 and 14 and
6177 * is not supported by kernel. The glibc fcntl call actually adjusts
6178 * them to 5, 6 and 7 before making the syscall(). Since we make the
6179 * syscall directly, adjust to what is supported by the kernel.
6181 if (ret >= F_GETLK64 && ret <= F_SETLKW64) {
6182 ret -= F_GETLK64 - 5;
6184 #endif
6186 return ret;
6189 #define FLOCK_TRANSTBL \
6190 switch (type) { \
6191 TRANSTBL_CONVERT(F_RDLCK); \
6192 TRANSTBL_CONVERT(F_WRLCK); \
6193 TRANSTBL_CONVERT(F_UNLCK); \
6194 TRANSTBL_CONVERT(F_EXLCK); \
6195 TRANSTBL_CONVERT(F_SHLCK); \
6198 static int target_to_host_flock(int type)
6200 #define TRANSTBL_CONVERT(a) case TARGET_##a: return a
6201 FLOCK_TRANSTBL
6202 #undef TRANSTBL_CONVERT
6203 return -TARGET_EINVAL;
6206 static int host_to_target_flock(int type)
6208 #define TRANSTBL_CONVERT(a) case a: return TARGET_##a
6209 FLOCK_TRANSTBL
6210 #undef TRANSTBL_CONVERT
6211 /* if we don't know how to convert the value coming
6212 * from the host we copy to the target field as-is
6214 return type;
6217 static inline abi_long copy_from_user_flock(struct flock64 *fl,
6218 abi_ulong target_flock_addr)
6220 struct target_flock *target_fl;
6221 int l_type;
6223 if (!lock_user_struct(VERIFY_READ, target_fl, target_flock_addr, 1)) {
6224 return -TARGET_EFAULT;
6227 __get_user(l_type, &target_fl->l_type);
6228 l_type = target_to_host_flock(l_type);
6229 if (l_type < 0) {
6230 return l_type;
6232 fl->l_type = l_type;
6233 __get_user(fl->l_whence, &target_fl->l_whence);
6234 __get_user(fl->l_start, &target_fl->l_start);
6235 __get_user(fl->l_len, &target_fl->l_len);
6236 __get_user(fl->l_pid, &target_fl->l_pid);
6237 unlock_user_struct(target_fl, target_flock_addr, 0);
6238 return 0;
6241 static inline abi_long copy_to_user_flock(abi_ulong target_flock_addr,
6242 const struct flock64 *fl)
6244 struct target_flock *target_fl;
6245 short l_type;
6247 if (!lock_user_struct(VERIFY_WRITE, target_fl, target_flock_addr, 0)) {
6248 return -TARGET_EFAULT;
6251 l_type = host_to_target_flock(fl->l_type);
6252 __put_user(l_type, &target_fl->l_type);
6253 __put_user(fl->l_whence, &target_fl->l_whence);
6254 __put_user(fl->l_start, &target_fl->l_start);
6255 __put_user(fl->l_len, &target_fl->l_len);
6256 __put_user(fl->l_pid, &target_fl->l_pid);
6257 unlock_user_struct(target_fl, target_flock_addr, 1);
6258 return 0;
6261 typedef abi_long from_flock64_fn(struct flock64 *fl, abi_ulong target_addr);
6262 typedef abi_long to_flock64_fn(abi_ulong target_addr, const struct flock64 *fl);
6264 #if defined(TARGET_ARM) && TARGET_ABI_BITS == 32
6265 static inline abi_long copy_from_user_oabi_flock64(struct flock64 *fl,
6266 abi_ulong target_flock_addr)
6268 struct target_oabi_flock64 *target_fl;
6269 int l_type;
6271 if (!lock_user_struct(VERIFY_READ, target_fl, target_flock_addr, 1)) {
6272 return -TARGET_EFAULT;
6275 __get_user(l_type, &target_fl->l_type);
6276 l_type = target_to_host_flock(l_type);
6277 if (l_type < 0) {
6278 return l_type;
6280 fl->l_type = l_type;
6281 __get_user(fl->l_whence, &target_fl->l_whence);
6282 __get_user(fl->l_start, &target_fl->l_start);
6283 __get_user(fl->l_len, &target_fl->l_len);
6284 __get_user(fl->l_pid, &target_fl->l_pid);
6285 unlock_user_struct(target_fl, target_flock_addr, 0);
6286 return 0;
6289 static inline abi_long copy_to_user_oabi_flock64(abi_ulong target_flock_addr,
6290 const struct flock64 *fl)
6292 struct target_oabi_flock64 *target_fl;
6293 short l_type;
6295 if (!lock_user_struct(VERIFY_WRITE, target_fl, target_flock_addr, 0)) {
6296 return -TARGET_EFAULT;
6299 l_type = host_to_target_flock(fl->l_type);
6300 __put_user(l_type, &target_fl->l_type);
6301 __put_user(fl->l_whence, &target_fl->l_whence);
6302 __put_user(fl->l_start, &target_fl->l_start);
6303 __put_user(fl->l_len, &target_fl->l_len);
6304 __put_user(fl->l_pid, &target_fl->l_pid);
6305 unlock_user_struct(target_fl, target_flock_addr, 1);
6306 return 0;
6308 #endif
6310 static inline abi_long copy_from_user_flock64(struct flock64 *fl,
6311 abi_ulong target_flock_addr)
6313 struct target_flock64 *target_fl;
6314 int l_type;
6316 if (!lock_user_struct(VERIFY_READ, target_fl, target_flock_addr, 1)) {
6317 return -TARGET_EFAULT;
6320 __get_user(l_type, &target_fl->l_type);
6321 l_type = target_to_host_flock(l_type);
6322 if (l_type < 0) {
6323 return l_type;
6325 fl->l_type = l_type;
6326 __get_user(fl->l_whence, &target_fl->l_whence);
6327 __get_user(fl->l_start, &target_fl->l_start);
6328 __get_user(fl->l_len, &target_fl->l_len);
6329 __get_user(fl->l_pid, &target_fl->l_pid);
6330 unlock_user_struct(target_fl, target_flock_addr, 0);
6331 return 0;
6334 static inline abi_long copy_to_user_flock64(abi_ulong target_flock_addr,
6335 const struct flock64 *fl)
6337 struct target_flock64 *target_fl;
6338 short l_type;
6340 if (!lock_user_struct(VERIFY_WRITE, target_fl, target_flock_addr, 0)) {
6341 return -TARGET_EFAULT;
6344 l_type = host_to_target_flock(fl->l_type);
6345 __put_user(l_type, &target_fl->l_type);
6346 __put_user(fl->l_whence, &target_fl->l_whence);
6347 __put_user(fl->l_start, &target_fl->l_start);
6348 __put_user(fl->l_len, &target_fl->l_len);
6349 __put_user(fl->l_pid, &target_fl->l_pid);
6350 unlock_user_struct(target_fl, target_flock_addr, 1);
6351 return 0;
6354 static abi_long do_fcntl(int fd, int cmd, abi_ulong arg)
6356 struct flock64 fl64;
6357 #ifdef F_GETOWN_EX
6358 struct f_owner_ex fox;
6359 struct target_f_owner_ex *target_fox;
6360 #endif
6361 abi_long ret;
6362 int host_cmd = target_to_host_fcntl_cmd(cmd);
6364 if (host_cmd == -TARGET_EINVAL)
6365 return host_cmd;
6367 switch(cmd) {
6368 case TARGET_F_GETLK:
6369 ret = copy_from_user_flock(&fl64, arg);
6370 if (ret) {
6371 return ret;
6373 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64));
6374 if (ret == 0) {
6375 ret = copy_to_user_flock(arg, &fl64);
6377 break;
6379 case TARGET_F_SETLK:
6380 case TARGET_F_SETLKW:
6381 ret = copy_from_user_flock(&fl64, arg);
6382 if (ret) {
6383 return ret;
6385 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64));
6386 break;
6388 case TARGET_F_GETLK64:
6389 case TARGET_F_OFD_GETLK:
6390 ret = copy_from_user_flock64(&fl64, arg);
6391 if (ret) {
6392 return ret;
6394 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64));
6395 if (ret == 0) {
6396 ret = copy_to_user_flock64(arg, &fl64);
6398 break;
6399 case TARGET_F_SETLK64:
6400 case TARGET_F_SETLKW64:
6401 case TARGET_F_OFD_SETLK:
6402 case TARGET_F_OFD_SETLKW:
6403 ret = copy_from_user_flock64(&fl64, arg);
6404 if (ret) {
6405 return ret;
6407 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64));
6408 break;
6410 case TARGET_F_GETFL:
6411 ret = get_errno(safe_fcntl(fd, host_cmd, arg));
6412 if (ret >= 0) {
6413 ret = host_to_target_bitmask(ret, fcntl_flags_tbl);
6415 break;
6417 case TARGET_F_SETFL:
6418 ret = get_errno(safe_fcntl(fd, host_cmd,
6419 target_to_host_bitmask(arg,
6420 fcntl_flags_tbl)));
6421 break;
6423 #ifdef F_GETOWN_EX
6424 case TARGET_F_GETOWN_EX:
6425 ret = get_errno(safe_fcntl(fd, host_cmd, &fox));
6426 if (ret >= 0) {
6427 if (!lock_user_struct(VERIFY_WRITE, target_fox, arg, 0))
6428 return -TARGET_EFAULT;
6429 target_fox->type = tswap32(fox.type);
6430 target_fox->pid = tswap32(fox.pid);
6431 unlock_user_struct(target_fox, arg, 1);
6433 break;
6434 #endif
6436 #ifdef F_SETOWN_EX
6437 case TARGET_F_SETOWN_EX:
6438 if (!lock_user_struct(VERIFY_READ, target_fox, arg, 1))
6439 return -TARGET_EFAULT;
6440 fox.type = tswap32(target_fox->type);
6441 fox.pid = tswap32(target_fox->pid);
6442 unlock_user_struct(target_fox, arg, 0);
6443 ret = get_errno(safe_fcntl(fd, host_cmd, &fox));
6444 break;
6445 #endif
6447 case TARGET_F_SETOWN:
6448 case TARGET_F_GETOWN:
6449 case TARGET_F_SETSIG:
6450 case TARGET_F_GETSIG:
6451 case TARGET_F_SETLEASE:
6452 case TARGET_F_GETLEASE:
6453 case TARGET_F_SETPIPE_SZ:
6454 case TARGET_F_GETPIPE_SZ:
6455 ret = get_errno(safe_fcntl(fd, host_cmd, arg));
6456 break;
6458 default:
6459 ret = get_errno(safe_fcntl(fd, cmd, arg));
6460 break;
6462 return ret;
6465 #ifdef USE_UID16
6467 static inline int high2lowuid(int uid)
6469 if (uid > 65535)
6470 return 65534;
6471 else
6472 return uid;
6475 static inline int high2lowgid(int gid)
6477 if (gid > 65535)
6478 return 65534;
6479 else
6480 return gid;
6483 static inline int low2highuid(int uid)
6485 if ((int16_t)uid == -1)
6486 return -1;
6487 else
6488 return uid;
6491 static inline int low2highgid(int gid)
6493 if ((int16_t)gid == -1)
6494 return -1;
6495 else
6496 return gid;
6498 static inline int tswapid(int id)
6500 return tswap16(id);
6503 #define put_user_id(x, gaddr) put_user_u16(x, gaddr)
6505 #else /* !USE_UID16 */
6506 static inline int high2lowuid(int uid)
6508 return uid;
6510 static inline int high2lowgid(int gid)
6512 return gid;
6514 static inline int low2highuid(int uid)
6516 return uid;
6518 static inline int low2highgid(int gid)
6520 return gid;
6522 static inline int tswapid(int id)
6524 return tswap32(id);
6527 #define put_user_id(x, gaddr) put_user_u32(x, gaddr)
6529 #endif /* USE_UID16 */
6531 /* We must do direct syscalls for setting UID/GID, because we want to
6532 * implement the Linux system call semantics of "change only for this thread",
6533 * not the libc/POSIX semantics of "change for all threads in process".
6534 * (See http://ewontfix.com/17/ for more details.)
6535 * We use the 32-bit version of the syscalls if present; if it is not
6536 * then either the host architecture supports 32-bit UIDs natively with
6537 * the standard syscall, or the 16-bit UID is the best we can do.
6539 #ifdef __NR_setuid32
6540 #define __NR_sys_setuid __NR_setuid32
6541 #else
6542 #define __NR_sys_setuid __NR_setuid
6543 #endif
6544 #ifdef __NR_setgid32
6545 #define __NR_sys_setgid __NR_setgid32
6546 #else
6547 #define __NR_sys_setgid __NR_setgid
6548 #endif
6549 #ifdef __NR_setresuid32
6550 #define __NR_sys_setresuid __NR_setresuid32
6551 #else
6552 #define __NR_sys_setresuid __NR_setresuid
6553 #endif
6554 #ifdef __NR_setresgid32
6555 #define __NR_sys_setresgid __NR_setresgid32
6556 #else
6557 #define __NR_sys_setresgid __NR_setresgid
6558 #endif
6560 _syscall1(int, sys_setuid, uid_t, uid)
6561 _syscall1(int, sys_setgid, gid_t, gid)
6562 _syscall3(int, sys_setresuid, uid_t, ruid, uid_t, euid, uid_t, suid)
6563 _syscall3(int, sys_setresgid, gid_t, rgid, gid_t, egid, gid_t, sgid)
6565 void syscall_init(void)
6567 IOCTLEntry *ie;
6568 const argtype *arg_type;
6569 int size;
6570 int i;
6572 thunk_init(STRUCT_MAX);
6574 #define STRUCT(name, ...) thunk_register_struct(STRUCT_ ## name, #name, struct_ ## name ## _def);
6575 #define STRUCT_SPECIAL(name) thunk_register_struct_direct(STRUCT_ ## name, #name, &struct_ ## name ## _def);
6576 #include "syscall_types.h"
6577 #undef STRUCT
6578 #undef STRUCT_SPECIAL
6580 /* Build target_to_host_errno_table[] table from
6581 * host_to_target_errno_table[]. */
6582 for (i = 0; i < ERRNO_TABLE_SIZE; i++) {
6583 target_to_host_errno_table[host_to_target_errno_table[i]] = i;
6586 /* we patch the ioctl size if necessary. We rely on the fact that
6587 no ioctl has all the bits at '1' in the size field */
6588 ie = ioctl_entries;
6589 while (ie->target_cmd != 0) {
6590 if (((ie->target_cmd >> TARGET_IOC_SIZESHIFT) & TARGET_IOC_SIZEMASK) ==
6591 TARGET_IOC_SIZEMASK) {
6592 arg_type = ie->arg_type;
6593 if (arg_type[0] != TYPE_PTR) {
6594 fprintf(stderr, "cannot patch size for ioctl 0x%x\n",
6595 ie->target_cmd);
6596 exit(1);
6598 arg_type++;
6599 size = thunk_type_size(arg_type, 0);
6600 ie->target_cmd = (ie->target_cmd &
6601 ~(TARGET_IOC_SIZEMASK << TARGET_IOC_SIZESHIFT)) |
6602 (size << TARGET_IOC_SIZESHIFT);
6605 /* automatic consistency check if same arch */
6606 #if (defined(__i386__) && defined(TARGET_I386) && defined(TARGET_ABI32)) || \
6607 (defined(__x86_64__) && defined(TARGET_X86_64))
6608 if (unlikely(ie->target_cmd != ie->host_cmd)) {
6609 fprintf(stderr, "ERROR: ioctl(%s): target=0x%x host=0x%x\n",
6610 ie->name, ie->target_cmd, ie->host_cmd);
6612 #endif
6613 ie++;
6617 #if TARGET_ABI_BITS == 32
6618 static inline uint64_t target_offset64(uint32_t word0, uint32_t word1)
6620 #ifdef TARGET_WORDS_BIGENDIAN
6621 return ((uint64_t)word0 << 32) | word1;
6622 #else
6623 return ((uint64_t)word1 << 32) | word0;
6624 #endif
6626 #else /* TARGET_ABI_BITS == 32 */
6627 static inline uint64_t target_offset64(uint64_t word0, uint64_t word1)
6629 return word0;
6631 #endif /* TARGET_ABI_BITS != 32 */
6633 #ifdef TARGET_NR_truncate64
6634 static inline abi_long target_truncate64(void *cpu_env, const char *arg1,
6635 abi_long arg2,
6636 abi_long arg3,
6637 abi_long arg4)
6639 if (regpairs_aligned(cpu_env, TARGET_NR_truncate64)) {
6640 arg2 = arg3;
6641 arg3 = arg4;
6643 return get_errno(truncate64(arg1, target_offset64(arg2, arg3)));
6645 #endif
6647 #ifdef TARGET_NR_ftruncate64
6648 static inline abi_long target_ftruncate64(void *cpu_env, abi_long arg1,
6649 abi_long arg2,
6650 abi_long arg3,
6651 abi_long arg4)
6653 if (regpairs_aligned(cpu_env, TARGET_NR_ftruncate64)) {
6654 arg2 = arg3;
6655 arg3 = arg4;
6657 return get_errno(ftruncate64(arg1, target_offset64(arg2, arg3)));
6659 #endif
6661 #if defined(TARGET_NR_timer_settime) || \
6662 (defined(TARGET_NR_timerfd_settime) && defined(CONFIG_TIMERFD))
6663 static inline abi_long target_to_host_itimerspec(struct itimerspec *host_itspec,
6664 abi_ulong target_addr)
6666 struct target_itimerspec *target_itspec;
6668 if (!lock_user_struct(VERIFY_READ, target_itspec, target_addr, 1)) {
6669 return -TARGET_EFAULT;
6672 host_itspec->it_interval.tv_sec =
6673 tswapal(target_itspec->it_interval.tv_sec);
6674 host_itspec->it_interval.tv_nsec =
6675 tswapal(target_itspec->it_interval.tv_nsec);
6676 host_itspec->it_value.tv_sec = tswapal(target_itspec->it_value.tv_sec);
6677 host_itspec->it_value.tv_nsec = tswapal(target_itspec->it_value.tv_nsec);
6679 unlock_user_struct(target_itspec, target_addr, 1);
6680 return 0;
6682 #endif
6684 #if ((defined(TARGET_NR_timerfd_gettime) || \
6685 defined(TARGET_NR_timerfd_settime)) && defined(CONFIG_TIMERFD)) || \
6686 defined(TARGET_NR_timer_gettime) || defined(TARGET_NR_timer_settime)
6687 static inline abi_long host_to_target_itimerspec(abi_ulong target_addr,
6688 struct itimerspec *host_its)
6690 struct target_itimerspec *target_itspec;
6692 if (!lock_user_struct(VERIFY_WRITE, target_itspec, target_addr, 0)) {
6693 return -TARGET_EFAULT;
6696 target_itspec->it_interval.tv_sec = tswapal(host_its->it_interval.tv_sec);
6697 target_itspec->it_interval.tv_nsec = tswapal(host_its->it_interval.tv_nsec);
6699 target_itspec->it_value.tv_sec = tswapal(host_its->it_value.tv_sec);
6700 target_itspec->it_value.tv_nsec = tswapal(host_its->it_value.tv_nsec);
6702 unlock_user_struct(target_itspec, target_addr, 0);
6703 return 0;
6705 #endif
6707 #if defined(TARGET_NR_adjtimex) || \
6708 (defined(TARGET_NR_clock_adjtime) && defined(CONFIG_CLOCK_ADJTIME))
6709 static inline abi_long target_to_host_timex(struct timex *host_tx,
6710 abi_long target_addr)
6712 struct target_timex *target_tx;
6714 if (!lock_user_struct(VERIFY_READ, target_tx, target_addr, 1)) {
6715 return -TARGET_EFAULT;
6718 __get_user(host_tx->modes, &target_tx->modes);
6719 __get_user(host_tx->offset, &target_tx->offset);
6720 __get_user(host_tx->freq, &target_tx->freq);
6721 __get_user(host_tx->maxerror, &target_tx->maxerror);
6722 __get_user(host_tx->esterror, &target_tx->esterror);
6723 __get_user(host_tx->status, &target_tx->status);
6724 __get_user(host_tx->constant, &target_tx->constant);
6725 __get_user(host_tx->precision, &target_tx->precision);
6726 __get_user(host_tx->tolerance, &target_tx->tolerance);
6727 __get_user(host_tx->time.tv_sec, &target_tx->time.tv_sec);
6728 __get_user(host_tx->time.tv_usec, &target_tx->time.tv_usec);
6729 __get_user(host_tx->tick, &target_tx->tick);
6730 __get_user(host_tx->ppsfreq, &target_tx->ppsfreq);
6731 __get_user(host_tx->jitter, &target_tx->jitter);
6732 __get_user(host_tx->shift, &target_tx->shift);
6733 __get_user(host_tx->stabil, &target_tx->stabil);
6734 __get_user(host_tx->jitcnt, &target_tx->jitcnt);
6735 __get_user(host_tx->calcnt, &target_tx->calcnt);
6736 __get_user(host_tx->errcnt, &target_tx->errcnt);
6737 __get_user(host_tx->stbcnt, &target_tx->stbcnt);
6738 __get_user(host_tx->tai, &target_tx->tai);
6740 unlock_user_struct(target_tx, target_addr, 0);
6741 return 0;
6744 static inline abi_long host_to_target_timex(abi_long target_addr,
6745 struct timex *host_tx)
6747 struct target_timex *target_tx;
6749 if (!lock_user_struct(VERIFY_WRITE, target_tx, target_addr, 0)) {
6750 return -TARGET_EFAULT;
6753 __put_user(host_tx->modes, &target_tx->modes);
6754 __put_user(host_tx->offset, &target_tx->offset);
6755 __put_user(host_tx->freq, &target_tx->freq);
6756 __put_user(host_tx->maxerror, &target_tx->maxerror);
6757 __put_user(host_tx->esterror, &target_tx->esterror);
6758 __put_user(host_tx->status, &target_tx->status);
6759 __put_user(host_tx->constant, &target_tx->constant);
6760 __put_user(host_tx->precision, &target_tx->precision);
6761 __put_user(host_tx->tolerance, &target_tx->tolerance);
6762 __put_user(host_tx->time.tv_sec, &target_tx->time.tv_sec);
6763 __put_user(host_tx->time.tv_usec, &target_tx->time.tv_usec);
6764 __put_user(host_tx->tick, &target_tx->tick);
6765 __put_user(host_tx->ppsfreq, &target_tx->ppsfreq);
6766 __put_user(host_tx->jitter, &target_tx->jitter);
6767 __put_user(host_tx->shift, &target_tx->shift);
6768 __put_user(host_tx->stabil, &target_tx->stabil);
6769 __put_user(host_tx->jitcnt, &target_tx->jitcnt);
6770 __put_user(host_tx->calcnt, &target_tx->calcnt);
6771 __put_user(host_tx->errcnt, &target_tx->errcnt);
6772 __put_user(host_tx->stbcnt, &target_tx->stbcnt);
6773 __put_user(host_tx->tai, &target_tx->tai);
6775 unlock_user_struct(target_tx, target_addr, 1);
6776 return 0;
6778 #endif
6780 static inline abi_long target_to_host_sigevent(struct sigevent *host_sevp,
6781 abi_ulong target_addr)
6783 struct target_sigevent *target_sevp;
6785 if (!lock_user_struct(VERIFY_READ, target_sevp, target_addr, 1)) {
6786 return -TARGET_EFAULT;
6789 /* This union is awkward on 64 bit systems because it has a 32 bit
6790 * integer and a pointer in it; we follow the conversion approach
6791 * used for handling sigval types in signal.c so the guest should get
6792 * the correct value back even if we did a 64 bit byteswap and it's
6793 * using the 32 bit integer.
6795 host_sevp->sigev_value.sival_ptr =
6796 (void *)(uintptr_t)tswapal(target_sevp->sigev_value.sival_ptr);
6797 host_sevp->sigev_signo =
6798 target_to_host_signal(tswap32(target_sevp->sigev_signo));
6799 host_sevp->sigev_notify = tswap32(target_sevp->sigev_notify);
6800 host_sevp->_sigev_un._tid = tswap32(target_sevp->_sigev_un._tid);
6802 unlock_user_struct(target_sevp, target_addr, 1);
6803 return 0;
6806 #if defined(TARGET_NR_mlockall)
6807 static inline int target_to_host_mlockall_arg(int arg)
6809 int result = 0;
6811 if (arg & TARGET_MLOCKALL_MCL_CURRENT) {
6812 result |= MCL_CURRENT;
6814 if (arg & TARGET_MLOCKALL_MCL_FUTURE) {
6815 result |= MCL_FUTURE;
6817 return result;
6819 #endif
6821 #if (defined(TARGET_NR_stat64) || defined(TARGET_NR_lstat64) || \
6822 defined(TARGET_NR_fstat64) || defined(TARGET_NR_fstatat64) || \
6823 defined(TARGET_NR_newfstatat))
6824 static inline abi_long host_to_target_stat64(void *cpu_env,
6825 abi_ulong target_addr,
6826 struct stat *host_st)
6828 #if defined(TARGET_ARM) && defined(TARGET_ABI32)
6829 if (((CPUARMState *)cpu_env)->eabi) {
6830 struct target_eabi_stat64 *target_st;
6832 if (!lock_user_struct(VERIFY_WRITE, target_st, target_addr, 0))
6833 return -TARGET_EFAULT;
6834 memset(target_st, 0, sizeof(struct target_eabi_stat64));
6835 __put_user(host_st->st_dev, &target_st->st_dev);
6836 __put_user(host_st->st_ino, &target_st->st_ino);
6837 #ifdef TARGET_STAT64_HAS_BROKEN_ST_INO
6838 __put_user(host_st->st_ino, &target_st->__st_ino);
6839 #endif
6840 __put_user(host_st->st_mode, &target_st->st_mode);
6841 __put_user(host_st->st_nlink, &target_st->st_nlink);
6842 __put_user(host_st->st_uid, &target_st->st_uid);
6843 __put_user(host_st->st_gid, &target_st->st_gid);
6844 __put_user(host_st->st_rdev, &target_st->st_rdev);
6845 __put_user(host_st->st_size, &target_st->st_size);
6846 __put_user(host_st->st_blksize, &target_st->st_blksize);
6847 __put_user(host_st->st_blocks, &target_st->st_blocks);
6848 __put_user(host_st->st_atime, &target_st->target_st_atime);
6849 __put_user(host_st->st_mtime, &target_st->target_st_mtime);
6850 __put_user(host_st->st_ctime, &target_st->target_st_ctime);
6851 #if _POSIX_C_SOURCE >= 200809L || _XOPEN_SOURCE >= 700
6852 __put_user(host_st->st_atim.tv_nsec, &target_st->target_st_atime_nsec);
6853 __put_user(host_st->st_mtim.tv_nsec, &target_st->target_st_mtime_nsec);
6854 __put_user(host_st->st_ctim.tv_nsec, &target_st->target_st_ctime_nsec);
6855 #endif
6856 unlock_user_struct(target_st, target_addr, 1);
6857 } else
6858 #endif
6860 #if defined(TARGET_HAS_STRUCT_STAT64)
6861 struct target_stat64 *target_st;
6862 #else
6863 struct target_stat *target_st;
6864 #endif
6866 if (!lock_user_struct(VERIFY_WRITE, target_st, target_addr, 0))
6867 return -TARGET_EFAULT;
6868 memset(target_st, 0, sizeof(*target_st));
6869 __put_user(host_st->st_dev, &target_st->st_dev);
6870 __put_user(host_st->st_ino, &target_st->st_ino);
6871 #ifdef TARGET_STAT64_HAS_BROKEN_ST_INO
6872 __put_user(host_st->st_ino, &target_st->__st_ino);
6873 #endif
6874 __put_user(host_st->st_mode, &target_st->st_mode);
6875 __put_user(host_st->st_nlink, &target_st->st_nlink);
6876 __put_user(host_st->st_uid, &target_st->st_uid);
6877 __put_user(host_st->st_gid, &target_st->st_gid);
6878 __put_user(host_st->st_rdev, &target_st->st_rdev);
6879 /* XXX: better use of kernel struct */
6880 __put_user(host_st->st_size, &target_st->st_size);
6881 __put_user(host_st->st_blksize, &target_st->st_blksize);
6882 __put_user(host_st->st_blocks, &target_st->st_blocks);
6883 __put_user(host_st->st_atime, &target_st->target_st_atime);
6884 __put_user(host_st->st_mtime, &target_st->target_st_mtime);
6885 __put_user(host_st->st_ctime, &target_st->target_st_ctime);
6886 #if _POSIX_C_SOURCE >= 200809L || _XOPEN_SOURCE >= 700
6887 __put_user(host_st->st_atim.tv_nsec, &target_st->target_st_atime_nsec);
6888 __put_user(host_st->st_mtim.tv_nsec, &target_st->target_st_mtime_nsec);
6889 __put_user(host_st->st_ctim.tv_nsec, &target_st->target_st_ctime_nsec);
6890 #endif
6891 unlock_user_struct(target_st, target_addr, 1);
6894 return 0;
6896 #endif
6898 #if defined(TARGET_NR_statx) && defined(__NR_statx)
6899 static inline abi_long host_to_target_statx(struct target_statx *host_stx,
6900 abi_ulong target_addr)
6902 struct target_statx *target_stx;
6904 if (!lock_user_struct(VERIFY_WRITE, target_stx, target_addr, 0)) {
6905 return -TARGET_EFAULT;
6907 memset(target_stx, 0, sizeof(*target_stx));
6909 __put_user(host_stx->stx_mask, &target_stx->stx_mask);
6910 __put_user(host_stx->stx_blksize, &target_stx->stx_blksize);
6911 __put_user(host_stx->stx_attributes, &target_stx->stx_attributes);
6912 __put_user(host_stx->stx_nlink, &target_stx->stx_nlink);
6913 __put_user(host_stx->stx_uid, &target_stx->stx_uid);
6914 __put_user(host_stx->stx_gid, &target_stx->stx_gid);
6915 __put_user(host_stx->stx_mode, &target_stx->stx_mode);
6916 __put_user(host_stx->stx_ino, &target_stx->stx_ino);
6917 __put_user(host_stx->stx_size, &target_stx->stx_size);
6918 __put_user(host_stx->stx_blocks, &target_stx->stx_blocks);
6919 __put_user(host_stx->stx_attributes_mask, &target_stx->stx_attributes_mask);
6920 __put_user(host_stx->stx_atime.tv_sec, &target_stx->stx_atime.tv_sec);
6921 __put_user(host_stx->stx_atime.tv_nsec, &target_stx->stx_atime.tv_nsec);
6922 __put_user(host_stx->stx_btime.tv_sec, &target_stx->stx_btime.tv_sec);
6923 __put_user(host_stx->stx_btime.tv_nsec, &target_stx->stx_btime.tv_nsec);
6924 __put_user(host_stx->stx_ctime.tv_sec, &target_stx->stx_ctime.tv_sec);
6925 __put_user(host_stx->stx_ctime.tv_nsec, &target_stx->stx_ctime.tv_nsec);
6926 __put_user(host_stx->stx_mtime.tv_sec, &target_stx->stx_mtime.tv_sec);
6927 __put_user(host_stx->stx_mtime.tv_nsec, &target_stx->stx_mtime.tv_nsec);
6928 __put_user(host_stx->stx_rdev_major, &target_stx->stx_rdev_major);
6929 __put_user(host_stx->stx_rdev_minor, &target_stx->stx_rdev_minor);
6930 __put_user(host_stx->stx_dev_major, &target_stx->stx_dev_major);
6931 __put_user(host_stx->stx_dev_minor, &target_stx->stx_dev_minor);
6933 unlock_user_struct(target_stx, target_addr, 1);
6935 return 0;
6937 #endif
6939 static int do_sys_futex(int *uaddr, int op, int val,
6940 const struct timespec *timeout, int *uaddr2,
6941 int val3)
6943 #if HOST_LONG_BITS == 64
6944 #if defined(__NR_futex)
6945 /* always a 64-bit time_t, it doesn't define _time64 version */
6946 return sys_futex(uaddr, op, val, timeout, uaddr2, val3);
6948 #endif
6949 #else /* HOST_LONG_BITS == 64 */
6950 #if defined(__NR_futex_time64)
6951 if (sizeof(timeout->tv_sec) == 8) {
6952 /* _time64 function on 32bit arch */
6953 return sys_futex_time64(uaddr, op, val, timeout, uaddr2, val3);
6955 #endif
6956 #if defined(__NR_futex)
6957 /* old function on 32bit arch */
6958 return sys_futex(uaddr, op, val, timeout, uaddr2, val3);
6959 #endif
6960 #endif /* HOST_LONG_BITS == 64 */
6961 g_assert_not_reached();
6964 static int do_safe_futex(int *uaddr, int op, int val,
6965 const struct timespec *timeout, int *uaddr2,
6966 int val3)
6968 #if HOST_LONG_BITS == 64
6969 #if defined(__NR_futex)
6970 /* always a 64-bit time_t, it doesn't define _time64 version */
6971 return get_errno(safe_futex(uaddr, op, val, timeout, uaddr2, val3));
6972 #endif
6973 #else /* HOST_LONG_BITS == 64 */
6974 #if defined(__NR_futex_time64)
6975 if (sizeof(timeout->tv_sec) == 8) {
6976 /* _time64 function on 32bit arch */
6977 return get_errno(safe_futex_time64(uaddr, op, val, timeout, uaddr2,
6978 val3));
6980 #endif
6981 #if defined(__NR_futex)
6982 /* old function on 32bit arch */
6983 return get_errno(safe_futex(uaddr, op, val, timeout, uaddr2, val3));
6984 #endif
6985 #endif /* HOST_LONG_BITS == 64 */
6986 return -TARGET_ENOSYS;
6989 /* ??? Using host futex calls even when target atomic operations
6990 are not really atomic probably breaks things. However implementing
6991 futexes locally would make futexes shared between multiple processes
6992 tricky. However they're probably useless because guest atomic
6993 operations won't work either. */
6994 #if defined(TARGET_NR_futex)
6995 static int do_futex(target_ulong uaddr, int op, int val, target_ulong timeout,
6996 target_ulong uaddr2, int val3)
6998 struct timespec ts, *pts;
6999 int base_op;
7001 /* ??? We assume FUTEX_* constants are the same on both host
7002 and target. */
7003 #ifdef FUTEX_CMD_MASK
7004 base_op = op & FUTEX_CMD_MASK;
7005 #else
7006 base_op = op;
7007 #endif
7008 switch (base_op) {
7009 case FUTEX_WAIT:
7010 case FUTEX_WAIT_BITSET:
7011 if (timeout) {
7012 pts = &ts;
7013 target_to_host_timespec(pts, timeout);
7014 } else {
7015 pts = NULL;
7017 return do_safe_futex(g2h(uaddr), op, tswap32(val), pts, NULL, val3);
7018 case FUTEX_WAKE:
7019 return do_safe_futex(g2h(uaddr), op, val, NULL, NULL, 0);
7020 case FUTEX_FD:
7021 return do_safe_futex(g2h(uaddr), op, val, NULL, NULL, 0);
7022 case FUTEX_REQUEUE:
7023 case FUTEX_CMP_REQUEUE:
7024 case FUTEX_WAKE_OP:
7025 /* For FUTEX_REQUEUE, FUTEX_CMP_REQUEUE, and FUTEX_WAKE_OP, the
7026 TIMEOUT parameter is interpreted as a uint32_t by the kernel.
7027 But the prototype takes a `struct timespec *'; insert casts
7028 to satisfy the compiler. We do not need to tswap TIMEOUT
7029 since it's not compared to guest memory. */
7030 pts = (struct timespec *)(uintptr_t) timeout;
7031 return do_safe_futex(g2h(uaddr), op, val, pts, g2h(uaddr2),
7032 (base_op == FUTEX_CMP_REQUEUE
7033 ? tswap32(val3)
7034 : val3));
7035 default:
7036 return -TARGET_ENOSYS;
7039 #endif
7041 #if defined(TARGET_NR_futex_time64)
7042 static int do_futex_time64(target_ulong uaddr, int op, int val, target_ulong timeout,
7043 target_ulong uaddr2, int val3)
7045 struct timespec ts, *pts;
7046 int base_op;
7048 /* ??? We assume FUTEX_* constants are the same on both host
7049 and target. */
7050 #ifdef FUTEX_CMD_MASK
7051 base_op = op & FUTEX_CMD_MASK;
7052 #else
7053 base_op = op;
7054 #endif
7055 switch (base_op) {
7056 case FUTEX_WAIT:
7057 case FUTEX_WAIT_BITSET:
7058 if (timeout) {
7059 pts = &ts;
7060 target_to_host_timespec64(pts, timeout);
7061 } else {
7062 pts = NULL;
7064 return do_safe_futex(g2h(uaddr), op, tswap32(val), pts, NULL, val3);
7065 case FUTEX_WAKE:
7066 return do_safe_futex(g2h(uaddr), op, val, NULL, NULL, 0);
7067 case FUTEX_FD:
7068 return do_safe_futex(g2h(uaddr), op, val, NULL, NULL, 0);
7069 case FUTEX_REQUEUE:
7070 case FUTEX_CMP_REQUEUE:
7071 case FUTEX_WAKE_OP:
7072 /* For FUTEX_REQUEUE, FUTEX_CMP_REQUEUE, and FUTEX_WAKE_OP, the
7073 TIMEOUT parameter is interpreted as a uint32_t by the kernel.
7074 But the prototype takes a `struct timespec *'; insert casts
7075 to satisfy the compiler. We do not need to tswap TIMEOUT
7076 since it's not compared to guest memory. */
7077 pts = (struct timespec *)(uintptr_t) timeout;
7078 return do_safe_futex(g2h(uaddr), op, val, pts, g2h(uaddr2),
7079 (base_op == FUTEX_CMP_REQUEUE
7080 ? tswap32(val3)
7081 : val3));
7082 default:
7083 return -TARGET_ENOSYS;
7086 #endif
7088 #if defined(TARGET_NR_name_to_handle_at) && defined(CONFIG_OPEN_BY_HANDLE)
7089 static abi_long do_name_to_handle_at(abi_long dirfd, abi_long pathname,
7090 abi_long handle, abi_long mount_id,
7091 abi_long flags)
7093 struct file_handle *target_fh;
7094 struct file_handle *fh;
7095 int mid = 0;
7096 abi_long ret;
7097 char *name;
7098 unsigned int size, total_size;
7100 if (get_user_s32(size, handle)) {
7101 return -TARGET_EFAULT;
7104 name = lock_user_string(pathname);
7105 if (!name) {
7106 return -TARGET_EFAULT;
7109 total_size = sizeof(struct file_handle) + size;
7110 target_fh = lock_user(VERIFY_WRITE, handle, total_size, 0);
7111 if (!target_fh) {
7112 unlock_user(name, pathname, 0);
7113 return -TARGET_EFAULT;
7116 fh = g_malloc0(total_size);
7117 fh->handle_bytes = size;
7119 ret = get_errno(name_to_handle_at(dirfd, path(name), fh, &mid, flags));
7120 unlock_user(name, pathname, 0);
7122 /* man name_to_handle_at(2):
7123 * Other than the use of the handle_bytes field, the caller should treat
7124 * the file_handle structure as an opaque data type
7127 memcpy(target_fh, fh, total_size);
7128 target_fh->handle_bytes = tswap32(fh->handle_bytes);
7129 target_fh->handle_type = tswap32(fh->handle_type);
7130 g_free(fh);
7131 unlock_user(target_fh, handle, total_size);
7133 if (put_user_s32(mid, mount_id)) {
7134 return -TARGET_EFAULT;
7137 return ret;
7140 #endif
7142 #if defined(TARGET_NR_open_by_handle_at) && defined(CONFIG_OPEN_BY_HANDLE)
7143 static abi_long do_open_by_handle_at(abi_long mount_fd, abi_long handle,
7144 abi_long flags)
7146 struct file_handle *target_fh;
7147 struct file_handle *fh;
7148 unsigned int size, total_size;
7149 abi_long ret;
7151 if (get_user_s32(size, handle)) {
7152 return -TARGET_EFAULT;
7155 total_size = sizeof(struct file_handle) + size;
7156 target_fh = lock_user(VERIFY_READ, handle, total_size, 1);
7157 if (!target_fh) {
7158 return -TARGET_EFAULT;
7161 fh = g_memdup(target_fh, total_size);
7162 fh->handle_bytes = size;
7163 fh->handle_type = tswap32(target_fh->handle_type);
7165 ret = get_errno(open_by_handle_at(mount_fd, fh,
7166 target_to_host_bitmask(flags, fcntl_flags_tbl)));
7168 g_free(fh);
7170 unlock_user(target_fh, handle, total_size);
7172 return ret;
7174 #endif
7176 #if defined(TARGET_NR_signalfd) || defined(TARGET_NR_signalfd4)
7178 static abi_long do_signalfd4(int fd, abi_long mask, int flags)
7180 int host_flags;
7181 target_sigset_t *target_mask;
7182 sigset_t host_mask;
7183 abi_long ret;
7185 if (flags & ~(TARGET_O_NONBLOCK | TARGET_O_CLOEXEC)) {
7186 return -TARGET_EINVAL;
7188 if (!lock_user_struct(VERIFY_READ, target_mask, mask, 1)) {
7189 return -TARGET_EFAULT;
7192 target_to_host_sigset(&host_mask, target_mask);
7194 host_flags = target_to_host_bitmask(flags, fcntl_flags_tbl);
7196 ret = get_errno(signalfd(fd, &host_mask, host_flags));
7197 if (ret >= 0) {
7198 fd_trans_register(ret, &target_signalfd_trans);
7201 unlock_user_struct(target_mask, mask, 0);
7203 return ret;
7205 #endif
7207 /* Map host to target signal numbers for the wait family of syscalls.
7208 Assume all other status bits are the same. */
7209 int host_to_target_waitstatus(int status)
7211 if (WIFSIGNALED(status)) {
7212 return host_to_target_signal(WTERMSIG(status)) | (status & ~0x7f);
7214 if (WIFSTOPPED(status)) {
7215 return (host_to_target_signal(WSTOPSIG(status)) << 8)
7216 | (status & 0xff);
7218 return status;
7221 static int open_self_cmdline(void *cpu_env, int fd)
7223 CPUState *cpu = env_cpu((CPUArchState *)cpu_env);
7224 struct linux_binprm *bprm = ((TaskState *)cpu->opaque)->bprm;
7225 int i;
7227 for (i = 0; i < bprm->argc; i++) {
7228 size_t len = strlen(bprm->argv[i]) + 1;
7230 if (write(fd, bprm->argv[i], len) != len) {
7231 return -1;
7235 return 0;
7238 static int open_self_maps(void *cpu_env, int fd)
7240 CPUState *cpu = env_cpu((CPUArchState *)cpu_env);
7241 TaskState *ts = cpu->opaque;
7242 GSList *map_info = read_self_maps();
7243 GSList *s;
7244 int count;
7246 for (s = map_info; s; s = g_slist_next(s)) {
7247 MapInfo *e = (MapInfo *) s->data;
7249 if (h2g_valid(e->start)) {
7250 unsigned long min = e->start;
7251 unsigned long max = e->end;
7252 int flags = page_get_flags(h2g(min));
7253 const char *path;
7255 max = h2g_valid(max - 1) ?
7256 max : (uintptr_t) g2h(GUEST_ADDR_MAX) + 1;
7258 if (page_check_range(h2g(min), max - min, flags) == -1) {
7259 continue;
7262 if (h2g(min) == ts->info->stack_limit) {
7263 path = "[stack]";
7264 } else {
7265 path = e->path;
7268 count = dprintf(fd, TARGET_ABI_FMT_ptr "-" TARGET_ABI_FMT_ptr
7269 " %c%c%c%c %08" PRIx64 " %s %"PRId64,
7270 h2g(min), h2g(max - 1) + 1,
7271 e->is_read ? 'r' : '-',
7272 e->is_write ? 'w' : '-',
7273 e->is_exec ? 'x' : '-',
7274 e->is_priv ? 'p' : '-',
7275 (uint64_t) e->offset, e->dev, e->inode);
7276 if (path) {
7277 dprintf(fd, "%*s%s\n", 73 - count, "", path);
7278 } else {
7279 dprintf(fd, "\n");
7284 free_self_maps(map_info);
7286 #ifdef TARGET_VSYSCALL_PAGE
7288 * We only support execution from the vsyscall page.
7289 * This is as if CONFIG_LEGACY_VSYSCALL_XONLY=y from v5.3.
7291 count = dprintf(fd, TARGET_FMT_lx "-" TARGET_FMT_lx
7292 " --xp 00000000 00:00 0",
7293 TARGET_VSYSCALL_PAGE, TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE);
7294 dprintf(fd, "%*s%s\n", 73 - count, "", "[vsyscall]");
7295 #endif
7297 return 0;
7300 static int open_self_stat(void *cpu_env, int fd)
7302 CPUState *cpu = env_cpu((CPUArchState *)cpu_env);
7303 TaskState *ts = cpu->opaque;
7304 g_autoptr(GString) buf = g_string_new(NULL);
7305 int i;
7307 for (i = 0; i < 44; i++) {
7308 if (i == 0) {
7309 /* pid */
7310 g_string_printf(buf, FMT_pid " ", getpid());
7311 } else if (i == 1) {
7312 /* app name */
7313 gchar *bin = g_strrstr(ts->bprm->argv[0], "/");
7314 bin = bin ? bin + 1 : ts->bprm->argv[0];
7315 g_string_printf(buf, "(%.15s) ", bin);
7316 } else if (i == 27) {
7317 /* stack bottom */
7318 g_string_printf(buf, TARGET_ABI_FMT_ld " ", ts->info->start_stack);
7319 } else {
7320 /* for the rest, there is MasterCard */
7321 g_string_printf(buf, "0%c", i == 43 ? '\n' : ' ');
7324 if (write(fd, buf->str, buf->len) != buf->len) {
7325 return -1;
7329 return 0;
7332 static int open_self_auxv(void *cpu_env, int fd)
7334 CPUState *cpu = env_cpu((CPUArchState *)cpu_env);
7335 TaskState *ts = cpu->opaque;
7336 abi_ulong auxv = ts->info->saved_auxv;
7337 abi_ulong len = ts->info->auxv_len;
7338 char *ptr;
7341 * Auxiliary vector is stored in target process stack.
7342 * read in whole auxv vector and copy it to file
7344 ptr = lock_user(VERIFY_READ, auxv, len, 0);
7345 if (ptr != NULL) {
7346 while (len > 0) {
7347 ssize_t r;
7348 r = write(fd, ptr, len);
7349 if (r <= 0) {
7350 break;
7352 len -= r;
7353 ptr += r;
7355 lseek(fd, 0, SEEK_SET);
7356 unlock_user(ptr, auxv, len);
7359 return 0;
7362 static int is_proc_myself(const char *filename, const char *entry)
7364 if (!strncmp(filename, "/proc/", strlen("/proc/"))) {
7365 filename += strlen("/proc/");
7366 if (!strncmp(filename, "self/", strlen("self/"))) {
7367 filename += strlen("self/");
7368 } else if (*filename >= '1' && *filename <= '9') {
7369 char myself[80];
7370 snprintf(myself, sizeof(myself), "%d/", getpid());
7371 if (!strncmp(filename, myself, strlen(myself))) {
7372 filename += strlen(myself);
7373 } else {
7374 return 0;
7376 } else {
7377 return 0;
7379 if (!strcmp(filename, entry)) {
7380 return 1;
7383 return 0;
7386 #if defined(HOST_WORDS_BIGENDIAN) != defined(TARGET_WORDS_BIGENDIAN) || \
7387 defined(TARGET_SPARC) || defined(TARGET_M68K) || defined(TARGET_HPPA)
7388 static int is_proc(const char *filename, const char *entry)
7390 return strcmp(filename, entry) == 0;
7392 #endif
7394 #if defined(HOST_WORDS_BIGENDIAN) != defined(TARGET_WORDS_BIGENDIAN)
7395 static int open_net_route(void *cpu_env, int fd)
7397 FILE *fp;
7398 char *line = NULL;
7399 size_t len = 0;
7400 ssize_t read;
7402 fp = fopen("/proc/net/route", "r");
7403 if (fp == NULL) {
7404 return -1;
7407 /* read header */
7409 read = getline(&line, &len, fp);
7410 dprintf(fd, "%s", line);
7412 /* read routes */
7414 while ((read = getline(&line, &len, fp)) != -1) {
7415 char iface[16];
7416 uint32_t dest, gw, mask;
7417 unsigned int flags, refcnt, use, metric, mtu, window, irtt;
7418 int fields;
7420 fields = sscanf(line,
7421 "%s\t%08x\t%08x\t%04x\t%d\t%d\t%d\t%08x\t%d\t%u\t%u\n",
7422 iface, &dest, &gw, &flags, &refcnt, &use, &metric,
7423 &mask, &mtu, &window, &irtt);
7424 if (fields != 11) {
7425 continue;
7427 dprintf(fd, "%s\t%08x\t%08x\t%04x\t%d\t%d\t%d\t%08x\t%d\t%u\t%u\n",
7428 iface, tswap32(dest), tswap32(gw), flags, refcnt, use,
7429 metric, tswap32(mask), mtu, window, irtt);
7432 free(line);
7433 fclose(fp);
7435 return 0;
7437 #endif
7439 #if defined(TARGET_SPARC)
7440 static int open_cpuinfo(void *cpu_env, int fd)
7442 dprintf(fd, "type\t\t: sun4u\n");
7443 return 0;
7445 #endif
7447 #if defined(TARGET_HPPA)
7448 static int open_cpuinfo(void *cpu_env, int fd)
7450 dprintf(fd, "cpu family\t: PA-RISC 1.1e\n");
7451 dprintf(fd, "cpu\t\t: PA7300LC (PCX-L2)\n");
7452 dprintf(fd, "capabilities\t: os32\n");
7453 dprintf(fd, "model\t\t: 9000/778/B160L\n");
7454 dprintf(fd, "model name\t: Merlin L2 160 QEMU (9000/778/B160L)\n");
7455 return 0;
7457 #endif
7459 #if defined(TARGET_M68K)
7460 static int open_hardware(void *cpu_env, int fd)
7462 dprintf(fd, "Model:\t\tqemu-m68k\n");
7463 return 0;
7465 #endif
7467 static int do_openat(void *cpu_env, int dirfd, const char *pathname, int flags, mode_t mode)
7469 struct fake_open {
7470 const char *filename;
7471 int (*fill)(void *cpu_env, int fd);
7472 int (*cmp)(const char *s1, const char *s2);
7474 const struct fake_open *fake_open;
7475 static const struct fake_open fakes[] = {
7476 { "maps", open_self_maps, is_proc_myself },
7477 { "stat", open_self_stat, is_proc_myself },
7478 { "auxv", open_self_auxv, is_proc_myself },
7479 { "cmdline", open_self_cmdline, is_proc_myself },
7480 #if defined(HOST_WORDS_BIGENDIAN) != defined(TARGET_WORDS_BIGENDIAN)
7481 { "/proc/net/route", open_net_route, is_proc },
7482 #endif
7483 #if defined(TARGET_SPARC) || defined(TARGET_HPPA)
7484 { "/proc/cpuinfo", open_cpuinfo, is_proc },
7485 #endif
7486 #if defined(TARGET_M68K)
7487 { "/proc/hardware", open_hardware, is_proc },
7488 #endif
7489 { NULL, NULL, NULL }
7492 if (is_proc_myself(pathname, "exe")) {
7493 int execfd = qemu_getauxval(AT_EXECFD);
7494 return execfd ? execfd : safe_openat(dirfd, exec_path, flags, mode);
7497 for (fake_open = fakes; fake_open->filename; fake_open++) {
7498 if (fake_open->cmp(pathname, fake_open->filename)) {
7499 break;
7503 if (fake_open->filename) {
7504 const char *tmpdir;
7505 char filename[PATH_MAX];
7506 int fd, r;
7508 /* create temporary file to map stat to */
7509 tmpdir = getenv("TMPDIR");
7510 if (!tmpdir)
7511 tmpdir = "/tmp";
7512 snprintf(filename, sizeof(filename), "%s/qemu-open.XXXXXX", tmpdir);
7513 fd = mkstemp(filename);
7514 if (fd < 0) {
7515 return fd;
7517 unlink(filename);
7519 if ((r = fake_open->fill(cpu_env, fd))) {
7520 int e = errno;
7521 close(fd);
7522 errno = e;
7523 return r;
7525 lseek(fd, 0, SEEK_SET);
7527 return fd;
7530 return safe_openat(dirfd, path(pathname), flags, mode);
7533 #define TIMER_MAGIC 0x0caf0000
7534 #define TIMER_MAGIC_MASK 0xffff0000
7536 /* Convert QEMU provided timer ID back to internal 16bit index format */
7537 static target_timer_t get_timer_id(abi_long arg)
7539 target_timer_t timerid = arg;
7541 if ((timerid & TIMER_MAGIC_MASK) != TIMER_MAGIC) {
7542 return -TARGET_EINVAL;
7545 timerid &= 0xffff;
7547 if (timerid >= ARRAY_SIZE(g_posix_timers)) {
7548 return -TARGET_EINVAL;
7551 return timerid;
7554 static int target_to_host_cpu_mask(unsigned long *host_mask,
7555 size_t host_size,
7556 abi_ulong target_addr,
7557 size_t target_size)
7559 unsigned target_bits = sizeof(abi_ulong) * 8;
7560 unsigned host_bits = sizeof(*host_mask) * 8;
7561 abi_ulong *target_mask;
7562 unsigned i, j;
7564 assert(host_size >= target_size);
7566 target_mask = lock_user(VERIFY_READ, target_addr, target_size, 1);
7567 if (!target_mask) {
7568 return -TARGET_EFAULT;
7570 memset(host_mask, 0, host_size);
7572 for (i = 0 ; i < target_size / sizeof(abi_ulong); i++) {
7573 unsigned bit = i * target_bits;
7574 abi_ulong val;
7576 __get_user(val, &target_mask[i]);
7577 for (j = 0; j < target_bits; j++, bit++) {
7578 if (val & (1UL << j)) {
7579 host_mask[bit / host_bits] |= 1UL << (bit % host_bits);
7584 unlock_user(target_mask, target_addr, 0);
7585 return 0;
7588 static int host_to_target_cpu_mask(const unsigned long *host_mask,
7589 size_t host_size,
7590 abi_ulong target_addr,
7591 size_t target_size)
7593 unsigned target_bits = sizeof(abi_ulong) * 8;
7594 unsigned host_bits = sizeof(*host_mask) * 8;
7595 abi_ulong *target_mask;
7596 unsigned i, j;
7598 assert(host_size >= target_size);
7600 target_mask = lock_user(VERIFY_WRITE, target_addr, target_size, 0);
7601 if (!target_mask) {
7602 return -TARGET_EFAULT;
7605 for (i = 0 ; i < target_size / sizeof(abi_ulong); i++) {
7606 unsigned bit = i * target_bits;
7607 abi_ulong val = 0;
7609 for (j = 0; j < target_bits; j++, bit++) {
7610 if (host_mask[bit / host_bits] & (1UL << (bit % host_bits))) {
7611 val |= 1UL << j;
7614 __put_user(val, &target_mask[i]);
7617 unlock_user(target_mask, target_addr, target_size);
7618 return 0;
7621 /* This is an internal helper for do_syscall so that it is easier
7622 * to have a single return point, so that actions, such as logging
7623 * of syscall results, can be performed.
7624 * All errnos that do_syscall() returns must be -TARGET_<errcode>.
7626 static abi_long do_syscall1(void *cpu_env, int num, abi_long arg1,
7627 abi_long arg2, abi_long arg3, abi_long arg4,
7628 abi_long arg5, abi_long arg6, abi_long arg7,
7629 abi_long arg8)
7631 CPUState *cpu = env_cpu(cpu_env);
7632 abi_long ret;
7633 #if defined(TARGET_NR_stat) || defined(TARGET_NR_stat64) \
7634 || defined(TARGET_NR_lstat) || defined(TARGET_NR_lstat64) \
7635 || defined(TARGET_NR_fstat) || defined(TARGET_NR_fstat64) \
7636 || defined(TARGET_NR_statx)
7637 struct stat st;
7638 #endif
7639 #if defined(TARGET_NR_statfs) || defined(TARGET_NR_statfs64) \
7640 || defined(TARGET_NR_fstatfs)
7641 struct statfs stfs;
7642 #endif
7643 void *p;
7645 switch(num) {
7646 case TARGET_NR_exit:
7647 /* In old applications this may be used to implement _exit(2).
7648 However in threaded applictions it is used for thread termination,
7649 and _exit_group is used for application termination.
7650 Do thread termination if we have more then one thread. */
7652 if (block_signals()) {
7653 return -TARGET_ERESTARTSYS;
7656 pthread_mutex_lock(&clone_lock);
7658 if (CPU_NEXT(first_cpu)) {
7659 TaskState *ts = cpu->opaque;
7661 object_property_set_bool(OBJECT(cpu), false, "realized", NULL);
7662 object_unref(OBJECT(cpu));
7664 * At this point the CPU should be unrealized and removed
7665 * from cpu lists. We can clean-up the rest of the thread
7666 * data without the lock held.
7669 pthread_mutex_unlock(&clone_lock);
7671 if (ts->child_tidptr) {
7672 put_user_u32(0, ts->child_tidptr);
7673 do_sys_futex(g2h(ts->child_tidptr), FUTEX_WAKE, INT_MAX,
7674 NULL, NULL, 0);
7676 thread_cpu = NULL;
7677 g_free(ts);
7678 rcu_unregister_thread();
7679 pthread_exit(NULL);
7682 pthread_mutex_unlock(&clone_lock);
7683 preexit_cleanup(cpu_env, arg1);
7684 _exit(arg1);
7685 return 0; /* avoid warning */
7686 case TARGET_NR_read:
7687 if (arg2 == 0 && arg3 == 0) {
7688 return get_errno(safe_read(arg1, 0, 0));
7689 } else {
7690 if (!(p = lock_user(VERIFY_WRITE, arg2, arg3, 0)))
7691 return -TARGET_EFAULT;
7692 ret = get_errno(safe_read(arg1, p, arg3));
7693 if (ret >= 0 &&
7694 fd_trans_host_to_target_data(arg1)) {
7695 ret = fd_trans_host_to_target_data(arg1)(p, ret);
7697 unlock_user(p, arg2, ret);
7699 return ret;
7700 case TARGET_NR_write:
7701 if (arg2 == 0 && arg3 == 0) {
7702 return get_errno(safe_write(arg1, 0, 0));
7704 if (!(p = lock_user(VERIFY_READ, arg2, arg3, 1)))
7705 return -TARGET_EFAULT;
7706 if (fd_trans_target_to_host_data(arg1)) {
7707 void *copy = g_malloc(arg3);
7708 memcpy(copy, p, arg3);
7709 ret = fd_trans_target_to_host_data(arg1)(copy, arg3);
7710 if (ret >= 0) {
7711 ret = get_errno(safe_write(arg1, copy, ret));
7713 g_free(copy);
7714 } else {
7715 ret = get_errno(safe_write(arg1, p, arg3));
7717 unlock_user(p, arg2, 0);
7718 return ret;
7720 #ifdef TARGET_NR_open
7721 case TARGET_NR_open:
7722 if (!(p = lock_user_string(arg1)))
7723 return -TARGET_EFAULT;
7724 ret = get_errno(do_openat(cpu_env, AT_FDCWD, p,
7725 target_to_host_bitmask(arg2, fcntl_flags_tbl),
7726 arg3));
7727 fd_trans_unregister(ret);
7728 unlock_user(p, arg1, 0);
7729 return ret;
7730 #endif
7731 case TARGET_NR_openat:
7732 if (!(p = lock_user_string(arg2)))
7733 return -TARGET_EFAULT;
7734 ret = get_errno(do_openat(cpu_env, arg1, p,
7735 target_to_host_bitmask(arg3, fcntl_flags_tbl),
7736 arg4));
7737 fd_trans_unregister(ret);
7738 unlock_user(p, arg2, 0);
7739 return ret;
7740 #if defined(TARGET_NR_name_to_handle_at) && defined(CONFIG_OPEN_BY_HANDLE)
7741 case TARGET_NR_name_to_handle_at:
7742 ret = do_name_to_handle_at(arg1, arg2, arg3, arg4, arg5);
7743 return ret;
7744 #endif
7745 #if defined(TARGET_NR_open_by_handle_at) && defined(CONFIG_OPEN_BY_HANDLE)
7746 case TARGET_NR_open_by_handle_at:
7747 ret = do_open_by_handle_at(arg1, arg2, arg3);
7748 fd_trans_unregister(ret);
7749 return ret;
7750 #endif
7751 case TARGET_NR_close:
7752 fd_trans_unregister(arg1);
7753 return get_errno(close(arg1));
7755 case TARGET_NR_brk:
7756 return do_brk(arg1);
7757 #ifdef TARGET_NR_fork
7758 case TARGET_NR_fork:
7759 return get_errno(do_fork(cpu_env, TARGET_SIGCHLD, 0, 0, 0, 0));
7760 #endif
7761 #ifdef TARGET_NR_waitpid
7762 case TARGET_NR_waitpid:
7764 int status;
7765 ret = get_errno(safe_wait4(arg1, &status, arg3, 0));
7766 if (!is_error(ret) && arg2 && ret
7767 && put_user_s32(host_to_target_waitstatus(status), arg2))
7768 return -TARGET_EFAULT;
7770 return ret;
7771 #endif
7772 #ifdef TARGET_NR_waitid
7773 case TARGET_NR_waitid:
7775 siginfo_t info;
7776 info.si_pid = 0;
7777 ret = get_errno(safe_waitid(arg1, arg2, &info, arg4, NULL));
7778 if (!is_error(ret) && arg3 && info.si_pid != 0) {
7779 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_siginfo_t), 0)))
7780 return -TARGET_EFAULT;
7781 host_to_target_siginfo(p, &info);
7782 unlock_user(p, arg3, sizeof(target_siginfo_t));
7785 return ret;
7786 #endif
7787 #ifdef TARGET_NR_creat /* not on alpha */
7788 case TARGET_NR_creat:
7789 if (!(p = lock_user_string(arg1)))
7790 return -TARGET_EFAULT;
7791 ret = get_errno(creat(p, arg2));
7792 fd_trans_unregister(ret);
7793 unlock_user(p, arg1, 0);
7794 return ret;
7795 #endif
7796 #ifdef TARGET_NR_link
7797 case TARGET_NR_link:
7799 void * p2;
7800 p = lock_user_string(arg1);
7801 p2 = lock_user_string(arg2);
7802 if (!p || !p2)
7803 ret = -TARGET_EFAULT;
7804 else
7805 ret = get_errno(link(p, p2));
7806 unlock_user(p2, arg2, 0);
7807 unlock_user(p, arg1, 0);
7809 return ret;
7810 #endif
7811 #if defined(TARGET_NR_linkat)
7812 case TARGET_NR_linkat:
7814 void * p2 = NULL;
7815 if (!arg2 || !arg4)
7816 return -TARGET_EFAULT;
7817 p = lock_user_string(arg2);
7818 p2 = lock_user_string(arg4);
7819 if (!p || !p2)
7820 ret = -TARGET_EFAULT;
7821 else
7822 ret = get_errno(linkat(arg1, p, arg3, p2, arg5));
7823 unlock_user(p, arg2, 0);
7824 unlock_user(p2, arg4, 0);
7826 return ret;
7827 #endif
7828 #ifdef TARGET_NR_unlink
7829 case TARGET_NR_unlink:
7830 if (!(p = lock_user_string(arg1)))
7831 return -TARGET_EFAULT;
7832 ret = get_errno(unlink(p));
7833 unlock_user(p, arg1, 0);
7834 return ret;
7835 #endif
7836 #if defined(TARGET_NR_unlinkat)
7837 case TARGET_NR_unlinkat:
7838 if (!(p = lock_user_string(arg2)))
7839 return -TARGET_EFAULT;
7840 ret = get_errno(unlinkat(arg1, p, arg3));
7841 unlock_user(p, arg2, 0);
7842 return ret;
7843 #endif
7844 case TARGET_NR_execve:
7846 char **argp, **envp;
7847 int argc, envc;
7848 abi_ulong gp;
7849 abi_ulong guest_argp;
7850 abi_ulong guest_envp;
7851 abi_ulong addr;
7852 char **q;
7853 int total_size = 0;
7855 argc = 0;
7856 guest_argp = arg2;
7857 for (gp = guest_argp; gp; gp += sizeof(abi_ulong)) {
7858 if (get_user_ual(addr, gp))
7859 return -TARGET_EFAULT;
7860 if (!addr)
7861 break;
7862 argc++;
7864 envc = 0;
7865 guest_envp = arg3;
7866 for (gp = guest_envp; gp; gp += sizeof(abi_ulong)) {
7867 if (get_user_ual(addr, gp))
7868 return -TARGET_EFAULT;
7869 if (!addr)
7870 break;
7871 envc++;
7874 argp = g_new0(char *, argc + 1);
7875 envp = g_new0(char *, envc + 1);
7877 for (gp = guest_argp, q = argp; gp;
7878 gp += sizeof(abi_ulong), q++) {
7879 if (get_user_ual(addr, gp))
7880 goto execve_efault;
7881 if (!addr)
7882 break;
7883 if (!(*q = lock_user_string(addr)))
7884 goto execve_efault;
7885 total_size += strlen(*q) + 1;
7887 *q = NULL;
7889 for (gp = guest_envp, q = envp; gp;
7890 gp += sizeof(abi_ulong), q++) {
7891 if (get_user_ual(addr, gp))
7892 goto execve_efault;
7893 if (!addr)
7894 break;
7895 if (!(*q = lock_user_string(addr)))
7896 goto execve_efault;
7897 total_size += strlen(*q) + 1;
7899 *q = NULL;
7901 if (!(p = lock_user_string(arg1)))
7902 goto execve_efault;
7903 /* Although execve() is not an interruptible syscall it is
7904 * a special case where we must use the safe_syscall wrapper:
7905 * if we allow a signal to happen before we make the host
7906 * syscall then we will 'lose' it, because at the point of
7907 * execve the process leaves QEMU's control. So we use the
7908 * safe syscall wrapper to ensure that we either take the
7909 * signal as a guest signal, or else it does not happen
7910 * before the execve completes and makes it the other
7911 * program's problem.
7913 ret = get_errno(safe_execve(p, argp, envp));
7914 unlock_user(p, arg1, 0);
7916 goto execve_end;
7918 execve_efault:
7919 ret = -TARGET_EFAULT;
7921 execve_end:
7922 for (gp = guest_argp, q = argp; *q;
7923 gp += sizeof(abi_ulong), q++) {
7924 if (get_user_ual(addr, gp)
7925 || !addr)
7926 break;
7927 unlock_user(*q, addr, 0);
7929 for (gp = guest_envp, q = envp; *q;
7930 gp += sizeof(abi_ulong), q++) {
7931 if (get_user_ual(addr, gp)
7932 || !addr)
7933 break;
7934 unlock_user(*q, addr, 0);
7937 g_free(argp);
7938 g_free(envp);
7940 return ret;
7941 case TARGET_NR_chdir:
7942 if (!(p = lock_user_string(arg1)))
7943 return -TARGET_EFAULT;
7944 ret = get_errno(chdir(p));
7945 unlock_user(p, arg1, 0);
7946 return ret;
7947 #ifdef TARGET_NR_time
7948 case TARGET_NR_time:
7950 time_t host_time;
7951 ret = get_errno(time(&host_time));
7952 if (!is_error(ret)
7953 && arg1
7954 && put_user_sal(host_time, arg1))
7955 return -TARGET_EFAULT;
7957 return ret;
7958 #endif
7959 #ifdef TARGET_NR_mknod
7960 case TARGET_NR_mknod:
7961 if (!(p = lock_user_string(arg1)))
7962 return -TARGET_EFAULT;
7963 ret = get_errno(mknod(p, arg2, arg3));
7964 unlock_user(p, arg1, 0);
7965 return ret;
7966 #endif
7967 #if defined(TARGET_NR_mknodat)
7968 case TARGET_NR_mknodat:
7969 if (!(p = lock_user_string(arg2)))
7970 return -TARGET_EFAULT;
7971 ret = get_errno(mknodat(arg1, p, arg3, arg4));
7972 unlock_user(p, arg2, 0);
7973 return ret;
7974 #endif
7975 #ifdef TARGET_NR_chmod
7976 case TARGET_NR_chmod:
7977 if (!(p = lock_user_string(arg1)))
7978 return -TARGET_EFAULT;
7979 ret = get_errno(chmod(p, arg2));
7980 unlock_user(p, arg1, 0);
7981 return ret;
7982 #endif
7983 #ifdef TARGET_NR_lseek
7984 case TARGET_NR_lseek:
7985 return get_errno(lseek(arg1, arg2, arg3));
7986 #endif
7987 #if defined(TARGET_NR_getxpid) && defined(TARGET_ALPHA)
7988 /* Alpha specific */
7989 case TARGET_NR_getxpid:
7990 ((CPUAlphaState *)cpu_env)->ir[IR_A4] = getppid();
7991 return get_errno(getpid());
7992 #endif
7993 #ifdef TARGET_NR_getpid
7994 case TARGET_NR_getpid:
7995 return get_errno(getpid());
7996 #endif
7997 case TARGET_NR_mount:
7999 /* need to look at the data field */
8000 void *p2, *p3;
8002 if (arg1) {
8003 p = lock_user_string(arg1);
8004 if (!p) {
8005 return -TARGET_EFAULT;
8007 } else {
8008 p = NULL;
8011 p2 = lock_user_string(arg2);
8012 if (!p2) {
8013 if (arg1) {
8014 unlock_user(p, arg1, 0);
8016 return -TARGET_EFAULT;
8019 if (arg3) {
8020 p3 = lock_user_string(arg3);
8021 if (!p3) {
8022 if (arg1) {
8023 unlock_user(p, arg1, 0);
8025 unlock_user(p2, arg2, 0);
8026 return -TARGET_EFAULT;
8028 } else {
8029 p3 = NULL;
8032 /* FIXME - arg5 should be locked, but it isn't clear how to
8033 * do that since it's not guaranteed to be a NULL-terminated
8034 * string.
8036 if (!arg5) {
8037 ret = mount(p, p2, p3, (unsigned long)arg4, NULL);
8038 } else {
8039 ret = mount(p, p2, p3, (unsigned long)arg4, g2h(arg5));
8041 ret = get_errno(ret);
8043 if (arg1) {
8044 unlock_user(p, arg1, 0);
8046 unlock_user(p2, arg2, 0);
8047 if (arg3) {
8048 unlock_user(p3, arg3, 0);
8051 return ret;
8052 #if defined(TARGET_NR_umount) || defined(TARGET_NR_oldumount)
8053 #if defined(TARGET_NR_umount)
8054 case TARGET_NR_umount:
8055 #endif
8056 #if defined(TARGET_NR_oldumount)
8057 case TARGET_NR_oldumount:
8058 #endif
8059 if (!(p = lock_user_string(arg1)))
8060 return -TARGET_EFAULT;
8061 ret = get_errno(umount(p));
8062 unlock_user(p, arg1, 0);
8063 return ret;
8064 #endif
8065 #ifdef TARGET_NR_stime /* not on alpha */
8066 case TARGET_NR_stime:
8068 struct timespec ts;
8069 ts.tv_nsec = 0;
8070 if (get_user_sal(ts.tv_sec, arg1)) {
8071 return -TARGET_EFAULT;
8073 return get_errno(clock_settime(CLOCK_REALTIME, &ts));
8075 #endif
8076 #ifdef TARGET_NR_alarm /* not on alpha */
8077 case TARGET_NR_alarm:
8078 return alarm(arg1);
8079 #endif
8080 #ifdef TARGET_NR_pause /* not on alpha */
8081 case TARGET_NR_pause:
8082 if (!block_signals()) {
8083 sigsuspend(&((TaskState *)cpu->opaque)->signal_mask);
8085 return -TARGET_EINTR;
8086 #endif
8087 #ifdef TARGET_NR_utime
8088 case TARGET_NR_utime:
8090 struct utimbuf tbuf, *host_tbuf;
8091 struct target_utimbuf *target_tbuf;
8092 if (arg2) {
8093 if (!lock_user_struct(VERIFY_READ, target_tbuf, arg2, 1))
8094 return -TARGET_EFAULT;
8095 tbuf.actime = tswapal(target_tbuf->actime);
8096 tbuf.modtime = tswapal(target_tbuf->modtime);
8097 unlock_user_struct(target_tbuf, arg2, 0);
8098 host_tbuf = &tbuf;
8099 } else {
8100 host_tbuf = NULL;
8102 if (!(p = lock_user_string(arg1)))
8103 return -TARGET_EFAULT;
8104 ret = get_errno(utime(p, host_tbuf));
8105 unlock_user(p, arg1, 0);
8107 return ret;
8108 #endif
8109 #ifdef TARGET_NR_utimes
8110 case TARGET_NR_utimes:
8112 struct timeval *tvp, tv[2];
8113 if (arg2) {
8114 if (copy_from_user_timeval(&tv[0], arg2)
8115 || copy_from_user_timeval(&tv[1],
8116 arg2 + sizeof(struct target_timeval)))
8117 return -TARGET_EFAULT;
8118 tvp = tv;
8119 } else {
8120 tvp = NULL;
8122 if (!(p = lock_user_string(arg1)))
8123 return -TARGET_EFAULT;
8124 ret = get_errno(utimes(p, tvp));
8125 unlock_user(p, arg1, 0);
8127 return ret;
8128 #endif
8129 #if defined(TARGET_NR_futimesat)
8130 case TARGET_NR_futimesat:
8132 struct timeval *tvp, tv[2];
8133 if (arg3) {
8134 if (copy_from_user_timeval(&tv[0], arg3)
8135 || copy_from_user_timeval(&tv[1],
8136 arg3 + sizeof(struct target_timeval)))
8137 return -TARGET_EFAULT;
8138 tvp = tv;
8139 } else {
8140 tvp = NULL;
8142 if (!(p = lock_user_string(arg2))) {
8143 return -TARGET_EFAULT;
8145 ret = get_errno(futimesat(arg1, path(p), tvp));
8146 unlock_user(p, arg2, 0);
8148 return ret;
8149 #endif
8150 #ifdef TARGET_NR_access
8151 case TARGET_NR_access:
8152 if (!(p = lock_user_string(arg1))) {
8153 return -TARGET_EFAULT;
8155 ret = get_errno(access(path(p), arg2));
8156 unlock_user(p, arg1, 0);
8157 return ret;
8158 #endif
8159 #if defined(TARGET_NR_faccessat) && defined(__NR_faccessat)
8160 case TARGET_NR_faccessat:
8161 if (!(p = lock_user_string(arg2))) {
8162 return -TARGET_EFAULT;
8164 ret = get_errno(faccessat(arg1, p, arg3, 0));
8165 unlock_user(p, arg2, 0);
8166 return ret;
8167 #endif
8168 #ifdef TARGET_NR_nice /* not on alpha */
8169 case TARGET_NR_nice:
8170 return get_errno(nice(arg1));
8171 #endif
8172 case TARGET_NR_sync:
8173 sync();
8174 return 0;
8175 #if defined(TARGET_NR_syncfs) && defined(CONFIG_SYNCFS)
8176 case TARGET_NR_syncfs:
8177 return get_errno(syncfs(arg1));
8178 #endif
8179 case TARGET_NR_kill:
8180 return get_errno(safe_kill(arg1, target_to_host_signal(arg2)));
8181 #ifdef TARGET_NR_rename
8182 case TARGET_NR_rename:
8184 void *p2;
8185 p = lock_user_string(arg1);
8186 p2 = lock_user_string(arg2);
8187 if (!p || !p2)
8188 ret = -TARGET_EFAULT;
8189 else
8190 ret = get_errno(rename(p, p2));
8191 unlock_user(p2, arg2, 0);
8192 unlock_user(p, arg1, 0);
8194 return ret;
8195 #endif
8196 #if defined(TARGET_NR_renameat)
8197 case TARGET_NR_renameat:
8199 void *p2;
8200 p = lock_user_string(arg2);
8201 p2 = lock_user_string(arg4);
8202 if (!p || !p2)
8203 ret = -TARGET_EFAULT;
8204 else
8205 ret = get_errno(renameat(arg1, p, arg3, p2));
8206 unlock_user(p2, arg4, 0);
8207 unlock_user(p, arg2, 0);
8209 return ret;
8210 #endif
8211 #if defined(TARGET_NR_renameat2)
8212 case TARGET_NR_renameat2:
8214 void *p2;
8215 p = lock_user_string(arg2);
8216 p2 = lock_user_string(arg4);
8217 if (!p || !p2) {
8218 ret = -TARGET_EFAULT;
8219 } else {
8220 ret = get_errno(sys_renameat2(arg1, p, arg3, p2, arg5));
8222 unlock_user(p2, arg4, 0);
8223 unlock_user(p, arg2, 0);
8225 return ret;
8226 #endif
8227 #ifdef TARGET_NR_mkdir
8228 case TARGET_NR_mkdir:
8229 if (!(p = lock_user_string(arg1)))
8230 return -TARGET_EFAULT;
8231 ret = get_errno(mkdir(p, arg2));
8232 unlock_user(p, arg1, 0);
8233 return ret;
8234 #endif
8235 #if defined(TARGET_NR_mkdirat)
8236 case TARGET_NR_mkdirat:
8237 if (!(p = lock_user_string(arg2)))
8238 return -TARGET_EFAULT;
8239 ret = get_errno(mkdirat(arg1, p, arg3));
8240 unlock_user(p, arg2, 0);
8241 return ret;
8242 #endif
8243 #ifdef TARGET_NR_rmdir
8244 case TARGET_NR_rmdir:
8245 if (!(p = lock_user_string(arg1)))
8246 return -TARGET_EFAULT;
8247 ret = get_errno(rmdir(p));
8248 unlock_user(p, arg1, 0);
8249 return ret;
8250 #endif
8251 case TARGET_NR_dup:
8252 ret = get_errno(dup(arg1));
8253 if (ret >= 0) {
8254 fd_trans_dup(arg1, ret);
8256 return ret;
8257 #ifdef TARGET_NR_pipe
8258 case TARGET_NR_pipe:
8259 return do_pipe(cpu_env, arg1, 0, 0);
8260 #endif
8261 #ifdef TARGET_NR_pipe2
8262 case TARGET_NR_pipe2:
8263 return do_pipe(cpu_env, arg1,
8264 target_to_host_bitmask(arg2, fcntl_flags_tbl), 1);
8265 #endif
8266 case TARGET_NR_times:
8268 struct target_tms *tmsp;
8269 struct tms tms;
8270 ret = get_errno(times(&tms));
8271 if (arg1) {
8272 tmsp = lock_user(VERIFY_WRITE, arg1, sizeof(struct target_tms), 0);
8273 if (!tmsp)
8274 return -TARGET_EFAULT;
8275 tmsp->tms_utime = tswapal(host_to_target_clock_t(tms.tms_utime));
8276 tmsp->tms_stime = tswapal(host_to_target_clock_t(tms.tms_stime));
8277 tmsp->tms_cutime = tswapal(host_to_target_clock_t(tms.tms_cutime));
8278 tmsp->tms_cstime = tswapal(host_to_target_clock_t(tms.tms_cstime));
8280 if (!is_error(ret))
8281 ret = host_to_target_clock_t(ret);
8283 return ret;
8284 case TARGET_NR_acct:
8285 if (arg1 == 0) {
8286 ret = get_errno(acct(NULL));
8287 } else {
8288 if (!(p = lock_user_string(arg1))) {
8289 return -TARGET_EFAULT;
8291 ret = get_errno(acct(path(p)));
8292 unlock_user(p, arg1, 0);
8294 return ret;
8295 #ifdef TARGET_NR_umount2
8296 case TARGET_NR_umount2:
8297 if (!(p = lock_user_string(arg1)))
8298 return -TARGET_EFAULT;
8299 ret = get_errno(umount2(p, arg2));
8300 unlock_user(p, arg1, 0);
8301 return ret;
8302 #endif
8303 case TARGET_NR_ioctl:
8304 return do_ioctl(arg1, arg2, arg3);
8305 #ifdef TARGET_NR_fcntl
8306 case TARGET_NR_fcntl:
8307 return do_fcntl(arg1, arg2, arg3);
8308 #endif
8309 case TARGET_NR_setpgid:
8310 return get_errno(setpgid(arg1, arg2));
8311 case TARGET_NR_umask:
8312 return get_errno(umask(arg1));
8313 case TARGET_NR_chroot:
8314 if (!(p = lock_user_string(arg1)))
8315 return -TARGET_EFAULT;
8316 ret = get_errno(chroot(p));
8317 unlock_user(p, arg1, 0);
8318 return ret;
8319 #ifdef TARGET_NR_dup2
8320 case TARGET_NR_dup2:
8321 ret = get_errno(dup2(arg1, arg2));
8322 if (ret >= 0) {
8323 fd_trans_dup(arg1, arg2);
8325 return ret;
8326 #endif
8327 #if defined(CONFIG_DUP3) && defined(TARGET_NR_dup3)
8328 case TARGET_NR_dup3:
8330 int host_flags;
8332 if ((arg3 & ~TARGET_O_CLOEXEC) != 0) {
8333 return -EINVAL;
8335 host_flags = target_to_host_bitmask(arg3, fcntl_flags_tbl);
8336 ret = get_errno(dup3(arg1, arg2, host_flags));
8337 if (ret >= 0) {
8338 fd_trans_dup(arg1, arg2);
8340 return ret;
8342 #endif
8343 #ifdef TARGET_NR_getppid /* not on alpha */
8344 case TARGET_NR_getppid:
8345 return get_errno(getppid());
8346 #endif
8347 #ifdef TARGET_NR_getpgrp
8348 case TARGET_NR_getpgrp:
8349 return get_errno(getpgrp());
8350 #endif
8351 case TARGET_NR_setsid:
8352 return get_errno(setsid());
8353 #ifdef TARGET_NR_sigaction
8354 case TARGET_NR_sigaction:
8356 #if defined(TARGET_ALPHA)
8357 struct target_sigaction act, oact, *pact = 0;
8358 struct target_old_sigaction *old_act;
8359 if (arg2) {
8360 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1))
8361 return -TARGET_EFAULT;
8362 act._sa_handler = old_act->_sa_handler;
8363 target_siginitset(&act.sa_mask, old_act->sa_mask);
8364 act.sa_flags = old_act->sa_flags;
8365 act.sa_restorer = 0;
8366 unlock_user_struct(old_act, arg2, 0);
8367 pact = &act;
8369 ret = get_errno(do_sigaction(arg1, pact, &oact));
8370 if (!is_error(ret) && arg3) {
8371 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0))
8372 return -TARGET_EFAULT;
8373 old_act->_sa_handler = oact._sa_handler;
8374 old_act->sa_mask = oact.sa_mask.sig[0];
8375 old_act->sa_flags = oact.sa_flags;
8376 unlock_user_struct(old_act, arg3, 1);
8378 #elif defined(TARGET_MIPS)
8379 struct target_sigaction act, oact, *pact, *old_act;
8381 if (arg2) {
8382 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1))
8383 return -TARGET_EFAULT;
8384 act._sa_handler = old_act->_sa_handler;
8385 target_siginitset(&act.sa_mask, old_act->sa_mask.sig[0]);
8386 act.sa_flags = old_act->sa_flags;
8387 unlock_user_struct(old_act, arg2, 0);
8388 pact = &act;
8389 } else {
8390 pact = NULL;
8393 ret = get_errno(do_sigaction(arg1, pact, &oact));
8395 if (!is_error(ret) && arg3) {
8396 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0))
8397 return -TARGET_EFAULT;
8398 old_act->_sa_handler = oact._sa_handler;
8399 old_act->sa_flags = oact.sa_flags;
8400 old_act->sa_mask.sig[0] = oact.sa_mask.sig[0];
8401 old_act->sa_mask.sig[1] = 0;
8402 old_act->sa_mask.sig[2] = 0;
8403 old_act->sa_mask.sig[3] = 0;
8404 unlock_user_struct(old_act, arg3, 1);
8406 #else
8407 struct target_old_sigaction *old_act;
8408 struct target_sigaction act, oact, *pact;
8409 if (arg2) {
8410 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1))
8411 return -TARGET_EFAULT;
8412 act._sa_handler = old_act->_sa_handler;
8413 target_siginitset(&act.sa_mask, old_act->sa_mask);
8414 act.sa_flags = old_act->sa_flags;
8415 act.sa_restorer = old_act->sa_restorer;
8416 #ifdef TARGET_ARCH_HAS_KA_RESTORER
8417 act.ka_restorer = 0;
8418 #endif
8419 unlock_user_struct(old_act, arg2, 0);
8420 pact = &act;
8421 } else {
8422 pact = NULL;
8424 ret = get_errno(do_sigaction(arg1, pact, &oact));
8425 if (!is_error(ret) && arg3) {
8426 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0))
8427 return -TARGET_EFAULT;
8428 old_act->_sa_handler = oact._sa_handler;
8429 old_act->sa_mask = oact.sa_mask.sig[0];
8430 old_act->sa_flags = oact.sa_flags;
8431 old_act->sa_restorer = oact.sa_restorer;
8432 unlock_user_struct(old_act, arg3, 1);
8434 #endif
8436 return ret;
8437 #endif
8438 case TARGET_NR_rt_sigaction:
8440 #if defined(TARGET_ALPHA)
8441 /* For Alpha and SPARC this is a 5 argument syscall, with
8442 * a 'restorer' parameter which must be copied into the
8443 * sa_restorer field of the sigaction struct.
8444 * For Alpha that 'restorer' is arg5; for SPARC it is arg4,
8445 * and arg5 is the sigsetsize.
8446 * Alpha also has a separate rt_sigaction struct that it uses
8447 * here; SPARC uses the usual sigaction struct.
8449 struct target_rt_sigaction *rt_act;
8450 struct target_sigaction act, oact, *pact = 0;
8452 if (arg4 != sizeof(target_sigset_t)) {
8453 return -TARGET_EINVAL;
8455 if (arg2) {
8456 if (!lock_user_struct(VERIFY_READ, rt_act, arg2, 1))
8457 return -TARGET_EFAULT;
8458 act._sa_handler = rt_act->_sa_handler;
8459 act.sa_mask = rt_act->sa_mask;
8460 act.sa_flags = rt_act->sa_flags;
8461 act.sa_restorer = arg5;
8462 unlock_user_struct(rt_act, arg2, 0);
8463 pact = &act;
8465 ret = get_errno(do_sigaction(arg1, pact, &oact));
8466 if (!is_error(ret) && arg3) {
8467 if (!lock_user_struct(VERIFY_WRITE, rt_act, arg3, 0))
8468 return -TARGET_EFAULT;
8469 rt_act->_sa_handler = oact._sa_handler;
8470 rt_act->sa_mask = oact.sa_mask;
8471 rt_act->sa_flags = oact.sa_flags;
8472 unlock_user_struct(rt_act, arg3, 1);
8474 #else
8475 #ifdef TARGET_SPARC
8476 target_ulong restorer = arg4;
8477 target_ulong sigsetsize = arg5;
8478 #else
8479 target_ulong sigsetsize = arg4;
8480 #endif
8481 struct target_sigaction *act;
8482 struct target_sigaction *oact;
8484 if (sigsetsize != sizeof(target_sigset_t)) {
8485 return -TARGET_EINVAL;
8487 if (arg2) {
8488 if (!lock_user_struct(VERIFY_READ, act, arg2, 1)) {
8489 return -TARGET_EFAULT;
8491 #ifdef TARGET_ARCH_HAS_KA_RESTORER
8492 act->ka_restorer = restorer;
8493 #endif
8494 } else {
8495 act = NULL;
8497 if (arg3) {
8498 if (!lock_user_struct(VERIFY_WRITE, oact, arg3, 0)) {
8499 ret = -TARGET_EFAULT;
8500 goto rt_sigaction_fail;
8502 } else
8503 oact = NULL;
8504 ret = get_errno(do_sigaction(arg1, act, oact));
8505 rt_sigaction_fail:
8506 if (act)
8507 unlock_user_struct(act, arg2, 0);
8508 if (oact)
8509 unlock_user_struct(oact, arg3, 1);
8510 #endif
8512 return ret;
8513 #ifdef TARGET_NR_sgetmask /* not on alpha */
8514 case TARGET_NR_sgetmask:
8516 sigset_t cur_set;
8517 abi_ulong target_set;
8518 ret = do_sigprocmask(0, NULL, &cur_set);
8519 if (!ret) {
8520 host_to_target_old_sigset(&target_set, &cur_set);
8521 ret = target_set;
8524 return ret;
8525 #endif
8526 #ifdef TARGET_NR_ssetmask /* not on alpha */
8527 case TARGET_NR_ssetmask:
8529 sigset_t set, oset;
8530 abi_ulong target_set = arg1;
8531 target_to_host_old_sigset(&set, &target_set);
8532 ret = do_sigprocmask(SIG_SETMASK, &set, &oset);
8533 if (!ret) {
8534 host_to_target_old_sigset(&target_set, &oset);
8535 ret = target_set;
8538 return ret;
8539 #endif
8540 #ifdef TARGET_NR_sigprocmask
8541 case TARGET_NR_sigprocmask:
8543 #if defined(TARGET_ALPHA)
8544 sigset_t set, oldset;
8545 abi_ulong mask;
8546 int how;
8548 switch (arg1) {
8549 case TARGET_SIG_BLOCK:
8550 how = SIG_BLOCK;
8551 break;
8552 case TARGET_SIG_UNBLOCK:
8553 how = SIG_UNBLOCK;
8554 break;
8555 case TARGET_SIG_SETMASK:
8556 how = SIG_SETMASK;
8557 break;
8558 default:
8559 return -TARGET_EINVAL;
8561 mask = arg2;
8562 target_to_host_old_sigset(&set, &mask);
8564 ret = do_sigprocmask(how, &set, &oldset);
8565 if (!is_error(ret)) {
8566 host_to_target_old_sigset(&mask, &oldset);
8567 ret = mask;
8568 ((CPUAlphaState *)cpu_env)->ir[IR_V0] = 0; /* force no error */
8570 #else
8571 sigset_t set, oldset, *set_ptr;
8572 int how;
8574 if (arg2) {
8575 switch (arg1) {
8576 case TARGET_SIG_BLOCK:
8577 how = SIG_BLOCK;
8578 break;
8579 case TARGET_SIG_UNBLOCK:
8580 how = SIG_UNBLOCK;
8581 break;
8582 case TARGET_SIG_SETMASK:
8583 how = SIG_SETMASK;
8584 break;
8585 default:
8586 return -TARGET_EINVAL;
8588 if (!(p = lock_user(VERIFY_READ, arg2, sizeof(target_sigset_t), 1)))
8589 return -TARGET_EFAULT;
8590 target_to_host_old_sigset(&set, p);
8591 unlock_user(p, arg2, 0);
8592 set_ptr = &set;
8593 } else {
8594 how = 0;
8595 set_ptr = NULL;
8597 ret = do_sigprocmask(how, set_ptr, &oldset);
8598 if (!is_error(ret) && arg3) {
8599 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_sigset_t), 0)))
8600 return -TARGET_EFAULT;
8601 host_to_target_old_sigset(p, &oldset);
8602 unlock_user(p, arg3, sizeof(target_sigset_t));
8604 #endif
8606 return ret;
8607 #endif
8608 case TARGET_NR_rt_sigprocmask:
8610 int how = arg1;
8611 sigset_t set, oldset, *set_ptr;
8613 if (arg4 != sizeof(target_sigset_t)) {
8614 return -TARGET_EINVAL;
8617 if (arg2) {
8618 switch(how) {
8619 case TARGET_SIG_BLOCK:
8620 how = SIG_BLOCK;
8621 break;
8622 case TARGET_SIG_UNBLOCK:
8623 how = SIG_UNBLOCK;
8624 break;
8625 case TARGET_SIG_SETMASK:
8626 how = SIG_SETMASK;
8627 break;
8628 default:
8629 return -TARGET_EINVAL;
8631 if (!(p = lock_user(VERIFY_READ, arg2, sizeof(target_sigset_t), 1)))
8632 return -TARGET_EFAULT;
8633 target_to_host_sigset(&set, p);
8634 unlock_user(p, arg2, 0);
8635 set_ptr = &set;
8636 } else {
8637 how = 0;
8638 set_ptr = NULL;
8640 ret = do_sigprocmask(how, set_ptr, &oldset);
8641 if (!is_error(ret) && arg3) {
8642 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_sigset_t), 0)))
8643 return -TARGET_EFAULT;
8644 host_to_target_sigset(p, &oldset);
8645 unlock_user(p, arg3, sizeof(target_sigset_t));
8648 return ret;
8649 #ifdef TARGET_NR_sigpending
8650 case TARGET_NR_sigpending:
8652 sigset_t set;
8653 ret = get_errno(sigpending(&set));
8654 if (!is_error(ret)) {
8655 if (!(p = lock_user(VERIFY_WRITE, arg1, sizeof(target_sigset_t), 0)))
8656 return -TARGET_EFAULT;
8657 host_to_target_old_sigset(p, &set);
8658 unlock_user(p, arg1, sizeof(target_sigset_t));
8661 return ret;
8662 #endif
8663 case TARGET_NR_rt_sigpending:
8665 sigset_t set;
8667 /* Yes, this check is >, not != like most. We follow the kernel's
8668 * logic and it does it like this because it implements
8669 * NR_sigpending through the same code path, and in that case
8670 * the old_sigset_t is smaller in size.
8672 if (arg2 > sizeof(target_sigset_t)) {
8673 return -TARGET_EINVAL;
8676 ret = get_errno(sigpending(&set));
8677 if (!is_error(ret)) {
8678 if (!(p = lock_user(VERIFY_WRITE, arg1, sizeof(target_sigset_t), 0)))
8679 return -TARGET_EFAULT;
8680 host_to_target_sigset(p, &set);
8681 unlock_user(p, arg1, sizeof(target_sigset_t));
8684 return ret;
8685 #ifdef TARGET_NR_sigsuspend
8686 case TARGET_NR_sigsuspend:
8688 TaskState *ts = cpu->opaque;
8689 #if defined(TARGET_ALPHA)
8690 abi_ulong mask = arg1;
8691 target_to_host_old_sigset(&ts->sigsuspend_mask, &mask);
8692 #else
8693 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1)))
8694 return -TARGET_EFAULT;
8695 target_to_host_old_sigset(&ts->sigsuspend_mask, p);
8696 unlock_user(p, arg1, 0);
8697 #endif
8698 ret = get_errno(safe_rt_sigsuspend(&ts->sigsuspend_mask,
8699 SIGSET_T_SIZE));
8700 if (ret != -TARGET_ERESTARTSYS) {
8701 ts->in_sigsuspend = 1;
8704 return ret;
8705 #endif
8706 case TARGET_NR_rt_sigsuspend:
8708 TaskState *ts = cpu->opaque;
8710 if (arg2 != sizeof(target_sigset_t)) {
8711 return -TARGET_EINVAL;
8713 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1)))
8714 return -TARGET_EFAULT;
8715 target_to_host_sigset(&ts->sigsuspend_mask, p);
8716 unlock_user(p, arg1, 0);
8717 ret = get_errno(safe_rt_sigsuspend(&ts->sigsuspend_mask,
8718 SIGSET_T_SIZE));
8719 if (ret != -TARGET_ERESTARTSYS) {
8720 ts->in_sigsuspend = 1;
8723 return ret;
8724 #ifdef TARGET_NR_rt_sigtimedwait
8725 case TARGET_NR_rt_sigtimedwait:
8727 sigset_t set;
8728 struct timespec uts, *puts;
8729 siginfo_t uinfo;
8731 if (arg4 != sizeof(target_sigset_t)) {
8732 return -TARGET_EINVAL;
8735 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1)))
8736 return -TARGET_EFAULT;
8737 target_to_host_sigset(&set, p);
8738 unlock_user(p, arg1, 0);
8739 if (arg3) {
8740 puts = &uts;
8741 target_to_host_timespec(puts, arg3);
8742 } else {
8743 puts = NULL;
8745 ret = get_errno(safe_rt_sigtimedwait(&set, &uinfo, puts,
8746 SIGSET_T_SIZE));
8747 if (!is_error(ret)) {
8748 if (arg2) {
8749 p = lock_user(VERIFY_WRITE, arg2, sizeof(target_siginfo_t),
8751 if (!p) {
8752 return -TARGET_EFAULT;
8754 host_to_target_siginfo(p, &uinfo);
8755 unlock_user(p, arg2, sizeof(target_siginfo_t));
8757 ret = host_to_target_signal(ret);
8760 return ret;
8761 #endif
8762 case TARGET_NR_rt_sigqueueinfo:
8764 siginfo_t uinfo;
8766 p = lock_user(VERIFY_READ, arg3, sizeof(target_siginfo_t), 1);
8767 if (!p) {
8768 return -TARGET_EFAULT;
8770 target_to_host_siginfo(&uinfo, p);
8771 unlock_user(p, arg3, 0);
8772 ret = get_errno(sys_rt_sigqueueinfo(arg1, arg2, &uinfo));
8774 return ret;
8775 case TARGET_NR_rt_tgsigqueueinfo:
8777 siginfo_t uinfo;
8779 p = lock_user(VERIFY_READ, arg4, sizeof(target_siginfo_t), 1);
8780 if (!p) {
8781 return -TARGET_EFAULT;
8783 target_to_host_siginfo(&uinfo, p);
8784 unlock_user(p, arg4, 0);
8785 ret = get_errno(sys_rt_tgsigqueueinfo(arg1, arg2, arg3, &uinfo));
8787 return ret;
8788 #ifdef TARGET_NR_sigreturn
8789 case TARGET_NR_sigreturn:
8790 if (block_signals()) {
8791 return -TARGET_ERESTARTSYS;
8793 return do_sigreturn(cpu_env);
8794 #endif
8795 case TARGET_NR_rt_sigreturn:
8796 if (block_signals()) {
8797 return -TARGET_ERESTARTSYS;
8799 return do_rt_sigreturn(cpu_env);
8800 case TARGET_NR_sethostname:
8801 if (!(p = lock_user_string(arg1)))
8802 return -TARGET_EFAULT;
8803 ret = get_errno(sethostname(p, arg2));
8804 unlock_user(p, arg1, 0);
8805 return ret;
8806 #ifdef TARGET_NR_setrlimit
8807 case TARGET_NR_setrlimit:
8809 int resource = target_to_host_resource(arg1);
8810 struct target_rlimit *target_rlim;
8811 struct rlimit rlim;
8812 if (!lock_user_struct(VERIFY_READ, target_rlim, arg2, 1))
8813 return -TARGET_EFAULT;
8814 rlim.rlim_cur = target_to_host_rlim(target_rlim->rlim_cur);
8815 rlim.rlim_max = target_to_host_rlim(target_rlim->rlim_max);
8816 unlock_user_struct(target_rlim, arg2, 0);
8818 * If we just passed through resource limit settings for memory then
8819 * they would also apply to QEMU's own allocations, and QEMU will
8820 * crash or hang or die if its allocations fail. Ideally we would
8821 * track the guest allocations in QEMU and apply the limits ourselves.
8822 * For now, just tell the guest the call succeeded but don't actually
8823 * limit anything.
8825 if (resource != RLIMIT_AS &&
8826 resource != RLIMIT_DATA &&
8827 resource != RLIMIT_STACK) {
8828 return get_errno(setrlimit(resource, &rlim));
8829 } else {
8830 return 0;
8833 #endif
8834 #ifdef TARGET_NR_getrlimit
8835 case TARGET_NR_getrlimit:
8837 int resource = target_to_host_resource(arg1);
8838 struct target_rlimit *target_rlim;
8839 struct rlimit rlim;
8841 ret = get_errno(getrlimit(resource, &rlim));
8842 if (!is_error(ret)) {
8843 if (!lock_user_struct(VERIFY_WRITE, target_rlim, arg2, 0))
8844 return -TARGET_EFAULT;
8845 target_rlim->rlim_cur = host_to_target_rlim(rlim.rlim_cur);
8846 target_rlim->rlim_max = host_to_target_rlim(rlim.rlim_max);
8847 unlock_user_struct(target_rlim, arg2, 1);
8850 return ret;
8851 #endif
8852 case TARGET_NR_getrusage:
8854 struct rusage rusage;
8855 ret = get_errno(getrusage(arg1, &rusage));
8856 if (!is_error(ret)) {
8857 ret = host_to_target_rusage(arg2, &rusage);
8860 return ret;
8861 #if defined(TARGET_NR_gettimeofday)
8862 case TARGET_NR_gettimeofday:
8864 struct timeval tv;
8865 struct timezone tz;
8867 ret = get_errno(gettimeofday(&tv, &tz));
8868 if (!is_error(ret)) {
8869 if (arg1 && copy_to_user_timeval(arg1, &tv)) {
8870 return -TARGET_EFAULT;
8872 if (arg2 && copy_to_user_timezone(arg2, &tz)) {
8873 return -TARGET_EFAULT;
8877 return ret;
8878 #endif
8879 #if defined(TARGET_NR_settimeofday)
8880 case TARGET_NR_settimeofday:
8882 struct timeval tv, *ptv = NULL;
8883 struct timezone tz, *ptz = NULL;
8885 if (arg1) {
8886 if (copy_from_user_timeval(&tv, arg1)) {
8887 return -TARGET_EFAULT;
8889 ptv = &tv;
8892 if (arg2) {
8893 if (copy_from_user_timezone(&tz, arg2)) {
8894 return -TARGET_EFAULT;
8896 ptz = &tz;
8899 return get_errno(settimeofday(ptv, ptz));
8901 #endif
8902 #if defined(TARGET_NR_select)
8903 case TARGET_NR_select:
8904 #if defined(TARGET_WANT_NI_OLD_SELECT)
8905 /* some architectures used to have old_select here
8906 * but now ENOSYS it.
8908 ret = -TARGET_ENOSYS;
8909 #elif defined(TARGET_WANT_OLD_SYS_SELECT)
8910 ret = do_old_select(arg1);
8911 #else
8912 ret = do_select(arg1, arg2, arg3, arg4, arg5);
8913 #endif
8914 return ret;
8915 #endif
8916 #ifdef TARGET_NR_pselect6
8917 case TARGET_NR_pselect6:
8919 abi_long rfd_addr, wfd_addr, efd_addr, n, ts_addr;
8920 fd_set rfds, wfds, efds;
8921 fd_set *rfds_ptr, *wfds_ptr, *efds_ptr;
8922 struct timespec ts, *ts_ptr;
8925 * The 6th arg is actually two args smashed together,
8926 * so we cannot use the C library.
8928 sigset_t set;
8929 struct {
8930 sigset_t *set;
8931 size_t size;
8932 } sig, *sig_ptr;
8934 abi_ulong arg_sigset, arg_sigsize, *arg7;
8935 target_sigset_t *target_sigset;
8937 n = arg1;
8938 rfd_addr = arg2;
8939 wfd_addr = arg3;
8940 efd_addr = arg4;
8941 ts_addr = arg5;
8943 ret = copy_from_user_fdset_ptr(&rfds, &rfds_ptr, rfd_addr, n);
8944 if (ret) {
8945 return ret;
8947 ret = copy_from_user_fdset_ptr(&wfds, &wfds_ptr, wfd_addr, n);
8948 if (ret) {
8949 return ret;
8951 ret = copy_from_user_fdset_ptr(&efds, &efds_ptr, efd_addr, n);
8952 if (ret) {
8953 return ret;
8957 * This takes a timespec, and not a timeval, so we cannot
8958 * use the do_select() helper ...
8960 if (ts_addr) {
8961 if (target_to_host_timespec(&ts, ts_addr)) {
8962 return -TARGET_EFAULT;
8964 ts_ptr = &ts;
8965 } else {
8966 ts_ptr = NULL;
8969 /* Extract the two packed args for the sigset */
8970 if (arg6) {
8971 sig_ptr = &sig;
8972 sig.size = SIGSET_T_SIZE;
8974 arg7 = lock_user(VERIFY_READ, arg6, sizeof(*arg7) * 2, 1);
8975 if (!arg7) {
8976 return -TARGET_EFAULT;
8978 arg_sigset = tswapal(arg7[0]);
8979 arg_sigsize = tswapal(arg7[1]);
8980 unlock_user(arg7, arg6, 0);
8982 if (arg_sigset) {
8983 sig.set = &set;
8984 if (arg_sigsize != sizeof(*target_sigset)) {
8985 /* Like the kernel, we enforce correct size sigsets */
8986 return -TARGET_EINVAL;
8988 target_sigset = lock_user(VERIFY_READ, arg_sigset,
8989 sizeof(*target_sigset), 1);
8990 if (!target_sigset) {
8991 return -TARGET_EFAULT;
8993 target_to_host_sigset(&set, target_sigset);
8994 unlock_user(target_sigset, arg_sigset, 0);
8995 } else {
8996 sig.set = NULL;
8998 } else {
8999 sig_ptr = NULL;
9002 ret = get_errno(safe_pselect6(n, rfds_ptr, wfds_ptr, efds_ptr,
9003 ts_ptr, sig_ptr));
9005 if (!is_error(ret)) {
9006 if (rfd_addr && copy_to_user_fdset(rfd_addr, &rfds, n))
9007 return -TARGET_EFAULT;
9008 if (wfd_addr && copy_to_user_fdset(wfd_addr, &wfds, n))
9009 return -TARGET_EFAULT;
9010 if (efd_addr && copy_to_user_fdset(efd_addr, &efds, n))
9011 return -TARGET_EFAULT;
9013 if (ts_addr && host_to_target_timespec(ts_addr, &ts))
9014 return -TARGET_EFAULT;
9017 return ret;
9018 #endif
9019 #ifdef TARGET_NR_symlink
9020 case TARGET_NR_symlink:
9022 void *p2;
9023 p = lock_user_string(arg1);
9024 p2 = lock_user_string(arg2);
9025 if (!p || !p2)
9026 ret = -TARGET_EFAULT;
9027 else
9028 ret = get_errno(symlink(p, p2));
9029 unlock_user(p2, arg2, 0);
9030 unlock_user(p, arg1, 0);
9032 return ret;
9033 #endif
9034 #if defined(TARGET_NR_symlinkat)
9035 case TARGET_NR_symlinkat:
9037 void *p2;
9038 p = lock_user_string(arg1);
9039 p2 = lock_user_string(arg3);
9040 if (!p || !p2)
9041 ret = -TARGET_EFAULT;
9042 else
9043 ret = get_errno(symlinkat(p, arg2, p2));
9044 unlock_user(p2, arg3, 0);
9045 unlock_user(p, arg1, 0);
9047 return ret;
9048 #endif
9049 #ifdef TARGET_NR_readlink
9050 case TARGET_NR_readlink:
9052 void *p2;
9053 p = lock_user_string(arg1);
9054 p2 = lock_user(VERIFY_WRITE, arg2, arg3, 0);
9055 if (!p || !p2) {
9056 ret = -TARGET_EFAULT;
9057 } else if (!arg3) {
9058 /* Short circuit this for the magic exe check. */
9059 ret = -TARGET_EINVAL;
9060 } else if (is_proc_myself((const char *)p, "exe")) {
9061 char real[PATH_MAX], *temp;
9062 temp = realpath(exec_path, real);
9063 /* Return value is # of bytes that we wrote to the buffer. */
9064 if (temp == NULL) {
9065 ret = get_errno(-1);
9066 } else {
9067 /* Don't worry about sign mismatch as earlier mapping
9068 * logic would have thrown a bad address error. */
9069 ret = MIN(strlen(real), arg3);
9070 /* We cannot NUL terminate the string. */
9071 memcpy(p2, real, ret);
9073 } else {
9074 ret = get_errno(readlink(path(p), p2, arg3));
9076 unlock_user(p2, arg2, ret);
9077 unlock_user(p, arg1, 0);
9079 return ret;
9080 #endif
9081 #if defined(TARGET_NR_readlinkat)
9082 case TARGET_NR_readlinkat:
9084 void *p2;
9085 p = lock_user_string(arg2);
9086 p2 = lock_user(VERIFY_WRITE, arg3, arg4, 0);
9087 if (!p || !p2) {
9088 ret = -TARGET_EFAULT;
9089 } else if (is_proc_myself((const char *)p, "exe")) {
9090 char real[PATH_MAX], *temp;
9091 temp = realpath(exec_path, real);
9092 ret = temp == NULL ? get_errno(-1) : strlen(real) ;
9093 snprintf((char *)p2, arg4, "%s", real);
9094 } else {
9095 ret = get_errno(readlinkat(arg1, path(p), p2, arg4));
9097 unlock_user(p2, arg3, ret);
9098 unlock_user(p, arg2, 0);
9100 return ret;
9101 #endif
9102 #ifdef TARGET_NR_swapon
9103 case TARGET_NR_swapon:
9104 if (!(p = lock_user_string(arg1)))
9105 return -TARGET_EFAULT;
9106 ret = get_errno(swapon(p, arg2));
9107 unlock_user(p, arg1, 0);
9108 return ret;
9109 #endif
9110 case TARGET_NR_reboot:
9111 if (arg3 == LINUX_REBOOT_CMD_RESTART2) {
9112 /* arg4 must be ignored in all other cases */
9113 p = lock_user_string(arg4);
9114 if (!p) {
9115 return -TARGET_EFAULT;
9117 ret = get_errno(reboot(arg1, arg2, arg3, p));
9118 unlock_user(p, arg4, 0);
9119 } else {
9120 ret = get_errno(reboot(arg1, arg2, arg3, NULL));
9122 return ret;
9123 #ifdef TARGET_NR_mmap
9124 case TARGET_NR_mmap:
9125 #if (defined(TARGET_I386) && defined(TARGET_ABI32)) || \
9126 (defined(TARGET_ARM) && defined(TARGET_ABI32)) || \
9127 defined(TARGET_M68K) || defined(TARGET_CRIS) || defined(TARGET_MICROBLAZE) \
9128 || defined(TARGET_S390X)
9130 abi_ulong *v;
9131 abi_ulong v1, v2, v3, v4, v5, v6;
9132 if (!(v = lock_user(VERIFY_READ, arg1, 6 * sizeof(abi_ulong), 1)))
9133 return -TARGET_EFAULT;
9134 v1 = tswapal(v[0]);
9135 v2 = tswapal(v[1]);
9136 v3 = tswapal(v[2]);
9137 v4 = tswapal(v[3]);
9138 v5 = tswapal(v[4]);
9139 v6 = tswapal(v[5]);
9140 unlock_user(v, arg1, 0);
9141 ret = get_errno(target_mmap(v1, v2, v3,
9142 target_to_host_bitmask(v4, mmap_flags_tbl),
9143 v5, v6));
9145 #else
9146 ret = get_errno(target_mmap(arg1, arg2, arg3,
9147 target_to_host_bitmask(arg4, mmap_flags_tbl),
9148 arg5,
9149 arg6));
9150 #endif
9151 return ret;
9152 #endif
9153 #ifdef TARGET_NR_mmap2
9154 case TARGET_NR_mmap2:
9155 #ifndef MMAP_SHIFT
9156 #define MMAP_SHIFT 12
9157 #endif
9158 ret = target_mmap(arg1, arg2, arg3,
9159 target_to_host_bitmask(arg4, mmap_flags_tbl),
9160 arg5, arg6 << MMAP_SHIFT);
9161 return get_errno(ret);
9162 #endif
9163 case TARGET_NR_munmap:
9164 return get_errno(target_munmap(arg1, arg2));
9165 case TARGET_NR_mprotect:
9167 TaskState *ts = cpu->opaque;
9168 /* Special hack to detect libc making the stack executable. */
9169 if ((arg3 & PROT_GROWSDOWN)
9170 && arg1 >= ts->info->stack_limit
9171 && arg1 <= ts->info->start_stack) {
9172 arg3 &= ~PROT_GROWSDOWN;
9173 arg2 = arg2 + arg1 - ts->info->stack_limit;
9174 arg1 = ts->info->stack_limit;
9177 return get_errno(target_mprotect(arg1, arg2, arg3));
9178 #ifdef TARGET_NR_mremap
9179 case TARGET_NR_mremap:
9180 return get_errno(target_mremap(arg1, arg2, arg3, arg4, arg5));
9181 #endif
9182 /* ??? msync/mlock/munlock are broken for softmmu. */
9183 #ifdef TARGET_NR_msync
9184 case TARGET_NR_msync:
9185 return get_errno(msync(g2h(arg1), arg2, arg3));
9186 #endif
9187 #ifdef TARGET_NR_mlock
9188 case TARGET_NR_mlock:
9189 return get_errno(mlock(g2h(arg1), arg2));
9190 #endif
9191 #ifdef TARGET_NR_munlock
9192 case TARGET_NR_munlock:
9193 return get_errno(munlock(g2h(arg1), arg2));
9194 #endif
9195 #ifdef TARGET_NR_mlockall
9196 case TARGET_NR_mlockall:
9197 return get_errno(mlockall(target_to_host_mlockall_arg(arg1)));
9198 #endif
9199 #ifdef TARGET_NR_munlockall
9200 case TARGET_NR_munlockall:
9201 return get_errno(munlockall());
9202 #endif
9203 #ifdef TARGET_NR_truncate
9204 case TARGET_NR_truncate:
9205 if (!(p = lock_user_string(arg1)))
9206 return -TARGET_EFAULT;
9207 ret = get_errno(truncate(p, arg2));
9208 unlock_user(p, arg1, 0);
9209 return ret;
9210 #endif
9211 #ifdef TARGET_NR_ftruncate
9212 case TARGET_NR_ftruncate:
9213 return get_errno(ftruncate(arg1, arg2));
9214 #endif
9215 case TARGET_NR_fchmod:
9216 return get_errno(fchmod(arg1, arg2));
9217 #if defined(TARGET_NR_fchmodat)
9218 case TARGET_NR_fchmodat:
9219 if (!(p = lock_user_string(arg2)))
9220 return -TARGET_EFAULT;
9221 ret = get_errno(fchmodat(arg1, p, arg3, 0));
9222 unlock_user(p, arg2, 0);
9223 return ret;
9224 #endif
9225 case TARGET_NR_getpriority:
9226 /* Note that negative values are valid for getpriority, so we must
9227 differentiate based on errno settings. */
9228 errno = 0;
9229 ret = getpriority(arg1, arg2);
9230 if (ret == -1 && errno != 0) {
9231 return -host_to_target_errno(errno);
9233 #ifdef TARGET_ALPHA
9234 /* Return value is the unbiased priority. Signal no error. */
9235 ((CPUAlphaState *)cpu_env)->ir[IR_V0] = 0;
9236 #else
9237 /* Return value is a biased priority to avoid negative numbers. */
9238 ret = 20 - ret;
9239 #endif
9240 return ret;
9241 case TARGET_NR_setpriority:
9242 return get_errno(setpriority(arg1, arg2, arg3));
9243 #ifdef TARGET_NR_statfs
9244 case TARGET_NR_statfs:
9245 if (!(p = lock_user_string(arg1))) {
9246 return -TARGET_EFAULT;
9248 ret = get_errno(statfs(path(p), &stfs));
9249 unlock_user(p, arg1, 0);
9250 convert_statfs:
9251 if (!is_error(ret)) {
9252 struct target_statfs *target_stfs;
9254 if (!lock_user_struct(VERIFY_WRITE, target_stfs, arg2, 0))
9255 return -TARGET_EFAULT;
9256 __put_user(stfs.f_type, &target_stfs->f_type);
9257 __put_user(stfs.f_bsize, &target_stfs->f_bsize);
9258 __put_user(stfs.f_blocks, &target_stfs->f_blocks);
9259 __put_user(stfs.f_bfree, &target_stfs->f_bfree);
9260 __put_user(stfs.f_bavail, &target_stfs->f_bavail);
9261 __put_user(stfs.f_files, &target_stfs->f_files);
9262 __put_user(stfs.f_ffree, &target_stfs->f_ffree);
9263 __put_user(stfs.f_fsid.__val[0], &target_stfs->f_fsid.val[0]);
9264 __put_user(stfs.f_fsid.__val[1], &target_stfs->f_fsid.val[1]);
9265 __put_user(stfs.f_namelen, &target_stfs->f_namelen);
9266 __put_user(stfs.f_frsize, &target_stfs->f_frsize);
9267 #ifdef _STATFS_F_FLAGS
9268 __put_user(stfs.f_flags, &target_stfs->f_flags);
9269 #else
9270 __put_user(0, &target_stfs->f_flags);
9271 #endif
9272 memset(target_stfs->f_spare, 0, sizeof(target_stfs->f_spare));
9273 unlock_user_struct(target_stfs, arg2, 1);
9275 return ret;
9276 #endif
9277 #ifdef TARGET_NR_fstatfs
9278 case TARGET_NR_fstatfs:
9279 ret = get_errno(fstatfs(arg1, &stfs));
9280 goto convert_statfs;
9281 #endif
9282 #ifdef TARGET_NR_statfs64
9283 case TARGET_NR_statfs64:
9284 if (!(p = lock_user_string(arg1))) {
9285 return -TARGET_EFAULT;
9287 ret = get_errno(statfs(path(p), &stfs));
9288 unlock_user(p, arg1, 0);
9289 convert_statfs64:
9290 if (!is_error(ret)) {
9291 struct target_statfs64 *target_stfs;
9293 if (!lock_user_struct(VERIFY_WRITE, target_stfs, arg3, 0))
9294 return -TARGET_EFAULT;
9295 __put_user(stfs.f_type, &target_stfs->f_type);
9296 __put_user(stfs.f_bsize, &target_stfs->f_bsize);
9297 __put_user(stfs.f_blocks, &target_stfs->f_blocks);
9298 __put_user(stfs.f_bfree, &target_stfs->f_bfree);
9299 __put_user(stfs.f_bavail, &target_stfs->f_bavail);
9300 __put_user(stfs.f_files, &target_stfs->f_files);
9301 __put_user(stfs.f_ffree, &target_stfs->f_ffree);
9302 __put_user(stfs.f_fsid.__val[0], &target_stfs->f_fsid.val[0]);
9303 __put_user(stfs.f_fsid.__val[1], &target_stfs->f_fsid.val[1]);
9304 __put_user(stfs.f_namelen, &target_stfs->f_namelen);
9305 __put_user(stfs.f_frsize, &target_stfs->f_frsize);
9306 memset(target_stfs->f_spare, 0, sizeof(target_stfs->f_spare));
9307 unlock_user_struct(target_stfs, arg3, 1);
9309 return ret;
9310 case TARGET_NR_fstatfs64:
9311 ret = get_errno(fstatfs(arg1, &stfs));
9312 goto convert_statfs64;
9313 #endif
9314 #ifdef TARGET_NR_socketcall
9315 case TARGET_NR_socketcall:
9316 return do_socketcall(arg1, arg2);
9317 #endif
9318 #ifdef TARGET_NR_accept
9319 case TARGET_NR_accept:
9320 return do_accept4(arg1, arg2, arg3, 0);
9321 #endif
9322 #ifdef TARGET_NR_accept4
9323 case TARGET_NR_accept4:
9324 return do_accept4(arg1, arg2, arg3, arg4);
9325 #endif
9326 #ifdef TARGET_NR_bind
9327 case TARGET_NR_bind:
9328 return do_bind(arg1, arg2, arg3);
9329 #endif
9330 #ifdef TARGET_NR_connect
9331 case TARGET_NR_connect:
9332 return do_connect(arg1, arg2, arg3);
9333 #endif
9334 #ifdef TARGET_NR_getpeername
9335 case TARGET_NR_getpeername:
9336 return do_getpeername(arg1, arg2, arg3);
9337 #endif
9338 #ifdef TARGET_NR_getsockname
9339 case TARGET_NR_getsockname:
9340 return do_getsockname(arg1, arg2, arg3);
9341 #endif
9342 #ifdef TARGET_NR_getsockopt
9343 case TARGET_NR_getsockopt:
9344 return do_getsockopt(arg1, arg2, arg3, arg4, arg5);
9345 #endif
9346 #ifdef TARGET_NR_listen
9347 case TARGET_NR_listen:
9348 return get_errno(listen(arg1, arg2));
9349 #endif
9350 #ifdef TARGET_NR_recv
9351 case TARGET_NR_recv:
9352 return do_recvfrom(arg1, arg2, arg3, arg4, 0, 0);
9353 #endif
9354 #ifdef TARGET_NR_recvfrom
9355 case TARGET_NR_recvfrom:
9356 return do_recvfrom(arg1, arg2, arg3, arg4, arg5, arg6);
9357 #endif
9358 #ifdef TARGET_NR_recvmsg
9359 case TARGET_NR_recvmsg:
9360 return do_sendrecvmsg(arg1, arg2, arg3, 0);
9361 #endif
9362 #ifdef TARGET_NR_send
9363 case TARGET_NR_send:
9364 return do_sendto(arg1, arg2, arg3, arg4, 0, 0);
9365 #endif
9366 #ifdef TARGET_NR_sendmsg
9367 case TARGET_NR_sendmsg:
9368 return do_sendrecvmsg(arg1, arg2, arg3, 1);
9369 #endif
9370 #ifdef TARGET_NR_sendmmsg
9371 case TARGET_NR_sendmmsg:
9372 return do_sendrecvmmsg(arg1, arg2, arg3, arg4, 1);
9373 #endif
9374 #ifdef TARGET_NR_recvmmsg
9375 case TARGET_NR_recvmmsg:
9376 return do_sendrecvmmsg(arg1, arg2, arg3, arg4, 0);
9377 #endif
9378 #ifdef TARGET_NR_sendto
9379 case TARGET_NR_sendto:
9380 return do_sendto(arg1, arg2, arg3, arg4, arg5, arg6);
9381 #endif
9382 #ifdef TARGET_NR_shutdown
9383 case TARGET_NR_shutdown:
9384 return get_errno(shutdown(arg1, arg2));
9385 #endif
9386 #if defined(TARGET_NR_getrandom) && defined(__NR_getrandom)
9387 case TARGET_NR_getrandom:
9388 p = lock_user(VERIFY_WRITE, arg1, arg2, 0);
9389 if (!p) {
9390 return -TARGET_EFAULT;
9392 ret = get_errno(getrandom(p, arg2, arg3));
9393 unlock_user(p, arg1, ret);
9394 return ret;
9395 #endif
9396 #ifdef TARGET_NR_socket
9397 case TARGET_NR_socket:
9398 return do_socket(arg1, arg2, arg3);
9399 #endif
9400 #ifdef TARGET_NR_socketpair
9401 case TARGET_NR_socketpair:
9402 return do_socketpair(arg1, arg2, arg3, arg4);
9403 #endif
9404 #ifdef TARGET_NR_setsockopt
9405 case TARGET_NR_setsockopt:
9406 return do_setsockopt(arg1, arg2, arg3, arg4, (socklen_t) arg5);
9407 #endif
9408 #if defined(TARGET_NR_syslog)
9409 case TARGET_NR_syslog:
9411 int len = arg2;
9413 switch (arg1) {
9414 case TARGET_SYSLOG_ACTION_CLOSE: /* Close log */
9415 case TARGET_SYSLOG_ACTION_OPEN: /* Open log */
9416 case TARGET_SYSLOG_ACTION_CLEAR: /* Clear ring buffer */
9417 case TARGET_SYSLOG_ACTION_CONSOLE_OFF: /* Disable logging */
9418 case TARGET_SYSLOG_ACTION_CONSOLE_ON: /* Enable logging */
9419 case TARGET_SYSLOG_ACTION_CONSOLE_LEVEL: /* Set messages level */
9420 case TARGET_SYSLOG_ACTION_SIZE_UNREAD: /* Number of chars */
9421 case TARGET_SYSLOG_ACTION_SIZE_BUFFER: /* Size of the buffer */
9422 return get_errno(sys_syslog((int)arg1, NULL, (int)arg3));
9423 case TARGET_SYSLOG_ACTION_READ: /* Read from log */
9424 case TARGET_SYSLOG_ACTION_READ_CLEAR: /* Read/clear msgs */
9425 case TARGET_SYSLOG_ACTION_READ_ALL: /* Read last messages */
9427 if (len < 0) {
9428 return -TARGET_EINVAL;
9430 if (len == 0) {
9431 return 0;
9433 p = lock_user(VERIFY_WRITE, arg2, arg3, 0);
9434 if (!p) {
9435 return -TARGET_EFAULT;
9437 ret = get_errno(sys_syslog((int)arg1, p, (int)arg3));
9438 unlock_user(p, arg2, arg3);
9440 return ret;
9441 default:
9442 return -TARGET_EINVAL;
9445 break;
9446 #endif
9447 case TARGET_NR_setitimer:
9449 struct itimerval value, ovalue, *pvalue;
9451 if (arg2) {
9452 pvalue = &value;
9453 if (copy_from_user_timeval(&pvalue->it_interval, arg2)
9454 || copy_from_user_timeval(&pvalue->it_value,
9455 arg2 + sizeof(struct target_timeval)))
9456 return -TARGET_EFAULT;
9457 } else {
9458 pvalue = NULL;
9460 ret = get_errno(setitimer(arg1, pvalue, &ovalue));
9461 if (!is_error(ret) && arg3) {
9462 if (copy_to_user_timeval(arg3,
9463 &ovalue.it_interval)
9464 || copy_to_user_timeval(arg3 + sizeof(struct target_timeval),
9465 &ovalue.it_value))
9466 return -TARGET_EFAULT;
9469 return ret;
9470 case TARGET_NR_getitimer:
9472 struct itimerval value;
9474 ret = get_errno(getitimer(arg1, &value));
9475 if (!is_error(ret) && arg2) {
9476 if (copy_to_user_timeval(arg2,
9477 &value.it_interval)
9478 || copy_to_user_timeval(arg2 + sizeof(struct target_timeval),
9479 &value.it_value))
9480 return -TARGET_EFAULT;
9483 return ret;
9484 #ifdef TARGET_NR_stat
9485 case TARGET_NR_stat:
9486 if (!(p = lock_user_string(arg1))) {
9487 return -TARGET_EFAULT;
9489 ret = get_errno(stat(path(p), &st));
9490 unlock_user(p, arg1, 0);
9491 goto do_stat;
9492 #endif
9493 #ifdef TARGET_NR_lstat
9494 case TARGET_NR_lstat:
9495 if (!(p = lock_user_string(arg1))) {
9496 return -TARGET_EFAULT;
9498 ret = get_errno(lstat(path(p), &st));
9499 unlock_user(p, arg1, 0);
9500 goto do_stat;
9501 #endif
9502 #ifdef TARGET_NR_fstat
9503 case TARGET_NR_fstat:
9505 ret = get_errno(fstat(arg1, &st));
9506 #if defined(TARGET_NR_stat) || defined(TARGET_NR_lstat)
9507 do_stat:
9508 #endif
9509 if (!is_error(ret)) {
9510 struct target_stat *target_st;
9512 if (!lock_user_struct(VERIFY_WRITE, target_st, arg2, 0))
9513 return -TARGET_EFAULT;
9514 memset(target_st, 0, sizeof(*target_st));
9515 __put_user(st.st_dev, &target_st->st_dev);
9516 __put_user(st.st_ino, &target_st->st_ino);
9517 __put_user(st.st_mode, &target_st->st_mode);
9518 __put_user(st.st_uid, &target_st->st_uid);
9519 __put_user(st.st_gid, &target_st->st_gid);
9520 __put_user(st.st_nlink, &target_st->st_nlink);
9521 __put_user(st.st_rdev, &target_st->st_rdev);
9522 __put_user(st.st_size, &target_st->st_size);
9523 __put_user(st.st_blksize, &target_st->st_blksize);
9524 __put_user(st.st_blocks, &target_st->st_blocks);
9525 __put_user(st.st_atime, &target_st->target_st_atime);
9526 __put_user(st.st_mtime, &target_st->target_st_mtime);
9527 __put_user(st.st_ctime, &target_st->target_st_ctime);
9528 #if (_POSIX_C_SOURCE >= 200809L || _XOPEN_SOURCE >= 700) && \
9529 defined(TARGET_STAT_HAVE_NSEC)
9530 __put_user(st.st_atim.tv_nsec,
9531 &target_st->target_st_atime_nsec);
9532 __put_user(st.st_mtim.tv_nsec,
9533 &target_st->target_st_mtime_nsec);
9534 __put_user(st.st_ctim.tv_nsec,
9535 &target_st->target_st_ctime_nsec);
9536 #endif
9537 unlock_user_struct(target_st, arg2, 1);
9540 return ret;
9541 #endif
9542 case TARGET_NR_vhangup:
9543 return get_errno(vhangup());
9544 #ifdef TARGET_NR_syscall
9545 case TARGET_NR_syscall:
9546 return do_syscall(cpu_env, arg1 & 0xffff, arg2, arg3, arg4, arg5,
9547 arg6, arg7, arg8, 0);
9548 #endif
9549 #if defined(TARGET_NR_wait4)
9550 case TARGET_NR_wait4:
9552 int status;
9553 abi_long status_ptr = arg2;
9554 struct rusage rusage, *rusage_ptr;
9555 abi_ulong target_rusage = arg4;
9556 abi_long rusage_err;
9557 if (target_rusage)
9558 rusage_ptr = &rusage;
9559 else
9560 rusage_ptr = NULL;
9561 ret = get_errno(safe_wait4(arg1, &status, arg3, rusage_ptr));
9562 if (!is_error(ret)) {
9563 if (status_ptr && ret) {
9564 status = host_to_target_waitstatus(status);
9565 if (put_user_s32(status, status_ptr))
9566 return -TARGET_EFAULT;
9568 if (target_rusage) {
9569 rusage_err = host_to_target_rusage(target_rusage, &rusage);
9570 if (rusage_err) {
9571 ret = rusage_err;
9576 return ret;
9577 #endif
9578 #ifdef TARGET_NR_swapoff
9579 case TARGET_NR_swapoff:
9580 if (!(p = lock_user_string(arg1)))
9581 return -TARGET_EFAULT;
9582 ret = get_errno(swapoff(p));
9583 unlock_user(p, arg1, 0);
9584 return ret;
9585 #endif
9586 case TARGET_NR_sysinfo:
9588 struct target_sysinfo *target_value;
9589 struct sysinfo value;
9590 ret = get_errno(sysinfo(&value));
9591 if (!is_error(ret) && arg1)
9593 if (!lock_user_struct(VERIFY_WRITE, target_value, arg1, 0))
9594 return -TARGET_EFAULT;
9595 __put_user(value.uptime, &target_value->uptime);
9596 __put_user(value.loads[0], &target_value->loads[0]);
9597 __put_user(value.loads[1], &target_value->loads[1]);
9598 __put_user(value.loads[2], &target_value->loads[2]);
9599 __put_user(value.totalram, &target_value->totalram);
9600 __put_user(value.freeram, &target_value->freeram);
9601 __put_user(value.sharedram, &target_value->sharedram);
9602 __put_user(value.bufferram, &target_value->bufferram);
9603 __put_user(value.totalswap, &target_value->totalswap);
9604 __put_user(value.freeswap, &target_value->freeswap);
9605 __put_user(value.procs, &target_value->procs);
9606 __put_user(value.totalhigh, &target_value->totalhigh);
9607 __put_user(value.freehigh, &target_value->freehigh);
9608 __put_user(value.mem_unit, &target_value->mem_unit);
9609 unlock_user_struct(target_value, arg1, 1);
9612 return ret;
9613 #ifdef TARGET_NR_ipc
9614 case TARGET_NR_ipc:
9615 return do_ipc(cpu_env, arg1, arg2, arg3, arg4, arg5, arg6);
9616 #endif
9617 #ifdef TARGET_NR_semget
9618 case TARGET_NR_semget:
9619 return get_errno(semget(arg1, arg2, arg3));
9620 #endif
9621 #ifdef TARGET_NR_semop
9622 case TARGET_NR_semop:
9623 return do_semop(arg1, arg2, arg3);
9624 #endif
9625 #ifdef TARGET_NR_semctl
9626 case TARGET_NR_semctl:
9627 return do_semctl(arg1, arg2, arg3, arg4);
9628 #endif
9629 #ifdef TARGET_NR_msgctl
9630 case TARGET_NR_msgctl:
9631 return do_msgctl(arg1, arg2, arg3);
9632 #endif
9633 #ifdef TARGET_NR_msgget
9634 case TARGET_NR_msgget:
9635 return get_errno(msgget(arg1, arg2));
9636 #endif
9637 #ifdef TARGET_NR_msgrcv
9638 case TARGET_NR_msgrcv:
9639 return do_msgrcv(arg1, arg2, arg3, arg4, arg5);
9640 #endif
9641 #ifdef TARGET_NR_msgsnd
9642 case TARGET_NR_msgsnd:
9643 return do_msgsnd(arg1, arg2, arg3, arg4);
9644 #endif
9645 #ifdef TARGET_NR_shmget
9646 case TARGET_NR_shmget:
9647 return get_errno(shmget(arg1, arg2, arg3));
9648 #endif
9649 #ifdef TARGET_NR_shmctl
9650 case TARGET_NR_shmctl:
9651 return do_shmctl(arg1, arg2, arg3);
9652 #endif
9653 #ifdef TARGET_NR_shmat
9654 case TARGET_NR_shmat:
9655 return do_shmat(cpu_env, arg1, arg2, arg3);
9656 #endif
9657 #ifdef TARGET_NR_shmdt
9658 case TARGET_NR_shmdt:
9659 return do_shmdt(arg1);
9660 #endif
9661 case TARGET_NR_fsync:
9662 return get_errno(fsync(arg1));
9663 case TARGET_NR_clone:
9664 /* Linux manages to have three different orderings for its
9665 * arguments to clone(); the BACKWARDS and BACKWARDS2 defines
9666 * match the kernel's CONFIG_CLONE_* settings.
9667 * Microblaze is further special in that it uses a sixth
9668 * implicit argument to clone for the TLS pointer.
9670 #if defined(TARGET_MICROBLAZE)
9671 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg4, arg6, arg5));
9672 #elif defined(TARGET_CLONE_BACKWARDS)
9673 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg3, arg4, arg5));
9674 #elif defined(TARGET_CLONE_BACKWARDS2)
9675 ret = get_errno(do_fork(cpu_env, arg2, arg1, arg3, arg5, arg4));
9676 #else
9677 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg3, arg5, arg4));
9678 #endif
9679 return ret;
9680 #ifdef __NR_exit_group
9681 /* new thread calls */
9682 case TARGET_NR_exit_group:
9683 preexit_cleanup(cpu_env, arg1);
9684 return get_errno(exit_group(arg1));
9685 #endif
9686 case TARGET_NR_setdomainname:
9687 if (!(p = lock_user_string(arg1)))
9688 return -TARGET_EFAULT;
9689 ret = get_errno(setdomainname(p, arg2));
9690 unlock_user(p, arg1, 0);
9691 return ret;
9692 case TARGET_NR_uname:
9693 /* no need to transcode because we use the linux syscall */
9695 struct new_utsname * buf;
9697 if (!lock_user_struct(VERIFY_WRITE, buf, arg1, 0))
9698 return -TARGET_EFAULT;
9699 ret = get_errno(sys_uname(buf));
9700 if (!is_error(ret)) {
9701 /* Overwrite the native machine name with whatever is being
9702 emulated. */
9703 g_strlcpy(buf->machine, cpu_to_uname_machine(cpu_env),
9704 sizeof(buf->machine));
9705 /* Allow the user to override the reported release. */
9706 if (qemu_uname_release && *qemu_uname_release) {
9707 g_strlcpy(buf->release, qemu_uname_release,
9708 sizeof(buf->release));
9711 unlock_user_struct(buf, arg1, 1);
9713 return ret;
9714 #ifdef TARGET_I386
9715 case TARGET_NR_modify_ldt:
9716 return do_modify_ldt(cpu_env, arg1, arg2, arg3);
9717 #if !defined(TARGET_X86_64)
9718 case TARGET_NR_vm86:
9719 return do_vm86(cpu_env, arg1, arg2);
9720 #endif
9721 #endif
9722 #if defined(TARGET_NR_adjtimex)
9723 case TARGET_NR_adjtimex:
9725 struct timex host_buf;
9727 if (target_to_host_timex(&host_buf, arg1) != 0) {
9728 return -TARGET_EFAULT;
9730 ret = get_errno(adjtimex(&host_buf));
9731 if (!is_error(ret)) {
9732 if (host_to_target_timex(arg1, &host_buf) != 0) {
9733 return -TARGET_EFAULT;
9737 return ret;
9738 #endif
9739 #if defined(TARGET_NR_clock_adjtime) && defined(CONFIG_CLOCK_ADJTIME)
9740 case TARGET_NR_clock_adjtime:
9742 struct timex htx, *phtx = &htx;
9744 if (target_to_host_timex(phtx, arg2) != 0) {
9745 return -TARGET_EFAULT;
9747 ret = get_errno(clock_adjtime(arg1, phtx));
9748 if (!is_error(ret) && phtx) {
9749 if (host_to_target_timex(arg2, phtx) != 0) {
9750 return -TARGET_EFAULT;
9754 return ret;
9755 #endif
9756 case TARGET_NR_getpgid:
9757 return get_errno(getpgid(arg1));
9758 case TARGET_NR_fchdir:
9759 return get_errno(fchdir(arg1));
9760 case TARGET_NR_personality:
9761 return get_errno(personality(arg1));
9762 #ifdef TARGET_NR__llseek /* Not on alpha */
9763 case TARGET_NR__llseek:
9765 int64_t res;
9766 #if !defined(__NR_llseek)
9767 res = lseek(arg1, ((uint64_t)arg2 << 32) | (abi_ulong)arg3, arg5);
9768 if (res == -1) {
9769 ret = get_errno(res);
9770 } else {
9771 ret = 0;
9773 #else
9774 ret = get_errno(_llseek(arg1, arg2, arg3, &res, arg5));
9775 #endif
9776 if ((ret == 0) && put_user_s64(res, arg4)) {
9777 return -TARGET_EFAULT;
9780 return ret;
9781 #endif
9782 #ifdef TARGET_NR_getdents
9783 case TARGET_NR_getdents:
9784 #ifdef EMULATE_GETDENTS_WITH_GETDENTS
9785 #if TARGET_ABI_BITS == 32 && HOST_LONG_BITS == 64
9787 struct target_dirent *target_dirp;
9788 struct linux_dirent *dirp;
9789 abi_long count = arg3;
9791 dirp = g_try_malloc(count);
9792 if (!dirp) {
9793 return -TARGET_ENOMEM;
9796 ret = get_errno(sys_getdents(arg1, dirp, count));
9797 if (!is_error(ret)) {
9798 struct linux_dirent *de;
9799 struct target_dirent *tde;
9800 int len = ret;
9801 int reclen, treclen;
9802 int count1, tnamelen;
9804 count1 = 0;
9805 de = dirp;
9806 if (!(target_dirp = lock_user(VERIFY_WRITE, arg2, count, 0)))
9807 return -TARGET_EFAULT;
9808 tde = target_dirp;
9809 while (len > 0) {
9810 reclen = de->d_reclen;
9811 tnamelen = reclen - offsetof(struct linux_dirent, d_name);
9812 assert(tnamelen >= 0);
9813 treclen = tnamelen + offsetof(struct target_dirent, d_name);
9814 assert(count1 + treclen <= count);
9815 tde->d_reclen = tswap16(treclen);
9816 tde->d_ino = tswapal(de->d_ino);
9817 tde->d_off = tswapal(de->d_off);
9818 memcpy(tde->d_name, de->d_name, tnamelen);
9819 de = (struct linux_dirent *)((char *)de + reclen);
9820 len -= reclen;
9821 tde = (struct target_dirent *)((char *)tde + treclen);
9822 count1 += treclen;
9824 ret = count1;
9825 unlock_user(target_dirp, arg2, ret);
9827 g_free(dirp);
9829 #else
9831 struct linux_dirent *dirp;
9832 abi_long count = arg3;
9834 if (!(dirp = lock_user(VERIFY_WRITE, arg2, count, 0)))
9835 return -TARGET_EFAULT;
9836 ret = get_errno(sys_getdents(arg1, dirp, count));
9837 if (!is_error(ret)) {
9838 struct linux_dirent *de;
9839 int len = ret;
9840 int reclen;
9841 de = dirp;
9842 while (len > 0) {
9843 reclen = de->d_reclen;
9844 if (reclen > len)
9845 break;
9846 de->d_reclen = tswap16(reclen);
9847 tswapls(&de->d_ino);
9848 tswapls(&de->d_off);
9849 de = (struct linux_dirent *)((char *)de + reclen);
9850 len -= reclen;
9853 unlock_user(dirp, arg2, ret);
9855 #endif
9856 #else
9857 /* Implement getdents in terms of getdents64 */
9859 struct linux_dirent64 *dirp;
9860 abi_long count = arg3;
9862 dirp = lock_user(VERIFY_WRITE, arg2, count, 0);
9863 if (!dirp) {
9864 return -TARGET_EFAULT;
9866 ret = get_errno(sys_getdents64(arg1, dirp, count));
9867 if (!is_error(ret)) {
9868 /* Convert the dirent64 structs to target dirent. We do this
9869 * in-place, since we can guarantee that a target_dirent is no
9870 * larger than a dirent64; however this means we have to be
9871 * careful to read everything before writing in the new format.
9873 struct linux_dirent64 *de;
9874 struct target_dirent *tde;
9875 int len = ret;
9876 int tlen = 0;
9878 de = dirp;
9879 tde = (struct target_dirent *)dirp;
9880 while (len > 0) {
9881 int namelen, treclen;
9882 int reclen = de->d_reclen;
9883 uint64_t ino = de->d_ino;
9884 int64_t off = de->d_off;
9885 uint8_t type = de->d_type;
9887 namelen = strlen(de->d_name);
9888 treclen = offsetof(struct target_dirent, d_name)
9889 + namelen + 2;
9890 treclen = QEMU_ALIGN_UP(treclen, sizeof(abi_long));
9892 memmove(tde->d_name, de->d_name, namelen + 1);
9893 tde->d_ino = tswapal(ino);
9894 tde->d_off = tswapal(off);
9895 tde->d_reclen = tswap16(treclen);
9896 /* The target_dirent type is in what was formerly a padding
9897 * byte at the end of the structure:
9899 *(((char *)tde) + treclen - 1) = type;
9901 de = (struct linux_dirent64 *)((char *)de + reclen);
9902 tde = (struct target_dirent *)((char *)tde + treclen);
9903 len -= reclen;
9904 tlen += treclen;
9906 ret = tlen;
9908 unlock_user(dirp, arg2, ret);
9910 #endif
9911 return ret;
9912 #endif /* TARGET_NR_getdents */
9913 #if defined(TARGET_NR_getdents64) && defined(__NR_getdents64)
9914 case TARGET_NR_getdents64:
9916 struct linux_dirent64 *dirp;
9917 abi_long count = arg3;
9918 if (!(dirp = lock_user(VERIFY_WRITE, arg2, count, 0)))
9919 return -TARGET_EFAULT;
9920 ret = get_errno(sys_getdents64(arg1, dirp, count));
9921 if (!is_error(ret)) {
9922 struct linux_dirent64 *de;
9923 int len = ret;
9924 int reclen;
9925 de = dirp;
9926 while (len > 0) {
9927 reclen = de->d_reclen;
9928 if (reclen > len)
9929 break;
9930 de->d_reclen = tswap16(reclen);
9931 tswap64s((uint64_t *)&de->d_ino);
9932 tswap64s((uint64_t *)&de->d_off);
9933 de = (struct linux_dirent64 *)((char *)de + reclen);
9934 len -= reclen;
9937 unlock_user(dirp, arg2, ret);
9939 return ret;
9940 #endif /* TARGET_NR_getdents64 */
9941 #if defined(TARGET_NR__newselect)
9942 case TARGET_NR__newselect:
9943 return do_select(arg1, arg2, arg3, arg4, arg5);
9944 #endif
9945 #if defined(TARGET_NR_poll) || defined(TARGET_NR_ppoll)
9946 # ifdef TARGET_NR_poll
9947 case TARGET_NR_poll:
9948 # endif
9949 # ifdef TARGET_NR_ppoll
9950 case TARGET_NR_ppoll:
9951 # endif
9953 struct target_pollfd *target_pfd;
9954 unsigned int nfds = arg2;
9955 struct pollfd *pfd;
9956 unsigned int i;
9958 pfd = NULL;
9959 target_pfd = NULL;
9960 if (nfds) {
9961 if (nfds > (INT_MAX / sizeof(struct target_pollfd))) {
9962 return -TARGET_EINVAL;
9965 target_pfd = lock_user(VERIFY_WRITE, arg1,
9966 sizeof(struct target_pollfd) * nfds, 1);
9967 if (!target_pfd) {
9968 return -TARGET_EFAULT;
9971 pfd = alloca(sizeof(struct pollfd) * nfds);
9972 for (i = 0; i < nfds; i++) {
9973 pfd[i].fd = tswap32(target_pfd[i].fd);
9974 pfd[i].events = tswap16(target_pfd[i].events);
9978 switch (num) {
9979 # ifdef TARGET_NR_ppoll
9980 case TARGET_NR_ppoll:
9982 struct timespec _timeout_ts, *timeout_ts = &_timeout_ts;
9983 target_sigset_t *target_set;
9984 sigset_t _set, *set = &_set;
9986 if (arg3) {
9987 if (target_to_host_timespec(timeout_ts, arg3)) {
9988 unlock_user(target_pfd, arg1, 0);
9989 return -TARGET_EFAULT;
9991 } else {
9992 timeout_ts = NULL;
9995 if (arg4) {
9996 if (arg5 != sizeof(target_sigset_t)) {
9997 unlock_user(target_pfd, arg1, 0);
9998 return -TARGET_EINVAL;
10001 target_set = lock_user(VERIFY_READ, arg4, sizeof(target_sigset_t), 1);
10002 if (!target_set) {
10003 unlock_user(target_pfd, arg1, 0);
10004 return -TARGET_EFAULT;
10006 target_to_host_sigset(set, target_set);
10007 } else {
10008 set = NULL;
10011 ret = get_errno(safe_ppoll(pfd, nfds, timeout_ts,
10012 set, SIGSET_T_SIZE));
10014 if (!is_error(ret) && arg3) {
10015 host_to_target_timespec(arg3, timeout_ts);
10017 if (arg4) {
10018 unlock_user(target_set, arg4, 0);
10020 break;
10022 # endif
10023 # ifdef TARGET_NR_poll
10024 case TARGET_NR_poll:
10026 struct timespec ts, *pts;
10028 if (arg3 >= 0) {
10029 /* Convert ms to secs, ns */
10030 ts.tv_sec = arg3 / 1000;
10031 ts.tv_nsec = (arg3 % 1000) * 1000000LL;
10032 pts = &ts;
10033 } else {
10034 /* -ve poll() timeout means "infinite" */
10035 pts = NULL;
10037 ret = get_errno(safe_ppoll(pfd, nfds, pts, NULL, 0));
10038 break;
10040 # endif
10041 default:
10042 g_assert_not_reached();
10045 if (!is_error(ret)) {
10046 for(i = 0; i < nfds; i++) {
10047 target_pfd[i].revents = tswap16(pfd[i].revents);
10050 unlock_user(target_pfd, arg1, sizeof(struct target_pollfd) * nfds);
10052 return ret;
10053 #endif
10054 case TARGET_NR_flock:
10055 /* NOTE: the flock constant seems to be the same for every
10056 Linux platform */
10057 return get_errno(safe_flock(arg1, arg2));
10058 case TARGET_NR_readv:
10060 struct iovec *vec = lock_iovec(VERIFY_WRITE, arg2, arg3, 0);
10061 if (vec != NULL) {
10062 ret = get_errno(safe_readv(arg1, vec, arg3));
10063 unlock_iovec(vec, arg2, arg3, 1);
10064 } else {
10065 ret = -host_to_target_errno(errno);
10068 return ret;
10069 case TARGET_NR_writev:
10071 struct iovec *vec = lock_iovec(VERIFY_READ, arg2, arg3, 1);
10072 if (vec != NULL) {
10073 ret = get_errno(safe_writev(arg1, vec, arg3));
10074 unlock_iovec(vec, arg2, arg3, 0);
10075 } else {
10076 ret = -host_to_target_errno(errno);
10079 return ret;
10080 #if defined(TARGET_NR_preadv)
10081 case TARGET_NR_preadv:
10083 struct iovec *vec = lock_iovec(VERIFY_WRITE, arg2, arg3, 0);
10084 if (vec != NULL) {
10085 unsigned long low, high;
10087 target_to_host_low_high(arg4, arg5, &low, &high);
10088 ret = get_errno(safe_preadv(arg1, vec, arg3, low, high));
10089 unlock_iovec(vec, arg2, arg3, 1);
10090 } else {
10091 ret = -host_to_target_errno(errno);
10094 return ret;
10095 #endif
10096 #if defined(TARGET_NR_pwritev)
10097 case TARGET_NR_pwritev:
10099 struct iovec *vec = lock_iovec(VERIFY_READ, arg2, arg3, 1);
10100 if (vec != NULL) {
10101 unsigned long low, high;
10103 target_to_host_low_high(arg4, arg5, &low, &high);
10104 ret = get_errno(safe_pwritev(arg1, vec, arg3, low, high));
10105 unlock_iovec(vec, arg2, arg3, 0);
10106 } else {
10107 ret = -host_to_target_errno(errno);
10110 return ret;
10111 #endif
10112 case TARGET_NR_getsid:
10113 return get_errno(getsid(arg1));
10114 #if defined(TARGET_NR_fdatasync) /* Not on alpha (osf_datasync ?) */
10115 case TARGET_NR_fdatasync:
10116 return get_errno(fdatasync(arg1));
10117 #endif
10118 #ifdef TARGET_NR__sysctl
10119 case TARGET_NR__sysctl:
10120 /* We don't implement this, but ENOTDIR is always a safe
10121 return value. */
10122 return -TARGET_ENOTDIR;
10123 #endif
10124 case TARGET_NR_sched_getaffinity:
10126 unsigned int mask_size;
10127 unsigned long *mask;
10130 * sched_getaffinity needs multiples of ulong, so need to take
10131 * care of mismatches between target ulong and host ulong sizes.
10133 if (arg2 & (sizeof(abi_ulong) - 1)) {
10134 return -TARGET_EINVAL;
10136 mask_size = (arg2 + (sizeof(*mask) - 1)) & ~(sizeof(*mask) - 1);
10138 mask = alloca(mask_size);
10139 memset(mask, 0, mask_size);
10140 ret = get_errno(sys_sched_getaffinity(arg1, mask_size, mask));
10142 if (!is_error(ret)) {
10143 if (ret > arg2) {
10144 /* More data returned than the caller's buffer will fit.
10145 * This only happens if sizeof(abi_long) < sizeof(long)
10146 * and the caller passed us a buffer holding an odd number
10147 * of abi_longs. If the host kernel is actually using the
10148 * extra 4 bytes then fail EINVAL; otherwise we can just
10149 * ignore them and only copy the interesting part.
10151 int numcpus = sysconf(_SC_NPROCESSORS_CONF);
10152 if (numcpus > arg2 * 8) {
10153 return -TARGET_EINVAL;
10155 ret = arg2;
10158 if (host_to_target_cpu_mask(mask, mask_size, arg3, ret)) {
10159 return -TARGET_EFAULT;
10163 return ret;
10164 case TARGET_NR_sched_setaffinity:
10166 unsigned int mask_size;
10167 unsigned long *mask;
10170 * sched_setaffinity needs multiples of ulong, so need to take
10171 * care of mismatches between target ulong and host ulong sizes.
10173 if (arg2 & (sizeof(abi_ulong) - 1)) {
10174 return -TARGET_EINVAL;
10176 mask_size = (arg2 + (sizeof(*mask) - 1)) & ~(sizeof(*mask) - 1);
10177 mask = alloca(mask_size);
10179 ret = target_to_host_cpu_mask(mask, mask_size, arg3, arg2);
10180 if (ret) {
10181 return ret;
10184 return get_errno(sys_sched_setaffinity(arg1, mask_size, mask));
10186 case TARGET_NR_getcpu:
10188 unsigned cpu, node;
10189 ret = get_errno(sys_getcpu(arg1 ? &cpu : NULL,
10190 arg2 ? &node : NULL,
10191 NULL));
10192 if (is_error(ret)) {
10193 return ret;
10195 if (arg1 && put_user_u32(cpu, arg1)) {
10196 return -TARGET_EFAULT;
10198 if (arg2 && put_user_u32(node, arg2)) {
10199 return -TARGET_EFAULT;
10202 return ret;
10203 case TARGET_NR_sched_setparam:
10205 struct sched_param *target_schp;
10206 struct sched_param schp;
10208 if (arg2 == 0) {
10209 return -TARGET_EINVAL;
10211 if (!lock_user_struct(VERIFY_READ, target_schp, arg2, 1))
10212 return -TARGET_EFAULT;
10213 schp.sched_priority = tswap32(target_schp->sched_priority);
10214 unlock_user_struct(target_schp, arg2, 0);
10215 return get_errno(sched_setparam(arg1, &schp));
10217 case TARGET_NR_sched_getparam:
10219 struct sched_param *target_schp;
10220 struct sched_param schp;
10222 if (arg2 == 0) {
10223 return -TARGET_EINVAL;
10225 ret = get_errno(sched_getparam(arg1, &schp));
10226 if (!is_error(ret)) {
10227 if (!lock_user_struct(VERIFY_WRITE, target_schp, arg2, 0))
10228 return -TARGET_EFAULT;
10229 target_schp->sched_priority = tswap32(schp.sched_priority);
10230 unlock_user_struct(target_schp, arg2, 1);
10233 return ret;
10234 case TARGET_NR_sched_setscheduler:
10236 struct sched_param *target_schp;
10237 struct sched_param schp;
10238 if (arg3 == 0) {
10239 return -TARGET_EINVAL;
10241 if (!lock_user_struct(VERIFY_READ, target_schp, arg3, 1))
10242 return -TARGET_EFAULT;
10243 schp.sched_priority = tswap32(target_schp->sched_priority);
10244 unlock_user_struct(target_schp, arg3, 0);
10245 return get_errno(sched_setscheduler(arg1, arg2, &schp));
10247 case TARGET_NR_sched_getscheduler:
10248 return get_errno(sched_getscheduler(arg1));
10249 case TARGET_NR_sched_yield:
10250 return get_errno(sched_yield());
10251 case TARGET_NR_sched_get_priority_max:
10252 return get_errno(sched_get_priority_max(arg1));
10253 case TARGET_NR_sched_get_priority_min:
10254 return get_errno(sched_get_priority_min(arg1));
10255 #ifdef TARGET_NR_sched_rr_get_interval
10256 case TARGET_NR_sched_rr_get_interval:
10258 struct timespec ts;
10259 ret = get_errno(sched_rr_get_interval(arg1, &ts));
10260 if (!is_error(ret)) {
10261 ret = host_to_target_timespec(arg2, &ts);
10264 return ret;
10265 #endif
10266 #if defined(TARGET_NR_nanosleep)
10267 case TARGET_NR_nanosleep:
10269 struct timespec req, rem;
10270 target_to_host_timespec(&req, arg1);
10271 ret = get_errno(safe_nanosleep(&req, &rem));
10272 if (is_error(ret) && arg2) {
10273 host_to_target_timespec(arg2, &rem);
10276 return ret;
10277 #endif
10278 case TARGET_NR_prctl:
10279 switch (arg1) {
10280 case PR_GET_PDEATHSIG:
10282 int deathsig;
10283 ret = get_errno(prctl(arg1, &deathsig, arg3, arg4, arg5));
10284 if (!is_error(ret) && arg2
10285 && put_user_ual(deathsig, arg2)) {
10286 return -TARGET_EFAULT;
10288 return ret;
10290 #ifdef PR_GET_NAME
10291 case PR_GET_NAME:
10293 void *name = lock_user(VERIFY_WRITE, arg2, 16, 1);
10294 if (!name) {
10295 return -TARGET_EFAULT;
10297 ret = get_errno(prctl(arg1, (unsigned long)name,
10298 arg3, arg4, arg5));
10299 unlock_user(name, arg2, 16);
10300 return ret;
10302 case PR_SET_NAME:
10304 void *name = lock_user(VERIFY_READ, arg2, 16, 1);
10305 if (!name) {
10306 return -TARGET_EFAULT;
10308 ret = get_errno(prctl(arg1, (unsigned long)name,
10309 arg3, arg4, arg5));
10310 unlock_user(name, arg2, 0);
10311 return ret;
10313 #endif
10314 #ifdef TARGET_MIPS
10315 case TARGET_PR_GET_FP_MODE:
10317 CPUMIPSState *env = ((CPUMIPSState *)cpu_env);
10318 ret = 0;
10319 if (env->CP0_Status & (1 << CP0St_FR)) {
10320 ret |= TARGET_PR_FP_MODE_FR;
10322 if (env->CP0_Config5 & (1 << CP0C5_FRE)) {
10323 ret |= TARGET_PR_FP_MODE_FRE;
10325 return ret;
10327 case TARGET_PR_SET_FP_MODE:
10329 CPUMIPSState *env = ((CPUMIPSState *)cpu_env);
10330 bool old_fr = env->CP0_Status & (1 << CP0St_FR);
10331 bool old_fre = env->CP0_Config5 & (1 << CP0C5_FRE);
10332 bool new_fr = arg2 & TARGET_PR_FP_MODE_FR;
10333 bool new_fre = arg2 & TARGET_PR_FP_MODE_FRE;
10335 const unsigned int known_bits = TARGET_PR_FP_MODE_FR |
10336 TARGET_PR_FP_MODE_FRE;
10338 /* If nothing to change, return right away, successfully. */
10339 if (old_fr == new_fr && old_fre == new_fre) {
10340 return 0;
10342 /* Check the value is valid */
10343 if (arg2 & ~known_bits) {
10344 return -TARGET_EOPNOTSUPP;
10346 /* Setting FRE without FR is not supported. */
10347 if (new_fre && !new_fr) {
10348 return -TARGET_EOPNOTSUPP;
10350 if (new_fr && !(env->active_fpu.fcr0 & (1 << FCR0_F64))) {
10351 /* FR1 is not supported */
10352 return -TARGET_EOPNOTSUPP;
10354 if (!new_fr && (env->active_fpu.fcr0 & (1 << FCR0_F64))
10355 && !(env->CP0_Status_rw_bitmask & (1 << CP0St_FR))) {
10356 /* cannot set FR=0 */
10357 return -TARGET_EOPNOTSUPP;
10359 if (new_fre && !(env->active_fpu.fcr0 & (1 << FCR0_FREP))) {
10360 /* Cannot set FRE=1 */
10361 return -TARGET_EOPNOTSUPP;
10364 int i;
10365 fpr_t *fpr = env->active_fpu.fpr;
10366 for (i = 0; i < 32 ; i += 2) {
10367 if (!old_fr && new_fr) {
10368 fpr[i].w[!FP_ENDIAN_IDX] = fpr[i + 1].w[FP_ENDIAN_IDX];
10369 } else if (old_fr && !new_fr) {
10370 fpr[i + 1].w[FP_ENDIAN_IDX] = fpr[i].w[!FP_ENDIAN_IDX];
10374 if (new_fr) {
10375 env->CP0_Status |= (1 << CP0St_FR);
10376 env->hflags |= MIPS_HFLAG_F64;
10377 } else {
10378 env->CP0_Status &= ~(1 << CP0St_FR);
10379 env->hflags &= ~MIPS_HFLAG_F64;
10381 if (new_fre) {
10382 env->CP0_Config5 |= (1 << CP0C5_FRE);
10383 if (env->active_fpu.fcr0 & (1 << FCR0_FREP)) {
10384 env->hflags |= MIPS_HFLAG_FRE;
10386 } else {
10387 env->CP0_Config5 &= ~(1 << CP0C5_FRE);
10388 env->hflags &= ~MIPS_HFLAG_FRE;
10391 return 0;
10393 #endif /* MIPS */
10394 #ifdef TARGET_AARCH64
10395 case TARGET_PR_SVE_SET_VL:
10397 * We cannot support either PR_SVE_SET_VL_ONEXEC or
10398 * PR_SVE_VL_INHERIT. Note the kernel definition
10399 * of sve_vl_valid allows for VQ=512, i.e. VL=8192,
10400 * even though the current architectural maximum is VQ=16.
10402 ret = -TARGET_EINVAL;
10403 if (cpu_isar_feature(aa64_sve, env_archcpu(cpu_env))
10404 && arg2 >= 0 && arg2 <= 512 * 16 && !(arg2 & 15)) {
10405 CPUARMState *env = cpu_env;
10406 ARMCPU *cpu = env_archcpu(env);
10407 uint32_t vq, old_vq;
10409 old_vq = (env->vfp.zcr_el[1] & 0xf) + 1;
10410 vq = MAX(arg2 / 16, 1);
10411 vq = MIN(vq, cpu->sve_max_vq);
10413 if (vq < old_vq) {
10414 aarch64_sve_narrow_vq(env, vq);
10416 env->vfp.zcr_el[1] = vq - 1;
10417 arm_rebuild_hflags(env);
10418 ret = vq * 16;
10420 return ret;
10421 case TARGET_PR_SVE_GET_VL:
10422 ret = -TARGET_EINVAL;
10424 ARMCPU *cpu = env_archcpu(cpu_env);
10425 if (cpu_isar_feature(aa64_sve, cpu)) {
10426 ret = ((cpu->env.vfp.zcr_el[1] & 0xf) + 1) * 16;
10429 return ret;
10430 case TARGET_PR_PAC_RESET_KEYS:
10432 CPUARMState *env = cpu_env;
10433 ARMCPU *cpu = env_archcpu(env);
10435 if (arg3 || arg4 || arg5) {
10436 return -TARGET_EINVAL;
10438 if (cpu_isar_feature(aa64_pauth, cpu)) {
10439 int all = (TARGET_PR_PAC_APIAKEY | TARGET_PR_PAC_APIBKEY |
10440 TARGET_PR_PAC_APDAKEY | TARGET_PR_PAC_APDBKEY |
10441 TARGET_PR_PAC_APGAKEY);
10442 int ret = 0;
10443 Error *err = NULL;
10445 if (arg2 == 0) {
10446 arg2 = all;
10447 } else if (arg2 & ~all) {
10448 return -TARGET_EINVAL;
10450 if (arg2 & TARGET_PR_PAC_APIAKEY) {
10451 ret |= qemu_guest_getrandom(&env->keys.apia,
10452 sizeof(ARMPACKey), &err);
10454 if (arg2 & TARGET_PR_PAC_APIBKEY) {
10455 ret |= qemu_guest_getrandom(&env->keys.apib,
10456 sizeof(ARMPACKey), &err);
10458 if (arg2 & TARGET_PR_PAC_APDAKEY) {
10459 ret |= qemu_guest_getrandom(&env->keys.apda,
10460 sizeof(ARMPACKey), &err);
10462 if (arg2 & TARGET_PR_PAC_APDBKEY) {
10463 ret |= qemu_guest_getrandom(&env->keys.apdb,
10464 sizeof(ARMPACKey), &err);
10466 if (arg2 & TARGET_PR_PAC_APGAKEY) {
10467 ret |= qemu_guest_getrandom(&env->keys.apga,
10468 sizeof(ARMPACKey), &err);
10470 if (ret != 0) {
10472 * Some unknown failure in the crypto. The best
10473 * we can do is log it and fail the syscall.
10474 * The real syscall cannot fail this way.
10476 qemu_log_mask(LOG_UNIMP,
10477 "PR_PAC_RESET_KEYS: Crypto failure: %s",
10478 error_get_pretty(err));
10479 error_free(err);
10480 return -TARGET_EIO;
10482 return 0;
10485 return -TARGET_EINVAL;
10486 #endif /* AARCH64 */
10487 case PR_GET_SECCOMP:
10488 case PR_SET_SECCOMP:
10489 /* Disable seccomp to prevent the target disabling syscalls we
10490 * need. */
10491 return -TARGET_EINVAL;
10492 default:
10493 /* Most prctl options have no pointer arguments */
10494 return get_errno(prctl(arg1, arg2, arg3, arg4, arg5));
10496 break;
10497 #ifdef TARGET_NR_arch_prctl
10498 case TARGET_NR_arch_prctl:
10499 return do_arch_prctl(cpu_env, arg1, arg2);
10500 #endif
10501 #ifdef TARGET_NR_pread64
10502 case TARGET_NR_pread64:
10503 if (regpairs_aligned(cpu_env, num)) {
10504 arg4 = arg5;
10505 arg5 = arg6;
10507 if (arg2 == 0 && arg3 == 0) {
10508 /* Special-case NULL buffer and zero length, which should succeed */
10509 p = 0;
10510 } else {
10511 p = lock_user(VERIFY_WRITE, arg2, arg3, 0);
10512 if (!p) {
10513 return -TARGET_EFAULT;
10516 ret = get_errno(pread64(arg1, p, arg3, target_offset64(arg4, arg5)));
10517 unlock_user(p, arg2, ret);
10518 return ret;
10519 case TARGET_NR_pwrite64:
10520 if (regpairs_aligned(cpu_env, num)) {
10521 arg4 = arg5;
10522 arg5 = arg6;
10524 if (arg2 == 0 && arg3 == 0) {
10525 /* Special-case NULL buffer and zero length, which should succeed */
10526 p = 0;
10527 } else {
10528 p = lock_user(VERIFY_READ, arg2, arg3, 1);
10529 if (!p) {
10530 return -TARGET_EFAULT;
10533 ret = get_errno(pwrite64(arg1, p, arg3, target_offset64(arg4, arg5)));
10534 unlock_user(p, arg2, 0);
10535 return ret;
10536 #endif
10537 case TARGET_NR_getcwd:
10538 if (!(p = lock_user(VERIFY_WRITE, arg1, arg2, 0)))
10539 return -TARGET_EFAULT;
10540 ret = get_errno(sys_getcwd1(p, arg2));
10541 unlock_user(p, arg1, ret);
10542 return ret;
10543 case TARGET_NR_capget:
10544 case TARGET_NR_capset:
10546 struct target_user_cap_header *target_header;
10547 struct target_user_cap_data *target_data = NULL;
10548 struct __user_cap_header_struct header;
10549 struct __user_cap_data_struct data[2];
10550 struct __user_cap_data_struct *dataptr = NULL;
10551 int i, target_datalen;
10552 int data_items = 1;
10554 if (!lock_user_struct(VERIFY_WRITE, target_header, arg1, 1)) {
10555 return -TARGET_EFAULT;
10557 header.version = tswap32(target_header->version);
10558 header.pid = tswap32(target_header->pid);
10560 if (header.version != _LINUX_CAPABILITY_VERSION) {
10561 /* Version 2 and up takes pointer to two user_data structs */
10562 data_items = 2;
10565 target_datalen = sizeof(*target_data) * data_items;
10567 if (arg2) {
10568 if (num == TARGET_NR_capget) {
10569 target_data = lock_user(VERIFY_WRITE, arg2, target_datalen, 0);
10570 } else {
10571 target_data = lock_user(VERIFY_READ, arg2, target_datalen, 1);
10573 if (!target_data) {
10574 unlock_user_struct(target_header, arg1, 0);
10575 return -TARGET_EFAULT;
10578 if (num == TARGET_NR_capset) {
10579 for (i = 0; i < data_items; i++) {
10580 data[i].effective = tswap32(target_data[i].effective);
10581 data[i].permitted = tswap32(target_data[i].permitted);
10582 data[i].inheritable = tswap32(target_data[i].inheritable);
10586 dataptr = data;
10589 if (num == TARGET_NR_capget) {
10590 ret = get_errno(capget(&header, dataptr));
10591 } else {
10592 ret = get_errno(capset(&header, dataptr));
10595 /* The kernel always updates version for both capget and capset */
10596 target_header->version = tswap32(header.version);
10597 unlock_user_struct(target_header, arg1, 1);
10599 if (arg2) {
10600 if (num == TARGET_NR_capget) {
10601 for (i = 0; i < data_items; i++) {
10602 target_data[i].effective = tswap32(data[i].effective);
10603 target_data[i].permitted = tswap32(data[i].permitted);
10604 target_data[i].inheritable = tswap32(data[i].inheritable);
10606 unlock_user(target_data, arg2, target_datalen);
10607 } else {
10608 unlock_user(target_data, arg2, 0);
10611 return ret;
10613 case TARGET_NR_sigaltstack:
10614 return do_sigaltstack(arg1, arg2,
10615 get_sp_from_cpustate((CPUArchState *)cpu_env));
10617 #ifdef CONFIG_SENDFILE
10618 #ifdef TARGET_NR_sendfile
10619 case TARGET_NR_sendfile:
10621 off_t *offp = NULL;
10622 off_t off;
10623 if (arg3) {
10624 ret = get_user_sal(off, arg3);
10625 if (is_error(ret)) {
10626 return ret;
10628 offp = &off;
10630 ret = get_errno(sendfile(arg1, arg2, offp, arg4));
10631 if (!is_error(ret) && arg3) {
10632 abi_long ret2 = put_user_sal(off, arg3);
10633 if (is_error(ret2)) {
10634 ret = ret2;
10637 return ret;
10639 #endif
10640 #ifdef TARGET_NR_sendfile64
10641 case TARGET_NR_sendfile64:
10643 off_t *offp = NULL;
10644 off_t off;
10645 if (arg3) {
10646 ret = get_user_s64(off, arg3);
10647 if (is_error(ret)) {
10648 return ret;
10650 offp = &off;
10652 ret = get_errno(sendfile(arg1, arg2, offp, arg4));
10653 if (!is_error(ret) && arg3) {
10654 abi_long ret2 = put_user_s64(off, arg3);
10655 if (is_error(ret2)) {
10656 ret = ret2;
10659 return ret;
10661 #endif
10662 #endif
10663 #ifdef TARGET_NR_vfork
10664 case TARGET_NR_vfork:
10665 return get_errno(do_fork(cpu_env,
10666 CLONE_VFORK | CLONE_VM | TARGET_SIGCHLD,
10667 0, 0, 0, 0));
10668 #endif
10669 #ifdef TARGET_NR_ugetrlimit
10670 case TARGET_NR_ugetrlimit:
10672 struct rlimit rlim;
10673 int resource = target_to_host_resource(arg1);
10674 ret = get_errno(getrlimit(resource, &rlim));
10675 if (!is_error(ret)) {
10676 struct target_rlimit *target_rlim;
10677 if (!lock_user_struct(VERIFY_WRITE, target_rlim, arg2, 0))
10678 return -TARGET_EFAULT;
10679 target_rlim->rlim_cur = host_to_target_rlim(rlim.rlim_cur);
10680 target_rlim->rlim_max = host_to_target_rlim(rlim.rlim_max);
10681 unlock_user_struct(target_rlim, arg2, 1);
10683 return ret;
10685 #endif
10686 #ifdef TARGET_NR_truncate64
10687 case TARGET_NR_truncate64:
10688 if (!(p = lock_user_string(arg1)))
10689 return -TARGET_EFAULT;
10690 ret = target_truncate64(cpu_env, p, arg2, arg3, arg4);
10691 unlock_user(p, arg1, 0);
10692 return ret;
10693 #endif
10694 #ifdef TARGET_NR_ftruncate64
10695 case TARGET_NR_ftruncate64:
10696 return target_ftruncate64(cpu_env, arg1, arg2, arg3, arg4);
10697 #endif
10698 #ifdef TARGET_NR_stat64
10699 case TARGET_NR_stat64:
10700 if (!(p = lock_user_string(arg1))) {
10701 return -TARGET_EFAULT;
10703 ret = get_errno(stat(path(p), &st));
10704 unlock_user(p, arg1, 0);
10705 if (!is_error(ret))
10706 ret = host_to_target_stat64(cpu_env, arg2, &st);
10707 return ret;
10708 #endif
10709 #ifdef TARGET_NR_lstat64
10710 case TARGET_NR_lstat64:
10711 if (!(p = lock_user_string(arg1))) {
10712 return -TARGET_EFAULT;
10714 ret = get_errno(lstat(path(p), &st));
10715 unlock_user(p, arg1, 0);
10716 if (!is_error(ret))
10717 ret = host_to_target_stat64(cpu_env, arg2, &st);
10718 return ret;
10719 #endif
10720 #ifdef TARGET_NR_fstat64
10721 case TARGET_NR_fstat64:
10722 ret = get_errno(fstat(arg1, &st));
10723 if (!is_error(ret))
10724 ret = host_to_target_stat64(cpu_env, arg2, &st);
10725 return ret;
10726 #endif
10727 #if (defined(TARGET_NR_fstatat64) || defined(TARGET_NR_newfstatat))
10728 #ifdef TARGET_NR_fstatat64
10729 case TARGET_NR_fstatat64:
10730 #endif
10731 #ifdef TARGET_NR_newfstatat
10732 case TARGET_NR_newfstatat:
10733 #endif
10734 if (!(p = lock_user_string(arg2))) {
10735 return -TARGET_EFAULT;
10737 ret = get_errno(fstatat(arg1, path(p), &st, arg4));
10738 unlock_user(p, arg2, 0);
10739 if (!is_error(ret))
10740 ret = host_to_target_stat64(cpu_env, arg3, &st);
10741 return ret;
10742 #endif
10743 #if defined(TARGET_NR_statx)
10744 case TARGET_NR_statx:
10746 struct target_statx *target_stx;
10747 int dirfd = arg1;
10748 int flags = arg3;
10750 p = lock_user_string(arg2);
10751 if (p == NULL) {
10752 return -TARGET_EFAULT;
10754 #if defined(__NR_statx)
10757 * It is assumed that struct statx is architecture independent.
10759 struct target_statx host_stx;
10760 int mask = arg4;
10762 ret = get_errno(sys_statx(dirfd, p, flags, mask, &host_stx));
10763 if (!is_error(ret)) {
10764 if (host_to_target_statx(&host_stx, arg5) != 0) {
10765 unlock_user(p, arg2, 0);
10766 return -TARGET_EFAULT;
10770 if (ret != -TARGET_ENOSYS) {
10771 unlock_user(p, arg2, 0);
10772 return ret;
10775 #endif
10776 ret = get_errno(fstatat(dirfd, path(p), &st, flags));
10777 unlock_user(p, arg2, 0);
10779 if (!is_error(ret)) {
10780 if (!lock_user_struct(VERIFY_WRITE, target_stx, arg5, 0)) {
10781 return -TARGET_EFAULT;
10783 memset(target_stx, 0, sizeof(*target_stx));
10784 __put_user(major(st.st_dev), &target_stx->stx_dev_major);
10785 __put_user(minor(st.st_dev), &target_stx->stx_dev_minor);
10786 __put_user(st.st_ino, &target_stx->stx_ino);
10787 __put_user(st.st_mode, &target_stx->stx_mode);
10788 __put_user(st.st_uid, &target_stx->stx_uid);
10789 __put_user(st.st_gid, &target_stx->stx_gid);
10790 __put_user(st.st_nlink, &target_stx->stx_nlink);
10791 __put_user(major(st.st_rdev), &target_stx->stx_rdev_major);
10792 __put_user(minor(st.st_rdev), &target_stx->stx_rdev_minor);
10793 __put_user(st.st_size, &target_stx->stx_size);
10794 __put_user(st.st_blksize, &target_stx->stx_blksize);
10795 __put_user(st.st_blocks, &target_stx->stx_blocks);
10796 __put_user(st.st_atime, &target_stx->stx_atime.tv_sec);
10797 __put_user(st.st_mtime, &target_stx->stx_mtime.tv_sec);
10798 __put_user(st.st_ctime, &target_stx->stx_ctime.tv_sec);
10799 unlock_user_struct(target_stx, arg5, 1);
10802 return ret;
10803 #endif
10804 #ifdef TARGET_NR_lchown
10805 case TARGET_NR_lchown:
10806 if (!(p = lock_user_string(arg1)))
10807 return -TARGET_EFAULT;
10808 ret = get_errno(lchown(p, low2highuid(arg2), low2highgid(arg3)));
10809 unlock_user(p, arg1, 0);
10810 return ret;
10811 #endif
10812 #ifdef TARGET_NR_getuid
10813 case TARGET_NR_getuid:
10814 return get_errno(high2lowuid(getuid()));
10815 #endif
10816 #ifdef TARGET_NR_getgid
10817 case TARGET_NR_getgid:
10818 return get_errno(high2lowgid(getgid()));
10819 #endif
10820 #ifdef TARGET_NR_geteuid
10821 case TARGET_NR_geteuid:
10822 return get_errno(high2lowuid(geteuid()));
10823 #endif
10824 #ifdef TARGET_NR_getegid
10825 case TARGET_NR_getegid:
10826 return get_errno(high2lowgid(getegid()));
10827 #endif
10828 case TARGET_NR_setreuid:
10829 return get_errno(setreuid(low2highuid(arg1), low2highuid(arg2)));
10830 case TARGET_NR_setregid:
10831 return get_errno(setregid(low2highgid(arg1), low2highgid(arg2)));
10832 case TARGET_NR_getgroups:
10834 int gidsetsize = arg1;
10835 target_id *target_grouplist;
10836 gid_t *grouplist;
10837 int i;
10839 grouplist = alloca(gidsetsize * sizeof(gid_t));
10840 ret = get_errno(getgroups(gidsetsize, grouplist));
10841 if (gidsetsize == 0)
10842 return ret;
10843 if (!is_error(ret)) {
10844 target_grouplist = lock_user(VERIFY_WRITE, arg2, gidsetsize * sizeof(target_id), 0);
10845 if (!target_grouplist)
10846 return -TARGET_EFAULT;
10847 for(i = 0;i < ret; i++)
10848 target_grouplist[i] = tswapid(high2lowgid(grouplist[i]));
10849 unlock_user(target_grouplist, arg2, gidsetsize * sizeof(target_id));
10852 return ret;
10853 case TARGET_NR_setgroups:
10855 int gidsetsize = arg1;
10856 target_id *target_grouplist;
10857 gid_t *grouplist = NULL;
10858 int i;
10859 if (gidsetsize) {
10860 grouplist = alloca(gidsetsize * sizeof(gid_t));
10861 target_grouplist = lock_user(VERIFY_READ, arg2, gidsetsize * sizeof(target_id), 1);
10862 if (!target_grouplist) {
10863 return -TARGET_EFAULT;
10865 for (i = 0; i < gidsetsize; i++) {
10866 grouplist[i] = low2highgid(tswapid(target_grouplist[i]));
10868 unlock_user(target_grouplist, arg2, 0);
10870 return get_errno(setgroups(gidsetsize, grouplist));
10872 case TARGET_NR_fchown:
10873 return get_errno(fchown(arg1, low2highuid(arg2), low2highgid(arg3)));
10874 #if defined(TARGET_NR_fchownat)
10875 case TARGET_NR_fchownat:
10876 if (!(p = lock_user_string(arg2)))
10877 return -TARGET_EFAULT;
10878 ret = get_errno(fchownat(arg1, p, low2highuid(arg3),
10879 low2highgid(arg4), arg5));
10880 unlock_user(p, arg2, 0);
10881 return ret;
10882 #endif
10883 #ifdef TARGET_NR_setresuid
10884 case TARGET_NR_setresuid:
10885 return get_errno(sys_setresuid(low2highuid(arg1),
10886 low2highuid(arg2),
10887 low2highuid(arg3)));
10888 #endif
10889 #ifdef TARGET_NR_getresuid
10890 case TARGET_NR_getresuid:
10892 uid_t ruid, euid, suid;
10893 ret = get_errno(getresuid(&ruid, &euid, &suid));
10894 if (!is_error(ret)) {
10895 if (put_user_id(high2lowuid(ruid), arg1)
10896 || put_user_id(high2lowuid(euid), arg2)
10897 || put_user_id(high2lowuid(suid), arg3))
10898 return -TARGET_EFAULT;
10901 return ret;
10902 #endif
10903 #ifdef TARGET_NR_getresgid
10904 case TARGET_NR_setresgid:
10905 return get_errno(sys_setresgid(low2highgid(arg1),
10906 low2highgid(arg2),
10907 low2highgid(arg3)));
10908 #endif
10909 #ifdef TARGET_NR_getresgid
10910 case TARGET_NR_getresgid:
10912 gid_t rgid, egid, sgid;
10913 ret = get_errno(getresgid(&rgid, &egid, &sgid));
10914 if (!is_error(ret)) {
10915 if (put_user_id(high2lowgid(rgid), arg1)
10916 || put_user_id(high2lowgid(egid), arg2)
10917 || put_user_id(high2lowgid(sgid), arg3))
10918 return -TARGET_EFAULT;
10921 return ret;
10922 #endif
10923 #ifdef TARGET_NR_chown
10924 case TARGET_NR_chown:
10925 if (!(p = lock_user_string(arg1)))
10926 return -TARGET_EFAULT;
10927 ret = get_errno(chown(p, low2highuid(arg2), low2highgid(arg3)));
10928 unlock_user(p, arg1, 0);
10929 return ret;
10930 #endif
10931 case TARGET_NR_setuid:
10932 return get_errno(sys_setuid(low2highuid(arg1)));
10933 case TARGET_NR_setgid:
10934 return get_errno(sys_setgid(low2highgid(arg1)));
10935 case TARGET_NR_setfsuid:
10936 return get_errno(setfsuid(arg1));
10937 case TARGET_NR_setfsgid:
10938 return get_errno(setfsgid(arg1));
10940 #ifdef TARGET_NR_lchown32
10941 case TARGET_NR_lchown32:
10942 if (!(p = lock_user_string(arg1)))
10943 return -TARGET_EFAULT;
10944 ret = get_errno(lchown(p, arg2, arg3));
10945 unlock_user(p, arg1, 0);
10946 return ret;
10947 #endif
10948 #ifdef TARGET_NR_getuid32
10949 case TARGET_NR_getuid32:
10950 return get_errno(getuid());
10951 #endif
10953 #if defined(TARGET_NR_getxuid) && defined(TARGET_ALPHA)
10954 /* Alpha specific */
10955 case TARGET_NR_getxuid:
10957 uid_t euid;
10958 euid=geteuid();
10959 ((CPUAlphaState *)cpu_env)->ir[IR_A4]=euid;
10961 return get_errno(getuid());
10962 #endif
10963 #if defined(TARGET_NR_getxgid) && defined(TARGET_ALPHA)
10964 /* Alpha specific */
10965 case TARGET_NR_getxgid:
10967 uid_t egid;
10968 egid=getegid();
10969 ((CPUAlphaState *)cpu_env)->ir[IR_A4]=egid;
10971 return get_errno(getgid());
10972 #endif
10973 #if defined(TARGET_NR_osf_getsysinfo) && defined(TARGET_ALPHA)
10974 /* Alpha specific */
10975 case TARGET_NR_osf_getsysinfo:
10976 ret = -TARGET_EOPNOTSUPP;
10977 switch (arg1) {
10978 case TARGET_GSI_IEEE_FP_CONTROL:
10980 uint64_t fpcr = cpu_alpha_load_fpcr(cpu_env);
10981 uint64_t swcr = ((CPUAlphaState *)cpu_env)->swcr;
10983 swcr &= ~SWCR_STATUS_MASK;
10984 swcr |= (fpcr >> 35) & SWCR_STATUS_MASK;
10986 if (put_user_u64 (swcr, arg2))
10987 return -TARGET_EFAULT;
10988 ret = 0;
10990 break;
10992 /* case GSI_IEEE_STATE_AT_SIGNAL:
10993 -- Not implemented in linux kernel.
10994 case GSI_UACPROC:
10995 -- Retrieves current unaligned access state; not much used.
10996 case GSI_PROC_TYPE:
10997 -- Retrieves implver information; surely not used.
10998 case GSI_GET_HWRPB:
10999 -- Grabs a copy of the HWRPB; surely not used.
11002 return ret;
11003 #endif
11004 #if defined(TARGET_NR_osf_setsysinfo) && defined(TARGET_ALPHA)
11005 /* Alpha specific */
11006 case TARGET_NR_osf_setsysinfo:
11007 ret = -TARGET_EOPNOTSUPP;
11008 switch (arg1) {
11009 case TARGET_SSI_IEEE_FP_CONTROL:
11011 uint64_t swcr, fpcr;
11013 if (get_user_u64 (swcr, arg2)) {
11014 return -TARGET_EFAULT;
11018 * The kernel calls swcr_update_status to update the
11019 * status bits from the fpcr at every point that it
11020 * could be queried. Therefore, we store the status
11021 * bits only in FPCR.
11023 ((CPUAlphaState *)cpu_env)->swcr
11024 = swcr & (SWCR_TRAP_ENABLE_MASK | SWCR_MAP_MASK);
11026 fpcr = cpu_alpha_load_fpcr(cpu_env);
11027 fpcr &= ((uint64_t)FPCR_DYN_MASK << 32);
11028 fpcr |= alpha_ieee_swcr_to_fpcr(swcr);
11029 cpu_alpha_store_fpcr(cpu_env, fpcr);
11030 ret = 0;
11032 break;
11034 case TARGET_SSI_IEEE_RAISE_EXCEPTION:
11036 uint64_t exc, fpcr, fex;
11038 if (get_user_u64(exc, arg2)) {
11039 return -TARGET_EFAULT;
11041 exc &= SWCR_STATUS_MASK;
11042 fpcr = cpu_alpha_load_fpcr(cpu_env);
11044 /* Old exceptions are not signaled. */
11045 fex = alpha_ieee_fpcr_to_swcr(fpcr);
11046 fex = exc & ~fex;
11047 fex >>= SWCR_STATUS_TO_EXCSUM_SHIFT;
11048 fex &= ((CPUArchState *)cpu_env)->swcr;
11050 /* Update the hardware fpcr. */
11051 fpcr |= alpha_ieee_swcr_to_fpcr(exc);
11052 cpu_alpha_store_fpcr(cpu_env, fpcr);
11054 if (fex) {
11055 int si_code = TARGET_FPE_FLTUNK;
11056 target_siginfo_t info;
11058 if (fex & SWCR_TRAP_ENABLE_DNO) {
11059 si_code = TARGET_FPE_FLTUND;
11061 if (fex & SWCR_TRAP_ENABLE_INE) {
11062 si_code = TARGET_FPE_FLTRES;
11064 if (fex & SWCR_TRAP_ENABLE_UNF) {
11065 si_code = TARGET_FPE_FLTUND;
11067 if (fex & SWCR_TRAP_ENABLE_OVF) {
11068 si_code = TARGET_FPE_FLTOVF;
11070 if (fex & SWCR_TRAP_ENABLE_DZE) {
11071 si_code = TARGET_FPE_FLTDIV;
11073 if (fex & SWCR_TRAP_ENABLE_INV) {
11074 si_code = TARGET_FPE_FLTINV;
11077 info.si_signo = SIGFPE;
11078 info.si_errno = 0;
11079 info.si_code = si_code;
11080 info._sifields._sigfault._addr
11081 = ((CPUArchState *)cpu_env)->pc;
11082 queue_signal((CPUArchState *)cpu_env, info.si_signo,
11083 QEMU_SI_FAULT, &info);
11085 ret = 0;
11087 break;
11089 /* case SSI_NVPAIRS:
11090 -- Used with SSIN_UACPROC to enable unaligned accesses.
11091 case SSI_IEEE_STATE_AT_SIGNAL:
11092 case SSI_IEEE_IGNORE_STATE_AT_SIGNAL:
11093 -- Not implemented in linux kernel
11096 return ret;
11097 #endif
11098 #ifdef TARGET_NR_osf_sigprocmask
11099 /* Alpha specific. */
11100 case TARGET_NR_osf_sigprocmask:
11102 abi_ulong mask;
11103 int how;
11104 sigset_t set, oldset;
11106 switch(arg1) {
11107 case TARGET_SIG_BLOCK:
11108 how = SIG_BLOCK;
11109 break;
11110 case TARGET_SIG_UNBLOCK:
11111 how = SIG_UNBLOCK;
11112 break;
11113 case TARGET_SIG_SETMASK:
11114 how = SIG_SETMASK;
11115 break;
11116 default:
11117 return -TARGET_EINVAL;
11119 mask = arg2;
11120 target_to_host_old_sigset(&set, &mask);
11121 ret = do_sigprocmask(how, &set, &oldset);
11122 if (!ret) {
11123 host_to_target_old_sigset(&mask, &oldset);
11124 ret = mask;
11127 return ret;
11128 #endif
11130 #ifdef TARGET_NR_getgid32
11131 case TARGET_NR_getgid32:
11132 return get_errno(getgid());
11133 #endif
11134 #ifdef TARGET_NR_geteuid32
11135 case TARGET_NR_geteuid32:
11136 return get_errno(geteuid());
11137 #endif
11138 #ifdef TARGET_NR_getegid32
11139 case TARGET_NR_getegid32:
11140 return get_errno(getegid());
11141 #endif
11142 #ifdef TARGET_NR_setreuid32
11143 case TARGET_NR_setreuid32:
11144 return get_errno(setreuid(arg1, arg2));
11145 #endif
11146 #ifdef TARGET_NR_setregid32
11147 case TARGET_NR_setregid32:
11148 return get_errno(setregid(arg1, arg2));
11149 #endif
11150 #ifdef TARGET_NR_getgroups32
11151 case TARGET_NR_getgroups32:
11153 int gidsetsize = arg1;
11154 uint32_t *target_grouplist;
11155 gid_t *grouplist;
11156 int i;
11158 grouplist = alloca(gidsetsize * sizeof(gid_t));
11159 ret = get_errno(getgroups(gidsetsize, grouplist));
11160 if (gidsetsize == 0)
11161 return ret;
11162 if (!is_error(ret)) {
11163 target_grouplist = lock_user(VERIFY_WRITE, arg2, gidsetsize * 4, 0);
11164 if (!target_grouplist) {
11165 return -TARGET_EFAULT;
11167 for(i = 0;i < ret; i++)
11168 target_grouplist[i] = tswap32(grouplist[i]);
11169 unlock_user(target_grouplist, arg2, gidsetsize * 4);
11172 return ret;
11173 #endif
11174 #ifdef TARGET_NR_setgroups32
11175 case TARGET_NR_setgroups32:
11177 int gidsetsize = arg1;
11178 uint32_t *target_grouplist;
11179 gid_t *grouplist;
11180 int i;
11182 grouplist = alloca(gidsetsize * sizeof(gid_t));
11183 target_grouplist = lock_user(VERIFY_READ, arg2, gidsetsize * 4, 1);
11184 if (!target_grouplist) {
11185 return -TARGET_EFAULT;
11187 for(i = 0;i < gidsetsize; i++)
11188 grouplist[i] = tswap32(target_grouplist[i]);
11189 unlock_user(target_grouplist, arg2, 0);
11190 return get_errno(setgroups(gidsetsize, grouplist));
11192 #endif
11193 #ifdef TARGET_NR_fchown32
11194 case TARGET_NR_fchown32:
11195 return get_errno(fchown(arg1, arg2, arg3));
11196 #endif
11197 #ifdef TARGET_NR_setresuid32
11198 case TARGET_NR_setresuid32:
11199 return get_errno(sys_setresuid(arg1, arg2, arg3));
11200 #endif
11201 #ifdef TARGET_NR_getresuid32
11202 case TARGET_NR_getresuid32:
11204 uid_t ruid, euid, suid;
11205 ret = get_errno(getresuid(&ruid, &euid, &suid));
11206 if (!is_error(ret)) {
11207 if (put_user_u32(ruid, arg1)
11208 || put_user_u32(euid, arg2)
11209 || put_user_u32(suid, arg3))
11210 return -TARGET_EFAULT;
11213 return ret;
11214 #endif
11215 #ifdef TARGET_NR_setresgid32
11216 case TARGET_NR_setresgid32:
11217 return get_errno(sys_setresgid(arg1, arg2, arg3));
11218 #endif
11219 #ifdef TARGET_NR_getresgid32
11220 case TARGET_NR_getresgid32:
11222 gid_t rgid, egid, sgid;
11223 ret = get_errno(getresgid(&rgid, &egid, &sgid));
11224 if (!is_error(ret)) {
11225 if (put_user_u32(rgid, arg1)
11226 || put_user_u32(egid, arg2)
11227 || put_user_u32(sgid, arg3))
11228 return -TARGET_EFAULT;
11231 return ret;
11232 #endif
11233 #ifdef TARGET_NR_chown32
11234 case TARGET_NR_chown32:
11235 if (!(p = lock_user_string(arg1)))
11236 return -TARGET_EFAULT;
11237 ret = get_errno(chown(p, arg2, arg3));
11238 unlock_user(p, arg1, 0);
11239 return ret;
11240 #endif
11241 #ifdef TARGET_NR_setuid32
11242 case TARGET_NR_setuid32:
11243 return get_errno(sys_setuid(arg1));
11244 #endif
11245 #ifdef TARGET_NR_setgid32
11246 case TARGET_NR_setgid32:
11247 return get_errno(sys_setgid(arg1));
11248 #endif
11249 #ifdef TARGET_NR_setfsuid32
11250 case TARGET_NR_setfsuid32:
11251 return get_errno(setfsuid(arg1));
11252 #endif
11253 #ifdef TARGET_NR_setfsgid32
11254 case TARGET_NR_setfsgid32:
11255 return get_errno(setfsgid(arg1));
11256 #endif
11257 #ifdef TARGET_NR_mincore
11258 case TARGET_NR_mincore:
11260 void *a = lock_user(VERIFY_READ, arg1, arg2, 0);
11261 if (!a) {
11262 return -TARGET_ENOMEM;
11264 p = lock_user_string(arg3);
11265 if (!p) {
11266 ret = -TARGET_EFAULT;
11267 } else {
11268 ret = get_errno(mincore(a, arg2, p));
11269 unlock_user(p, arg3, ret);
11271 unlock_user(a, arg1, 0);
11273 return ret;
11274 #endif
11275 #ifdef TARGET_NR_arm_fadvise64_64
11276 case TARGET_NR_arm_fadvise64_64:
11277 /* arm_fadvise64_64 looks like fadvise64_64 but
11278 * with different argument order: fd, advice, offset, len
11279 * rather than the usual fd, offset, len, advice.
11280 * Note that offset and len are both 64-bit so appear as
11281 * pairs of 32-bit registers.
11283 ret = posix_fadvise(arg1, target_offset64(arg3, arg4),
11284 target_offset64(arg5, arg6), arg2);
11285 return -host_to_target_errno(ret);
11286 #endif
11288 #if TARGET_ABI_BITS == 32
11290 #ifdef TARGET_NR_fadvise64_64
11291 case TARGET_NR_fadvise64_64:
11292 #if defined(TARGET_PPC) || defined(TARGET_XTENSA)
11293 /* 6 args: fd, advice, offset (high, low), len (high, low) */
11294 ret = arg2;
11295 arg2 = arg3;
11296 arg3 = arg4;
11297 arg4 = arg5;
11298 arg5 = arg6;
11299 arg6 = ret;
11300 #else
11301 /* 6 args: fd, offset (high, low), len (high, low), advice */
11302 if (regpairs_aligned(cpu_env, num)) {
11303 /* offset is in (3,4), len in (5,6) and advice in 7 */
11304 arg2 = arg3;
11305 arg3 = arg4;
11306 arg4 = arg5;
11307 arg5 = arg6;
11308 arg6 = arg7;
11310 #endif
11311 ret = posix_fadvise(arg1, target_offset64(arg2, arg3),
11312 target_offset64(arg4, arg5), arg6);
11313 return -host_to_target_errno(ret);
11314 #endif
11316 #ifdef TARGET_NR_fadvise64
11317 case TARGET_NR_fadvise64:
11318 /* 5 args: fd, offset (high, low), len, advice */
11319 if (regpairs_aligned(cpu_env, num)) {
11320 /* offset is in (3,4), len in 5 and advice in 6 */
11321 arg2 = arg3;
11322 arg3 = arg4;
11323 arg4 = arg5;
11324 arg5 = arg6;
11326 ret = posix_fadvise(arg1, target_offset64(arg2, arg3), arg4, arg5);
11327 return -host_to_target_errno(ret);
11328 #endif
11330 #else /* not a 32-bit ABI */
11331 #if defined(TARGET_NR_fadvise64_64) || defined(TARGET_NR_fadvise64)
11332 #ifdef TARGET_NR_fadvise64_64
11333 case TARGET_NR_fadvise64_64:
11334 #endif
11335 #ifdef TARGET_NR_fadvise64
11336 case TARGET_NR_fadvise64:
11337 #endif
11338 #ifdef TARGET_S390X
11339 switch (arg4) {
11340 case 4: arg4 = POSIX_FADV_NOREUSE + 1; break; /* make sure it's an invalid value */
11341 case 5: arg4 = POSIX_FADV_NOREUSE + 2; break; /* ditto */
11342 case 6: arg4 = POSIX_FADV_DONTNEED; break;
11343 case 7: arg4 = POSIX_FADV_NOREUSE; break;
11344 default: break;
11346 #endif
11347 return -host_to_target_errno(posix_fadvise(arg1, arg2, arg3, arg4));
11348 #endif
11349 #endif /* end of 64-bit ABI fadvise handling */
11351 #ifdef TARGET_NR_madvise
11352 case TARGET_NR_madvise:
11353 /* A straight passthrough may not be safe because qemu sometimes
11354 turns private file-backed mappings into anonymous mappings.
11355 This will break MADV_DONTNEED.
11356 This is a hint, so ignoring and returning success is ok. */
11357 return 0;
11358 #endif
11359 #ifdef TARGET_NR_fcntl64
11360 case TARGET_NR_fcntl64:
11362 int cmd;
11363 struct flock64 fl;
11364 from_flock64_fn *copyfrom = copy_from_user_flock64;
11365 to_flock64_fn *copyto = copy_to_user_flock64;
11367 #ifdef TARGET_ARM
11368 if (!((CPUARMState *)cpu_env)->eabi) {
11369 copyfrom = copy_from_user_oabi_flock64;
11370 copyto = copy_to_user_oabi_flock64;
11372 #endif
11374 cmd = target_to_host_fcntl_cmd(arg2);
11375 if (cmd == -TARGET_EINVAL) {
11376 return cmd;
11379 switch(arg2) {
11380 case TARGET_F_GETLK64:
11381 ret = copyfrom(&fl, arg3);
11382 if (ret) {
11383 break;
11385 ret = get_errno(safe_fcntl(arg1, cmd, &fl));
11386 if (ret == 0) {
11387 ret = copyto(arg3, &fl);
11389 break;
11391 case TARGET_F_SETLK64:
11392 case TARGET_F_SETLKW64:
11393 ret = copyfrom(&fl, arg3);
11394 if (ret) {
11395 break;
11397 ret = get_errno(safe_fcntl(arg1, cmd, &fl));
11398 break;
11399 default:
11400 ret = do_fcntl(arg1, arg2, arg3);
11401 break;
11403 return ret;
11405 #endif
11406 #ifdef TARGET_NR_cacheflush
11407 case TARGET_NR_cacheflush:
11408 /* self-modifying code is handled automatically, so nothing needed */
11409 return 0;
11410 #endif
11411 #ifdef TARGET_NR_getpagesize
11412 case TARGET_NR_getpagesize:
11413 return TARGET_PAGE_SIZE;
11414 #endif
11415 case TARGET_NR_gettid:
11416 return get_errno(sys_gettid());
11417 #ifdef TARGET_NR_readahead
11418 case TARGET_NR_readahead:
11419 #if TARGET_ABI_BITS == 32
11420 if (regpairs_aligned(cpu_env, num)) {
11421 arg2 = arg3;
11422 arg3 = arg4;
11423 arg4 = arg5;
11425 ret = get_errno(readahead(arg1, target_offset64(arg2, arg3) , arg4));
11426 #else
11427 ret = get_errno(readahead(arg1, arg2, arg3));
11428 #endif
11429 return ret;
11430 #endif
11431 #ifdef CONFIG_ATTR
11432 #ifdef TARGET_NR_setxattr
11433 case TARGET_NR_listxattr:
11434 case TARGET_NR_llistxattr:
11436 void *p, *b = 0;
11437 if (arg2) {
11438 b = lock_user(VERIFY_WRITE, arg2, arg3, 0);
11439 if (!b) {
11440 return -TARGET_EFAULT;
11443 p = lock_user_string(arg1);
11444 if (p) {
11445 if (num == TARGET_NR_listxattr) {
11446 ret = get_errno(listxattr(p, b, arg3));
11447 } else {
11448 ret = get_errno(llistxattr(p, b, arg3));
11450 } else {
11451 ret = -TARGET_EFAULT;
11453 unlock_user(p, arg1, 0);
11454 unlock_user(b, arg2, arg3);
11455 return ret;
11457 case TARGET_NR_flistxattr:
11459 void *b = 0;
11460 if (arg2) {
11461 b = lock_user(VERIFY_WRITE, arg2, arg3, 0);
11462 if (!b) {
11463 return -TARGET_EFAULT;
11466 ret = get_errno(flistxattr(arg1, b, arg3));
11467 unlock_user(b, arg2, arg3);
11468 return ret;
11470 case TARGET_NR_setxattr:
11471 case TARGET_NR_lsetxattr:
11473 void *p, *n, *v = 0;
11474 if (arg3) {
11475 v = lock_user(VERIFY_READ, arg3, arg4, 1);
11476 if (!v) {
11477 return -TARGET_EFAULT;
11480 p = lock_user_string(arg1);
11481 n = lock_user_string(arg2);
11482 if (p && n) {
11483 if (num == TARGET_NR_setxattr) {
11484 ret = get_errno(setxattr(p, n, v, arg4, arg5));
11485 } else {
11486 ret = get_errno(lsetxattr(p, n, v, arg4, arg5));
11488 } else {
11489 ret = -TARGET_EFAULT;
11491 unlock_user(p, arg1, 0);
11492 unlock_user(n, arg2, 0);
11493 unlock_user(v, arg3, 0);
11495 return ret;
11496 case TARGET_NR_fsetxattr:
11498 void *n, *v = 0;
11499 if (arg3) {
11500 v = lock_user(VERIFY_READ, arg3, arg4, 1);
11501 if (!v) {
11502 return -TARGET_EFAULT;
11505 n = lock_user_string(arg2);
11506 if (n) {
11507 ret = get_errno(fsetxattr(arg1, n, v, arg4, arg5));
11508 } else {
11509 ret = -TARGET_EFAULT;
11511 unlock_user(n, arg2, 0);
11512 unlock_user(v, arg3, 0);
11514 return ret;
11515 case TARGET_NR_getxattr:
11516 case TARGET_NR_lgetxattr:
11518 void *p, *n, *v = 0;
11519 if (arg3) {
11520 v = lock_user(VERIFY_WRITE, arg3, arg4, 0);
11521 if (!v) {
11522 return -TARGET_EFAULT;
11525 p = lock_user_string(arg1);
11526 n = lock_user_string(arg2);
11527 if (p && n) {
11528 if (num == TARGET_NR_getxattr) {
11529 ret = get_errno(getxattr(p, n, v, arg4));
11530 } else {
11531 ret = get_errno(lgetxattr(p, n, v, arg4));
11533 } else {
11534 ret = -TARGET_EFAULT;
11536 unlock_user(p, arg1, 0);
11537 unlock_user(n, arg2, 0);
11538 unlock_user(v, arg3, arg4);
11540 return ret;
11541 case TARGET_NR_fgetxattr:
11543 void *n, *v = 0;
11544 if (arg3) {
11545 v = lock_user(VERIFY_WRITE, arg3, arg4, 0);
11546 if (!v) {
11547 return -TARGET_EFAULT;
11550 n = lock_user_string(arg2);
11551 if (n) {
11552 ret = get_errno(fgetxattr(arg1, n, v, arg4));
11553 } else {
11554 ret = -TARGET_EFAULT;
11556 unlock_user(n, arg2, 0);
11557 unlock_user(v, arg3, arg4);
11559 return ret;
11560 case TARGET_NR_removexattr:
11561 case TARGET_NR_lremovexattr:
11563 void *p, *n;
11564 p = lock_user_string(arg1);
11565 n = lock_user_string(arg2);
11566 if (p && n) {
11567 if (num == TARGET_NR_removexattr) {
11568 ret = get_errno(removexattr(p, n));
11569 } else {
11570 ret = get_errno(lremovexattr(p, n));
11572 } else {
11573 ret = -TARGET_EFAULT;
11575 unlock_user(p, arg1, 0);
11576 unlock_user(n, arg2, 0);
11578 return ret;
11579 case TARGET_NR_fremovexattr:
11581 void *n;
11582 n = lock_user_string(arg2);
11583 if (n) {
11584 ret = get_errno(fremovexattr(arg1, n));
11585 } else {
11586 ret = -TARGET_EFAULT;
11588 unlock_user(n, arg2, 0);
11590 return ret;
11591 #endif
11592 #endif /* CONFIG_ATTR */
11593 #ifdef TARGET_NR_set_thread_area
11594 case TARGET_NR_set_thread_area:
11595 #if defined(TARGET_MIPS)
11596 ((CPUMIPSState *) cpu_env)->active_tc.CP0_UserLocal = arg1;
11597 return 0;
11598 #elif defined(TARGET_CRIS)
11599 if (arg1 & 0xff)
11600 ret = -TARGET_EINVAL;
11601 else {
11602 ((CPUCRISState *) cpu_env)->pregs[PR_PID] = arg1;
11603 ret = 0;
11605 return ret;
11606 #elif defined(TARGET_I386) && defined(TARGET_ABI32)
11607 return do_set_thread_area(cpu_env, arg1);
11608 #elif defined(TARGET_M68K)
11610 TaskState *ts = cpu->opaque;
11611 ts->tp_value = arg1;
11612 return 0;
11614 #else
11615 return -TARGET_ENOSYS;
11616 #endif
11617 #endif
11618 #ifdef TARGET_NR_get_thread_area
11619 case TARGET_NR_get_thread_area:
11620 #if defined(TARGET_I386) && defined(TARGET_ABI32)
11621 return do_get_thread_area(cpu_env, arg1);
11622 #elif defined(TARGET_M68K)
11624 TaskState *ts = cpu->opaque;
11625 return ts->tp_value;
11627 #else
11628 return -TARGET_ENOSYS;
11629 #endif
11630 #endif
11631 #ifdef TARGET_NR_getdomainname
11632 case TARGET_NR_getdomainname:
11633 return -TARGET_ENOSYS;
11634 #endif
11636 #ifdef TARGET_NR_clock_settime
11637 case TARGET_NR_clock_settime:
11639 struct timespec ts;
11641 ret = target_to_host_timespec(&ts, arg2);
11642 if (!is_error(ret)) {
11643 ret = get_errno(clock_settime(arg1, &ts));
11645 return ret;
11647 #endif
11648 #ifdef TARGET_NR_clock_settime64
11649 case TARGET_NR_clock_settime64:
11651 struct timespec ts;
11653 ret = target_to_host_timespec64(&ts, arg2);
11654 if (!is_error(ret)) {
11655 ret = get_errno(clock_settime(arg1, &ts));
11657 return ret;
11659 #endif
11660 #ifdef TARGET_NR_clock_gettime
11661 case TARGET_NR_clock_gettime:
11663 struct timespec ts;
11664 ret = get_errno(clock_gettime(arg1, &ts));
11665 if (!is_error(ret)) {
11666 ret = host_to_target_timespec(arg2, &ts);
11668 return ret;
11670 #endif
11671 #ifdef TARGET_NR_clock_gettime64
11672 case TARGET_NR_clock_gettime64:
11674 struct timespec ts;
11675 ret = get_errno(clock_gettime(arg1, &ts));
11676 if (!is_error(ret)) {
11677 ret = host_to_target_timespec64(arg2, &ts);
11679 return ret;
11681 #endif
11682 #ifdef TARGET_NR_clock_getres
11683 case TARGET_NR_clock_getres:
11685 struct timespec ts;
11686 ret = get_errno(clock_getres(arg1, &ts));
11687 if (!is_error(ret)) {
11688 host_to_target_timespec(arg2, &ts);
11690 return ret;
11692 #endif
11693 #ifdef TARGET_NR_clock_nanosleep
11694 case TARGET_NR_clock_nanosleep:
11696 struct timespec ts;
11697 target_to_host_timespec(&ts, arg3);
11698 ret = get_errno(safe_clock_nanosleep(arg1, arg2,
11699 &ts, arg4 ? &ts : NULL));
11700 if (arg4)
11701 host_to_target_timespec(arg4, &ts);
11703 #if defined(TARGET_PPC)
11704 /* clock_nanosleep is odd in that it returns positive errno values.
11705 * On PPC, CR0 bit 3 should be set in such a situation. */
11706 if (ret && ret != -TARGET_ERESTARTSYS) {
11707 ((CPUPPCState *)cpu_env)->crf[0] |= 1;
11709 #endif
11710 return ret;
11712 #endif
11714 #if defined(TARGET_NR_set_tid_address) && defined(__NR_set_tid_address)
11715 case TARGET_NR_set_tid_address:
11716 return get_errno(set_tid_address((int *)g2h(arg1)));
11717 #endif
11719 case TARGET_NR_tkill:
11720 return get_errno(safe_tkill((int)arg1, target_to_host_signal(arg2)));
11722 case TARGET_NR_tgkill:
11723 return get_errno(safe_tgkill((int)arg1, (int)arg2,
11724 target_to_host_signal(arg3)));
11726 #ifdef TARGET_NR_set_robust_list
11727 case TARGET_NR_set_robust_list:
11728 case TARGET_NR_get_robust_list:
11729 /* The ABI for supporting robust futexes has userspace pass
11730 * the kernel a pointer to a linked list which is updated by
11731 * userspace after the syscall; the list is walked by the kernel
11732 * when the thread exits. Since the linked list in QEMU guest
11733 * memory isn't a valid linked list for the host and we have
11734 * no way to reliably intercept the thread-death event, we can't
11735 * support these. Silently return ENOSYS so that guest userspace
11736 * falls back to a non-robust futex implementation (which should
11737 * be OK except in the corner case of the guest crashing while
11738 * holding a mutex that is shared with another process via
11739 * shared memory).
11741 return -TARGET_ENOSYS;
11742 #endif
11744 #if defined(TARGET_NR_utimensat)
11745 case TARGET_NR_utimensat:
11747 struct timespec *tsp, ts[2];
11748 if (!arg3) {
11749 tsp = NULL;
11750 } else {
11751 target_to_host_timespec(ts, arg3);
11752 target_to_host_timespec(ts+1, arg3+sizeof(struct target_timespec));
11753 tsp = ts;
11755 if (!arg2)
11756 ret = get_errno(sys_utimensat(arg1, NULL, tsp, arg4));
11757 else {
11758 if (!(p = lock_user_string(arg2))) {
11759 return -TARGET_EFAULT;
11761 ret = get_errno(sys_utimensat(arg1, path(p), tsp, arg4));
11762 unlock_user(p, arg2, 0);
11765 return ret;
11766 #endif
11767 #ifdef TARGET_NR_futex
11768 case TARGET_NR_futex:
11769 return do_futex(arg1, arg2, arg3, arg4, arg5, arg6);
11770 #endif
11771 #ifdef TARGET_NR_futex_time64
11772 case TARGET_NR_futex_time64:
11773 return do_futex_time64(arg1, arg2, arg3, arg4, arg5, arg6);
11774 #endif
11775 #if defined(TARGET_NR_inotify_init) && defined(__NR_inotify_init)
11776 case TARGET_NR_inotify_init:
11777 ret = get_errno(sys_inotify_init());
11778 if (ret >= 0) {
11779 fd_trans_register(ret, &target_inotify_trans);
11781 return ret;
11782 #endif
11783 #ifdef CONFIG_INOTIFY1
11784 #if defined(TARGET_NR_inotify_init1) && defined(__NR_inotify_init1)
11785 case TARGET_NR_inotify_init1:
11786 ret = get_errno(sys_inotify_init1(target_to_host_bitmask(arg1,
11787 fcntl_flags_tbl)));
11788 if (ret >= 0) {
11789 fd_trans_register(ret, &target_inotify_trans);
11791 return ret;
11792 #endif
11793 #endif
11794 #if defined(TARGET_NR_inotify_add_watch) && defined(__NR_inotify_add_watch)
11795 case TARGET_NR_inotify_add_watch:
11796 p = lock_user_string(arg2);
11797 ret = get_errno(sys_inotify_add_watch(arg1, path(p), arg3));
11798 unlock_user(p, arg2, 0);
11799 return ret;
11800 #endif
11801 #if defined(TARGET_NR_inotify_rm_watch) && defined(__NR_inotify_rm_watch)
11802 case TARGET_NR_inotify_rm_watch:
11803 return get_errno(sys_inotify_rm_watch(arg1, arg2));
11804 #endif
11806 #if defined(TARGET_NR_mq_open) && defined(__NR_mq_open)
11807 case TARGET_NR_mq_open:
11809 struct mq_attr posix_mq_attr;
11810 struct mq_attr *pposix_mq_attr;
11811 int host_flags;
11813 host_flags = target_to_host_bitmask(arg2, fcntl_flags_tbl);
11814 pposix_mq_attr = NULL;
11815 if (arg4) {
11816 if (copy_from_user_mq_attr(&posix_mq_attr, arg4) != 0) {
11817 return -TARGET_EFAULT;
11819 pposix_mq_attr = &posix_mq_attr;
11821 p = lock_user_string(arg1 - 1);
11822 if (!p) {
11823 return -TARGET_EFAULT;
11825 ret = get_errno(mq_open(p, host_flags, arg3, pposix_mq_attr));
11826 unlock_user (p, arg1, 0);
11828 return ret;
11830 case TARGET_NR_mq_unlink:
11831 p = lock_user_string(arg1 - 1);
11832 if (!p) {
11833 return -TARGET_EFAULT;
11835 ret = get_errno(mq_unlink(p));
11836 unlock_user (p, arg1, 0);
11837 return ret;
11839 #ifdef TARGET_NR_mq_timedsend
11840 case TARGET_NR_mq_timedsend:
11842 struct timespec ts;
11844 p = lock_user (VERIFY_READ, arg2, arg3, 1);
11845 if (arg5 != 0) {
11846 target_to_host_timespec(&ts, arg5);
11847 ret = get_errno(safe_mq_timedsend(arg1, p, arg3, arg4, &ts));
11848 host_to_target_timespec(arg5, &ts);
11849 } else {
11850 ret = get_errno(safe_mq_timedsend(arg1, p, arg3, arg4, NULL));
11852 unlock_user (p, arg2, arg3);
11854 return ret;
11855 #endif
11857 #ifdef TARGET_NR_mq_timedreceive
11858 case TARGET_NR_mq_timedreceive:
11860 struct timespec ts;
11861 unsigned int prio;
11863 p = lock_user (VERIFY_READ, arg2, arg3, 1);
11864 if (arg5 != 0) {
11865 target_to_host_timespec(&ts, arg5);
11866 ret = get_errno(safe_mq_timedreceive(arg1, p, arg3,
11867 &prio, &ts));
11868 host_to_target_timespec(arg5, &ts);
11869 } else {
11870 ret = get_errno(safe_mq_timedreceive(arg1, p, arg3,
11871 &prio, NULL));
11873 unlock_user (p, arg2, arg3);
11874 if (arg4 != 0)
11875 put_user_u32(prio, arg4);
11877 return ret;
11878 #endif
11880 /* Not implemented for now... */
11881 /* case TARGET_NR_mq_notify: */
11882 /* break; */
11884 case TARGET_NR_mq_getsetattr:
11886 struct mq_attr posix_mq_attr_in, posix_mq_attr_out;
11887 ret = 0;
11888 if (arg2 != 0) {
11889 copy_from_user_mq_attr(&posix_mq_attr_in, arg2);
11890 ret = get_errno(mq_setattr(arg1, &posix_mq_attr_in,
11891 &posix_mq_attr_out));
11892 } else if (arg3 != 0) {
11893 ret = get_errno(mq_getattr(arg1, &posix_mq_attr_out));
11895 if (ret == 0 && arg3 != 0) {
11896 copy_to_user_mq_attr(arg3, &posix_mq_attr_out);
11899 return ret;
11900 #endif
11902 #ifdef CONFIG_SPLICE
11903 #ifdef TARGET_NR_tee
11904 case TARGET_NR_tee:
11906 ret = get_errno(tee(arg1,arg2,arg3,arg4));
11908 return ret;
11909 #endif
11910 #ifdef TARGET_NR_splice
11911 case TARGET_NR_splice:
11913 loff_t loff_in, loff_out;
11914 loff_t *ploff_in = NULL, *ploff_out = NULL;
11915 if (arg2) {
11916 if (get_user_u64(loff_in, arg2)) {
11917 return -TARGET_EFAULT;
11919 ploff_in = &loff_in;
11921 if (arg4) {
11922 if (get_user_u64(loff_out, arg4)) {
11923 return -TARGET_EFAULT;
11925 ploff_out = &loff_out;
11927 ret = get_errno(splice(arg1, ploff_in, arg3, ploff_out, arg5, arg6));
11928 if (arg2) {
11929 if (put_user_u64(loff_in, arg2)) {
11930 return -TARGET_EFAULT;
11933 if (arg4) {
11934 if (put_user_u64(loff_out, arg4)) {
11935 return -TARGET_EFAULT;
11939 return ret;
11940 #endif
11941 #ifdef TARGET_NR_vmsplice
11942 case TARGET_NR_vmsplice:
11944 struct iovec *vec = lock_iovec(VERIFY_READ, arg2, arg3, 1);
11945 if (vec != NULL) {
11946 ret = get_errno(vmsplice(arg1, vec, arg3, arg4));
11947 unlock_iovec(vec, arg2, arg3, 0);
11948 } else {
11949 ret = -host_to_target_errno(errno);
11952 return ret;
11953 #endif
11954 #endif /* CONFIG_SPLICE */
11955 #ifdef CONFIG_EVENTFD
11956 #if defined(TARGET_NR_eventfd)
11957 case TARGET_NR_eventfd:
11958 ret = get_errno(eventfd(arg1, 0));
11959 if (ret >= 0) {
11960 fd_trans_register(ret, &target_eventfd_trans);
11962 return ret;
11963 #endif
11964 #if defined(TARGET_NR_eventfd2)
11965 case TARGET_NR_eventfd2:
11967 int host_flags = arg2 & (~(TARGET_O_NONBLOCK | TARGET_O_CLOEXEC));
11968 if (arg2 & TARGET_O_NONBLOCK) {
11969 host_flags |= O_NONBLOCK;
11971 if (arg2 & TARGET_O_CLOEXEC) {
11972 host_flags |= O_CLOEXEC;
11974 ret = get_errno(eventfd(arg1, host_flags));
11975 if (ret >= 0) {
11976 fd_trans_register(ret, &target_eventfd_trans);
11978 return ret;
11980 #endif
11981 #endif /* CONFIG_EVENTFD */
11982 #if defined(CONFIG_FALLOCATE) && defined(TARGET_NR_fallocate)
11983 case TARGET_NR_fallocate:
11984 #if TARGET_ABI_BITS == 32
11985 ret = get_errno(fallocate(arg1, arg2, target_offset64(arg3, arg4),
11986 target_offset64(arg5, arg6)));
11987 #else
11988 ret = get_errno(fallocate(arg1, arg2, arg3, arg4));
11989 #endif
11990 return ret;
11991 #endif
11992 #if defined(CONFIG_SYNC_FILE_RANGE)
11993 #if defined(TARGET_NR_sync_file_range)
11994 case TARGET_NR_sync_file_range:
11995 #if TARGET_ABI_BITS == 32
11996 #if defined(TARGET_MIPS)
11997 ret = get_errno(sync_file_range(arg1, target_offset64(arg3, arg4),
11998 target_offset64(arg5, arg6), arg7));
11999 #else
12000 ret = get_errno(sync_file_range(arg1, target_offset64(arg2, arg3),
12001 target_offset64(arg4, arg5), arg6));
12002 #endif /* !TARGET_MIPS */
12003 #else
12004 ret = get_errno(sync_file_range(arg1, arg2, arg3, arg4));
12005 #endif
12006 return ret;
12007 #endif
12008 #if defined(TARGET_NR_sync_file_range2) || \
12009 defined(TARGET_NR_arm_sync_file_range)
12010 #if defined(TARGET_NR_sync_file_range2)
12011 case TARGET_NR_sync_file_range2:
12012 #endif
12013 #if defined(TARGET_NR_arm_sync_file_range)
12014 case TARGET_NR_arm_sync_file_range:
12015 #endif
12016 /* This is like sync_file_range but the arguments are reordered */
12017 #if TARGET_ABI_BITS == 32
12018 ret = get_errno(sync_file_range(arg1, target_offset64(arg3, arg4),
12019 target_offset64(arg5, arg6), arg2));
12020 #else
12021 ret = get_errno(sync_file_range(arg1, arg3, arg4, arg2));
12022 #endif
12023 return ret;
12024 #endif
12025 #endif
12026 #if defined(TARGET_NR_signalfd4)
12027 case TARGET_NR_signalfd4:
12028 return do_signalfd4(arg1, arg2, arg4);
12029 #endif
12030 #if defined(TARGET_NR_signalfd)
12031 case TARGET_NR_signalfd:
12032 return do_signalfd4(arg1, arg2, 0);
12033 #endif
12034 #if defined(CONFIG_EPOLL)
12035 #if defined(TARGET_NR_epoll_create)
12036 case TARGET_NR_epoll_create:
12037 return get_errno(epoll_create(arg1));
12038 #endif
12039 #if defined(TARGET_NR_epoll_create1) && defined(CONFIG_EPOLL_CREATE1)
12040 case TARGET_NR_epoll_create1:
12041 return get_errno(epoll_create1(target_to_host_bitmask(arg1, fcntl_flags_tbl)));
12042 #endif
12043 #if defined(TARGET_NR_epoll_ctl)
12044 case TARGET_NR_epoll_ctl:
12046 struct epoll_event ep;
12047 struct epoll_event *epp = 0;
12048 if (arg4) {
12049 struct target_epoll_event *target_ep;
12050 if (!lock_user_struct(VERIFY_READ, target_ep, arg4, 1)) {
12051 return -TARGET_EFAULT;
12053 ep.events = tswap32(target_ep->events);
12054 /* The epoll_data_t union is just opaque data to the kernel,
12055 * so we transfer all 64 bits across and need not worry what
12056 * actual data type it is.
12058 ep.data.u64 = tswap64(target_ep->data.u64);
12059 unlock_user_struct(target_ep, arg4, 0);
12060 epp = &ep;
12062 return get_errno(epoll_ctl(arg1, arg2, arg3, epp));
12064 #endif
12066 #if defined(TARGET_NR_epoll_wait) || defined(TARGET_NR_epoll_pwait)
12067 #if defined(TARGET_NR_epoll_wait)
12068 case TARGET_NR_epoll_wait:
12069 #endif
12070 #if defined(TARGET_NR_epoll_pwait)
12071 case TARGET_NR_epoll_pwait:
12072 #endif
12074 struct target_epoll_event *target_ep;
12075 struct epoll_event *ep;
12076 int epfd = arg1;
12077 int maxevents = arg3;
12078 int timeout = arg4;
12080 if (maxevents <= 0 || maxevents > TARGET_EP_MAX_EVENTS) {
12081 return -TARGET_EINVAL;
12084 target_ep = lock_user(VERIFY_WRITE, arg2,
12085 maxevents * sizeof(struct target_epoll_event), 1);
12086 if (!target_ep) {
12087 return -TARGET_EFAULT;
12090 ep = g_try_new(struct epoll_event, maxevents);
12091 if (!ep) {
12092 unlock_user(target_ep, arg2, 0);
12093 return -TARGET_ENOMEM;
12096 switch (num) {
12097 #if defined(TARGET_NR_epoll_pwait)
12098 case TARGET_NR_epoll_pwait:
12100 target_sigset_t *target_set;
12101 sigset_t _set, *set = &_set;
12103 if (arg5) {
12104 if (arg6 != sizeof(target_sigset_t)) {
12105 ret = -TARGET_EINVAL;
12106 break;
12109 target_set = lock_user(VERIFY_READ, arg5,
12110 sizeof(target_sigset_t), 1);
12111 if (!target_set) {
12112 ret = -TARGET_EFAULT;
12113 break;
12115 target_to_host_sigset(set, target_set);
12116 unlock_user(target_set, arg5, 0);
12117 } else {
12118 set = NULL;
12121 ret = get_errno(safe_epoll_pwait(epfd, ep, maxevents, timeout,
12122 set, SIGSET_T_SIZE));
12123 break;
12125 #endif
12126 #if defined(TARGET_NR_epoll_wait)
12127 case TARGET_NR_epoll_wait:
12128 ret = get_errno(safe_epoll_pwait(epfd, ep, maxevents, timeout,
12129 NULL, 0));
12130 break;
12131 #endif
12132 default:
12133 ret = -TARGET_ENOSYS;
12135 if (!is_error(ret)) {
12136 int i;
12137 for (i = 0; i < ret; i++) {
12138 target_ep[i].events = tswap32(ep[i].events);
12139 target_ep[i].data.u64 = tswap64(ep[i].data.u64);
12141 unlock_user(target_ep, arg2,
12142 ret * sizeof(struct target_epoll_event));
12143 } else {
12144 unlock_user(target_ep, arg2, 0);
12146 g_free(ep);
12147 return ret;
12149 #endif
12150 #endif
12151 #ifdef TARGET_NR_prlimit64
12152 case TARGET_NR_prlimit64:
12154 /* args: pid, resource number, ptr to new rlimit, ptr to old rlimit */
12155 struct target_rlimit64 *target_rnew, *target_rold;
12156 struct host_rlimit64 rnew, rold, *rnewp = 0;
12157 int resource = target_to_host_resource(arg2);
12159 if (arg3 && (resource != RLIMIT_AS &&
12160 resource != RLIMIT_DATA &&
12161 resource != RLIMIT_STACK)) {
12162 if (!lock_user_struct(VERIFY_READ, target_rnew, arg3, 1)) {
12163 return -TARGET_EFAULT;
12165 rnew.rlim_cur = tswap64(target_rnew->rlim_cur);
12166 rnew.rlim_max = tswap64(target_rnew->rlim_max);
12167 unlock_user_struct(target_rnew, arg3, 0);
12168 rnewp = &rnew;
12171 ret = get_errno(sys_prlimit64(arg1, resource, rnewp, arg4 ? &rold : 0));
12172 if (!is_error(ret) && arg4) {
12173 if (!lock_user_struct(VERIFY_WRITE, target_rold, arg4, 1)) {
12174 return -TARGET_EFAULT;
12176 target_rold->rlim_cur = tswap64(rold.rlim_cur);
12177 target_rold->rlim_max = tswap64(rold.rlim_max);
12178 unlock_user_struct(target_rold, arg4, 1);
12180 return ret;
12182 #endif
12183 #ifdef TARGET_NR_gethostname
12184 case TARGET_NR_gethostname:
12186 char *name = lock_user(VERIFY_WRITE, arg1, arg2, 0);
12187 if (name) {
12188 ret = get_errno(gethostname(name, arg2));
12189 unlock_user(name, arg1, arg2);
12190 } else {
12191 ret = -TARGET_EFAULT;
12193 return ret;
12195 #endif
12196 #ifdef TARGET_NR_atomic_cmpxchg_32
12197 case TARGET_NR_atomic_cmpxchg_32:
12199 /* should use start_exclusive from main.c */
12200 abi_ulong mem_value;
12201 if (get_user_u32(mem_value, arg6)) {
12202 target_siginfo_t info;
12203 info.si_signo = SIGSEGV;
12204 info.si_errno = 0;
12205 info.si_code = TARGET_SEGV_MAPERR;
12206 info._sifields._sigfault._addr = arg6;
12207 queue_signal((CPUArchState *)cpu_env, info.si_signo,
12208 QEMU_SI_FAULT, &info);
12209 ret = 0xdeadbeef;
12212 if (mem_value == arg2)
12213 put_user_u32(arg1, arg6);
12214 return mem_value;
12216 #endif
12217 #ifdef TARGET_NR_atomic_barrier
12218 case TARGET_NR_atomic_barrier:
12219 /* Like the kernel implementation and the
12220 qemu arm barrier, no-op this? */
12221 return 0;
12222 #endif
12224 #ifdef TARGET_NR_timer_create
12225 case TARGET_NR_timer_create:
12227 /* args: clockid_t clockid, struct sigevent *sevp, timer_t *timerid */
12229 struct sigevent host_sevp = { {0}, }, *phost_sevp = NULL;
12231 int clkid = arg1;
12232 int timer_index = next_free_host_timer();
12234 if (timer_index < 0) {
12235 ret = -TARGET_EAGAIN;
12236 } else {
12237 timer_t *phtimer = g_posix_timers + timer_index;
12239 if (arg2) {
12240 phost_sevp = &host_sevp;
12241 ret = target_to_host_sigevent(phost_sevp, arg2);
12242 if (ret != 0) {
12243 return ret;
12247 ret = get_errno(timer_create(clkid, phost_sevp, phtimer));
12248 if (ret) {
12249 phtimer = NULL;
12250 } else {
12251 if (put_user(TIMER_MAGIC | timer_index, arg3, target_timer_t)) {
12252 return -TARGET_EFAULT;
12256 return ret;
12258 #endif
12260 #ifdef TARGET_NR_timer_settime
12261 case TARGET_NR_timer_settime:
12263 /* args: timer_t timerid, int flags, const struct itimerspec *new_value,
12264 * struct itimerspec * old_value */
12265 target_timer_t timerid = get_timer_id(arg1);
12267 if (timerid < 0) {
12268 ret = timerid;
12269 } else if (arg3 == 0) {
12270 ret = -TARGET_EINVAL;
12271 } else {
12272 timer_t htimer = g_posix_timers[timerid];
12273 struct itimerspec hspec_new = {{0},}, hspec_old = {{0},};
12275 if (target_to_host_itimerspec(&hspec_new, arg3)) {
12276 return -TARGET_EFAULT;
12278 ret = get_errno(
12279 timer_settime(htimer, arg2, &hspec_new, &hspec_old));
12280 if (arg4 && host_to_target_itimerspec(arg4, &hspec_old)) {
12281 return -TARGET_EFAULT;
12284 return ret;
12286 #endif
12288 #ifdef TARGET_NR_timer_gettime
12289 case TARGET_NR_timer_gettime:
12291 /* args: timer_t timerid, struct itimerspec *curr_value */
12292 target_timer_t timerid = get_timer_id(arg1);
12294 if (timerid < 0) {
12295 ret = timerid;
12296 } else if (!arg2) {
12297 ret = -TARGET_EFAULT;
12298 } else {
12299 timer_t htimer = g_posix_timers[timerid];
12300 struct itimerspec hspec;
12301 ret = get_errno(timer_gettime(htimer, &hspec));
12303 if (host_to_target_itimerspec(arg2, &hspec)) {
12304 ret = -TARGET_EFAULT;
12307 return ret;
12309 #endif
12311 #ifdef TARGET_NR_timer_getoverrun
12312 case TARGET_NR_timer_getoverrun:
12314 /* args: timer_t timerid */
12315 target_timer_t timerid = get_timer_id(arg1);
12317 if (timerid < 0) {
12318 ret = timerid;
12319 } else {
12320 timer_t htimer = g_posix_timers[timerid];
12321 ret = get_errno(timer_getoverrun(htimer));
12323 return ret;
12325 #endif
12327 #ifdef TARGET_NR_timer_delete
12328 case TARGET_NR_timer_delete:
12330 /* args: timer_t timerid */
12331 target_timer_t timerid = get_timer_id(arg1);
12333 if (timerid < 0) {
12334 ret = timerid;
12335 } else {
12336 timer_t htimer = g_posix_timers[timerid];
12337 ret = get_errno(timer_delete(htimer));
12338 g_posix_timers[timerid] = 0;
12340 return ret;
12342 #endif
12344 #if defined(TARGET_NR_timerfd_create) && defined(CONFIG_TIMERFD)
12345 case TARGET_NR_timerfd_create:
12346 return get_errno(timerfd_create(arg1,
12347 target_to_host_bitmask(arg2, fcntl_flags_tbl)));
12348 #endif
12350 #if defined(TARGET_NR_timerfd_gettime) && defined(CONFIG_TIMERFD)
12351 case TARGET_NR_timerfd_gettime:
12353 struct itimerspec its_curr;
12355 ret = get_errno(timerfd_gettime(arg1, &its_curr));
12357 if (arg2 && host_to_target_itimerspec(arg2, &its_curr)) {
12358 return -TARGET_EFAULT;
12361 return ret;
12362 #endif
12364 #if defined(TARGET_NR_timerfd_settime) && defined(CONFIG_TIMERFD)
12365 case TARGET_NR_timerfd_settime:
12367 struct itimerspec its_new, its_old, *p_new;
12369 if (arg3) {
12370 if (target_to_host_itimerspec(&its_new, arg3)) {
12371 return -TARGET_EFAULT;
12373 p_new = &its_new;
12374 } else {
12375 p_new = NULL;
12378 ret = get_errno(timerfd_settime(arg1, arg2, p_new, &its_old));
12380 if (arg4 && host_to_target_itimerspec(arg4, &its_old)) {
12381 return -TARGET_EFAULT;
12384 return ret;
12385 #endif
12387 #if defined(TARGET_NR_ioprio_get) && defined(__NR_ioprio_get)
12388 case TARGET_NR_ioprio_get:
12389 return get_errno(ioprio_get(arg1, arg2));
12390 #endif
12392 #if defined(TARGET_NR_ioprio_set) && defined(__NR_ioprio_set)
12393 case TARGET_NR_ioprio_set:
12394 return get_errno(ioprio_set(arg1, arg2, arg3));
12395 #endif
12397 #if defined(TARGET_NR_setns) && defined(CONFIG_SETNS)
12398 case TARGET_NR_setns:
12399 return get_errno(setns(arg1, arg2));
12400 #endif
12401 #if defined(TARGET_NR_unshare) && defined(CONFIG_SETNS)
12402 case TARGET_NR_unshare:
12403 return get_errno(unshare(arg1));
12404 #endif
12405 #if defined(TARGET_NR_kcmp) && defined(__NR_kcmp)
12406 case TARGET_NR_kcmp:
12407 return get_errno(kcmp(arg1, arg2, arg3, arg4, arg5));
12408 #endif
12409 #ifdef TARGET_NR_swapcontext
12410 case TARGET_NR_swapcontext:
12411 /* PowerPC specific. */
12412 return do_swapcontext(cpu_env, arg1, arg2, arg3);
12413 #endif
12414 #ifdef TARGET_NR_memfd_create
12415 case TARGET_NR_memfd_create:
12416 p = lock_user_string(arg1);
12417 if (!p) {
12418 return -TARGET_EFAULT;
12420 ret = get_errno(memfd_create(p, arg2));
12421 fd_trans_unregister(ret);
12422 unlock_user(p, arg1, 0);
12423 return ret;
12424 #endif
12425 #if defined TARGET_NR_membarrier && defined __NR_membarrier
12426 case TARGET_NR_membarrier:
12427 return get_errno(membarrier(arg1, arg2));
12428 #endif
12430 default:
12431 qemu_log_mask(LOG_UNIMP, "Unsupported syscall: %d\n", num);
12432 return -TARGET_ENOSYS;
12434 return ret;
12437 abi_long do_syscall(void *cpu_env, int num, abi_long arg1,
12438 abi_long arg2, abi_long arg3, abi_long arg4,
12439 abi_long arg5, abi_long arg6, abi_long arg7,
12440 abi_long arg8)
12442 CPUState *cpu = env_cpu(cpu_env);
12443 abi_long ret;
12445 #ifdef DEBUG_ERESTARTSYS
12446 /* Debug-only code for exercising the syscall-restart code paths
12447 * in the per-architecture cpu main loops: restart every syscall
12448 * the guest makes once before letting it through.
12451 static bool flag;
12452 flag = !flag;
12453 if (flag) {
12454 return -TARGET_ERESTARTSYS;
12457 #endif
12459 record_syscall_start(cpu, num, arg1,
12460 arg2, arg3, arg4, arg5, arg6, arg7, arg8);
12462 if (unlikely(qemu_loglevel_mask(LOG_STRACE))) {
12463 print_syscall(num, arg1, arg2, arg3, arg4, arg5, arg6);
12466 ret = do_syscall1(cpu_env, num, arg1, arg2, arg3, arg4,
12467 arg5, arg6, arg7, arg8);
12469 if (unlikely(qemu_loglevel_mask(LOG_STRACE))) {
12470 print_syscall_ret(num, ret);
12473 record_syscall_return(cpu, num, ret);
12474 return ret;