semihosting: Create semihost_sys_poll_one
[qemu/rayw.git] / linux-user / syscall.c
blob669add74c11ae19bacdfdd3c6715700b72ae2292
1 /*
2 * Linux syscalls
4 * Copyright (c) 2003 Fabrice Bellard
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, see <http://www.gnu.org/licenses/>.
19 #define _ATFILE_SOURCE
20 #include "qemu/osdep.h"
21 #include "qemu/cutils.h"
22 #include "qemu/path.h"
23 #include "qemu/memfd.h"
24 #include "qemu/queue.h"
25 #include <elf.h>
26 #include <endian.h>
27 #include <grp.h>
28 #include <sys/ipc.h>
29 #include <sys/msg.h>
30 #include <sys/wait.h>
31 #include <sys/mount.h>
32 #include <sys/file.h>
33 #include <sys/fsuid.h>
34 #include <sys/personality.h>
35 #include <sys/prctl.h>
36 #include <sys/resource.h>
37 #include <sys/swap.h>
38 #include <linux/capability.h>
39 #include <sched.h>
40 #include <sys/timex.h>
41 #include <sys/socket.h>
42 #include <linux/sockios.h>
43 #include <sys/un.h>
44 #include <sys/uio.h>
45 #include <poll.h>
46 #include <sys/times.h>
47 #include <sys/shm.h>
48 #include <sys/sem.h>
49 #include <sys/statfs.h>
50 #include <utime.h>
51 #include <sys/sysinfo.h>
52 #include <sys/signalfd.h>
53 //#include <sys/user.h>
54 #include <netinet/in.h>
55 #include <netinet/ip.h>
56 #include <netinet/tcp.h>
57 #include <netinet/udp.h>
58 #include <linux/wireless.h>
59 #include <linux/icmp.h>
60 #include <linux/icmpv6.h>
61 #include <linux/if_tun.h>
62 #include <linux/in6.h>
63 #include <linux/errqueue.h>
64 #include <linux/random.h>
65 #ifdef CONFIG_TIMERFD
66 #include <sys/timerfd.h>
67 #endif
68 #ifdef CONFIG_EVENTFD
69 #include <sys/eventfd.h>
70 #endif
71 #ifdef CONFIG_EPOLL
72 #include <sys/epoll.h>
73 #endif
74 #ifdef CONFIG_ATTR
75 #include "qemu/xattr.h"
76 #endif
77 #ifdef CONFIG_SENDFILE
78 #include <sys/sendfile.h>
79 #endif
80 #ifdef HAVE_SYS_KCOV_H
81 #include <sys/kcov.h>
82 #endif
84 #define termios host_termios
85 #define winsize host_winsize
86 #define termio host_termio
87 #define sgttyb host_sgttyb /* same as target */
88 #define tchars host_tchars /* same as target */
89 #define ltchars host_ltchars /* same as target */
91 #include <linux/termios.h>
92 #include <linux/unistd.h>
93 #include <linux/cdrom.h>
94 #include <linux/hdreg.h>
95 #include <linux/soundcard.h>
96 #include <linux/kd.h>
97 #include <linux/mtio.h>
98 #include <linux/fs.h>
99 #include <linux/fd.h>
100 #if defined(CONFIG_FIEMAP)
101 #include <linux/fiemap.h>
102 #endif
103 #include <linux/fb.h>
104 #if defined(CONFIG_USBFS)
105 #include <linux/usbdevice_fs.h>
106 #include <linux/usb/ch9.h>
107 #endif
108 #include <linux/vt.h>
109 #include <linux/dm-ioctl.h>
110 #include <linux/reboot.h>
111 #include <linux/route.h>
112 #include <linux/filter.h>
113 #include <linux/blkpg.h>
114 #include <netpacket/packet.h>
115 #include <linux/netlink.h>
116 #include <linux/if_alg.h>
117 #include <linux/rtc.h>
118 #include <sound/asound.h>
119 #ifdef HAVE_BTRFS_H
120 #include <linux/btrfs.h>
121 #endif
122 #ifdef HAVE_DRM_H
123 #include <libdrm/drm.h>
124 #include <libdrm/i915_drm.h>
125 #endif
126 #include "linux_loop.h"
127 #include "uname.h"
129 #include "qemu.h"
130 #include "user-internals.h"
131 #include "strace.h"
132 #include "signal-common.h"
133 #include "loader.h"
134 #include "user-mmap.h"
135 #include "user/safe-syscall.h"
136 #include "qemu/guest-random.h"
137 #include "qemu/selfmap.h"
138 #include "user/syscall-trace.h"
139 #include "special-errno.h"
140 #include "qapi/error.h"
141 #include "fd-trans.h"
142 #include "tcg/tcg.h"
144 #ifndef CLONE_IO
145 #define CLONE_IO 0x80000000 /* Clone io context */
146 #endif
148 /* We can't directly call the host clone syscall, because this will
149 * badly confuse libc (breaking mutexes, for example). So we must
150 * divide clone flags into:
151 * * flag combinations that look like pthread_create()
152 * * flag combinations that look like fork()
153 * * flags we can implement within QEMU itself
154 * * flags we can't support and will return an error for
156 /* For thread creation, all these flags must be present; for
157 * fork, none must be present.
159 #define CLONE_THREAD_FLAGS \
160 (CLONE_VM | CLONE_FS | CLONE_FILES | \
161 CLONE_SIGHAND | CLONE_THREAD | CLONE_SYSVSEM)
163 /* These flags are ignored:
164 * CLONE_DETACHED is now ignored by the kernel;
165 * CLONE_IO is just an optimisation hint to the I/O scheduler
167 #define CLONE_IGNORED_FLAGS \
168 (CLONE_DETACHED | CLONE_IO)
170 /* Flags for fork which we can implement within QEMU itself */
171 #define CLONE_OPTIONAL_FORK_FLAGS \
172 (CLONE_SETTLS | CLONE_PARENT_SETTID | \
173 CLONE_CHILD_CLEARTID | CLONE_CHILD_SETTID)
175 /* Flags for thread creation which we can implement within QEMU itself */
176 #define CLONE_OPTIONAL_THREAD_FLAGS \
177 (CLONE_SETTLS | CLONE_PARENT_SETTID | \
178 CLONE_CHILD_CLEARTID | CLONE_CHILD_SETTID | CLONE_PARENT)
180 #define CLONE_INVALID_FORK_FLAGS \
181 (~(CSIGNAL | CLONE_OPTIONAL_FORK_FLAGS | CLONE_IGNORED_FLAGS))
183 #define CLONE_INVALID_THREAD_FLAGS \
184 (~(CSIGNAL | CLONE_THREAD_FLAGS | CLONE_OPTIONAL_THREAD_FLAGS | \
185 CLONE_IGNORED_FLAGS))
187 /* CLONE_VFORK is special cased early in do_fork(). The other flag bits
188 * have almost all been allocated. We cannot support any of
189 * CLONE_NEWNS, CLONE_NEWCGROUP, CLONE_NEWUTS, CLONE_NEWIPC,
190 * CLONE_NEWUSER, CLONE_NEWPID, CLONE_NEWNET, CLONE_PTRACE, CLONE_UNTRACED.
191 * The checks against the invalid thread masks above will catch these.
192 * (The one remaining unallocated bit is 0x1000 which used to be CLONE_PID.)
195 /* Define DEBUG_ERESTARTSYS to force every syscall to be restarted
196 * once. This exercises the codepaths for restart.
198 //#define DEBUG_ERESTARTSYS
200 //#include <linux/msdos_fs.h>
201 #define VFAT_IOCTL_READDIR_BOTH \
202 _IOC(_IOC_READ, 'r', 1, (sizeof(struct linux_dirent) + 256) * 2)
203 #define VFAT_IOCTL_READDIR_SHORT \
204 _IOC(_IOC_READ, 'r', 2, (sizeof(struct linux_dirent) + 256) * 2)
206 #undef _syscall0
207 #undef _syscall1
208 #undef _syscall2
209 #undef _syscall3
210 #undef _syscall4
211 #undef _syscall5
212 #undef _syscall6
214 #define _syscall0(type,name) \
215 static type name (void) \
217 return syscall(__NR_##name); \
220 #define _syscall1(type,name,type1,arg1) \
221 static type name (type1 arg1) \
223 return syscall(__NR_##name, arg1); \
226 #define _syscall2(type,name,type1,arg1,type2,arg2) \
227 static type name (type1 arg1,type2 arg2) \
229 return syscall(__NR_##name, arg1, arg2); \
232 #define _syscall3(type,name,type1,arg1,type2,arg2,type3,arg3) \
233 static type name (type1 arg1,type2 arg2,type3 arg3) \
235 return syscall(__NR_##name, arg1, arg2, arg3); \
238 #define _syscall4(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4) \
239 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4) \
241 return syscall(__NR_##name, arg1, arg2, arg3, arg4); \
244 #define _syscall5(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
245 type5,arg5) \
246 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4,type5 arg5) \
248 return syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5); \
252 #define _syscall6(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
253 type5,arg5,type6,arg6) \
254 static type name (type1 arg1,type2 arg2,type3 arg3,type4 arg4,type5 arg5, \
255 type6 arg6) \
257 return syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5, arg6); \
261 #define __NR_sys_uname __NR_uname
262 #define __NR_sys_getcwd1 __NR_getcwd
263 #define __NR_sys_getdents __NR_getdents
264 #define __NR_sys_getdents64 __NR_getdents64
265 #define __NR_sys_getpriority __NR_getpriority
266 #define __NR_sys_rt_sigqueueinfo __NR_rt_sigqueueinfo
267 #define __NR_sys_rt_tgsigqueueinfo __NR_rt_tgsigqueueinfo
268 #define __NR_sys_syslog __NR_syslog
269 #if defined(__NR_futex)
270 # define __NR_sys_futex __NR_futex
271 #endif
272 #if defined(__NR_futex_time64)
273 # define __NR_sys_futex_time64 __NR_futex_time64
274 #endif
275 #define __NR_sys_statx __NR_statx
277 #if defined(__alpha__) || defined(__x86_64__) || defined(__s390x__)
278 #define __NR__llseek __NR_lseek
279 #endif
281 /* Newer kernel ports have llseek() instead of _llseek() */
282 #if defined(TARGET_NR_llseek) && !defined(TARGET_NR__llseek)
283 #define TARGET_NR__llseek TARGET_NR_llseek
284 #endif
286 /* some platforms need to mask more bits than just TARGET_O_NONBLOCK */
287 #ifndef TARGET_O_NONBLOCK_MASK
288 #define TARGET_O_NONBLOCK_MASK TARGET_O_NONBLOCK
289 #endif
291 #define __NR_sys_gettid __NR_gettid
292 _syscall0(int, sys_gettid)
294 /* For the 64-bit guest on 32-bit host case we must emulate
295 * getdents using getdents64, because otherwise the host
296 * might hand us back more dirent records than we can fit
297 * into the guest buffer after structure format conversion.
298 * Otherwise we emulate getdents with getdents if the host has it.
300 #if defined(__NR_getdents) && HOST_LONG_BITS >= TARGET_ABI_BITS
301 #define EMULATE_GETDENTS_WITH_GETDENTS
302 #endif
304 #if defined(TARGET_NR_getdents) && defined(EMULATE_GETDENTS_WITH_GETDENTS)
305 _syscall3(int, sys_getdents, uint, fd, struct linux_dirent *, dirp, uint, count);
306 #endif
307 #if (defined(TARGET_NR_getdents) && \
308 !defined(EMULATE_GETDENTS_WITH_GETDENTS)) || \
309 (defined(TARGET_NR_getdents64) && defined(__NR_getdents64))
310 _syscall3(int, sys_getdents64, uint, fd, struct linux_dirent64 *, dirp, uint, count);
311 #endif
312 #if defined(TARGET_NR__llseek) && defined(__NR_llseek)
313 _syscall5(int, _llseek, uint, fd, ulong, hi, ulong, lo,
314 loff_t *, res, uint, wh);
315 #endif
316 _syscall3(int, sys_rt_sigqueueinfo, pid_t, pid, int, sig, siginfo_t *, uinfo)
317 _syscall4(int, sys_rt_tgsigqueueinfo, pid_t, pid, pid_t, tid, int, sig,
318 siginfo_t *, uinfo)
319 _syscall3(int,sys_syslog,int,type,char*,bufp,int,len)
320 #ifdef __NR_exit_group
321 _syscall1(int,exit_group,int,error_code)
322 #endif
323 #if defined(__NR_futex)
324 _syscall6(int,sys_futex,int *,uaddr,int,op,int,val,
325 const struct timespec *,timeout,int *,uaddr2,int,val3)
326 #endif
327 #if defined(__NR_futex_time64)
328 _syscall6(int,sys_futex_time64,int *,uaddr,int,op,int,val,
329 const struct timespec *,timeout,int *,uaddr2,int,val3)
330 #endif
331 #define __NR_sys_sched_getaffinity __NR_sched_getaffinity
332 _syscall3(int, sys_sched_getaffinity, pid_t, pid, unsigned int, len,
333 unsigned long *, user_mask_ptr);
334 #define __NR_sys_sched_setaffinity __NR_sched_setaffinity
335 _syscall3(int, sys_sched_setaffinity, pid_t, pid, unsigned int, len,
336 unsigned long *, user_mask_ptr);
337 /* sched_attr is not defined in glibc */
338 struct sched_attr {
339 uint32_t size;
340 uint32_t sched_policy;
341 uint64_t sched_flags;
342 int32_t sched_nice;
343 uint32_t sched_priority;
344 uint64_t sched_runtime;
345 uint64_t sched_deadline;
346 uint64_t sched_period;
347 uint32_t sched_util_min;
348 uint32_t sched_util_max;
350 #define __NR_sys_sched_getattr __NR_sched_getattr
351 _syscall4(int, sys_sched_getattr, pid_t, pid, struct sched_attr *, attr,
352 unsigned int, size, unsigned int, flags);
353 #define __NR_sys_sched_setattr __NR_sched_setattr
354 _syscall3(int, sys_sched_setattr, pid_t, pid, struct sched_attr *, attr,
355 unsigned int, flags);
356 #define __NR_sys_sched_getscheduler __NR_sched_getscheduler
357 _syscall1(int, sys_sched_getscheduler, pid_t, pid);
358 #define __NR_sys_sched_setscheduler __NR_sched_setscheduler
359 _syscall3(int, sys_sched_setscheduler, pid_t, pid, int, policy,
360 const struct sched_param *, param);
361 #define __NR_sys_sched_getparam __NR_sched_getparam
362 _syscall2(int, sys_sched_getparam, pid_t, pid,
363 struct sched_param *, param);
364 #define __NR_sys_sched_setparam __NR_sched_setparam
365 _syscall2(int, sys_sched_setparam, pid_t, pid,
366 const struct sched_param *, param);
367 #define __NR_sys_getcpu __NR_getcpu
368 _syscall3(int, sys_getcpu, unsigned *, cpu, unsigned *, node, void *, tcache);
369 _syscall4(int, reboot, int, magic1, int, magic2, unsigned int, cmd,
370 void *, arg);
371 _syscall2(int, capget, struct __user_cap_header_struct *, header,
372 struct __user_cap_data_struct *, data);
373 _syscall2(int, capset, struct __user_cap_header_struct *, header,
374 struct __user_cap_data_struct *, data);
375 #if defined(TARGET_NR_ioprio_get) && defined(__NR_ioprio_get)
376 _syscall2(int, ioprio_get, int, which, int, who)
377 #endif
378 #if defined(TARGET_NR_ioprio_set) && defined(__NR_ioprio_set)
379 _syscall3(int, ioprio_set, int, which, int, who, int, ioprio)
380 #endif
381 #if defined(TARGET_NR_getrandom) && defined(__NR_getrandom)
382 _syscall3(int, getrandom, void *, buf, size_t, buflen, unsigned int, flags)
383 #endif
385 #if defined(TARGET_NR_kcmp) && defined(__NR_kcmp)
386 _syscall5(int, kcmp, pid_t, pid1, pid_t, pid2, int, type,
387 unsigned long, idx1, unsigned long, idx2)
388 #endif
391 * It is assumed that struct statx is architecture independent.
393 #if defined(TARGET_NR_statx) && defined(__NR_statx)
394 _syscall5(int, sys_statx, int, dirfd, const char *, pathname, int, flags,
395 unsigned int, mask, struct target_statx *, statxbuf)
396 #endif
397 #if defined(TARGET_NR_membarrier) && defined(__NR_membarrier)
398 _syscall2(int, membarrier, int, cmd, int, flags)
399 #endif
401 static const bitmask_transtbl fcntl_flags_tbl[] = {
402 { TARGET_O_ACCMODE, TARGET_O_WRONLY, O_ACCMODE, O_WRONLY, },
403 { TARGET_O_ACCMODE, TARGET_O_RDWR, O_ACCMODE, O_RDWR, },
404 { TARGET_O_CREAT, TARGET_O_CREAT, O_CREAT, O_CREAT, },
405 { TARGET_O_EXCL, TARGET_O_EXCL, O_EXCL, O_EXCL, },
406 { TARGET_O_NOCTTY, TARGET_O_NOCTTY, O_NOCTTY, O_NOCTTY, },
407 { TARGET_O_TRUNC, TARGET_O_TRUNC, O_TRUNC, O_TRUNC, },
408 { TARGET_O_APPEND, TARGET_O_APPEND, O_APPEND, O_APPEND, },
409 { TARGET_O_NONBLOCK, TARGET_O_NONBLOCK, O_NONBLOCK, O_NONBLOCK, },
410 { TARGET_O_SYNC, TARGET_O_DSYNC, O_SYNC, O_DSYNC, },
411 { TARGET_O_SYNC, TARGET_O_SYNC, O_SYNC, O_SYNC, },
412 { TARGET_FASYNC, TARGET_FASYNC, FASYNC, FASYNC, },
413 { TARGET_O_DIRECTORY, TARGET_O_DIRECTORY, O_DIRECTORY, O_DIRECTORY, },
414 { TARGET_O_NOFOLLOW, TARGET_O_NOFOLLOW, O_NOFOLLOW, O_NOFOLLOW, },
415 #if defined(O_DIRECT)
416 { TARGET_O_DIRECT, TARGET_O_DIRECT, O_DIRECT, O_DIRECT, },
417 #endif
418 #if defined(O_NOATIME)
419 { TARGET_O_NOATIME, TARGET_O_NOATIME, O_NOATIME, O_NOATIME },
420 #endif
421 #if defined(O_CLOEXEC)
422 { TARGET_O_CLOEXEC, TARGET_O_CLOEXEC, O_CLOEXEC, O_CLOEXEC },
423 #endif
424 #if defined(O_PATH)
425 { TARGET_O_PATH, TARGET_O_PATH, O_PATH, O_PATH },
426 #endif
427 #if defined(O_TMPFILE)
428 { TARGET_O_TMPFILE, TARGET_O_TMPFILE, O_TMPFILE, O_TMPFILE },
429 #endif
430 /* Don't terminate the list prematurely on 64-bit host+guest. */
431 #if TARGET_O_LARGEFILE != 0 || O_LARGEFILE != 0
432 { TARGET_O_LARGEFILE, TARGET_O_LARGEFILE, O_LARGEFILE, O_LARGEFILE, },
433 #endif
434 { 0, 0, 0, 0 }
437 _syscall2(int, sys_getcwd1, char *, buf, size_t, size)
439 #if defined(TARGET_NR_utimensat) || defined(TARGET_NR_utimensat_time64)
440 #if defined(__NR_utimensat)
441 #define __NR_sys_utimensat __NR_utimensat
442 _syscall4(int,sys_utimensat,int,dirfd,const char *,pathname,
443 const struct timespec *,tsp,int,flags)
444 #else
445 static int sys_utimensat(int dirfd, const char *pathname,
446 const struct timespec times[2], int flags)
448 errno = ENOSYS;
449 return -1;
451 #endif
452 #endif /* TARGET_NR_utimensat */
454 #ifdef TARGET_NR_renameat2
455 #if defined(__NR_renameat2)
456 #define __NR_sys_renameat2 __NR_renameat2
457 _syscall5(int, sys_renameat2, int, oldfd, const char *, old, int, newfd,
458 const char *, new, unsigned int, flags)
459 #else
460 static int sys_renameat2(int oldfd, const char *old,
461 int newfd, const char *new, int flags)
463 if (flags == 0) {
464 return renameat(oldfd, old, newfd, new);
466 errno = ENOSYS;
467 return -1;
469 #endif
470 #endif /* TARGET_NR_renameat2 */
472 #ifdef CONFIG_INOTIFY
473 #include <sys/inotify.h>
474 #else
475 /* Userspace can usually survive runtime without inotify */
476 #undef TARGET_NR_inotify_init
477 #undef TARGET_NR_inotify_init1
478 #undef TARGET_NR_inotify_add_watch
479 #undef TARGET_NR_inotify_rm_watch
480 #endif /* CONFIG_INOTIFY */
482 #if defined(TARGET_NR_prlimit64)
483 #ifndef __NR_prlimit64
484 # define __NR_prlimit64 -1
485 #endif
486 #define __NR_sys_prlimit64 __NR_prlimit64
487 /* The glibc rlimit structure may not be that used by the underlying syscall */
488 struct host_rlimit64 {
489 uint64_t rlim_cur;
490 uint64_t rlim_max;
492 _syscall4(int, sys_prlimit64, pid_t, pid, int, resource,
493 const struct host_rlimit64 *, new_limit,
494 struct host_rlimit64 *, old_limit)
495 #endif
498 #if defined(TARGET_NR_timer_create)
499 /* Maximum of 32 active POSIX timers allowed at any one time. */
500 static timer_t g_posix_timers[32] = { 0, } ;
502 static inline int next_free_host_timer(void)
504 int k ;
505 /* FIXME: Does finding the next free slot require a lock? */
506 for (k = 0; k < ARRAY_SIZE(g_posix_timers); k++) {
507 if (g_posix_timers[k] == 0) {
508 g_posix_timers[k] = (timer_t) 1;
509 return k;
512 return -1;
514 #endif
516 static inline int host_to_target_errno(int host_errno)
518 switch (host_errno) {
519 #define E(X) case X: return TARGET_##X;
520 #include "errnos.c.inc"
521 #undef E
522 default:
523 return host_errno;
527 static inline int target_to_host_errno(int target_errno)
529 switch (target_errno) {
530 #define E(X) case TARGET_##X: return X;
531 #include "errnos.c.inc"
532 #undef E
533 default:
534 return target_errno;
538 abi_long get_errno(abi_long ret)
540 if (ret == -1)
541 return -host_to_target_errno(errno);
542 else
543 return ret;
546 const char *target_strerror(int err)
548 if (err == QEMU_ERESTARTSYS) {
549 return "To be restarted";
551 if (err == QEMU_ESIGRETURN) {
552 return "Successful exit from sigreturn";
555 return strerror(target_to_host_errno(err));
558 static int check_zeroed_user(abi_long addr, size_t ksize, size_t usize)
560 int i;
561 uint8_t b;
562 if (usize <= ksize) {
563 return 1;
565 for (i = ksize; i < usize; i++) {
566 if (get_user_u8(b, addr + i)) {
567 return -TARGET_EFAULT;
569 if (b != 0) {
570 return 0;
573 return 1;
576 #define safe_syscall0(type, name) \
577 static type safe_##name(void) \
579 return safe_syscall(__NR_##name); \
582 #define safe_syscall1(type, name, type1, arg1) \
583 static type safe_##name(type1 arg1) \
585 return safe_syscall(__NR_##name, arg1); \
588 #define safe_syscall2(type, name, type1, arg1, type2, arg2) \
589 static type safe_##name(type1 arg1, type2 arg2) \
591 return safe_syscall(__NR_##name, arg1, arg2); \
594 #define safe_syscall3(type, name, type1, arg1, type2, arg2, type3, arg3) \
595 static type safe_##name(type1 arg1, type2 arg2, type3 arg3) \
597 return safe_syscall(__NR_##name, arg1, arg2, arg3); \
600 #define safe_syscall4(type, name, type1, arg1, type2, arg2, type3, arg3, \
601 type4, arg4) \
602 static type safe_##name(type1 arg1, type2 arg2, type3 arg3, type4 arg4) \
604 return safe_syscall(__NR_##name, arg1, arg2, arg3, arg4); \
607 #define safe_syscall5(type, name, type1, arg1, type2, arg2, type3, arg3, \
608 type4, arg4, type5, arg5) \
609 static type safe_##name(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
610 type5 arg5) \
612 return safe_syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5); \
615 #define safe_syscall6(type, name, type1, arg1, type2, arg2, type3, arg3, \
616 type4, arg4, type5, arg5, type6, arg6) \
617 static type safe_##name(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
618 type5 arg5, type6 arg6) \
620 return safe_syscall(__NR_##name, arg1, arg2, arg3, arg4, arg5, arg6); \
623 safe_syscall3(ssize_t, read, int, fd, void *, buff, size_t, count)
624 safe_syscall3(ssize_t, write, int, fd, const void *, buff, size_t, count)
625 safe_syscall4(int, openat, int, dirfd, const char *, pathname, \
626 int, flags, mode_t, mode)
627 #if defined(TARGET_NR_wait4) || defined(TARGET_NR_waitpid)
628 safe_syscall4(pid_t, wait4, pid_t, pid, int *, status, int, options, \
629 struct rusage *, rusage)
630 #endif
631 safe_syscall5(int, waitid, idtype_t, idtype, id_t, id, siginfo_t *, infop, \
632 int, options, struct rusage *, rusage)
633 safe_syscall3(int, execve, const char *, filename, char **, argv, char **, envp)
634 #if defined(TARGET_NR_select) || defined(TARGET_NR__newselect) || \
635 defined(TARGET_NR_pselect6) || defined(TARGET_NR_pselect6_time64)
636 safe_syscall6(int, pselect6, int, nfds, fd_set *, readfds, fd_set *, writefds, \
637 fd_set *, exceptfds, struct timespec *, timeout, void *, sig)
638 #endif
639 #if defined(TARGET_NR_ppoll) || defined(TARGET_NR_ppoll_time64)
640 safe_syscall5(int, ppoll, struct pollfd *, ufds, unsigned int, nfds,
641 struct timespec *, tsp, const sigset_t *, sigmask,
642 size_t, sigsetsize)
643 #endif
644 safe_syscall6(int, epoll_pwait, int, epfd, struct epoll_event *, events,
645 int, maxevents, int, timeout, const sigset_t *, sigmask,
646 size_t, sigsetsize)
647 #if defined(__NR_futex)
648 safe_syscall6(int,futex,int *,uaddr,int,op,int,val, \
649 const struct timespec *,timeout,int *,uaddr2,int,val3)
650 #endif
651 #if defined(__NR_futex_time64)
652 safe_syscall6(int,futex_time64,int *,uaddr,int,op,int,val, \
653 const struct timespec *,timeout,int *,uaddr2,int,val3)
654 #endif
655 safe_syscall2(int, rt_sigsuspend, sigset_t *, newset, size_t, sigsetsize)
656 safe_syscall2(int, kill, pid_t, pid, int, sig)
657 safe_syscall2(int, tkill, int, tid, int, sig)
658 safe_syscall3(int, tgkill, int, tgid, int, pid, int, sig)
659 safe_syscall3(ssize_t, readv, int, fd, const struct iovec *, iov, int, iovcnt)
660 safe_syscall3(ssize_t, writev, int, fd, const struct iovec *, iov, int, iovcnt)
661 safe_syscall5(ssize_t, preadv, int, fd, const struct iovec *, iov, int, iovcnt,
662 unsigned long, pos_l, unsigned long, pos_h)
663 safe_syscall5(ssize_t, pwritev, int, fd, const struct iovec *, iov, int, iovcnt,
664 unsigned long, pos_l, unsigned long, pos_h)
665 safe_syscall3(int, connect, int, fd, const struct sockaddr *, addr,
666 socklen_t, addrlen)
667 safe_syscall6(ssize_t, sendto, int, fd, const void *, buf, size_t, len,
668 int, flags, const struct sockaddr *, addr, socklen_t, addrlen)
669 safe_syscall6(ssize_t, recvfrom, int, fd, void *, buf, size_t, len,
670 int, flags, struct sockaddr *, addr, socklen_t *, addrlen)
671 safe_syscall3(ssize_t, sendmsg, int, fd, const struct msghdr *, msg, int, flags)
672 safe_syscall3(ssize_t, recvmsg, int, fd, struct msghdr *, msg, int, flags)
673 safe_syscall2(int, flock, int, fd, int, operation)
674 #if defined(TARGET_NR_rt_sigtimedwait) || defined(TARGET_NR_rt_sigtimedwait_time64)
675 safe_syscall4(int, rt_sigtimedwait, const sigset_t *, these, siginfo_t *, uinfo,
676 const struct timespec *, uts, size_t, sigsetsize)
677 #endif
678 safe_syscall4(int, accept4, int, fd, struct sockaddr *, addr, socklen_t *, len,
679 int, flags)
680 #if defined(TARGET_NR_nanosleep)
681 safe_syscall2(int, nanosleep, const struct timespec *, req,
682 struct timespec *, rem)
683 #endif
684 #if defined(TARGET_NR_clock_nanosleep) || \
685 defined(TARGET_NR_clock_nanosleep_time64)
686 safe_syscall4(int, clock_nanosleep, const clockid_t, clock, int, flags,
687 const struct timespec *, req, struct timespec *, rem)
688 #endif
689 #ifdef __NR_ipc
690 #ifdef __s390x__
691 safe_syscall5(int, ipc, int, call, long, first, long, second, long, third,
692 void *, ptr)
693 #else
694 safe_syscall6(int, ipc, int, call, long, first, long, second, long, third,
695 void *, ptr, long, fifth)
696 #endif
697 #endif
698 #ifdef __NR_msgsnd
699 safe_syscall4(int, msgsnd, int, msgid, const void *, msgp, size_t, sz,
700 int, flags)
701 #endif
702 #ifdef __NR_msgrcv
703 safe_syscall5(int, msgrcv, int, msgid, void *, msgp, size_t, sz,
704 long, msgtype, int, flags)
705 #endif
706 #ifdef __NR_semtimedop
707 safe_syscall4(int, semtimedop, int, semid, struct sembuf *, tsops,
708 unsigned, nsops, const struct timespec *, timeout)
709 #endif
710 #if defined(TARGET_NR_mq_timedsend) || \
711 defined(TARGET_NR_mq_timedsend_time64)
712 safe_syscall5(int, mq_timedsend, int, mqdes, const char *, msg_ptr,
713 size_t, len, unsigned, prio, const struct timespec *, timeout)
714 #endif
715 #if defined(TARGET_NR_mq_timedreceive) || \
716 defined(TARGET_NR_mq_timedreceive_time64)
717 safe_syscall5(int, mq_timedreceive, int, mqdes, char *, msg_ptr,
718 size_t, len, unsigned *, prio, const struct timespec *, timeout)
719 #endif
720 #if defined(TARGET_NR_copy_file_range) && defined(__NR_copy_file_range)
721 safe_syscall6(ssize_t, copy_file_range, int, infd, loff_t *, pinoff,
722 int, outfd, loff_t *, poutoff, size_t, length,
723 unsigned int, flags)
724 #endif
726 /* We do ioctl like this rather than via safe_syscall3 to preserve the
727 * "third argument might be integer or pointer or not present" behaviour of
728 * the libc function.
730 #define safe_ioctl(...) safe_syscall(__NR_ioctl, __VA_ARGS__)
731 /* Similarly for fcntl. Note that callers must always:
732 * pass the F_GETLK64 etc constants rather than the unsuffixed F_GETLK
733 * use the flock64 struct rather than unsuffixed flock
734 * This will then work and use a 64-bit offset for both 32-bit and 64-bit hosts.
736 #ifdef __NR_fcntl64
737 #define safe_fcntl(...) safe_syscall(__NR_fcntl64, __VA_ARGS__)
738 #else
739 #define safe_fcntl(...) safe_syscall(__NR_fcntl, __VA_ARGS__)
740 #endif
742 static inline int host_to_target_sock_type(int host_type)
744 int target_type;
746 switch (host_type & 0xf /* SOCK_TYPE_MASK */) {
747 case SOCK_DGRAM:
748 target_type = TARGET_SOCK_DGRAM;
749 break;
750 case SOCK_STREAM:
751 target_type = TARGET_SOCK_STREAM;
752 break;
753 default:
754 target_type = host_type & 0xf /* SOCK_TYPE_MASK */;
755 break;
758 #if defined(SOCK_CLOEXEC)
759 if (host_type & SOCK_CLOEXEC) {
760 target_type |= TARGET_SOCK_CLOEXEC;
762 #endif
764 #if defined(SOCK_NONBLOCK)
765 if (host_type & SOCK_NONBLOCK) {
766 target_type |= TARGET_SOCK_NONBLOCK;
768 #endif
770 return target_type;
773 static abi_ulong target_brk;
774 static abi_ulong target_original_brk;
775 static abi_ulong brk_page;
777 void target_set_brk(abi_ulong new_brk)
779 target_original_brk = target_brk = HOST_PAGE_ALIGN(new_brk);
780 brk_page = HOST_PAGE_ALIGN(target_brk);
783 //#define DEBUGF_BRK(message, args...) do { fprintf(stderr, (message), ## args); } while (0)
784 #define DEBUGF_BRK(message, args...)
786 /* do_brk() must return target values and target errnos. */
787 abi_long do_brk(abi_ulong new_brk)
789 abi_long mapped_addr;
790 abi_ulong new_alloc_size;
792 /* brk pointers are always untagged */
794 DEBUGF_BRK("do_brk(" TARGET_ABI_FMT_lx ") -> ", new_brk);
796 if (!new_brk) {
797 DEBUGF_BRK(TARGET_ABI_FMT_lx " (!new_brk)\n", target_brk);
798 return target_brk;
800 if (new_brk < target_original_brk) {
801 DEBUGF_BRK(TARGET_ABI_FMT_lx " (new_brk < target_original_brk)\n",
802 target_brk);
803 return target_brk;
806 /* If the new brk is less than the highest page reserved to the
807 * target heap allocation, set it and we're almost done... */
808 if (new_brk <= brk_page) {
809 /* Heap contents are initialized to zero, as for anonymous
810 * mapped pages. */
811 if (new_brk > target_brk) {
812 memset(g2h_untagged(target_brk), 0, new_brk - target_brk);
814 target_brk = new_brk;
815 DEBUGF_BRK(TARGET_ABI_FMT_lx " (new_brk <= brk_page)\n", target_brk);
816 return target_brk;
819 /* We need to allocate more memory after the brk... Note that
820 * we don't use MAP_FIXED because that will map over the top of
821 * any existing mapping (like the one with the host libc or qemu
822 * itself); instead we treat "mapped but at wrong address" as
823 * a failure and unmap again.
825 new_alloc_size = HOST_PAGE_ALIGN(new_brk - brk_page);
826 mapped_addr = get_errno(target_mmap(brk_page, new_alloc_size,
827 PROT_READ|PROT_WRITE,
828 MAP_ANON|MAP_PRIVATE, 0, 0));
830 if (mapped_addr == brk_page) {
831 /* Heap contents are initialized to zero, as for anonymous
832 * mapped pages. Technically the new pages are already
833 * initialized to zero since they *are* anonymous mapped
834 * pages, however we have to take care with the contents that
835 * come from the remaining part of the previous page: it may
836 * contains garbage data due to a previous heap usage (grown
837 * then shrunken). */
838 memset(g2h_untagged(target_brk), 0, brk_page - target_brk);
840 target_brk = new_brk;
841 brk_page = HOST_PAGE_ALIGN(target_brk);
842 DEBUGF_BRK(TARGET_ABI_FMT_lx " (mapped_addr == brk_page)\n",
843 target_brk);
844 return target_brk;
845 } else if (mapped_addr != -1) {
846 /* Mapped but at wrong address, meaning there wasn't actually
847 * enough space for this brk.
849 target_munmap(mapped_addr, new_alloc_size);
850 mapped_addr = -1;
851 DEBUGF_BRK(TARGET_ABI_FMT_lx " (mapped_addr != -1)\n", target_brk);
853 else {
854 DEBUGF_BRK(TARGET_ABI_FMT_lx " (otherwise)\n", target_brk);
857 #if defined(TARGET_ALPHA)
858 /* We (partially) emulate OSF/1 on Alpha, which requires we
859 return a proper errno, not an unchanged brk value. */
860 return -TARGET_ENOMEM;
861 #endif
862 /* For everything else, return the previous break. */
863 return target_brk;
866 #if defined(TARGET_NR_select) || defined(TARGET_NR__newselect) || \
867 defined(TARGET_NR_pselect6) || defined(TARGET_NR_pselect6_time64)
868 static inline abi_long copy_from_user_fdset(fd_set *fds,
869 abi_ulong target_fds_addr,
870 int n)
872 int i, nw, j, k;
873 abi_ulong b, *target_fds;
875 nw = DIV_ROUND_UP(n, TARGET_ABI_BITS);
876 if (!(target_fds = lock_user(VERIFY_READ,
877 target_fds_addr,
878 sizeof(abi_ulong) * nw,
879 1)))
880 return -TARGET_EFAULT;
882 FD_ZERO(fds);
883 k = 0;
884 for (i = 0; i < nw; i++) {
885 /* grab the abi_ulong */
886 __get_user(b, &target_fds[i]);
887 for (j = 0; j < TARGET_ABI_BITS; j++) {
888 /* check the bit inside the abi_ulong */
889 if ((b >> j) & 1)
890 FD_SET(k, fds);
891 k++;
895 unlock_user(target_fds, target_fds_addr, 0);
897 return 0;
900 static inline abi_ulong copy_from_user_fdset_ptr(fd_set *fds, fd_set **fds_ptr,
901 abi_ulong target_fds_addr,
902 int n)
904 if (target_fds_addr) {
905 if (copy_from_user_fdset(fds, target_fds_addr, n))
906 return -TARGET_EFAULT;
907 *fds_ptr = fds;
908 } else {
909 *fds_ptr = NULL;
911 return 0;
914 static inline abi_long copy_to_user_fdset(abi_ulong target_fds_addr,
915 const fd_set *fds,
916 int n)
918 int i, nw, j, k;
919 abi_long v;
920 abi_ulong *target_fds;
922 nw = DIV_ROUND_UP(n, TARGET_ABI_BITS);
923 if (!(target_fds = lock_user(VERIFY_WRITE,
924 target_fds_addr,
925 sizeof(abi_ulong) * nw,
926 0)))
927 return -TARGET_EFAULT;
929 k = 0;
930 for (i = 0; i < nw; i++) {
931 v = 0;
932 for (j = 0; j < TARGET_ABI_BITS; j++) {
933 v |= ((abi_ulong)(FD_ISSET(k, fds) != 0) << j);
934 k++;
936 __put_user(v, &target_fds[i]);
939 unlock_user(target_fds, target_fds_addr, sizeof(abi_ulong) * nw);
941 return 0;
943 #endif
945 #if defined(__alpha__)
946 #define HOST_HZ 1024
947 #else
948 #define HOST_HZ 100
949 #endif
951 static inline abi_long host_to_target_clock_t(long ticks)
953 #if HOST_HZ == TARGET_HZ
954 return ticks;
955 #else
956 return ((int64_t)ticks * TARGET_HZ) / HOST_HZ;
957 #endif
960 static inline abi_long host_to_target_rusage(abi_ulong target_addr,
961 const struct rusage *rusage)
963 struct target_rusage *target_rusage;
965 if (!lock_user_struct(VERIFY_WRITE, target_rusage, target_addr, 0))
966 return -TARGET_EFAULT;
967 target_rusage->ru_utime.tv_sec = tswapal(rusage->ru_utime.tv_sec);
968 target_rusage->ru_utime.tv_usec = tswapal(rusage->ru_utime.tv_usec);
969 target_rusage->ru_stime.tv_sec = tswapal(rusage->ru_stime.tv_sec);
970 target_rusage->ru_stime.tv_usec = tswapal(rusage->ru_stime.tv_usec);
971 target_rusage->ru_maxrss = tswapal(rusage->ru_maxrss);
972 target_rusage->ru_ixrss = tswapal(rusage->ru_ixrss);
973 target_rusage->ru_idrss = tswapal(rusage->ru_idrss);
974 target_rusage->ru_isrss = tswapal(rusage->ru_isrss);
975 target_rusage->ru_minflt = tswapal(rusage->ru_minflt);
976 target_rusage->ru_majflt = tswapal(rusage->ru_majflt);
977 target_rusage->ru_nswap = tswapal(rusage->ru_nswap);
978 target_rusage->ru_inblock = tswapal(rusage->ru_inblock);
979 target_rusage->ru_oublock = tswapal(rusage->ru_oublock);
980 target_rusage->ru_msgsnd = tswapal(rusage->ru_msgsnd);
981 target_rusage->ru_msgrcv = tswapal(rusage->ru_msgrcv);
982 target_rusage->ru_nsignals = tswapal(rusage->ru_nsignals);
983 target_rusage->ru_nvcsw = tswapal(rusage->ru_nvcsw);
984 target_rusage->ru_nivcsw = tswapal(rusage->ru_nivcsw);
985 unlock_user_struct(target_rusage, target_addr, 1);
987 return 0;
990 #ifdef TARGET_NR_setrlimit
991 static inline rlim_t target_to_host_rlim(abi_ulong target_rlim)
993 abi_ulong target_rlim_swap;
994 rlim_t result;
996 target_rlim_swap = tswapal(target_rlim);
997 if (target_rlim_swap == TARGET_RLIM_INFINITY)
998 return RLIM_INFINITY;
1000 result = target_rlim_swap;
1001 if (target_rlim_swap != (rlim_t)result)
1002 return RLIM_INFINITY;
1004 return result;
1006 #endif
1008 #if defined(TARGET_NR_getrlimit) || defined(TARGET_NR_ugetrlimit)
1009 static inline abi_ulong host_to_target_rlim(rlim_t rlim)
1011 abi_ulong target_rlim_swap;
1012 abi_ulong result;
1014 if (rlim == RLIM_INFINITY || rlim != (abi_long)rlim)
1015 target_rlim_swap = TARGET_RLIM_INFINITY;
1016 else
1017 target_rlim_swap = rlim;
1018 result = tswapal(target_rlim_swap);
1020 return result;
1022 #endif
1024 static inline int target_to_host_resource(int code)
1026 switch (code) {
1027 case TARGET_RLIMIT_AS:
1028 return RLIMIT_AS;
1029 case TARGET_RLIMIT_CORE:
1030 return RLIMIT_CORE;
1031 case TARGET_RLIMIT_CPU:
1032 return RLIMIT_CPU;
1033 case TARGET_RLIMIT_DATA:
1034 return RLIMIT_DATA;
1035 case TARGET_RLIMIT_FSIZE:
1036 return RLIMIT_FSIZE;
1037 case TARGET_RLIMIT_LOCKS:
1038 return RLIMIT_LOCKS;
1039 case TARGET_RLIMIT_MEMLOCK:
1040 return RLIMIT_MEMLOCK;
1041 case TARGET_RLIMIT_MSGQUEUE:
1042 return RLIMIT_MSGQUEUE;
1043 case TARGET_RLIMIT_NICE:
1044 return RLIMIT_NICE;
1045 case TARGET_RLIMIT_NOFILE:
1046 return RLIMIT_NOFILE;
1047 case TARGET_RLIMIT_NPROC:
1048 return RLIMIT_NPROC;
1049 case TARGET_RLIMIT_RSS:
1050 return RLIMIT_RSS;
1051 case TARGET_RLIMIT_RTPRIO:
1052 return RLIMIT_RTPRIO;
1053 #ifdef RLIMIT_RTTIME
1054 case TARGET_RLIMIT_RTTIME:
1055 return RLIMIT_RTTIME;
1056 #endif
1057 case TARGET_RLIMIT_SIGPENDING:
1058 return RLIMIT_SIGPENDING;
1059 case TARGET_RLIMIT_STACK:
1060 return RLIMIT_STACK;
1061 default:
1062 return code;
1066 static inline abi_long copy_from_user_timeval(struct timeval *tv,
1067 abi_ulong target_tv_addr)
1069 struct target_timeval *target_tv;
1071 if (!lock_user_struct(VERIFY_READ, target_tv, target_tv_addr, 1)) {
1072 return -TARGET_EFAULT;
1075 __get_user(tv->tv_sec, &target_tv->tv_sec);
1076 __get_user(tv->tv_usec, &target_tv->tv_usec);
1078 unlock_user_struct(target_tv, target_tv_addr, 0);
1080 return 0;
1083 static inline abi_long copy_to_user_timeval(abi_ulong target_tv_addr,
1084 const struct timeval *tv)
1086 struct target_timeval *target_tv;
1088 if (!lock_user_struct(VERIFY_WRITE, target_tv, target_tv_addr, 0)) {
1089 return -TARGET_EFAULT;
1092 __put_user(tv->tv_sec, &target_tv->tv_sec);
1093 __put_user(tv->tv_usec, &target_tv->tv_usec);
1095 unlock_user_struct(target_tv, target_tv_addr, 1);
1097 return 0;
1100 #if defined(TARGET_NR_clock_adjtime64) && defined(CONFIG_CLOCK_ADJTIME)
1101 static inline abi_long copy_from_user_timeval64(struct timeval *tv,
1102 abi_ulong target_tv_addr)
1104 struct target__kernel_sock_timeval *target_tv;
1106 if (!lock_user_struct(VERIFY_READ, target_tv, target_tv_addr, 1)) {
1107 return -TARGET_EFAULT;
1110 __get_user(tv->tv_sec, &target_tv->tv_sec);
1111 __get_user(tv->tv_usec, &target_tv->tv_usec);
1113 unlock_user_struct(target_tv, target_tv_addr, 0);
1115 return 0;
1117 #endif
1119 static inline abi_long copy_to_user_timeval64(abi_ulong target_tv_addr,
1120 const struct timeval *tv)
1122 struct target__kernel_sock_timeval *target_tv;
1124 if (!lock_user_struct(VERIFY_WRITE, target_tv, target_tv_addr, 0)) {
1125 return -TARGET_EFAULT;
1128 __put_user(tv->tv_sec, &target_tv->tv_sec);
1129 __put_user(tv->tv_usec, &target_tv->tv_usec);
1131 unlock_user_struct(target_tv, target_tv_addr, 1);
1133 return 0;
1136 #if defined(TARGET_NR_futex) || \
1137 defined(TARGET_NR_rt_sigtimedwait) || \
1138 defined(TARGET_NR_pselect6) || defined(TARGET_NR_pselect6) || \
1139 defined(TARGET_NR_nanosleep) || defined(TARGET_NR_clock_settime) || \
1140 defined(TARGET_NR_utimensat) || defined(TARGET_NR_mq_timedsend) || \
1141 defined(TARGET_NR_mq_timedreceive) || defined(TARGET_NR_ipc) || \
1142 defined(TARGET_NR_semop) || defined(TARGET_NR_semtimedop) || \
1143 defined(TARGET_NR_timer_settime) || \
1144 (defined(TARGET_NR_timerfd_settime) && defined(CONFIG_TIMERFD))
1145 static inline abi_long target_to_host_timespec(struct timespec *host_ts,
1146 abi_ulong target_addr)
1148 struct target_timespec *target_ts;
1150 if (!lock_user_struct(VERIFY_READ, target_ts, target_addr, 1)) {
1151 return -TARGET_EFAULT;
1153 __get_user(host_ts->tv_sec, &target_ts->tv_sec);
1154 __get_user(host_ts->tv_nsec, &target_ts->tv_nsec);
1155 unlock_user_struct(target_ts, target_addr, 0);
1156 return 0;
1158 #endif
1160 #if defined(TARGET_NR_clock_settime64) || defined(TARGET_NR_futex_time64) || \
1161 defined(TARGET_NR_timer_settime64) || \
1162 defined(TARGET_NR_mq_timedsend_time64) || \
1163 defined(TARGET_NR_mq_timedreceive_time64) || \
1164 (defined(TARGET_NR_timerfd_settime64) && defined(CONFIG_TIMERFD)) || \
1165 defined(TARGET_NR_clock_nanosleep_time64) || \
1166 defined(TARGET_NR_rt_sigtimedwait_time64) || \
1167 defined(TARGET_NR_utimensat) || \
1168 defined(TARGET_NR_utimensat_time64) || \
1169 defined(TARGET_NR_semtimedop_time64) || \
1170 defined(TARGET_NR_pselect6_time64) || defined(TARGET_NR_ppoll_time64)
1171 static inline abi_long target_to_host_timespec64(struct timespec *host_ts,
1172 abi_ulong target_addr)
1174 struct target__kernel_timespec *target_ts;
1176 if (!lock_user_struct(VERIFY_READ, target_ts, target_addr, 1)) {
1177 return -TARGET_EFAULT;
1179 __get_user(host_ts->tv_sec, &target_ts->tv_sec);
1180 __get_user(host_ts->tv_nsec, &target_ts->tv_nsec);
1181 /* in 32bit mode, this drops the padding */
1182 host_ts->tv_nsec = (long)(abi_long)host_ts->tv_nsec;
1183 unlock_user_struct(target_ts, target_addr, 0);
1184 return 0;
1186 #endif
1188 static inline abi_long host_to_target_timespec(abi_ulong target_addr,
1189 struct timespec *host_ts)
1191 struct target_timespec *target_ts;
1193 if (!lock_user_struct(VERIFY_WRITE, target_ts, target_addr, 0)) {
1194 return -TARGET_EFAULT;
1196 __put_user(host_ts->tv_sec, &target_ts->tv_sec);
1197 __put_user(host_ts->tv_nsec, &target_ts->tv_nsec);
1198 unlock_user_struct(target_ts, target_addr, 1);
1199 return 0;
1202 static inline abi_long host_to_target_timespec64(abi_ulong target_addr,
1203 struct timespec *host_ts)
1205 struct target__kernel_timespec *target_ts;
1207 if (!lock_user_struct(VERIFY_WRITE, target_ts, target_addr, 0)) {
1208 return -TARGET_EFAULT;
1210 __put_user(host_ts->tv_sec, &target_ts->tv_sec);
1211 __put_user(host_ts->tv_nsec, &target_ts->tv_nsec);
1212 unlock_user_struct(target_ts, target_addr, 1);
1213 return 0;
1216 #if defined(TARGET_NR_gettimeofday)
1217 static inline abi_long copy_to_user_timezone(abi_ulong target_tz_addr,
1218 struct timezone *tz)
1220 struct target_timezone *target_tz;
1222 if (!lock_user_struct(VERIFY_WRITE, target_tz, target_tz_addr, 1)) {
1223 return -TARGET_EFAULT;
1226 __put_user(tz->tz_minuteswest, &target_tz->tz_minuteswest);
1227 __put_user(tz->tz_dsttime, &target_tz->tz_dsttime);
1229 unlock_user_struct(target_tz, target_tz_addr, 1);
1231 return 0;
1233 #endif
1235 #if defined(TARGET_NR_settimeofday)
1236 static inline abi_long copy_from_user_timezone(struct timezone *tz,
1237 abi_ulong target_tz_addr)
1239 struct target_timezone *target_tz;
1241 if (!lock_user_struct(VERIFY_READ, target_tz, target_tz_addr, 1)) {
1242 return -TARGET_EFAULT;
1245 __get_user(tz->tz_minuteswest, &target_tz->tz_minuteswest);
1246 __get_user(tz->tz_dsttime, &target_tz->tz_dsttime);
1248 unlock_user_struct(target_tz, target_tz_addr, 0);
1250 return 0;
1252 #endif
1254 #if defined(TARGET_NR_mq_open) && defined(__NR_mq_open)
1255 #include <mqueue.h>
1257 static inline abi_long copy_from_user_mq_attr(struct mq_attr *attr,
1258 abi_ulong target_mq_attr_addr)
1260 struct target_mq_attr *target_mq_attr;
1262 if (!lock_user_struct(VERIFY_READ, target_mq_attr,
1263 target_mq_attr_addr, 1))
1264 return -TARGET_EFAULT;
1266 __get_user(attr->mq_flags, &target_mq_attr->mq_flags);
1267 __get_user(attr->mq_maxmsg, &target_mq_attr->mq_maxmsg);
1268 __get_user(attr->mq_msgsize, &target_mq_attr->mq_msgsize);
1269 __get_user(attr->mq_curmsgs, &target_mq_attr->mq_curmsgs);
1271 unlock_user_struct(target_mq_attr, target_mq_attr_addr, 0);
1273 return 0;
1276 static inline abi_long copy_to_user_mq_attr(abi_ulong target_mq_attr_addr,
1277 const struct mq_attr *attr)
1279 struct target_mq_attr *target_mq_attr;
1281 if (!lock_user_struct(VERIFY_WRITE, target_mq_attr,
1282 target_mq_attr_addr, 0))
1283 return -TARGET_EFAULT;
1285 __put_user(attr->mq_flags, &target_mq_attr->mq_flags);
1286 __put_user(attr->mq_maxmsg, &target_mq_attr->mq_maxmsg);
1287 __put_user(attr->mq_msgsize, &target_mq_attr->mq_msgsize);
1288 __put_user(attr->mq_curmsgs, &target_mq_attr->mq_curmsgs);
1290 unlock_user_struct(target_mq_attr, target_mq_attr_addr, 1);
1292 return 0;
1294 #endif
1296 #if defined(TARGET_NR_select) || defined(TARGET_NR__newselect)
1297 /* do_select() must return target values and target errnos. */
1298 static abi_long do_select(int n,
1299 abi_ulong rfd_addr, abi_ulong wfd_addr,
1300 abi_ulong efd_addr, abi_ulong target_tv_addr)
1302 fd_set rfds, wfds, efds;
1303 fd_set *rfds_ptr, *wfds_ptr, *efds_ptr;
1304 struct timeval tv;
1305 struct timespec ts, *ts_ptr;
1306 abi_long ret;
1308 ret = copy_from_user_fdset_ptr(&rfds, &rfds_ptr, rfd_addr, n);
1309 if (ret) {
1310 return ret;
1312 ret = copy_from_user_fdset_ptr(&wfds, &wfds_ptr, wfd_addr, n);
1313 if (ret) {
1314 return ret;
1316 ret = copy_from_user_fdset_ptr(&efds, &efds_ptr, efd_addr, n);
1317 if (ret) {
1318 return ret;
1321 if (target_tv_addr) {
1322 if (copy_from_user_timeval(&tv, target_tv_addr))
1323 return -TARGET_EFAULT;
1324 ts.tv_sec = tv.tv_sec;
1325 ts.tv_nsec = tv.tv_usec * 1000;
1326 ts_ptr = &ts;
1327 } else {
1328 ts_ptr = NULL;
1331 ret = get_errno(safe_pselect6(n, rfds_ptr, wfds_ptr, efds_ptr,
1332 ts_ptr, NULL));
1334 if (!is_error(ret)) {
1335 if (rfd_addr && copy_to_user_fdset(rfd_addr, &rfds, n))
1336 return -TARGET_EFAULT;
1337 if (wfd_addr && copy_to_user_fdset(wfd_addr, &wfds, n))
1338 return -TARGET_EFAULT;
1339 if (efd_addr && copy_to_user_fdset(efd_addr, &efds, n))
1340 return -TARGET_EFAULT;
1342 if (target_tv_addr) {
1343 tv.tv_sec = ts.tv_sec;
1344 tv.tv_usec = ts.tv_nsec / 1000;
1345 if (copy_to_user_timeval(target_tv_addr, &tv)) {
1346 return -TARGET_EFAULT;
1351 return ret;
1354 #if defined(TARGET_WANT_OLD_SYS_SELECT)
1355 static abi_long do_old_select(abi_ulong arg1)
1357 struct target_sel_arg_struct *sel;
1358 abi_ulong inp, outp, exp, tvp;
1359 long nsel;
1361 if (!lock_user_struct(VERIFY_READ, sel, arg1, 1)) {
1362 return -TARGET_EFAULT;
1365 nsel = tswapal(sel->n);
1366 inp = tswapal(sel->inp);
1367 outp = tswapal(sel->outp);
1368 exp = tswapal(sel->exp);
1369 tvp = tswapal(sel->tvp);
1371 unlock_user_struct(sel, arg1, 0);
1373 return do_select(nsel, inp, outp, exp, tvp);
1375 #endif
1376 #endif
1378 #if defined(TARGET_NR_pselect6) || defined(TARGET_NR_pselect6_time64)
1379 static abi_long do_pselect6(abi_long arg1, abi_long arg2, abi_long arg3,
1380 abi_long arg4, abi_long arg5, abi_long arg6,
1381 bool time64)
1383 abi_long rfd_addr, wfd_addr, efd_addr, n, ts_addr;
1384 fd_set rfds, wfds, efds;
1385 fd_set *rfds_ptr, *wfds_ptr, *efds_ptr;
1386 struct timespec ts, *ts_ptr;
1387 abi_long ret;
1390 * The 6th arg is actually two args smashed together,
1391 * so we cannot use the C library.
1393 struct {
1394 sigset_t *set;
1395 size_t size;
1396 } sig, *sig_ptr;
1398 abi_ulong arg_sigset, arg_sigsize, *arg7;
1400 n = arg1;
1401 rfd_addr = arg2;
1402 wfd_addr = arg3;
1403 efd_addr = arg4;
1404 ts_addr = arg5;
1406 ret = copy_from_user_fdset_ptr(&rfds, &rfds_ptr, rfd_addr, n);
1407 if (ret) {
1408 return ret;
1410 ret = copy_from_user_fdset_ptr(&wfds, &wfds_ptr, wfd_addr, n);
1411 if (ret) {
1412 return ret;
1414 ret = copy_from_user_fdset_ptr(&efds, &efds_ptr, efd_addr, n);
1415 if (ret) {
1416 return ret;
1420 * This takes a timespec, and not a timeval, so we cannot
1421 * use the do_select() helper ...
1423 if (ts_addr) {
1424 if (time64) {
1425 if (target_to_host_timespec64(&ts, ts_addr)) {
1426 return -TARGET_EFAULT;
1428 } else {
1429 if (target_to_host_timespec(&ts, ts_addr)) {
1430 return -TARGET_EFAULT;
1433 ts_ptr = &ts;
1434 } else {
1435 ts_ptr = NULL;
1438 /* Extract the two packed args for the sigset */
1439 sig_ptr = NULL;
1440 if (arg6) {
1441 arg7 = lock_user(VERIFY_READ, arg6, sizeof(*arg7) * 2, 1);
1442 if (!arg7) {
1443 return -TARGET_EFAULT;
1445 arg_sigset = tswapal(arg7[0]);
1446 arg_sigsize = tswapal(arg7[1]);
1447 unlock_user(arg7, arg6, 0);
1449 if (arg_sigset) {
1450 ret = process_sigsuspend_mask(&sig.set, arg_sigset, arg_sigsize);
1451 if (ret != 0) {
1452 return ret;
1454 sig_ptr = &sig;
1455 sig.size = SIGSET_T_SIZE;
1459 ret = get_errno(safe_pselect6(n, rfds_ptr, wfds_ptr, efds_ptr,
1460 ts_ptr, sig_ptr));
1462 if (sig_ptr) {
1463 finish_sigsuspend_mask(ret);
1466 if (!is_error(ret)) {
1467 if (rfd_addr && copy_to_user_fdset(rfd_addr, &rfds, n)) {
1468 return -TARGET_EFAULT;
1470 if (wfd_addr && copy_to_user_fdset(wfd_addr, &wfds, n)) {
1471 return -TARGET_EFAULT;
1473 if (efd_addr && copy_to_user_fdset(efd_addr, &efds, n)) {
1474 return -TARGET_EFAULT;
1476 if (time64) {
1477 if (ts_addr && host_to_target_timespec64(ts_addr, &ts)) {
1478 return -TARGET_EFAULT;
1480 } else {
1481 if (ts_addr && host_to_target_timespec(ts_addr, &ts)) {
1482 return -TARGET_EFAULT;
1486 return ret;
1488 #endif
1490 #if defined(TARGET_NR_poll) || defined(TARGET_NR_ppoll) || \
1491 defined(TARGET_NR_ppoll_time64)
1492 static abi_long do_ppoll(abi_long arg1, abi_long arg2, abi_long arg3,
1493 abi_long arg4, abi_long arg5, bool ppoll, bool time64)
1495 struct target_pollfd *target_pfd;
1496 unsigned int nfds = arg2;
1497 struct pollfd *pfd;
1498 unsigned int i;
1499 abi_long ret;
1501 pfd = NULL;
1502 target_pfd = NULL;
1503 if (nfds) {
1504 if (nfds > (INT_MAX / sizeof(struct target_pollfd))) {
1505 return -TARGET_EINVAL;
1507 target_pfd = lock_user(VERIFY_WRITE, arg1,
1508 sizeof(struct target_pollfd) * nfds, 1);
1509 if (!target_pfd) {
1510 return -TARGET_EFAULT;
1513 pfd = alloca(sizeof(struct pollfd) * nfds);
1514 for (i = 0; i < nfds; i++) {
1515 pfd[i].fd = tswap32(target_pfd[i].fd);
1516 pfd[i].events = tswap16(target_pfd[i].events);
1519 if (ppoll) {
1520 struct timespec _timeout_ts, *timeout_ts = &_timeout_ts;
1521 sigset_t *set = NULL;
1523 if (arg3) {
1524 if (time64) {
1525 if (target_to_host_timespec64(timeout_ts, arg3)) {
1526 unlock_user(target_pfd, arg1, 0);
1527 return -TARGET_EFAULT;
1529 } else {
1530 if (target_to_host_timespec(timeout_ts, arg3)) {
1531 unlock_user(target_pfd, arg1, 0);
1532 return -TARGET_EFAULT;
1535 } else {
1536 timeout_ts = NULL;
1539 if (arg4) {
1540 ret = process_sigsuspend_mask(&set, arg4, arg5);
1541 if (ret != 0) {
1542 unlock_user(target_pfd, arg1, 0);
1543 return ret;
1547 ret = get_errno(safe_ppoll(pfd, nfds, timeout_ts,
1548 set, SIGSET_T_SIZE));
1550 if (set) {
1551 finish_sigsuspend_mask(ret);
1553 if (!is_error(ret) && arg3) {
1554 if (time64) {
1555 if (host_to_target_timespec64(arg3, timeout_ts)) {
1556 return -TARGET_EFAULT;
1558 } else {
1559 if (host_to_target_timespec(arg3, timeout_ts)) {
1560 return -TARGET_EFAULT;
1564 } else {
1565 struct timespec ts, *pts;
1567 if (arg3 >= 0) {
1568 /* Convert ms to secs, ns */
1569 ts.tv_sec = arg3 / 1000;
1570 ts.tv_nsec = (arg3 % 1000) * 1000000LL;
1571 pts = &ts;
1572 } else {
1573 /* -ve poll() timeout means "infinite" */
1574 pts = NULL;
1576 ret = get_errno(safe_ppoll(pfd, nfds, pts, NULL, 0));
1579 if (!is_error(ret)) {
1580 for (i = 0; i < nfds; i++) {
1581 target_pfd[i].revents = tswap16(pfd[i].revents);
1584 unlock_user(target_pfd, arg1, sizeof(struct target_pollfd) * nfds);
1585 return ret;
1587 #endif
1589 static abi_long do_pipe2(int host_pipe[], int flags)
1591 #ifdef CONFIG_PIPE2
1592 return pipe2(host_pipe, flags);
1593 #else
1594 return -ENOSYS;
1595 #endif
1598 static abi_long do_pipe(CPUArchState *cpu_env, abi_ulong pipedes,
1599 int flags, int is_pipe2)
1601 int host_pipe[2];
1602 abi_long ret;
1603 ret = flags ? do_pipe2(host_pipe, flags) : pipe(host_pipe);
1605 if (is_error(ret))
1606 return get_errno(ret);
1608 /* Several targets have special calling conventions for the original
1609 pipe syscall, but didn't replicate this into the pipe2 syscall. */
1610 if (!is_pipe2) {
1611 #if defined(TARGET_ALPHA)
1612 cpu_env->ir[IR_A4] = host_pipe[1];
1613 return host_pipe[0];
1614 #elif defined(TARGET_MIPS)
1615 cpu_env->active_tc.gpr[3] = host_pipe[1];
1616 return host_pipe[0];
1617 #elif defined(TARGET_SH4)
1618 cpu_env->gregs[1] = host_pipe[1];
1619 return host_pipe[0];
1620 #elif defined(TARGET_SPARC)
1621 cpu_env->regwptr[1] = host_pipe[1];
1622 return host_pipe[0];
1623 #endif
1626 if (put_user_s32(host_pipe[0], pipedes)
1627 || put_user_s32(host_pipe[1], pipedes + sizeof(host_pipe[0])))
1628 return -TARGET_EFAULT;
1629 return get_errno(ret);
1632 static inline abi_long target_to_host_ip_mreq(struct ip_mreqn *mreqn,
1633 abi_ulong target_addr,
1634 socklen_t len)
1636 struct target_ip_mreqn *target_smreqn;
1638 target_smreqn = lock_user(VERIFY_READ, target_addr, len, 1);
1639 if (!target_smreqn)
1640 return -TARGET_EFAULT;
1641 mreqn->imr_multiaddr.s_addr = target_smreqn->imr_multiaddr.s_addr;
1642 mreqn->imr_address.s_addr = target_smreqn->imr_address.s_addr;
1643 if (len == sizeof(struct target_ip_mreqn))
1644 mreqn->imr_ifindex = tswapal(target_smreqn->imr_ifindex);
1645 unlock_user(target_smreqn, target_addr, 0);
1647 return 0;
1650 static inline abi_long target_to_host_sockaddr(int fd, struct sockaddr *addr,
1651 abi_ulong target_addr,
1652 socklen_t len)
1654 const socklen_t unix_maxlen = sizeof (struct sockaddr_un);
1655 sa_family_t sa_family;
1656 struct target_sockaddr *target_saddr;
1658 if (fd_trans_target_to_host_addr(fd)) {
1659 return fd_trans_target_to_host_addr(fd)(addr, target_addr, len);
1662 target_saddr = lock_user(VERIFY_READ, target_addr, len, 1);
1663 if (!target_saddr)
1664 return -TARGET_EFAULT;
1666 sa_family = tswap16(target_saddr->sa_family);
1668 /* Oops. The caller might send a incomplete sun_path; sun_path
1669 * must be terminated by \0 (see the manual page), but
1670 * unfortunately it is quite common to specify sockaddr_un
1671 * length as "strlen(x->sun_path)" while it should be
1672 * "strlen(...) + 1". We'll fix that here if needed.
1673 * Linux kernel has a similar feature.
1676 if (sa_family == AF_UNIX) {
1677 if (len < unix_maxlen && len > 0) {
1678 char *cp = (char*)target_saddr;
1680 if ( cp[len-1] && !cp[len] )
1681 len++;
1683 if (len > unix_maxlen)
1684 len = unix_maxlen;
1687 memcpy(addr, target_saddr, len);
1688 addr->sa_family = sa_family;
1689 if (sa_family == AF_NETLINK) {
1690 struct sockaddr_nl *nladdr;
1692 nladdr = (struct sockaddr_nl *)addr;
1693 nladdr->nl_pid = tswap32(nladdr->nl_pid);
1694 nladdr->nl_groups = tswap32(nladdr->nl_groups);
1695 } else if (sa_family == AF_PACKET) {
1696 struct target_sockaddr_ll *lladdr;
1698 lladdr = (struct target_sockaddr_ll *)addr;
1699 lladdr->sll_ifindex = tswap32(lladdr->sll_ifindex);
1700 lladdr->sll_hatype = tswap16(lladdr->sll_hatype);
1702 unlock_user(target_saddr, target_addr, 0);
1704 return 0;
1707 static inline abi_long host_to_target_sockaddr(abi_ulong target_addr,
1708 struct sockaddr *addr,
1709 socklen_t len)
1711 struct target_sockaddr *target_saddr;
1713 if (len == 0) {
1714 return 0;
1716 assert(addr);
1718 target_saddr = lock_user(VERIFY_WRITE, target_addr, len, 0);
1719 if (!target_saddr)
1720 return -TARGET_EFAULT;
1721 memcpy(target_saddr, addr, len);
1722 if (len >= offsetof(struct target_sockaddr, sa_family) +
1723 sizeof(target_saddr->sa_family)) {
1724 target_saddr->sa_family = tswap16(addr->sa_family);
1726 if (addr->sa_family == AF_NETLINK &&
1727 len >= sizeof(struct target_sockaddr_nl)) {
1728 struct target_sockaddr_nl *target_nl =
1729 (struct target_sockaddr_nl *)target_saddr;
1730 target_nl->nl_pid = tswap32(target_nl->nl_pid);
1731 target_nl->nl_groups = tswap32(target_nl->nl_groups);
1732 } else if (addr->sa_family == AF_PACKET) {
1733 struct sockaddr_ll *target_ll = (struct sockaddr_ll *)target_saddr;
1734 target_ll->sll_ifindex = tswap32(target_ll->sll_ifindex);
1735 target_ll->sll_hatype = tswap16(target_ll->sll_hatype);
1736 } else if (addr->sa_family == AF_INET6 &&
1737 len >= sizeof(struct target_sockaddr_in6)) {
1738 struct target_sockaddr_in6 *target_in6 =
1739 (struct target_sockaddr_in6 *)target_saddr;
1740 target_in6->sin6_scope_id = tswap16(target_in6->sin6_scope_id);
1742 unlock_user(target_saddr, target_addr, len);
1744 return 0;
1747 static inline abi_long target_to_host_cmsg(struct msghdr *msgh,
1748 struct target_msghdr *target_msgh)
1750 struct cmsghdr *cmsg = CMSG_FIRSTHDR(msgh);
1751 abi_long msg_controllen;
1752 abi_ulong target_cmsg_addr;
1753 struct target_cmsghdr *target_cmsg, *target_cmsg_start;
1754 socklen_t space = 0;
1756 msg_controllen = tswapal(target_msgh->msg_controllen);
1757 if (msg_controllen < sizeof (struct target_cmsghdr))
1758 goto the_end;
1759 target_cmsg_addr = tswapal(target_msgh->msg_control);
1760 target_cmsg = lock_user(VERIFY_READ, target_cmsg_addr, msg_controllen, 1);
1761 target_cmsg_start = target_cmsg;
1762 if (!target_cmsg)
1763 return -TARGET_EFAULT;
1765 while (cmsg && target_cmsg) {
1766 void *data = CMSG_DATA(cmsg);
1767 void *target_data = TARGET_CMSG_DATA(target_cmsg);
1769 int len = tswapal(target_cmsg->cmsg_len)
1770 - sizeof(struct target_cmsghdr);
1772 space += CMSG_SPACE(len);
1773 if (space > msgh->msg_controllen) {
1774 space -= CMSG_SPACE(len);
1775 /* This is a QEMU bug, since we allocated the payload
1776 * area ourselves (unlike overflow in host-to-target
1777 * conversion, which is just the guest giving us a buffer
1778 * that's too small). It can't happen for the payload types
1779 * we currently support; if it becomes an issue in future
1780 * we would need to improve our allocation strategy to
1781 * something more intelligent than "twice the size of the
1782 * target buffer we're reading from".
1784 qemu_log_mask(LOG_UNIMP,
1785 ("Unsupported ancillary data %d/%d: "
1786 "unhandled msg size\n"),
1787 tswap32(target_cmsg->cmsg_level),
1788 tswap32(target_cmsg->cmsg_type));
1789 break;
1792 if (tswap32(target_cmsg->cmsg_level) == TARGET_SOL_SOCKET) {
1793 cmsg->cmsg_level = SOL_SOCKET;
1794 } else {
1795 cmsg->cmsg_level = tswap32(target_cmsg->cmsg_level);
1797 cmsg->cmsg_type = tswap32(target_cmsg->cmsg_type);
1798 cmsg->cmsg_len = CMSG_LEN(len);
1800 if (cmsg->cmsg_level == SOL_SOCKET && cmsg->cmsg_type == SCM_RIGHTS) {
1801 int *fd = (int *)data;
1802 int *target_fd = (int *)target_data;
1803 int i, numfds = len / sizeof(int);
1805 for (i = 0; i < numfds; i++) {
1806 __get_user(fd[i], target_fd + i);
1808 } else if (cmsg->cmsg_level == SOL_SOCKET
1809 && cmsg->cmsg_type == SCM_CREDENTIALS) {
1810 struct ucred *cred = (struct ucred *)data;
1811 struct target_ucred *target_cred =
1812 (struct target_ucred *)target_data;
1814 __get_user(cred->pid, &target_cred->pid);
1815 __get_user(cred->uid, &target_cred->uid);
1816 __get_user(cred->gid, &target_cred->gid);
1817 } else {
1818 qemu_log_mask(LOG_UNIMP, "Unsupported ancillary data: %d/%d\n",
1819 cmsg->cmsg_level, cmsg->cmsg_type);
1820 memcpy(data, target_data, len);
1823 cmsg = CMSG_NXTHDR(msgh, cmsg);
1824 target_cmsg = TARGET_CMSG_NXTHDR(target_msgh, target_cmsg,
1825 target_cmsg_start);
1827 unlock_user(target_cmsg, target_cmsg_addr, 0);
1828 the_end:
1829 msgh->msg_controllen = space;
1830 return 0;
1833 static inline abi_long host_to_target_cmsg(struct target_msghdr *target_msgh,
1834 struct msghdr *msgh)
1836 struct cmsghdr *cmsg = CMSG_FIRSTHDR(msgh);
1837 abi_long msg_controllen;
1838 abi_ulong target_cmsg_addr;
1839 struct target_cmsghdr *target_cmsg, *target_cmsg_start;
1840 socklen_t space = 0;
1842 msg_controllen = tswapal(target_msgh->msg_controllen);
1843 if (msg_controllen < sizeof (struct target_cmsghdr))
1844 goto the_end;
1845 target_cmsg_addr = tswapal(target_msgh->msg_control);
1846 target_cmsg = lock_user(VERIFY_WRITE, target_cmsg_addr, msg_controllen, 0);
1847 target_cmsg_start = target_cmsg;
1848 if (!target_cmsg)
1849 return -TARGET_EFAULT;
1851 while (cmsg && target_cmsg) {
1852 void *data = CMSG_DATA(cmsg);
1853 void *target_data = TARGET_CMSG_DATA(target_cmsg);
1855 int len = cmsg->cmsg_len - sizeof(struct cmsghdr);
1856 int tgt_len, tgt_space;
1858 /* We never copy a half-header but may copy half-data;
1859 * this is Linux's behaviour in put_cmsg(). Note that
1860 * truncation here is a guest problem (which we report
1861 * to the guest via the CTRUNC bit), unlike truncation
1862 * in target_to_host_cmsg, which is a QEMU bug.
1864 if (msg_controllen < sizeof(struct target_cmsghdr)) {
1865 target_msgh->msg_flags |= tswap32(MSG_CTRUNC);
1866 break;
1869 if (cmsg->cmsg_level == SOL_SOCKET) {
1870 target_cmsg->cmsg_level = tswap32(TARGET_SOL_SOCKET);
1871 } else {
1872 target_cmsg->cmsg_level = tswap32(cmsg->cmsg_level);
1874 target_cmsg->cmsg_type = tswap32(cmsg->cmsg_type);
1876 /* Payload types which need a different size of payload on
1877 * the target must adjust tgt_len here.
1879 tgt_len = len;
1880 switch (cmsg->cmsg_level) {
1881 case SOL_SOCKET:
1882 switch (cmsg->cmsg_type) {
1883 case SO_TIMESTAMP:
1884 tgt_len = sizeof(struct target_timeval);
1885 break;
1886 default:
1887 break;
1889 break;
1890 default:
1891 break;
1894 if (msg_controllen < TARGET_CMSG_LEN(tgt_len)) {
1895 target_msgh->msg_flags |= tswap32(MSG_CTRUNC);
1896 tgt_len = msg_controllen - sizeof(struct target_cmsghdr);
1899 /* We must now copy-and-convert len bytes of payload
1900 * into tgt_len bytes of destination space. Bear in mind
1901 * that in both source and destination we may be dealing
1902 * with a truncated value!
1904 switch (cmsg->cmsg_level) {
1905 case SOL_SOCKET:
1906 switch (cmsg->cmsg_type) {
1907 case SCM_RIGHTS:
1909 int *fd = (int *)data;
1910 int *target_fd = (int *)target_data;
1911 int i, numfds = tgt_len / sizeof(int);
1913 for (i = 0; i < numfds; i++) {
1914 __put_user(fd[i], target_fd + i);
1916 break;
1918 case SO_TIMESTAMP:
1920 struct timeval *tv = (struct timeval *)data;
1921 struct target_timeval *target_tv =
1922 (struct target_timeval *)target_data;
1924 if (len != sizeof(struct timeval) ||
1925 tgt_len != sizeof(struct target_timeval)) {
1926 goto unimplemented;
1929 /* copy struct timeval to target */
1930 __put_user(tv->tv_sec, &target_tv->tv_sec);
1931 __put_user(tv->tv_usec, &target_tv->tv_usec);
1932 break;
1934 case SCM_CREDENTIALS:
1936 struct ucred *cred = (struct ucred *)data;
1937 struct target_ucred *target_cred =
1938 (struct target_ucred *)target_data;
1940 __put_user(cred->pid, &target_cred->pid);
1941 __put_user(cred->uid, &target_cred->uid);
1942 __put_user(cred->gid, &target_cred->gid);
1943 break;
1945 default:
1946 goto unimplemented;
1948 break;
1950 case SOL_IP:
1951 switch (cmsg->cmsg_type) {
1952 case IP_TTL:
1954 uint32_t *v = (uint32_t *)data;
1955 uint32_t *t_int = (uint32_t *)target_data;
1957 if (len != sizeof(uint32_t) ||
1958 tgt_len != sizeof(uint32_t)) {
1959 goto unimplemented;
1961 __put_user(*v, t_int);
1962 break;
1964 case IP_RECVERR:
1966 struct errhdr_t {
1967 struct sock_extended_err ee;
1968 struct sockaddr_in offender;
1970 struct errhdr_t *errh = (struct errhdr_t *)data;
1971 struct errhdr_t *target_errh =
1972 (struct errhdr_t *)target_data;
1974 if (len != sizeof(struct errhdr_t) ||
1975 tgt_len != sizeof(struct errhdr_t)) {
1976 goto unimplemented;
1978 __put_user(errh->ee.ee_errno, &target_errh->ee.ee_errno);
1979 __put_user(errh->ee.ee_origin, &target_errh->ee.ee_origin);
1980 __put_user(errh->ee.ee_type, &target_errh->ee.ee_type);
1981 __put_user(errh->ee.ee_code, &target_errh->ee.ee_code);
1982 __put_user(errh->ee.ee_pad, &target_errh->ee.ee_pad);
1983 __put_user(errh->ee.ee_info, &target_errh->ee.ee_info);
1984 __put_user(errh->ee.ee_data, &target_errh->ee.ee_data);
1985 host_to_target_sockaddr((unsigned long) &target_errh->offender,
1986 (void *) &errh->offender, sizeof(errh->offender));
1987 break;
1989 default:
1990 goto unimplemented;
1992 break;
1994 case SOL_IPV6:
1995 switch (cmsg->cmsg_type) {
1996 case IPV6_HOPLIMIT:
1998 uint32_t *v = (uint32_t *)data;
1999 uint32_t *t_int = (uint32_t *)target_data;
2001 if (len != sizeof(uint32_t) ||
2002 tgt_len != sizeof(uint32_t)) {
2003 goto unimplemented;
2005 __put_user(*v, t_int);
2006 break;
2008 case IPV6_RECVERR:
2010 struct errhdr6_t {
2011 struct sock_extended_err ee;
2012 struct sockaddr_in6 offender;
2014 struct errhdr6_t *errh = (struct errhdr6_t *)data;
2015 struct errhdr6_t *target_errh =
2016 (struct errhdr6_t *)target_data;
2018 if (len != sizeof(struct errhdr6_t) ||
2019 tgt_len != sizeof(struct errhdr6_t)) {
2020 goto unimplemented;
2022 __put_user(errh->ee.ee_errno, &target_errh->ee.ee_errno);
2023 __put_user(errh->ee.ee_origin, &target_errh->ee.ee_origin);
2024 __put_user(errh->ee.ee_type, &target_errh->ee.ee_type);
2025 __put_user(errh->ee.ee_code, &target_errh->ee.ee_code);
2026 __put_user(errh->ee.ee_pad, &target_errh->ee.ee_pad);
2027 __put_user(errh->ee.ee_info, &target_errh->ee.ee_info);
2028 __put_user(errh->ee.ee_data, &target_errh->ee.ee_data);
2029 host_to_target_sockaddr((unsigned long) &target_errh->offender,
2030 (void *) &errh->offender, sizeof(errh->offender));
2031 break;
2033 default:
2034 goto unimplemented;
2036 break;
2038 default:
2039 unimplemented:
2040 qemu_log_mask(LOG_UNIMP, "Unsupported ancillary data: %d/%d\n",
2041 cmsg->cmsg_level, cmsg->cmsg_type);
2042 memcpy(target_data, data, MIN(len, tgt_len));
2043 if (tgt_len > len) {
2044 memset(target_data + len, 0, tgt_len - len);
2048 target_cmsg->cmsg_len = tswapal(TARGET_CMSG_LEN(tgt_len));
2049 tgt_space = TARGET_CMSG_SPACE(tgt_len);
2050 if (msg_controllen < tgt_space) {
2051 tgt_space = msg_controllen;
2053 msg_controllen -= tgt_space;
2054 space += tgt_space;
2055 cmsg = CMSG_NXTHDR(msgh, cmsg);
2056 target_cmsg = TARGET_CMSG_NXTHDR(target_msgh, target_cmsg,
2057 target_cmsg_start);
2059 unlock_user(target_cmsg, target_cmsg_addr, space);
2060 the_end:
2061 target_msgh->msg_controllen = tswapal(space);
2062 return 0;
2065 /* do_setsockopt() Must return target values and target errnos. */
2066 static abi_long do_setsockopt(int sockfd, int level, int optname,
2067 abi_ulong optval_addr, socklen_t optlen)
2069 abi_long ret;
2070 int val;
2071 struct ip_mreqn *ip_mreq;
2072 struct ip_mreq_source *ip_mreq_source;
2074 switch(level) {
2075 case SOL_TCP:
2076 case SOL_UDP:
2077 /* TCP and UDP options all take an 'int' value. */
2078 if (optlen < sizeof(uint32_t))
2079 return -TARGET_EINVAL;
2081 if (get_user_u32(val, optval_addr))
2082 return -TARGET_EFAULT;
2083 ret = get_errno(setsockopt(sockfd, level, optname, &val, sizeof(val)));
2084 break;
2085 case SOL_IP:
2086 switch(optname) {
2087 case IP_TOS:
2088 case IP_TTL:
2089 case IP_HDRINCL:
2090 case IP_ROUTER_ALERT:
2091 case IP_RECVOPTS:
2092 case IP_RETOPTS:
2093 case IP_PKTINFO:
2094 case IP_MTU_DISCOVER:
2095 case IP_RECVERR:
2096 case IP_RECVTTL:
2097 case IP_RECVTOS:
2098 #ifdef IP_FREEBIND
2099 case IP_FREEBIND:
2100 #endif
2101 case IP_MULTICAST_TTL:
2102 case IP_MULTICAST_LOOP:
2103 val = 0;
2104 if (optlen >= sizeof(uint32_t)) {
2105 if (get_user_u32(val, optval_addr))
2106 return -TARGET_EFAULT;
2107 } else if (optlen >= 1) {
2108 if (get_user_u8(val, optval_addr))
2109 return -TARGET_EFAULT;
2111 ret = get_errno(setsockopt(sockfd, level, optname, &val, sizeof(val)));
2112 break;
2113 case IP_ADD_MEMBERSHIP:
2114 case IP_DROP_MEMBERSHIP:
2115 if (optlen < sizeof (struct target_ip_mreq) ||
2116 optlen > sizeof (struct target_ip_mreqn))
2117 return -TARGET_EINVAL;
2119 ip_mreq = (struct ip_mreqn *) alloca(optlen);
2120 target_to_host_ip_mreq(ip_mreq, optval_addr, optlen);
2121 ret = get_errno(setsockopt(sockfd, level, optname, ip_mreq, optlen));
2122 break;
2124 case IP_BLOCK_SOURCE:
2125 case IP_UNBLOCK_SOURCE:
2126 case IP_ADD_SOURCE_MEMBERSHIP:
2127 case IP_DROP_SOURCE_MEMBERSHIP:
2128 if (optlen != sizeof (struct target_ip_mreq_source))
2129 return -TARGET_EINVAL;
2131 ip_mreq_source = lock_user(VERIFY_READ, optval_addr, optlen, 1);
2132 if (!ip_mreq_source) {
2133 return -TARGET_EFAULT;
2135 ret = get_errno(setsockopt(sockfd, level, optname, ip_mreq_source, optlen));
2136 unlock_user (ip_mreq_source, optval_addr, 0);
2137 break;
2139 default:
2140 goto unimplemented;
2142 break;
2143 case SOL_IPV6:
2144 switch (optname) {
2145 case IPV6_MTU_DISCOVER:
2146 case IPV6_MTU:
2147 case IPV6_V6ONLY:
2148 case IPV6_RECVPKTINFO:
2149 case IPV6_UNICAST_HOPS:
2150 case IPV6_MULTICAST_HOPS:
2151 case IPV6_MULTICAST_LOOP:
2152 case IPV6_RECVERR:
2153 case IPV6_RECVHOPLIMIT:
2154 case IPV6_2292HOPLIMIT:
2155 case IPV6_CHECKSUM:
2156 case IPV6_ADDRFORM:
2157 case IPV6_2292PKTINFO:
2158 case IPV6_RECVTCLASS:
2159 case IPV6_RECVRTHDR:
2160 case IPV6_2292RTHDR:
2161 case IPV6_RECVHOPOPTS:
2162 case IPV6_2292HOPOPTS:
2163 case IPV6_RECVDSTOPTS:
2164 case IPV6_2292DSTOPTS:
2165 case IPV6_TCLASS:
2166 case IPV6_ADDR_PREFERENCES:
2167 #ifdef IPV6_RECVPATHMTU
2168 case IPV6_RECVPATHMTU:
2169 #endif
2170 #ifdef IPV6_TRANSPARENT
2171 case IPV6_TRANSPARENT:
2172 #endif
2173 #ifdef IPV6_FREEBIND
2174 case IPV6_FREEBIND:
2175 #endif
2176 #ifdef IPV6_RECVORIGDSTADDR
2177 case IPV6_RECVORIGDSTADDR:
2178 #endif
2179 val = 0;
2180 if (optlen < sizeof(uint32_t)) {
2181 return -TARGET_EINVAL;
2183 if (get_user_u32(val, optval_addr)) {
2184 return -TARGET_EFAULT;
2186 ret = get_errno(setsockopt(sockfd, level, optname,
2187 &val, sizeof(val)));
2188 break;
2189 case IPV6_PKTINFO:
2191 struct in6_pktinfo pki;
2193 if (optlen < sizeof(pki)) {
2194 return -TARGET_EINVAL;
2197 if (copy_from_user(&pki, optval_addr, sizeof(pki))) {
2198 return -TARGET_EFAULT;
2201 pki.ipi6_ifindex = tswap32(pki.ipi6_ifindex);
2203 ret = get_errno(setsockopt(sockfd, level, optname,
2204 &pki, sizeof(pki)));
2205 break;
2207 case IPV6_ADD_MEMBERSHIP:
2208 case IPV6_DROP_MEMBERSHIP:
2210 struct ipv6_mreq ipv6mreq;
2212 if (optlen < sizeof(ipv6mreq)) {
2213 return -TARGET_EINVAL;
2216 if (copy_from_user(&ipv6mreq, optval_addr, sizeof(ipv6mreq))) {
2217 return -TARGET_EFAULT;
2220 ipv6mreq.ipv6mr_interface = tswap32(ipv6mreq.ipv6mr_interface);
2222 ret = get_errno(setsockopt(sockfd, level, optname,
2223 &ipv6mreq, sizeof(ipv6mreq)));
2224 break;
2226 default:
2227 goto unimplemented;
2229 break;
2230 case SOL_ICMPV6:
2231 switch (optname) {
2232 case ICMPV6_FILTER:
2234 struct icmp6_filter icmp6f;
2236 if (optlen > sizeof(icmp6f)) {
2237 optlen = sizeof(icmp6f);
2240 if (copy_from_user(&icmp6f, optval_addr, optlen)) {
2241 return -TARGET_EFAULT;
2244 for (val = 0; val < 8; val++) {
2245 icmp6f.data[val] = tswap32(icmp6f.data[val]);
2248 ret = get_errno(setsockopt(sockfd, level, optname,
2249 &icmp6f, optlen));
2250 break;
2252 default:
2253 goto unimplemented;
2255 break;
2256 case SOL_RAW:
2257 switch (optname) {
2258 case ICMP_FILTER:
2259 case IPV6_CHECKSUM:
2260 /* those take an u32 value */
2261 if (optlen < sizeof(uint32_t)) {
2262 return -TARGET_EINVAL;
2265 if (get_user_u32(val, optval_addr)) {
2266 return -TARGET_EFAULT;
2268 ret = get_errno(setsockopt(sockfd, level, optname,
2269 &val, sizeof(val)));
2270 break;
2272 default:
2273 goto unimplemented;
2275 break;
2276 #if defined(SOL_ALG) && defined(ALG_SET_KEY) && defined(ALG_SET_AEAD_AUTHSIZE)
2277 case SOL_ALG:
2278 switch (optname) {
2279 case ALG_SET_KEY:
2281 char *alg_key = g_malloc(optlen);
2283 if (!alg_key) {
2284 return -TARGET_ENOMEM;
2286 if (copy_from_user(alg_key, optval_addr, optlen)) {
2287 g_free(alg_key);
2288 return -TARGET_EFAULT;
2290 ret = get_errno(setsockopt(sockfd, level, optname,
2291 alg_key, optlen));
2292 g_free(alg_key);
2293 break;
2295 case ALG_SET_AEAD_AUTHSIZE:
2297 ret = get_errno(setsockopt(sockfd, level, optname,
2298 NULL, optlen));
2299 break;
2301 default:
2302 goto unimplemented;
2304 break;
2305 #endif
2306 case TARGET_SOL_SOCKET:
2307 switch (optname) {
2308 case TARGET_SO_RCVTIMEO:
2310 struct timeval tv;
2312 optname = SO_RCVTIMEO;
2314 set_timeout:
2315 if (optlen != sizeof(struct target_timeval)) {
2316 return -TARGET_EINVAL;
2319 if (copy_from_user_timeval(&tv, optval_addr)) {
2320 return -TARGET_EFAULT;
2323 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname,
2324 &tv, sizeof(tv)));
2325 return ret;
2327 case TARGET_SO_SNDTIMEO:
2328 optname = SO_SNDTIMEO;
2329 goto set_timeout;
2330 case TARGET_SO_ATTACH_FILTER:
2332 struct target_sock_fprog *tfprog;
2333 struct target_sock_filter *tfilter;
2334 struct sock_fprog fprog;
2335 struct sock_filter *filter;
2336 int i;
2338 if (optlen != sizeof(*tfprog)) {
2339 return -TARGET_EINVAL;
2341 if (!lock_user_struct(VERIFY_READ, tfprog, optval_addr, 0)) {
2342 return -TARGET_EFAULT;
2344 if (!lock_user_struct(VERIFY_READ, tfilter,
2345 tswapal(tfprog->filter), 0)) {
2346 unlock_user_struct(tfprog, optval_addr, 1);
2347 return -TARGET_EFAULT;
2350 fprog.len = tswap16(tfprog->len);
2351 filter = g_try_new(struct sock_filter, fprog.len);
2352 if (filter == NULL) {
2353 unlock_user_struct(tfilter, tfprog->filter, 1);
2354 unlock_user_struct(tfprog, optval_addr, 1);
2355 return -TARGET_ENOMEM;
2357 for (i = 0; i < fprog.len; i++) {
2358 filter[i].code = tswap16(tfilter[i].code);
2359 filter[i].jt = tfilter[i].jt;
2360 filter[i].jf = tfilter[i].jf;
2361 filter[i].k = tswap32(tfilter[i].k);
2363 fprog.filter = filter;
2365 ret = get_errno(setsockopt(sockfd, SOL_SOCKET,
2366 SO_ATTACH_FILTER, &fprog, sizeof(fprog)));
2367 g_free(filter);
2369 unlock_user_struct(tfilter, tfprog->filter, 1);
2370 unlock_user_struct(tfprog, optval_addr, 1);
2371 return ret;
2373 case TARGET_SO_BINDTODEVICE:
2375 char *dev_ifname, *addr_ifname;
2377 if (optlen > IFNAMSIZ - 1) {
2378 optlen = IFNAMSIZ - 1;
2380 dev_ifname = lock_user(VERIFY_READ, optval_addr, optlen, 1);
2381 if (!dev_ifname) {
2382 return -TARGET_EFAULT;
2384 optname = SO_BINDTODEVICE;
2385 addr_ifname = alloca(IFNAMSIZ);
2386 memcpy(addr_ifname, dev_ifname, optlen);
2387 addr_ifname[optlen] = 0;
2388 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname,
2389 addr_ifname, optlen));
2390 unlock_user (dev_ifname, optval_addr, 0);
2391 return ret;
2393 case TARGET_SO_LINGER:
2395 struct linger lg;
2396 struct target_linger *tlg;
2398 if (optlen != sizeof(struct target_linger)) {
2399 return -TARGET_EINVAL;
2401 if (!lock_user_struct(VERIFY_READ, tlg, optval_addr, 1)) {
2402 return -TARGET_EFAULT;
2404 __get_user(lg.l_onoff, &tlg->l_onoff);
2405 __get_user(lg.l_linger, &tlg->l_linger);
2406 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, SO_LINGER,
2407 &lg, sizeof(lg)));
2408 unlock_user_struct(tlg, optval_addr, 0);
2409 return ret;
2411 /* Options with 'int' argument. */
2412 case TARGET_SO_DEBUG:
2413 optname = SO_DEBUG;
2414 break;
2415 case TARGET_SO_REUSEADDR:
2416 optname = SO_REUSEADDR;
2417 break;
2418 #ifdef SO_REUSEPORT
2419 case TARGET_SO_REUSEPORT:
2420 optname = SO_REUSEPORT;
2421 break;
2422 #endif
2423 case TARGET_SO_TYPE:
2424 optname = SO_TYPE;
2425 break;
2426 case TARGET_SO_ERROR:
2427 optname = SO_ERROR;
2428 break;
2429 case TARGET_SO_DONTROUTE:
2430 optname = SO_DONTROUTE;
2431 break;
2432 case TARGET_SO_BROADCAST:
2433 optname = SO_BROADCAST;
2434 break;
2435 case TARGET_SO_SNDBUF:
2436 optname = SO_SNDBUF;
2437 break;
2438 case TARGET_SO_SNDBUFFORCE:
2439 optname = SO_SNDBUFFORCE;
2440 break;
2441 case TARGET_SO_RCVBUF:
2442 optname = SO_RCVBUF;
2443 break;
2444 case TARGET_SO_RCVBUFFORCE:
2445 optname = SO_RCVBUFFORCE;
2446 break;
2447 case TARGET_SO_KEEPALIVE:
2448 optname = SO_KEEPALIVE;
2449 break;
2450 case TARGET_SO_OOBINLINE:
2451 optname = SO_OOBINLINE;
2452 break;
2453 case TARGET_SO_NO_CHECK:
2454 optname = SO_NO_CHECK;
2455 break;
2456 case TARGET_SO_PRIORITY:
2457 optname = SO_PRIORITY;
2458 break;
2459 #ifdef SO_BSDCOMPAT
2460 case TARGET_SO_BSDCOMPAT:
2461 optname = SO_BSDCOMPAT;
2462 break;
2463 #endif
2464 case TARGET_SO_PASSCRED:
2465 optname = SO_PASSCRED;
2466 break;
2467 case TARGET_SO_PASSSEC:
2468 optname = SO_PASSSEC;
2469 break;
2470 case TARGET_SO_TIMESTAMP:
2471 optname = SO_TIMESTAMP;
2472 break;
2473 case TARGET_SO_RCVLOWAT:
2474 optname = SO_RCVLOWAT;
2475 break;
2476 default:
2477 goto unimplemented;
2479 if (optlen < sizeof(uint32_t))
2480 return -TARGET_EINVAL;
2482 if (get_user_u32(val, optval_addr))
2483 return -TARGET_EFAULT;
2484 ret = get_errno(setsockopt(sockfd, SOL_SOCKET, optname, &val, sizeof(val)));
2485 break;
2486 #ifdef SOL_NETLINK
2487 case SOL_NETLINK:
2488 switch (optname) {
2489 case NETLINK_PKTINFO:
2490 case NETLINK_ADD_MEMBERSHIP:
2491 case NETLINK_DROP_MEMBERSHIP:
2492 case NETLINK_BROADCAST_ERROR:
2493 case NETLINK_NO_ENOBUFS:
2494 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0)
2495 case NETLINK_LISTEN_ALL_NSID:
2496 case NETLINK_CAP_ACK:
2497 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0) */
2498 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0)
2499 case NETLINK_EXT_ACK:
2500 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0) */
2501 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 20, 0)
2502 case NETLINK_GET_STRICT_CHK:
2503 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0) */
2504 break;
2505 default:
2506 goto unimplemented;
2508 val = 0;
2509 if (optlen < sizeof(uint32_t)) {
2510 return -TARGET_EINVAL;
2512 if (get_user_u32(val, optval_addr)) {
2513 return -TARGET_EFAULT;
2515 ret = get_errno(setsockopt(sockfd, SOL_NETLINK, optname, &val,
2516 sizeof(val)));
2517 break;
2518 #endif /* SOL_NETLINK */
2519 default:
2520 unimplemented:
2521 qemu_log_mask(LOG_UNIMP, "Unsupported setsockopt level=%d optname=%d\n",
2522 level, optname);
2523 ret = -TARGET_ENOPROTOOPT;
2525 return ret;
2528 /* do_getsockopt() Must return target values and target errnos. */
2529 static abi_long do_getsockopt(int sockfd, int level, int optname,
2530 abi_ulong optval_addr, abi_ulong optlen)
2532 abi_long ret;
2533 int len, val;
2534 socklen_t lv;
2536 switch(level) {
2537 case TARGET_SOL_SOCKET:
2538 level = SOL_SOCKET;
2539 switch (optname) {
2540 /* These don't just return a single integer */
2541 case TARGET_SO_PEERNAME:
2542 goto unimplemented;
2543 case TARGET_SO_RCVTIMEO: {
2544 struct timeval tv;
2545 socklen_t tvlen;
2547 optname = SO_RCVTIMEO;
2549 get_timeout:
2550 if (get_user_u32(len, optlen)) {
2551 return -TARGET_EFAULT;
2553 if (len < 0) {
2554 return -TARGET_EINVAL;
2557 tvlen = sizeof(tv);
2558 ret = get_errno(getsockopt(sockfd, level, optname,
2559 &tv, &tvlen));
2560 if (ret < 0) {
2561 return ret;
2563 if (len > sizeof(struct target_timeval)) {
2564 len = sizeof(struct target_timeval);
2566 if (copy_to_user_timeval(optval_addr, &tv)) {
2567 return -TARGET_EFAULT;
2569 if (put_user_u32(len, optlen)) {
2570 return -TARGET_EFAULT;
2572 break;
2574 case TARGET_SO_SNDTIMEO:
2575 optname = SO_SNDTIMEO;
2576 goto get_timeout;
2577 case TARGET_SO_PEERCRED: {
2578 struct ucred cr;
2579 socklen_t crlen;
2580 struct target_ucred *tcr;
2582 if (get_user_u32(len, optlen)) {
2583 return -TARGET_EFAULT;
2585 if (len < 0) {
2586 return -TARGET_EINVAL;
2589 crlen = sizeof(cr);
2590 ret = get_errno(getsockopt(sockfd, level, SO_PEERCRED,
2591 &cr, &crlen));
2592 if (ret < 0) {
2593 return ret;
2595 if (len > crlen) {
2596 len = crlen;
2598 if (!lock_user_struct(VERIFY_WRITE, tcr, optval_addr, 0)) {
2599 return -TARGET_EFAULT;
2601 __put_user(cr.pid, &tcr->pid);
2602 __put_user(cr.uid, &tcr->uid);
2603 __put_user(cr.gid, &tcr->gid);
2604 unlock_user_struct(tcr, optval_addr, 1);
2605 if (put_user_u32(len, optlen)) {
2606 return -TARGET_EFAULT;
2608 break;
2610 case TARGET_SO_PEERSEC: {
2611 char *name;
2613 if (get_user_u32(len, optlen)) {
2614 return -TARGET_EFAULT;
2616 if (len < 0) {
2617 return -TARGET_EINVAL;
2619 name = lock_user(VERIFY_WRITE, optval_addr, len, 0);
2620 if (!name) {
2621 return -TARGET_EFAULT;
2623 lv = len;
2624 ret = get_errno(getsockopt(sockfd, level, SO_PEERSEC,
2625 name, &lv));
2626 if (put_user_u32(lv, optlen)) {
2627 ret = -TARGET_EFAULT;
2629 unlock_user(name, optval_addr, lv);
2630 break;
2632 case TARGET_SO_LINGER:
2634 struct linger lg;
2635 socklen_t lglen;
2636 struct target_linger *tlg;
2638 if (get_user_u32(len, optlen)) {
2639 return -TARGET_EFAULT;
2641 if (len < 0) {
2642 return -TARGET_EINVAL;
2645 lglen = sizeof(lg);
2646 ret = get_errno(getsockopt(sockfd, level, SO_LINGER,
2647 &lg, &lglen));
2648 if (ret < 0) {
2649 return ret;
2651 if (len > lglen) {
2652 len = lglen;
2654 if (!lock_user_struct(VERIFY_WRITE, tlg, optval_addr, 0)) {
2655 return -TARGET_EFAULT;
2657 __put_user(lg.l_onoff, &tlg->l_onoff);
2658 __put_user(lg.l_linger, &tlg->l_linger);
2659 unlock_user_struct(tlg, optval_addr, 1);
2660 if (put_user_u32(len, optlen)) {
2661 return -TARGET_EFAULT;
2663 break;
2665 /* Options with 'int' argument. */
2666 case TARGET_SO_DEBUG:
2667 optname = SO_DEBUG;
2668 goto int_case;
2669 case TARGET_SO_REUSEADDR:
2670 optname = SO_REUSEADDR;
2671 goto int_case;
2672 #ifdef SO_REUSEPORT
2673 case TARGET_SO_REUSEPORT:
2674 optname = SO_REUSEPORT;
2675 goto int_case;
2676 #endif
2677 case TARGET_SO_TYPE:
2678 optname = SO_TYPE;
2679 goto int_case;
2680 case TARGET_SO_ERROR:
2681 optname = SO_ERROR;
2682 goto int_case;
2683 case TARGET_SO_DONTROUTE:
2684 optname = SO_DONTROUTE;
2685 goto int_case;
2686 case TARGET_SO_BROADCAST:
2687 optname = SO_BROADCAST;
2688 goto int_case;
2689 case TARGET_SO_SNDBUF:
2690 optname = SO_SNDBUF;
2691 goto int_case;
2692 case TARGET_SO_RCVBUF:
2693 optname = SO_RCVBUF;
2694 goto int_case;
2695 case TARGET_SO_KEEPALIVE:
2696 optname = SO_KEEPALIVE;
2697 goto int_case;
2698 case TARGET_SO_OOBINLINE:
2699 optname = SO_OOBINLINE;
2700 goto int_case;
2701 case TARGET_SO_NO_CHECK:
2702 optname = SO_NO_CHECK;
2703 goto int_case;
2704 case TARGET_SO_PRIORITY:
2705 optname = SO_PRIORITY;
2706 goto int_case;
2707 #ifdef SO_BSDCOMPAT
2708 case TARGET_SO_BSDCOMPAT:
2709 optname = SO_BSDCOMPAT;
2710 goto int_case;
2711 #endif
2712 case TARGET_SO_PASSCRED:
2713 optname = SO_PASSCRED;
2714 goto int_case;
2715 case TARGET_SO_TIMESTAMP:
2716 optname = SO_TIMESTAMP;
2717 goto int_case;
2718 case TARGET_SO_RCVLOWAT:
2719 optname = SO_RCVLOWAT;
2720 goto int_case;
2721 case TARGET_SO_ACCEPTCONN:
2722 optname = SO_ACCEPTCONN;
2723 goto int_case;
2724 case TARGET_SO_PROTOCOL:
2725 optname = SO_PROTOCOL;
2726 goto int_case;
2727 case TARGET_SO_DOMAIN:
2728 optname = SO_DOMAIN;
2729 goto int_case;
2730 default:
2731 goto int_case;
2733 break;
2734 case SOL_TCP:
2735 case SOL_UDP:
2736 /* TCP and UDP options all take an 'int' value. */
2737 int_case:
2738 if (get_user_u32(len, optlen))
2739 return -TARGET_EFAULT;
2740 if (len < 0)
2741 return -TARGET_EINVAL;
2742 lv = sizeof(lv);
2743 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv));
2744 if (ret < 0)
2745 return ret;
2746 if (optname == SO_TYPE) {
2747 val = host_to_target_sock_type(val);
2749 if (len > lv)
2750 len = lv;
2751 if (len == 4) {
2752 if (put_user_u32(val, optval_addr))
2753 return -TARGET_EFAULT;
2754 } else {
2755 if (put_user_u8(val, optval_addr))
2756 return -TARGET_EFAULT;
2758 if (put_user_u32(len, optlen))
2759 return -TARGET_EFAULT;
2760 break;
2761 case SOL_IP:
2762 switch(optname) {
2763 case IP_TOS:
2764 case IP_TTL:
2765 case IP_HDRINCL:
2766 case IP_ROUTER_ALERT:
2767 case IP_RECVOPTS:
2768 case IP_RETOPTS:
2769 case IP_PKTINFO:
2770 case IP_MTU_DISCOVER:
2771 case IP_RECVERR:
2772 case IP_RECVTOS:
2773 #ifdef IP_FREEBIND
2774 case IP_FREEBIND:
2775 #endif
2776 case IP_MULTICAST_TTL:
2777 case IP_MULTICAST_LOOP:
2778 if (get_user_u32(len, optlen))
2779 return -TARGET_EFAULT;
2780 if (len < 0)
2781 return -TARGET_EINVAL;
2782 lv = sizeof(lv);
2783 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv));
2784 if (ret < 0)
2785 return ret;
2786 if (len < sizeof(int) && len > 0 && val >= 0 && val < 255) {
2787 len = 1;
2788 if (put_user_u32(len, optlen)
2789 || put_user_u8(val, optval_addr))
2790 return -TARGET_EFAULT;
2791 } else {
2792 if (len > sizeof(int))
2793 len = sizeof(int);
2794 if (put_user_u32(len, optlen)
2795 || put_user_u32(val, optval_addr))
2796 return -TARGET_EFAULT;
2798 break;
2799 default:
2800 ret = -TARGET_ENOPROTOOPT;
2801 break;
2803 break;
2804 case SOL_IPV6:
2805 switch (optname) {
2806 case IPV6_MTU_DISCOVER:
2807 case IPV6_MTU:
2808 case IPV6_V6ONLY:
2809 case IPV6_RECVPKTINFO:
2810 case IPV6_UNICAST_HOPS:
2811 case IPV6_MULTICAST_HOPS:
2812 case IPV6_MULTICAST_LOOP:
2813 case IPV6_RECVERR:
2814 case IPV6_RECVHOPLIMIT:
2815 case IPV6_2292HOPLIMIT:
2816 case IPV6_CHECKSUM:
2817 case IPV6_ADDRFORM:
2818 case IPV6_2292PKTINFO:
2819 case IPV6_RECVTCLASS:
2820 case IPV6_RECVRTHDR:
2821 case IPV6_2292RTHDR:
2822 case IPV6_RECVHOPOPTS:
2823 case IPV6_2292HOPOPTS:
2824 case IPV6_RECVDSTOPTS:
2825 case IPV6_2292DSTOPTS:
2826 case IPV6_TCLASS:
2827 case IPV6_ADDR_PREFERENCES:
2828 #ifdef IPV6_RECVPATHMTU
2829 case IPV6_RECVPATHMTU:
2830 #endif
2831 #ifdef IPV6_TRANSPARENT
2832 case IPV6_TRANSPARENT:
2833 #endif
2834 #ifdef IPV6_FREEBIND
2835 case IPV6_FREEBIND:
2836 #endif
2837 #ifdef IPV6_RECVORIGDSTADDR
2838 case IPV6_RECVORIGDSTADDR:
2839 #endif
2840 if (get_user_u32(len, optlen))
2841 return -TARGET_EFAULT;
2842 if (len < 0)
2843 return -TARGET_EINVAL;
2844 lv = sizeof(lv);
2845 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv));
2846 if (ret < 0)
2847 return ret;
2848 if (len < sizeof(int) && len > 0 && val >= 0 && val < 255) {
2849 len = 1;
2850 if (put_user_u32(len, optlen)
2851 || put_user_u8(val, optval_addr))
2852 return -TARGET_EFAULT;
2853 } else {
2854 if (len > sizeof(int))
2855 len = sizeof(int);
2856 if (put_user_u32(len, optlen)
2857 || put_user_u32(val, optval_addr))
2858 return -TARGET_EFAULT;
2860 break;
2861 default:
2862 ret = -TARGET_ENOPROTOOPT;
2863 break;
2865 break;
2866 #ifdef SOL_NETLINK
2867 case SOL_NETLINK:
2868 switch (optname) {
2869 case NETLINK_PKTINFO:
2870 case NETLINK_BROADCAST_ERROR:
2871 case NETLINK_NO_ENOBUFS:
2872 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0)
2873 case NETLINK_LISTEN_ALL_NSID:
2874 case NETLINK_CAP_ACK:
2875 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0) */
2876 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0)
2877 case NETLINK_EXT_ACK:
2878 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0) */
2879 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 20, 0)
2880 case NETLINK_GET_STRICT_CHK:
2881 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 12, 0) */
2882 if (get_user_u32(len, optlen)) {
2883 return -TARGET_EFAULT;
2885 if (len != sizeof(val)) {
2886 return -TARGET_EINVAL;
2888 lv = len;
2889 ret = get_errno(getsockopt(sockfd, level, optname, &val, &lv));
2890 if (ret < 0) {
2891 return ret;
2893 if (put_user_u32(lv, optlen)
2894 || put_user_u32(val, optval_addr)) {
2895 return -TARGET_EFAULT;
2897 break;
2898 #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0)
2899 case NETLINK_LIST_MEMBERSHIPS:
2901 uint32_t *results;
2902 int i;
2903 if (get_user_u32(len, optlen)) {
2904 return -TARGET_EFAULT;
2906 if (len < 0) {
2907 return -TARGET_EINVAL;
2909 results = lock_user(VERIFY_WRITE, optval_addr, len, 1);
2910 if (!results && len > 0) {
2911 return -TARGET_EFAULT;
2913 lv = len;
2914 ret = get_errno(getsockopt(sockfd, level, optname, results, &lv));
2915 if (ret < 0) {
2916 unlock_user(results, optval_addr, 0);
2917 return ret;
2919 /* swap host endianess to target endianess. */
2920 for (i = 0; i < (len / sizeof(uint32_t)); i++) {
2921 results[i] = tswap32(results[i]);
2923 if (put_user_u32(lv, optlen)) {
2924 return -TARGET_EFAULT;
2926 unlock_user(results, optval_addr, 0);
2927 break;
2929 #endif /* LINUX_VERSION_CODE >= KERNEL_VERSION(4, 2, 0) */
2930 default:
2931 goto unimplemented;
2933 break;
2934 #endif /* SOL_NETLINK */
2935 default:
2936 unimplemented:
2937 qemu_log_mask(LOG_UNIMP,
2938 "getsockopt level=%d optname=%d not yet supported\n",
2939 level, optname);
2940 ret = -TARGET_EOPNOTSUPP;
2941 break;
2943 return ret;
2946 /* Convert target low/high pair representing file offset into the host
2947 * low/high pair. This function doesn't handle offsets bigger than 64 bits
2948 * as the kernel doesn't handle them either.
2950 static void target_to_host_low_high(abi_ulong tlow,
2951 abi_ulong thigh,
2952 unsigned long *hlow,
2953 unsigned long *hhigh)
2955 uint64_t off = tlow |
2956 ((unsigned long long)thigh << TARGET_LONG_BITS / 2) <<
2957 TARGET_LONG_BITS / 2;
2959 *hlow = off;
2960 *hhigh = (off >> HOST_LONG_BITS / 2) >> HOST_LONG_BITS / 2;
2963 static struct iovec *lock_iovec(int type, abi_ulong target_addr,
2964 abi_ulong count, int copy)
2966 struct target_iovec *target_vec;
2967 struct iovec *vec;
2968 abi_ulong total_len, max_len;
2969 int i;
2970 int err = 0;
2971 bool bad_address = false;
2973 if (count == 0) {
2974 errno = 0;
2975 return NULL;
2977 if (count > IOV_MAX) {
2978 errno = EINVAL;
2979 return NULL;
2982 vec = g_try_new0(struct iovec, count);
2983 if (vec == NULL) {
2984 errno = ENOMEM;
2985 return NULL;
2988 target_vec = lock_user(VERIFY_READ, target_addr,
2989 count * sizeof(struct target_iovec), 1);
2990 if (target_vec == NULL) {
2991 err = EFAULT;
2992 goto fail2;
2995 /* ??? If host page size > target page size, this will result in a
2996 value larger than what we can actually support. */
2997 max_len = 0x7fffffff & TARGET_PAGE_MASK;
2998 total_len = 0;
3000 for (i = 0; i < count; i++) {
3001 abi_ulong base = tswapal(target_vec[i].iov_base);
3002 abi_long len = tswapal(target_vec[i].iov_len);
3004 if (len < 0) {
3005 err = EINVAL;
3006 goto fail;
3007 } else if (len == 0) {
3008 /* Zero length pointer is ignored. */
3009 vec[i].iov_base = 0;
3010 } else {
3011 vec[i].iov_base = lock_user(type, base, len, copy);
3012 /* If the first buffer pointer is bad, this is a fault. But
3013 * subsequent bad buffers will result in a partial write; this
3014 * is realized by filling the vector with null pointers and
3015 * zero lengths. */
3016 if (!vec[i].iov_base) {
3017 if (i == 0) {
3018 err = EFAULT;
3019 goto fail;
3020 } else {
3021 bad_address = true;
3024 if (bad_address) {
3025 len = 0;
3027 if (len > max_len - total_len) {
3028 len = max_len - total_len;
3031 vec[i].iov_len = len;
3032 total_len += len;
3035 unlock_user(target_vec, target_addr, 0);
3036 return vec;
3038 fail:
3039 while (--i >= 0) {
3040 if (tswapal(target_vec[i].iov_len) > 0) {
3041 unlock_user(vec[i].iov_base, tswapal(target_vec[i].iov_base), 0);
3044 unlock_user(target_vec, target_addr, 0);
3045 fail2:
3046 g_free(vec);
3047 errno = err;
3048 return NULL;
3051 static void unlock_iovec(struct iovec *vec, abi_ulong target_addr,
3052 abi_ulong count, int copy)
3054 struct target_iovec *target_vec;
3055 int i;
3057 target_vec = lock_user(VERIFY_READ, target_addr,
3058 count * sizeof(struct target_iovec), 1);
3059 if (target_vec) {
3060 for (i = 0; i < count; i++) {
3061 abi_ulong base = tswapal(target_vec[i].iov_base);
3062 abi_long len = tswapal(target_vec[i].iov_len);
3063 if (len < 0) {
3064 break;
3066 unlock_user(vec[i].iov_base, base, copy ? vec[i].iov_len : 0);
3068 unlock_user(target_vec, target_addr, 0);
3071 g_free(vec);
3074 static inline int target_to_host_sock_type(int *type)
3076 int host_type = 0;
3077 int target_type = *type;
3079 switch (target_type & TARGET_SOCK_TYPE_MASK) {
3080 case TARGET_SOCK_DGRAM:
3081 host_type = SOCK_DGRAM;
3082 break;
3083 case TARGET_SOCK_STREAM:
3084 host_type = SOCK_STREAM;
3085 break;
3086 default:
3087 host_type = target_type & TARGET_SOCK_TYPE_MASK;
3088 break;
3090 if (target_type & TARGET_SOCK_CLOEXEC) {
3091 #if defined(SOCK_CLOEXEC)
3092 host_type |= SOCK_CLOEXEC;
3093 #else
3094 return -TARGET_EINVAL;
3095 #endif
3097 if (target_type & TARGET_SOCK_NONBLOCK) {
3098 #if defined(SOCK_NONBLOCK)
3099 host_type |= SOCK_NONBLOCK;
3100 #elif !defined(O_NONBLOCK)
3101 return -TARGET_EINVAL;
3102 #endif
3104 *type = host_type;
3105 return 0;
3108 /* Try to emulate socket type flags after socket creation. */
3109 static int sock_flags_fixup(int fd, int target_type)
3111 #if !defined(SOCK_NONBLOCK) && defined(O_NONBLOCK)
3112 if (target_type & TARGET_SOCK_NONBLOCK) {
3113 int flags = fcntl(fd, F_GETFL);
3114 if (fcntl(fd, F_SETFL, O_NONBLOCK | flags) == -1) {
3115 close(fd);
3116 return -TARGET_EINVAL;
3119 #endif
3120 return fd;
3123 /* do_socket() Must return target values and target errnos. */
3124 static abi_long do_socket(int domain, int type, int protocol)
3126 int target_type = type;
3127 int ret;
3129 ret = target_to_host_sock_type(&type);
3130 if (ret) {
3131 return ret;
3134 if (domain == PF_NETLINK && !(
3135 #ifdef CONFIG_RTNETLINK
3136 protocol == NETLINK_ROUTE ||
3137 #endif
3138 protocol == NETLINK_KOBJECT_UEVENT ||
3139 protocol == NETLINK_AUDIT)) {
3140 return -TARGET_EPROTONOSUPPORT;
3143 if (domain == AF_PACKET ||
3144 (domain == AF_INET && type == SOCK_PACKET)) {
3145 protocol = tswap16(protocol);
3148 ret = get_errno(socket(domain, type, protocol));
3149 if (ret >= 0) {
3150 ret = sock_flags_fixup(ret, target_type);
3151 if (type == SOCK_PACKET) {
3152 /* Manage an obsolete case :
3153 * if socket type is SOCK_PACKET, bind by name
3155 fd_trans_register(ret, &target_packet_trans);
3156 } else if (domain == PF_NETLINK) {
3157 switch (protocol) {
3158 #ifdef CONFIG_RTNETLINK
3159 case NETLINK_ROUTE:
3160 fd_trans_register(ret, &target_netlink_route_trans);
3161 break;
3162 #endif
3163 case NETLINK_KOBJECT_UEVENT:
3164 /* nothing to do: messages are strings */
3165 break;
3166 case NETLINK_AUDIT:
3167 fd_trans_register(ret, &target_netlink_audit_trans);
3168 break;
3169 default:
3170 g_assert_not_reached();
3174 return ret;
3177 /* do_bind() Must return target values and target errnos. */
3178 static abi_long do_bind(int sockfd, abi_ulong target_addr,
3179 socklen_t addrlen)
3181 void *addr;
3182 abi_long ret;
3184 if ((int)addrlen < 0) {
3185 return -TARGET_EINVAL;
3188 addr = alloca(addrlen+1);
3190 ret = target_to_host_sockaddr(sockfd, addr, target_addr, addrlen);
3191 if (ret)
3192 return ret;
3194 return get_errno(bind(sockfd, addr, addrlen));
3197 /* do_connect() Must return target values and target errnos. */
3198 static abi_long do_connect(int sockfd, abi_ulong target_addr,
3199 socklen_t addrlen)
3201 void *addr;
3202 abi_long ret;
3204 if ((int)addrlen < 0) {
3205 return -TARGET_EINVAL;
3208 addr = alloca(addrlen+1);
3210 ret = target_to_host_sockaddr(sockfd, addr, target_addr, addrlen);
3211 if (ret)
3212 return ret;
3214 return get_errno(safe_connect(sockfd, addr, addrlen));
3217 /* do_sendrecvmsg_locked() Must return target values and target errnos. */
3218 static abi_long do_sendrecvmsg_locked(int fd, struct target_msghdr *msgp,
3219 int flags, int send)
3221 abi_long ret, len;
3222 struct msghdr msg;
3223 abi_ulong count;
3224 struct iovec *vec;
3225 abi_ulong target_vec;
3227 if (msgp->msg_name) {
3228 msg.msg_namelen = tswap32(msgp->msg_namelen);
3229 msg.msg_name = alloca(msg.msg_namelen+1);
3230 ret = target_to_host_sockaddr(fd, msg.msg_name,
3231 tswapal(msgp->msg_name),
3232 msg.msg_namelen);
3233 if (ret == -TARGET_EFAULT) {
3234 /* For connected sockets msg_name and msg_namelen must
3235 * be ignored, so returning EFAULT immediately is wrong.
3236 * Instead, pass a bad msg_name to the host kernel, and
3237 * let it decide whether to return EFAULT or not.
3239 msg.msg_name = (void *)-1;
3240 } else if (ret) {
3241 goto out2;
3243 } else {
3244 msg.msg_name = NULL;
3245 msg.msg_namelen = 0;
3247 msg.msg_controllen = 2 * tswapal(msgp->msg_controllen);
3248 msg.msg_control = alloca(msg.msg_controllen);
3249 memset(msg.msg_control, 0, msg.msg_controllen);
3251 msg.msg_flags = tswap32(msgp->msg_flags);
3253 count = tswapal(msgp->msg_iovlen);
3254 target_vec = tswapal(msgp->msg_iov);
3256 if (count > IOV_MAX) {
3257 /* sendrcvmsg returns a different errno for this condition than
3258 * readv/writev, so we must catch it here before lock_iovec() does.
3260 ret = -TARGET_EMSGSIZE;
3261 goto out2;
3264 vec = lock_iovec(send ? VERIFY_READ : VERIFY_WRITE,
3265 target_vec, count, send);
3266 if (vec == NULL) {
3267 ret = -host_to_target_errno(errno);
3268 goto out2;
3270 msg.msg_iovlen = count;
3271 msg.msg_iov = vec;
3273 if (send) {
3274 if (fd_trans_target_to_host_data(fd)) {
3275 void *host_msg;
3277 host_msg = g_malloc(msg.msg_iov->iov_len);
3278 memcpy(host_msg, msg.msg_iov->iov_base, msg.msg_iov->iov_len);
3279 ret = fd_trans_target_to_host_data(fd)(host_msg,
3280 msg.msg_iov->iov_len);
3281 if (ret >= 0) {
3282 msg.msg_iov->iov_base = host_msg;
3283 ret = get_errno(safe_sendmsg(fd, &msg, flags));
3285 g_free(host_msg);
3286 } else {
3287 ret = target_to_host_cmsg(&msg, msgp);
3288 if (ret == 0) {
3289 ret = get_errno(safe_sendmsg(fd, &msg, flags));
3292 } else {
3293 ret = get_errno(safe_recvmsg(fd, &msg, flags));
3294 if (!is_error(ret)) {
3295 len = ret;
3296 if (fd_trans_host_to_target_data(fd)) {
3297 ret = fd_trans_host_to_target_data(fd)(msg.msg_iov->iov_base,
3298 MIN(msg.msg_iov->iov_len, len));
3299 } else {
3300 ret = host_to_target_cmsg(msgp, &msg);
3302 if (!is_error(ret)) {
3303 msgp->msg_namelen = tswap32(msg.msg_namelen);
3304 msgp->msg_flags = tswap32(msg.msg_flags);
3305 if (msg.msg_name != NULL && msg.msg_name != (void *)-1) {
3306 ret = host_to_target_sockaddr(tswapal(msgp->msg_name),
3307 msg.msg_name, msg.msg_namelen);
3308 if (ret) {
3309 goto out;
3313 ret = len;
3318 out:
3319 unlock_iovec(vec, target_vec, count, !send);
3320 out2:
3321 return ret;
3324 static abi_long do_sendrecvmsg(int fd, abi_ulong target_msg,
3325 int flags, int send)
3327 abi_long ret;
3328 struct target_msghdr *msgp;
3330 if (!lock_user_struct(send ? VERIFY_READ : VERIFY_WRITE,
3331 msgp,
3332 target_msg,
3333 send ? 1 : 0)) {
3334 return -TARGET_EFAULT;
3336 ret = do_sendrecvmsg_locked(fd, msgp, flags, send);
3337 unlock_user_struct(msgp, target_msg, send ? 0 : 1);
3338 return ret;
3341 /* We don't rely on the C library to have sendmmsg/recvmmsg support,
3342 * so it might not have this *mmsg-specific flag either.
3344 #ifndef MSG_WAITFORONE
3345 #define MSG_WAITFORONE 0x10000
3346 #endif
3348 static abi_long do_sendrecvmmsg(int fd, abi_ulong target_msgvec,
3349 unsigned int vlen, unsigned int flags,
3350 int send)
3352 struct target_mmsghdr *mmsgp;
3353 abi_long ret = 0;
3354 int i;
3356 if (vlen > UIO_MAXIOV) {
3357 vlen = UIO_MAXIOV;
3360 mmsgp = lock_user(VERIFY_WRITE, target_msgvec, sizeof(*mmsgp) * vlen, 1);
3361 if (!mmsgp) {
3362 return -TARGET_EFAULT;
3365 for (i = 0; i < vlen; i++) {
3366 ret = do_sendrecvmsg_locked(fd, &mmsgp[i].msg_hdr, flags, send);
3367 if (is_error(ret)) {
3368 break;
3370 mmsgp[i].msg_len = tswap32(ret);
3371 /* MSG_WAITFORONE turns on MSG_DONTWAIT after one packet */
3372 if (flags & MSG_WAITFORONE) {
3373 flags |= MSG_DONTWAIT;
3377 unlock_user(mmsgp, target_msgvec, sizeof(*mmsgp) * i);
3379 /* Return number of datagrams sent if we sent any at all;
3380 * otherwise return the error.
3382 if (i) {
3383 return i;
3385 return ret;
3388 /* do_accept4() Must return target values and target errnos. */
3389 static abi_long do_accept4(int fd, abi_ulong target_addr,
3390 abi_ulong target_addrlen_addr, int flags)
3392 socklen_t addrlen, ret_addrlen;
3393 void *addr;
3394 abi_long ret;
3395 int host_flags;
3397 host_flags = target_to_host_bitmask(flags, fcntl_flags_tbl);
3399 if (target_addr == 0) {
3400 return get_errno(safe_accept4(fd, NULL, NULL, host_flags));
3403 /* linux returns EFAULT if addrlen pointer is invalid */
3404 if (get_user_u32(addrlen, target_addrlen_addr))
3405 return -TARGET_EFAULT;
3407 if ((int)addrlen < 0) {
3408 return -TARGET_EINVAL;
3411 if (!access_ok(thread_cpu, VERIFY_WRITE, target_addr, addrlen)) {
3412 return -TARGET_EFAULT;
3415 addr = alloca(addrlen);
3417 ret_addrlen = addrlen;
3418 ret = get_errno(safe_accept4(fd, addr, &ret_addrlen, host_flags));
3419 if (!is_error(ret)) {
3420 host_to_target_sockaddr(target_addr, addr, MIN(addrlen, ret_addrlen));
3421 if (put_user_u32(ret_addrlen, target_addrlen_addr)) {
3422 ret = -TARGET_EFAULT;
3425 return ret;
3428 /* do_getpeername() Must return target values and target errnos. */
3429 static abi_long do_getpeername(int fd, abi_ulong target_addr,
3430 abi_ulong target_addrlen_addr)
3432 socklen_t addrlen, ret_addrlen;
3433 void *addr;
3434 abi_long ret;
3436 if (get_user_u32(addrlen, target_addrlen_addr))
3437 return -TARGET_EFAULT;
3439 if ((int)addrlen < 0) {
3440 return -TARGET_EINVAL;
3443 if (!access_ok(thread_cpu, VERIFY_WRITE, target_addr, addrlen)) {
3444 return -TARGET_EFAULT;
3447 addr = alloca(addrlen);
3449 ret_addrlen = addrlen;
3450 ret = get_errno(getpeername(fd, addr, &ret_addrlen));
3451 if (!is_error(ret)) {
3452 host_to_target_sockaddr(target_addr, addr, MIN(addrlen, ret_addrlen));
3453 if (put_user_u32(ret_addrlen, target_addrlen_addr)) {
3454 ret = -TARGET_EFAULT;
3457 return ret;
3460 /* do_getsockname() Must return target values and target errnos. */
3461 static abi_long do_getsockname(int fd, abi_ulong target_addr,
3462 abi_ulong target_addrlen_addr)
3464 socklen_t addrlen, ret_addrlen;
3465 void *addr;
3466 abi_long ret;
3468 if (get_user_u32(addrlen, target_addrlen_addr))
3469 return -TARGET_EFAULT;
3471 if ((int)addrlen < 0) {
3472 return -TARGET_EINVAL;
3475 if (!access_ok(thread_cpu, VERIFY_WRITE, target_addr, addrlen)) {
3476 return -TARGET_EFAULT;
3479 addr = alloca(addrlen);
3481 ret_addrlen = addrlen;
3482 ret = get_errno(getsockname(fd, addr, &ret_addrlen));
3483 if (!is_error(ret)) {
3484 host_to_target_sockaddr(target_addr, addr, MIN(addrlen, ret_addrlen));
3485 if (put_user_u32(ret_addrlen, target_addrlen_addr)) {
3486 ret = -TARGET_EFAULT;
3489 return ret;
3492 /* do_socketpair() Must return target values and target errnos. */
3493 static abi_long do_socketpair(int domain, int type, int protocol,
3494 abi_ulong target_tab_addr)
3496 int tab[2];
3497 abi_long ret;
3499 target_to_host_sock_type(&type);
3501 ret = get_errno(socketpair(domain, type, protocol, tab));
3502 if (!is_error(ret)) {
3503 if (put_user_s32(tab[0], target_tab_addr)
3504 || put_user_s32(tab[1], target_tab_addr + sizeof(tab[0])))
3505 ret = -TARGET_EFAULT;
3507 return ret;
3510 /* do_sendto() Must return target values and target errnos. */
3511 static abi_long do_sendto(int fd, abi_ulong msg, size_t len, int flags,
3512 abi_ulong target_addr, socklen_t addrlen)
3514 void *addr;
3515 void *host_msg;
3516 void *copy_msg = NULL;
3517 abi_long ret;
3519 if ((int)addrlen < 0) {
3520 return -TARGET_EINVAL;
3523 host_msg = lock_user(VERIFY_READ, msg, len, 1);
3524 if (!host_msg)
3525 return -TARGET_EFAULT;
3526 if (fd_trans_target_to_host_data(fd)) {
3527 copy_msg = host_msg;
3528 host_msg = g_malloc(len);
3529 memcpy(host_msg, copy_msg, len);
3530 ret = fd_trans_target_to_host_data(fd)(host_msg, len);
3531 if (ret < 0) {
3532 goto fail;
3535 if (target_addr) {
3536 addr = alloca(addrlen+1);
3537 ret = target_to_host_sockaddr(fd, addr, target_addr, addrlen);
3538 if (ret) {
3539 goto fail;
3541 ret = get_errno(safe_sendto(fd, host_msg, len, flags, addr, addrlen));
3542 } else {
3543 ret = get_errno(safe_sendto(fd, host_msg, len, flags, NULL, 0));
3545 fail:
3546 if (copy_msg) {
3547 g_free(host_msg);
3548 host_msg = copy_msg;
3550 unlock_user(host_msg, msg, 0);
3551 return ret;
3554 /* do_recvfrom() Must return target values and target errnos. */
3555 static abi_long do_recvfrom(int fd, abi_ulong msg, size_t len, int flags,
3556 abi_ulong target_addr,
3557 abi_ulong target_addrlen)
3559 socklen_t addrlen, ret_addrlen;
3560 void *addr;
3561 void *host_msg;
3562 abi_long ret;
3564 if (!msg) {
3565 host_msg = NULL;
3566 } else {
3567 host_msg = lock_user(VERIFY_WRITE, msg, len, 0);
3568 if (!host_msg) {
3569 return -TARGET_EFAULT;
3572 if (target_addr) {
3573 if (get_user_u32(addrlen, target_addrlen)) {
3574 ret = -TARGET_EFAULT;
3575 goto fail;
3577 if ((int)addrlen < 0) {
3578 ret = -TARGET_EINVAL;
3579 goto fail;
3581 addr = alloca(addrlen);
3582 ret_addrlen = addrlen;
3583 ret = get_errno(safe_recvfrom(fd, host_msg, len, flags,
3584 addr, &ret_addrlen));
3585 } else {
3586 addr = NULL; /* To keep compiler quiet. */
3587 addrlen = 0; /* To keep compiler quiet. */
3588 ret = get_errno(safe_recvfrom(fd, host_msg, len, flags, NULL, 0));
3590 if (!is_error(ret)) {
3591 if (fd_trans_host_to_target_data(fd)) {
3592 abi_long trans;
3593 trans = fd_trans_host_to_target_data(fd)(host_msg, MIN(ret, len));
3594 if (is_error(trans)) {
3595 ret = trans;
3596 goto fail;
3599 if (target_addr) {
3600 host_to_target_sockaddr(target_addr, addr,
3601 MIN(addrlen, ret_addrlen));
3602 if (put_user_u32(ret_addrlen, target_addrlen)) {
3603 ret = -TARGET_EFAULT;
3604 goto fail;
3607 unlock_user(host_msg, msg, len);
3608 } else {
3609 fail:
3610 unlock_user(host_msg, msg, 0);
3612 return ret;
3615 #ifdef TARGET_NR_socketcall
3616 /* do_socketcall() must return target values and target errnos. */
3617 static abi_long do_socketcall(int num, abi_ulong vptr)
3619 static const unsigned nargs[] = { /* number of arguments per operation */
3620 [TARGET_SYS_SOCKET] = 3, /* domain, type, protocol */
3621 [TARGET_SYS_BIND] = 3, /* fd, addr, addrlen */
3622 [TARGET_SYS_CONNECT] = 3, /* fd, addr, addrlen */
3623 [TARGET_SYS_LISTEN] = 2, /* fd, backlog */
3624 [TARGET_SYS_ACCEPT] = 3, /* fd, addr, addrlen */
3625 [TARGET_SYS_GETSOCKNAME] = 3, /* fd, addr, addrlen */
3626 [TARGET_SYS_GETPEERNAME] = 3, /* fd, addr, addrlen */
3627 [TARGET_SYS_SOCKETPAIR] = 4, /* domain, type, protocol, tab */
3628 [TARGET_SYS_SEND] = 4, /* fd, msg, len, flags */
3629 [TARGET_SYS_RECV] = 4, /* fd, msg, len, flags */
3630 [TARGET_SYS_SENDTO] = 6, /* fd, msg, len, flags, addr, addrlen */
3631 [TARGET_SYS_RECVFROM] = 6, /* fd, msg, len, flags, addr, addrlen */
3632 [TARGET_SYS_SHUTDOWN] = 2, /* fd, how */
3633 [TARGET_SYS_SETSOCKOPT] = 5, /* fd, level, optname, optval, optlen */
3634 [TARGET_SYS_GETSOCKOPT] = 5, /* fd, level, optname, optval, optlen */
3635 [TARGET_SYS_SENDMSG] = 3, /* fd, msg, flags */
3636 [TARGET_SYS_RECVMSG] = 3, /* fd, msg, flags */
3637 [TARGET_SYS_ACCEPT4] = 4, /* fd, addr, addrlen, flags */
3638 [TARGET_SYS_RECVMMSG] = 4, /* fd, msgvec, vlen, flags */
3639 [TARGET_SYS_SENDMMSG] = 4, /* fd, msgvec, vlen, flags */
3641 abi_long a[6]; /* max 6 args */
3642 unsigned i;
3644 /* check the range of the first argument num */
3645 /* (TARGET_SYS_SENDMMSG is the highest among TARGET_SYS_xxx) */
3646 if (num < 1 || num > TARGET_SYS_SENDMMSG) {
3647 return -TARGET_EINVAL;
3649 /* ensure we have space for args */
3650 if (nargs[num] > ARRAY_SIZE(a)) {
3651 return -TARGET_EINVAL;
3653 /* collect the arguments in a[] according to nargs[] */
3654 for (i = 0; i < nargs[num]; ++i) {
3655 if (get_user_ual(a[i], vptr + i * sizeof(abi_long)) != 0) {
3656 return -TARGET_EFAULT;
3659 /* now when we have the args, invoke the appropriate underlying function */
3660 switch (num) {
3661 case TARGET_SYS_SOCKET: /* domain, type, protocol */
3662 return do_socket(a[0], a[1], a[2]);
3663 case TARGET_SYS_BIND: /* sockfd, addr, addrlen */
3664 return do_bind(a[0], a[1], a[2]);
3665 case TARGET_SYS_CONNECT: /* sockfd, addr, addrlen */
3666 return do_connect(a[0], a[1], a[2]);
3667 case TARGET_SYS_LISTEN: /* sockfd, backlog */
3668 return get_errno(listen(a[0], a[1]));
3669 case TARGET_SYS_ACCEPT: /* sockfd, addr, addrlen */
3670 return do_accept4(a[0], a[1], a[2], 0);
3671 case TARGET_SYS_GETSOCKNAME: /* sockfd, addr, addrlen */
3672 return do_getsockname(a[0], a[1], a[2]);
3673 case TARGET_SYS_GETPEERNAME: /* sockfd, addr, addrlen */
3674 return do_getpeername(a[0], a[1], a[2]);
3675 case TARGET_SYS_SOCKETPAIR: /* domain, type, protocol, tab */
3676 return do_socketpair(a[0], a[1], a[2], a[3]);
3677 case TARGET_SYS_SEND: /* sockfd, msg, len, flags */
3678 return do_sendto(a[0], a[1], a[2], a[3], 0, 0);
3679 case TARGET_SYS_RECV: /* sockfd, msg, len, flags */
3680 return do_recvfrom(a[0], a[1], a[2], a[3], 0, 0);
3681 case TARGET_SYS_SENDTO: /* sockfd, msg, len, flags, addr, addrlen */
3682 return do_sendto(a[0], a[1], a[2], a[3], a[4], a[5]);
3683 case TARGET_SYS_RECVFROM: /* sockfd, msg, len, flags, addr, addrlen */
3684 return do_recvfrom(a[0], a[1], a[2], a[3], a[4], a[5]);
3685 case TARGET_SYS_SHUTDOWN: /* sockfd, how */
3686 return get_errno(shutdown(a[0], a[1]));
3687 case TARGET_SYS_SETSOCKOPT: /* sockfd, level, optname, optval, optlen */
3688 return do_setsockopt(a[0], a[1], a[2], a[3], a[4]);
3689 case TARGET_SYS_GETSOCKOPT: /* sockfd, level, optname, optval, optlen */
3690 return do_getsockopt(a[0], a[1], a[2], a[3], a[4]);
3691 case TARGET_SYS_SENDMSG: /* sockfd, msg, flags */
3692 return do_sendrecvmsg(a[0], a[1], a[2], 1);
3693 case TARGET_SYS_RECVMSG: /* sockfd, msg, flags */
3694 return do_sendrecvmsg(a[0], a[1], a[2], 0);
3695 case TARGET_SYS_ACCEPT4: /* sockfd, addr, addrlen, flags */
3696 return do_accept4(a[0], a[1], a[2], a[3]);
3697 case TARGET_SYS_RECVMMSG: /* sockfd, msgvec, vlen, flags */
3698 return do_sendrecvmmsg(a[0], a[1], a[2], a[3], 0);
3699 case TARGET_SYS_SENDMMSG: /* sockfd, msgvec, vlen, flags */
3700 return do_sendrecvmmsg(a[0], a[1], a[2], a[3], 1);
3701 default:
3702 qemu_log_mask(LOG_UNIMP, "Unsupported socketcall: %d\n", num);
3703 return -TARGET_EINVAL;
3706 #endif
3708 #define N_SHM_REGIONS 32
3710 static struct shm_region {
3711 abi_ulong start;
3712 abi_ulong size;
3713 bool in_use;
3714 } shm_regions[N_SHM_REGIONS];
3716 #ifndef TARGET_SEMID64_DS
3717 /* asm-generic version of this struct */
3718 struct target_semid64_ds
3720 struct target_ipc_perm sem_perm;
3721 abi_ulong sem_otime;
3722 #if TARGET_ABI_BITS == 32
3723 abi_ulong __unused1;
3724 #endif
3725 abi_ulong sem_ctime;
3726 #if TARGET_ABI_BITS == 32
3727 abi_ulong __unused2;
3728 #endif
3729 abi_ulong sem_nsems;
3730 abi_ulong __unused3;
3731 abi_ulong __unused4;
3733 #endif
3735 static inline abi_long target_to_host_ipc_perm(struct ipc_perm *host_ip,
3736 abi_ulong target_addr)
3738 struct target_ipc_perm *target_ip;
3739 struct target_semid64_ds *target_sd;
3741 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1))
3742 return -TARGET_EFAULT;
3743 target_ip = &(target_sd->sem_perm);
3744 host_ip->__key = tswap32(target_ip->__key);
3745 host_ip->uid = tswap32(target_ip->uid);
3746 host_ip->gid = tswap32(target_ip->gid);
3747 host_ip->cuid = tswap32(target_ip->cuid);
3748 host_ip->cgid = tswap32(target_ip->cgid);
3749 #if defined(TARGET_ALPHA) || defined(TARGET_MIPS) || defined(TARGET_PPC)
3750 host_ip->mode = tswap32(target_ip->mode);
3751 #else
3752 host_ip->mode = tswap16(target_ip->mode);
3753 #endif
3754 #if defined(TARGET_PPC)
3755 host_ip->__seq = tswap32(target_ip->__seq);
3756 #else
3757 host_ip->__seq = tswap16(target_ip->__seq);
3758 #endif
3759 unlock_user_struct(target_sd, target_addr, 0);
3760 return 0;
3763 static inline abi_long host_to_target_ipc_perm(abi_ulong target_addr,
3764 struct ipc_perm *host_ip)
3766 struct target_ipc_perm *target_ip;
3767 struct target_semid64_ds *target_sd;
3769 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0))
3770 return -TARGET_EFAULT;
3771 target_ip = &(target_sd->sem_perm);
3772 target_ip->__key = tswap32(host_ip->__key);
3773 target_ip->uid = tswap32(host_ip->uid);
3774 target_ip->gid = tswap32(host_ip->gid);
3775 target_ip->cuid = tswap32(host_ip->cuid);
3776 target_ip->cgid = tswap32(host_ip->cgid);
3777 #if defined(TARGET_ALPHA) || defined(TARGET_MIPS) || defined(TARGET_PPC)
3778 target_ip->mode = tswap32(host_ip->mode);
3779 #else
3780 target_ip->mode = tswap16(host_ip->mode);
3781 #endif
3782 #if defined(TARGET_PPC)
3783 target_ip->__seq = tswap32(host_ip->__seq);
3784 #else
3785 target_ip->__seq = tswap16(host_ip->__seq);
3786 #endif
3787 unlock_user_struct(target_sd, target_addr, 1);
3788 return 0;
3791 static inline abi_long target_to_host_semid_ds(struct semid_ds *host_sd,
3792 abi_ulong target_addr)
3794 struct target_semid64_ds *target_sd;
3796 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1))
3797 return -TARGET_EFAULT;
3798 if (target_to_host_ipc_perm(&(host_sd->sem_perm),target_addr))
3799 return -TARGET_EFAULT;
3800 host_sd->sem_nsems = tswapal(target_sd->sem_nsems);
3801 host_sd->sem_otime = tswapal(target_sd->sem_otime);
3802 host_sd->sem_ctime = tswapal(target_sd->sem_ctime);
3803 unlock_user_struct(target_sd, target_addr, 0);
3804 return 0;
3807 static inline abi_long host_to_target_semid_ds(abi_ulong target_addr,
3808 struct semid_ds *host_sd)
3810 struct target_semid64_ds *target_sd;
3812 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0))
3813 return -TARGET_EFAULT;
3814 if (host_to_target_ipc_perm(target_addr,&(host_sd->sem_perm)))
3815 return -TARGET_EFAULT;
3816 target_sd->sem_nsems = tswapal(host_sd->sem_nsems);
3817 target_sd->sem_otime = tswapal(host_sd->sem_otime);
3818 target_sd->sem_ctime = tswapal(host_sd->sem_ctime);
3819 unlock_user_struct(target_sd, target_addr, 1);
3820 return 0;
3823 struct target_seminfo {
3824 int semmap;
3825 int semmni;
3826 int semmns;
3827 int semmnu;
3828 int semmsl;
3829 int semopm;
3830 int semume;
3831 int semusz;
3832 int semvmx;
3833 int semaem;
3836 static inline abi_long host_to_target_seminfo(abi_ulong target_addr,
3837 struct seminfo *host_seminfo)
3839 struct target_seminfo *target_seminfo;
3840 if (!lock_user_struct(VERIFY_WRITE, target_seminfo, target_addr, 0))
3841 return -TARGET_EFAULT;
3842 __put_user(host_seminfo->semmap, &target_seminfo->semmap);
3843 __put_user(host_seminfo->semmni, &target_seminfo->semmni);
3844 __put_user(host_seminfo->semmns, &target_seminfo->semmns);
3845 __put_user(host_seminfo->semmnu, &target_seminfo->semmnu);
3846 __put_user(host_seminfo->semmsl, &target_seminfo->semmsl);
3847 __put_user(host_seminfo->semopm, &target_seminfo->semopm);
3848 __put_user(host_seminfo->semume, &target_seminfo->semume);
3849 __put_user(host_seminfo->semusz, &target_seminfo->semusz);
3850 __put_user(host_seminfo->semvmx, &target_seminfo->semvmx);
3851 __put_user(host_seminfo->semaem, &target_seminfo->semaem);
3852 unlock_user_struct(target_seminfo, target_addr, 1);
3853 return 0;
3856 union semun {
3857 int val;
3858 struct semid_ds *buf;
3859 unsigned short *array;
3860 struct seminfo *__buf;
3863 union target_semun {
3864 int val;
3865 abi_ulong buf;
3866 abi_ulong array;
3867 abi_ulong __buf;
3870 static inline abi_long target_to_host_semarray(int semid, unsigned short **host_array,
3871 abi_ulong target_addr)
3873 int nsems;
3874 unsigned short *array;
3875 union semun semun;
3876 struct semid_ds semid_ds;
3877 int i, ret;
3879 semun.buf = &semid_ds;
3881 ret = semctl(semid, 0, IPC_STAT, semun);
3882 if (ret == -1)
3883 return get_errno(ret);
3885 nsems = semid_ds.sem_nsems;
3887 *host_array = g_try_new(unsigned short, nsems);
3888 if (!*host_array) {
3889 return -TARGET_ENOMEM;
3891 array = lock_user(VERIFY_READ, target_addr,
3892 nsems*sizeof(unsigned short), 1);
3893 if (!array) {
3894 g_free(*host_array);
3895 return -TARGET_EFAULT;
3898 for(i=0; i<nsems; i++) {
3899 __get_user((*host_array)[i], &array[i]);
3901 unlock_user(array, target_addr, 0);
3903 return 0;
3906 static inline abi_long host_to_target_semarray(int semid, abi_ulong target_addr,
3907 unsigned short **host_array)
3909 int nsems;
3910 unsigned short *array;
3911 union semun semun;
3912 struct semid_ds semid_ds;
3913 int i, ret;
3915 semun.buf = &semid_ds;
3917 ret = semctl(semid, 0, IPC_STAT, semun);
3918 if (ret == -1)
3919 return get_errno(ret);
3921 nsems = semid_ds.sem_nsems;
3923 array = lock_user(VERIFY_WRITE, target_addr,
3924 nsems*sizeof(unsigned short), 0);
3925 if (!array)
3926 return -TARGET_EFAULT;
3928 for(i=0; i<nsems; i++) {
3929 __put_user((*host_array)[i], &array[i]);
3931 g_free(*host_array);
3932 unlock_user(array, target_addr, 1);
3934 return 0;
3937 static inline abi_long do_semctl(int semid, int semnum, int cmd,
3938 abi_ulong target_arg)
3940 union target_semun target_su = { .buf = target_arg };
3941 union semun arg;
3942 struct semid_ds dsarg;
3943 unsigned short *array = NULL;
3944 struct seminfo seminfo;
3945 abi_long ret = -TARGET_EINVAL;
3946 abi_long err;
3947 cmd &= 0xff;
3949 switch( cmd ) {
3950 case GETVAL:
3951 case SETVAL:
3952 /* In 64 bit cross-endian situations, we will erroneously pick up
3953 * the wrong half of the union for the "val" element. To rectify
3954 * this, the entire 8-byte structure is byteswapped, followed by
3955 * a swap of the 4 byte val field. In other cases, the data is
3956 * already in proper host byte order. */
3957 if (sizeof(target_su.val) != (sizeof(target_su.buf))) {
3958 target_su.buf = tswapal(target_su.buf);
3959 arg.val = tswap32(target_su.val);
3960 } else {
3961 arg.val = target_su.val;
3963 ret = get_errno(semctl(semid, semnum, cmd, arg));
3964 break;
3965 case GETALL:
3966 case SETALL:
3967 err = target_to_host_semarray(semid, &array, target_su.array);
3968 if (err)
3969 return err;
3970 arg.array = array;
3971 ret = get_errno(semctl(semid, semnum, cmd, arg));
3972 err = host_to_target_semarray(semid, target_su.array, &array);
3973 if (err)
3974 return err;
3975 break;
3976 case IPC_STAT:
3977 case IPC_SET:
3978 case SEM_STAT:
3979 err = target_to_host_semid_ds(&dsarg, target_su.buf);
3980 if (err)
3981 return err;
3982 arg.buf = &dsarg;
3983 ret = get_errno(semctl(semid, semnum, cmd, arg));
3984 err = host_to_target_semid_ds(target_su.buf, &dsarg);
3985 if (err)
3986 return err;
3987 break;
3988 case IPC_INFO:
3989 case SEM_INFO:
3990 arg.__buf = &seminfo;
3991 ret = get_errno(semctl(semid, semnum, cmd, arg));
3992 err = host_to_target_seminfo(target_su.__buf, &seminfo);
3993 if (err)
3994 return err;
3995 break;
3996 case IPC_RMID:
3997 case GETPID:
3998 case GETNCNT:
3999 case GETZCNT:
4000 ret = get_errno(semctl(semid, semnum, cmd, NULL));
4001 break;
4004 return ret;
4007 struct target_sembuf {
4008 unsigned short sem_num;
4009 short sem_op;
4010 short sem_flg;
4013 static inline abi_long target_to_host_sembuf(struct sembuf *host_sembuf,
4014 abi_ulong target_addr,
4015 unsigned nsops)
4017 struct target_sembuf *target_sembuf;
4018 int i;
4020 target_sembuf = lock_user(VERIFY_READ, target_addr,
4021 nsops*sizeof(struct target_sembuf), 1);
4022 if (!target_sembuf)
4023 return -TARGET_EFAULT;
4025 for(i=0; i<nsops; i++) {
4026 __get_user(host_sembuf[i].sem_num, &target_sembuf[i].sem_num);
4027 __get_user(host_sembuf[i].sem_op, &target_sembuf[i].sem_op);
4028 __get_user(host_sembuf[i].sem_flg, &target_sembuf[i].sem_flg);
4031 unlock_user(target_sembuf, target_addr, 0);
4033 return 0;
4036 #if defined(TARGET_NR_ipc) || defined(TARGET_NR_semop) || \
4037 defined(TARGET_NR_semtimedop) || defined(TARGET_NR_semtimedop_time64)
4040 * This macro is required to handle the s390 variants, which passes the
4041 * arguments in a different order than default.
4043 #ifdef __s390x__
4044 #define SEMTIMEDOP_IPC_ARGS(__nsops, __sops, __timeout) \
4045 (__nsops), (__timeout), (__sops)
4046 #else
4047 #define SEMTIMEDOP_IPC_ARGS(__nsops, __sops, __timeout) \
4048 (__nsops), 0, (__sops), (__timeout)
4049 #endif
4051 static inline abi_long do_semtimedop(int semid,
4052 abi_long ptr,
4053 unsigned nsops,
4054 abi_long timeout, bool time64)
4056 struct sembuf *sops;
4057 struct timespec ts, *pts = NULL;
4058 abi_long ret;
4060 if (timeout) {
4061 pts = &ts;
4062 if (time64) {
4063 if (target_to_host_timespec64(pts, timeout)) {
4064 return -TARGET_EFAULT;
4066 } else {
4067 if (target_to_host_timespec(pts, timeout)) {
4068 return -TARGET_EFAULT;
4073 if (nsops > TARGET_SEMOPM) {
4074 return -TARGET_E2BIG;
4077 sops = g_new(struct sembuf, nsops);
4079 if (target_to_host_sembuf(sops, ptr, nsops)) {
4080 g_free(sops);
4081 return -TARGET_EFAULT;
4084 ret = -TARGET_ENOSYS;
4085 #ifdef __NR_semtimedop
4086 ret = get_errno(safe_semtimedop(semid, sops, nsops, pts));
4087 #endif
4088 #ifdef __NR_ipc
4089 if (ret == -TARGET_ENOSYS) {
4090 ret = get_errno(safe_ipc(IPCOP_semtimedop, semid,
4091 SEMTIMEDOP_IPC_ARGS(nsops, sops, (long)pts)));
4093 #endif
4094 g_free(sops);
4095 return ret;
4097 #endif
4099 struct target_msqid_ds
4101 struct target_ipc_perm msg_perm;
4102 abi_ulong msg_stime;
4103 #if TARGET_ABI_BITS == 32
4104 abi_ulong __unused1;
4105 #endif
4106 abi_ulong msg_rtime;
4107 #if TARGET_ABI_BITS == 32
4108 abi_ulong __unused2;
4109 #endif
4110 abi_ulong msg_ctime;
4111 #if TARGET_ABI_BITS == 32
4112 abi_ulong __unused3;
4113 #endif
4114 abi_ulong __msg_cbytes;
4115 abi_ulong msg_qnum;
4116 abi_ulong msg_qbytes;
4117 abi_ulong msg_lspid;
4118 abi_ulong msg_lrpid;
4119 abi_ulong __unused4;
4120 abi_ulong __unused5;
4123 static inline abi_long target_to_host_msqid_ds(struct msqid_ds *host_md,
4124 abi_ulong target_addr)
4126 struct target_msqid_ds *target_md;
4128 if (!lock_user_struct(VERIFY_READ, target_md, target_addr, 1))
4129 return -TARGET_EFAULT;
4130 if (target_to_host_ipc_perm(&(host_md->msg_perm),target_addr))
4131 return -TARGET_EFAULT;
4132 host_md->msg_stime = tswapal(target_md->msg_stime);
4133 host_md->msg_rtime = tswapal(target_md->msg_rtime);
4134 host_md->msg_ctime = tswapal(target_md->msg_ctime);
4135 host_md->__msg_cbytes = tswapal(target_md->__msg_cbytes);
4136 host_md->msg_qnum = tswapal(target_md->msg_qnum);
4137 host_md->msg_qbytes = tswapal(target_md->msg_qbytes);
4138 host_md->msg_lspid = tswapal(target_md->msg_lspid);
4139 host_md->msg_lrpid = tswapal(target_md->msg_lrpid);
4140 unlock_user_struct(target_md, target_addr, 0);
4141 return 0;
4144 static inline abi_long host_to_target_msqid_ds(abi_ulong target_addr,
4145 struct msqid_ds *host_md)
4147 struct target_msqid_ds *target_md;
4149 if (!lock_user_struct(VERIFY_WRITE, target_md, target_addr, 0))
4150 return -TARGET_EFAULT;
4151 if (host_to_target_ipc_perm(target_addr,&(host_md->msg_perm)))
4152 return -TARGET_EFAULT;
4153 target_md->msg_stime = tswapal(host_md->msg_stime);
4154 target_md->msg_rtime = tswapal(host_md->msg_rtime);
4155 target_md->msg_ctime = tswapal(host_md->msg_ctime);
4156 target_md->__msg_cbytes = tswapal(host_md->__msg_cbytes);
4157 target_md->msg_qnum = tswapal(host_md->msg_qnum);
4158 target_md->msg_qbytes = tswapal(host_md->msg_qbytes);
4159 target_md->msg_lspid = tswapal(host_md->msg_lspid);
4160 target_md->msg_lrpid = tswapal(host_md->msg_lrpid);
4161 unlock_user_struct(target_md, target_addr, 1);
4162 return 0;
4165 struct target_msginfo {
4166 int msgpool;
4167 int msgmap;
4168 int msgmax;
4169 int msgmnb;
4170 int msgmni;
4171 int msgssz;
4172 int msgtql;
4173 unsigned short int msgseg;
4176 static inline abi_long host_to_target_msginfo(abi_ulong target_addr,
4177 struct msginfo *host_msginfo)
4179 struct target_msginfo *target_msginfo;
4180 if (!lock_user_struct(VERIFY_WRITE, target_msginfo, target_addr, 0))
4181 return -TARGET_EFAULT;
4182 __put_user(host_msginfo->msgpool, &target_msginfo->msgpool);
4183 __put_user(host_msginfo->msgmap, &target_msginfo->msgmap);
4184 __put_user(host_msginfo->msgmax, &target_msginfo->msgmax);
4185 __put_user(host_msginfo->msgmnb, &target_msginfo->msgmnb);
4186 __put_user(host_msginfo->msgmni, &target_msginfo->msgmni);
4187 __put_user(host_msginfo->msgssz, &target_msginfo->msgssz);
4188 __put_user(host_msginfo->msgtql, &target_msginfo->msgtql);
4189 __put_user(host_msginfo->msgseg, &target_msginfo->msgseg);
4190 unlock_user_struct(target_msginfo, target_addr, 1);
4191 return 0;
4194 static inline abi_long do_msgctl(int msgid, int cmd, abi_long ptr)
4196 struct msqid_ds dsarg;
4197 struct msginfo msginfo;
4198 abi_long ret = -TARGET_EINVAL;
4200 cmd &= 0xff;
4202 switch (cmd) {
4203 case IPC_STAT:
4204 case IPC_SET:
4205 case MSG_STAT:
4206 if (target_to_host_msqid_ds(&dsarg,ptr))
4207 return -TARGET_EFAULT;
4208 ret = get_errno(msgctl(msgid, cmd, &dsarg));
4209 if (host_to_target_msqid_ds(ptr,&dsarg))
4210 return -TARGET_EFAULT;
4211 break;
4212 case IPC_RMID:
4213 ret = get_errno(msgctl(msgid, cmd, NULL));
4214 break;
4215 case IPC_INFO:
4216 case MSG_INFO:
4217 ret = get_errno(msgctl(msgid, cmd, (struct msqid_ds *)&msginfo));
4218 if (host_to_target_msginfo(ptr, &msginfo))
4219 return -TARGET_EFAULT;
4220 break;
4223 return ret;
4226 struct target_msgbuf {
4227 abi_long mtype;
4228 char mtext[1];
4231 static inline abi_long do_msgsnd(int msqid, abi_long msgp,
4232 ssize_t msgsz, int msgflg)
4234 struct target_msgbuf *target_mb;
4235 struct msgbuf *host_mb;
4236 abi_long ret = 0;
4238 if (msgsz < 0) {
4239 return -TARGET_EINVAL;
4242 if (!lock_user_struct(VERIFY_READ, target_mb, msgp, 0))
4243 return -TARGET_EFAULT;
4244 host_mb = g_try_malloc(msgsz + sizeof(long));
4245 if (!host_mb) {
4246 unlock_user_struct(target_mb, msgp, 0);
4247 return -TARGET_ENOMEM;
4249 host_mb->mtype = (abi_long) tswapal(target_mb->mtype);
4250 memcpy(host_mb->mtext, target_mb->mtext, msgsz);
4251 ret = -TARGET_ENOSYS;
4252 #ifdef __NR_msgsnd
4253 ret = get_errno(safe_msgsnd(msqid, host_mb, msgsz, msgflg));
4254 #endif
4255 #ifdef __NR_ipc
4256 if (ret == -TARGET_ENOSYS) {
4257 #ifdef __s390x__
4258 ret = get_errno(safe_ipc(IPCOP_msgsnd, msqid, msgsz, msgflg,
4259 host_mb));
4260 #else
4261 ret = get_errno(safe_ipc(IPCOP_msgsnd, msqid, msgsz, msgflg,
4262 host_mb, 0));
4263 #endif
4265 #endif
4266 g_free(host_mb);
4267 unlock_user_struct(target_mb, msgp, 0);
4269 return ret;
4272 #ifdef __NR_ipc
4273 #if defined(__sparc__)
4274 /* SPARC for msgrcv it does not use the kludge on final 2 arguments. */
4275 #define MSGRCV_ARGS(__msgp, __msgtyp) __msgp, __msgtyp
4276 #elif defined(__s390x__)
4277 /* The s390 sys_ipc variant has only five parameters. */
4278 #define MSGRCV_ARGS(__msgp, __msgtyp) \
4279 ((long int[]){(long int)__msgp, __msgtyp})
4280 #else
4281 #define MSGRCV_ARGS(__msgp, __msgtyp) \
4282 ((long int[]){(long int)__msgp, __msgtyp}), 0
4283 #endif
4284 #endif
4286 static inline abi_long do_msgrcv(int msqid, abi_long msgp,
4287 ssize_t msgsz, abi_long msgtyp,
4288 int msgflg)
4290 struct target_msgbuf *target_mb;
4291 char *target_mtext;
4292 struct msgbuf *host_mb;
4293 abi_long ret = 0;
4295 if (msgsz < 0) {
4296 return -TARGET_EINVAL;
4299 if (!lock_user_struct(VERIFY_WRITE, target_mb, msgp, 0))
4300 return -TARGET_EFAULT;
4302 host_mb = g_try_malloc(msgsz + sizeof(long));
4303 if (!host_mb) {
4304 ret = -TARGET_ENOMEM;
4305 goto end;
4307 ret = -TARGET_ENOSYS;
4308 #ifdef __NR_msgrcv
4309 ret = get_errno(safe_msgrcv(msqid, host_mb, msgsz, msgtyp, msgflg));
4310 #endif
4311 #ifdef __NR_ipc
4312 if (ret == -TARGET_ENOSYS) {
4313 ret = get_errno(safe_ipc(IPCOP_CALL(1, IPCOP_msgrcv), msqid, msgsz,
4314 msgflg, MSGRCV_ARGS(host_mb, msgtyp)));
4316 #endif
4318 if (ret > 0) {
4319 abi_ulong target_mtext_addr = msgp + sizeof(abi_ulong);
4320 target_mtext = lock_user(VERIFY_WRITE, target_mtext_addr, ret, 0);
4321 if (!target_mtext) {
4322 ret = -TARGET_EFAULT;
4323 goto end;
4325 memcpy(target_mb->mtext, host_mb->mtext, ret);
4326 unlock_user(target_mtext, target_mtext_addr, ret);
4329 target_mb->mtype = tswapal(host_mb->mtype);
4331 end:
4332 if (target_mb)
4333 unlock_user_struct(target_mb, msgp, 1);
4334 g_free(host_mb);
4335 return ret;
4338 static inline abi_long target_to_host_shmid_ds(struct shmid_ds *host_sd,
4339 abi_ulong target_addr)
4341 struct target_shmid_ds *target_sd;
4343 if (!lock_user_struct(VERIFY_READ, target_sd, target_addr, 1))
4344 return -TARGET_EFAULT;
4345 if (target_to_host_ipc_perm(&(host_sd->shm_perm), target_addr))
4346 return -TARGET_EFAULT;
4347 __get_user(host_sd->shm_segsz, &target_sd->shm_segsz);
4348 __get_user(host_sd->shm_atime, &target_sd->shm_atime);
4349 __get_user(host_sd->shm_dtime, &target_sd->shm_dtime);
4350 __get_user(host_sd->shm_ctime, &target_sd->shm_ctime);
4351 __get_user(host_sd->shm_cpid, &target_sd->shm_cpid);
4352 __get_user(host_sd->shm_lpid, &target_sd->shm_lpid);
4353 __get_user(host_sd->shm_nattch, &target_sd->shm_nattch);
4354 unlock_user_struct(target_sd, target_addr, 0);
4355 return 0;
4358 static inline abi_long host_to_target_shmid_ds(abi_ulong target_addr,
4359 struct shmid_ds *host_sd)
4361 struct target_shmid_ds *target_sd;
4363 if (!lock_user_struct(VERIFY_WRITE, target_sd, target_addr, 0))
4364 return -TARGET_EFAULT;
4365 if (host_to_target_ipc_perm(target_addr, &(host_sd->shm_perm)))
4366 return -TARGET_EFAULT;
4367 __put_user(host_sd->shm_segsz, &target_sd->shm_segsz);
4368 __put_user(host_sd->shm_atime, &target_sd->shm_atime);
4369 __put_user(host_sd->shm_dtime, &target_sd->shm_dtime);
4370 __put_user(host_sd->shm_ctime, &target_sd->shm_ctime);
4371 __put_user(host_sd->shm_cpid, &target_sd->shm_cpid);
4372 __put_user(host_sd->shm_lpid, &target_sd->shm_lpid);
4373 __put_user(host_sd->shm_nattch, &target_sd->shm_nattch);
4374 unlock_user_struct(target_sd, target_addr, 1);
4375 return 0;
4378 struct target_shminfo {
4379 abi_ulong shmmax;
4380 abi_ulong shmmin;
4381 abi_ulong shmmni;
4382 abi_ulong shmseg;
4383 abi_ulong shmall;
4386 static inline abi_long host_to_target_shminfo(abi_ulong target_addr,
4387 struct shminfo *host_shminfo)
4389 struct target_shminfo *target_shminfo;
4390 if (!lock_user_struct(VERIFY_WRITE, target_shminfo, target_addr, 0))
4391 return -TARGET_EFAULT;
4392 __put_user(host_shminfo->shmmax, &target_shminfo->shmmax);
4393 __put_user(host_shminfo->shmmin, &target_shminfo->shmmin);
4394 __put_user(host_shminfo->shmmni, &target_shminfo->shmmni);
4395 __put_user(host_shminfo->shmseg, &target_shminfo->shmseg);
4396 __put_user(host_shminfo->shmall, &target_shminfo->shmall);
4397 unlock_user_struct(target_shminfo, target_addr, 1);
4398 return 0;
4401 struct target_shm_info {
4402 int used_ids;
4403 abi_ulong shm_tot;
4404 abi_ulong shm_rss;
4405 abi_ulong shm_swp;
4406 abi_ulong swap_attempts;
4407 abi_ulong swap_successes;
4410 static inline abi_long host_to_target_shm_info(abi_ulong target_addr,
4411 struct shm_info *host_shm_info)
4413 struct target_shm_info *target_shm_info;
4414 if (!lock_user_struct(VERIFY_WRITE, target_shm_info, target_addr, 0))
4415 return -TARGET_EFAULT;
4416 __put_user(host_shm_info->used_ids, &target_shm_info->used_ids);
4417 __put_user(host_shm_info->shm_tot, &target_shm_info->shm_tot);
4418 __put_user(host_shm_info->shm_rss, &target_shm_info->shm_rss);
4419 __put_user(host_shm_info->shm_swp, &target_shm_info->shm_swp);
4420 __put_user(host_shm_info->swap_attempts, &target_shm_info->swap_attempts);
4421 __put_user(host_shm_info->swap_successes, &target_shm_info->swap_successes);
4422 unlock_user_struct(target_shm_info, target_addr, 1);
4423 return 0;
4426 static inline abi_long do_shmctl(int shmid, int cmd, abi_long buf)
4428 struct shmid_ds dsarg;
4429 struct shminfo shminfo;
4430 struct shm_info shm_info;
4431 abi_long ret = -TARGET_EINVAL;
4433 cmd &= 0xff;
4435 switch(cmd) {
4436 case IPC_STAT:
4437 case IPC_SET:
4438 case SHM_STAT:
4439 if (target_to_host_shmid_ds(&dsarg, buf))
4440 return -TARGET_EFAULT;
4441 ret = get_errno(shmctl(shmid, cmd, &dsarg));
4442 if (host_to_target_shmid_ds(buf, &dsarg))
4443 return -TARGET_EFAULT;
4444 break;
4445 case IPC_INFO:
4446 ret = get_errno(shmctl(shmid, cmd, (struct shmid_ds *)&shminfo));
4447 if (host_to_target_shminfo(buf, &shminfo))
4448 return -TARGET_EFAULT;
4449 break;
4450 case SHM_INFO:
4451 ret = get_errno(shmctl(shmid, cmd, (struct shmid_ds *)&shm_info));
4452 if (host_to_target_shm_info(buf, &shm_info))
4453 return -TARGET_EFAULT;
4454 break;
4455 case IPC_RMID:
4456 case SHM_LOCK:
4457 case SHM_UNLOCK:
4458 ret = get_errno(shmctl(shmid, cmd, NULL));
4459 break;
4462 return ret;
4465 #ifndef TARGET_FORCE_SHMLBA
4466 /* For most architectures, SHMLBA is the same as the page size;
4467 * some architectures have larger values, in which case they should
4468 * define TARGET_FORCE_SHMLBA and provide a target_shmlba() function.
4469 * This corresponds to the kernel arch code defining __ARCH_FORCE_SHMLBA
4470 * and defining its own value for SHMLBA.
4472 * The kernel also permits SHMLBA to be set by the architecture to a
4473 * value larger than the page size without setting __ARCH_FORCE_SHMLBA;
4474 * this means that addresses are rounded to the large size if
4475 * SHM_RND is set but addresses not aligned to that size are not rejected
4476 * as long as they are at least page-aligned. Since the only architecture
4477 * which uses this is ia64 this code doesn't provide for that oddity.
4479 static inline abi_ulong target_shmlba(CPUArchState *cpu_env)
4481 return TARGET_PAGE_SIZE;
4483 #endif
4485 static inline abi_ulong do_shmat(CPUArchState *cpu_env,
4486 int shmid, abi_ulong shmaddr, int shmflg)
4488 CPUState *cpu = env_cpu(cpu_env);
4489 abi_long raddr;
4490 void *host_raddr;
4491 struct shmid_ds shm_info;
4492 int i,ret;
4493 abi_ulong shmlba;
4495 /* shmat pointers are always untagged */
4497 /* find out the length of the shared memory segment */
4498 ret = get_errno(shmctl(shmid, IPC_STAT, &shm_info));
4499 if (is_error(ret)) {
4500 /* can't get length, bail out */
4501 return ret;
4504 shmlba = target_shmlba(cpu_env);
4506 if (shmaddr & (shmlba - 1)) {
4507 if (shmflg & SHM_RND) {
4508 shmaddr &= ~(shmlba - 1);
4509 } else {
4510 return -TARGET_EINVAL;
4513 if (!guest_range_valid_untagged(shmaddr, shm_info.shm_segsz)) {
4514 return -TARGET_EINVAL;
4517 mmap_lock();
4520 * We're mapping shared memory, so ensure we generate code for parallel
4521 * execution and flush old translations. This will work up to the level
4522 * supported by the host -- anything that requires EXCP_ATOMIC will not
4523 * be atomic with respect to an external process.
4525 if (!(cpu->tcg_cflags & CF_PARALLEL)) {
4526 cpu->tcg_cflags |= CF_PARALLEL;
4527 tb_flush(cpu);
4530 if (shmaddr)
4531 host_raddr = shmat(shmid, (void *)g2h_untagged(shmaddr), shmflg);
4532 else {
4533 abi_ulong mmap_start;
4535 /* In order to use the host shmat, we need to honor host SHMLBA. */
4536 mmap_start = mmap_find_vma(0, shm_info.shm_segsz, MAX(SHMLBA, shmlba));
4538 if (mmap_start == -1) {
4539 errno = ENOMEM;
4540 host_raddr = (void *)-1;
4541 } else
4542 host_raddr = shmat(shmid, g2h_untagged(mmap_start),
4543 shmflg | SHM_REMAP);
4546 if (host_raddr == (void *)-1) {
4547 mmap_unlock();
4548 return get_errno((long)host_raddr);
4550 raddr=h2g((unsigned long)host_raddr);
4552 page_set_flags(raddr, raddr + shm_info.shm_segsz,
4553 PAGE_VALID | PAGE_RESET | PAGE_READ |
4554 (shmflg & SHM_RDONLY ? 0 : PAGE_WRITE));
4556 for (i = 0; i < N_SHM_REGIONS; i++) {
4557 if (!shm_regions[i].in_use) {
4558 shm_regions[i].in_use = true;
4559 shm_regions[i].start = raddr;
4560 shm_regions[i].size = shm_info.shm_segsz;
4561 break;
4565 mmap_unlock();
4566 return raddr;
4570 static inline abi_long do_shmdt(abi_ulong shmaddr)
4572 int i;
4573 abi_long rv;
4575 /* shmdt pointers are always untagged */
4577 mmap_lock();
4579 for (i = 0; i < N_SHM_REGIONS; ++i) {
4580 if (shm_regions[i].in_use && shm_regions[i].start == shmaddr) {
4581 shm_regions[i].in_use = false;
4582 page_set_flags(shmaddr, shmaddr + shm_regions[i].size, 0);
4583 break;
4586 rv = get_errno(shmdt(g2h_untagged(shmaddr)));
4588 mmap_unlock();
4590 return rv;
4593 #ifdef TARGET_NR_ipc
4594 /* ??? This only works with linear mappings. */
4595 /* do_ipc() must return target values and target errnos. */
4596 static abi_long do_ipc(CPUArchState *cpu_env,
4597 unsigned int call, abi_long first,
4598 abi_long second, abi_long third,
4599 abi_long ptr, abi_long fifth)
4601 int version;
4602 abi_long ret = 0;
4604 version = call >> 16;
4605 call &= 0xffff;
4607 switch (call) {
4608 case IPCOP_semop:
4609 ret = do_semtimedop(first, ptr, second, 0, false);
4610 break;
4611 case IPCOP_semtimedop:
4613 * The s390 sys_ipc variant has only five parameters instead of six
4614 * (as for default variant) and the only difference is the handling of
4615 * SEMTIMEDOP where on s390 the third parameter is used as a pointer
4616 * to a struct timespec where the generic variant uses fifth parameter.
4618 #if defined(TARGET_S390X)
4619 ret = do_semtimedop(first, ptr, second, third, TARGET_ABI_BITS == 64);
4620 #else
4621 ret = do_semtimedop(first, ptr, second, fifth, TARGET_ABI_BITS == 64);
4622 #endif
4623 break;
4625 case IPCOP_semget:
4626 ret = get_errno(semget(first, second, third));
4627 break;
4629 case IPCOP_semctl: {
4630 /* The semun argument to semctl is passed by value, so dereference the
4631 * ptr argument. */
4632 abi_ulong atptr;
4633 get_user_ual(atptr, ptr);
4634 ret = do_semctl(first, second, third, atptr);
4635 break;
4638 case IPCOP_msgget:
4639 ret = get_errno(msgget(first, second));
4640 break;
4642 case IPCOP_msgsnd:
4643 ret = do_msgsnd(first, ptr, second, third);
4644 break;
4646 case IPCOP_msgctl:
4647 ret = do_msgctl(first, second, ptr);
4648 break;
4650 case IPCOP_msgrcv:
4651 switch (version) {
4652 case 0:
4654 struct target_ipc_kludge {
4655 abi_long msgp;
4656 abi_long msgtyp;
4657 } *tmp;
4659 if (!lock_user_struct(VERIFY_READ, tmp, ptr, 1)) {
4660 ret = -TARGET_EFAULT;
4661 break;
4664 ret = do_msgrcv(first, tswapal(tmp->msgp), second, tswapal(tmp->msgtyp), third);
4666 unlock_user_struct(tmp, ptr, 0);
4667 break;
4669 default:
4670 ret = do_msgrcv(first, ptr, second, fifth, third);
4672 break;
4674 case IPCOP_shmat:
4675 switch (version) {
4676 default:
4678 abi_ulong raddr;
4679 raddr = do_shmat(cpu_env, first, ptr, second);
4680 if (is_error(raddr))
4681 return get_errno(raddr);
4682 if (put_user_ual(raddr, third))
4683 return -TARGET_EFAULT;
4684 break;
4686 case 1:
4687 ret = -TARGET_EINVAL;
4688 break;
4690 break;
4691 case IPCOP_shmdt:
4692 ret = do_shmdt(ptr);
4693 break;
4695 case IPCOP_shmget:
4696 /* IPC_* flag values are the same on all linux platforms */
4697 ret = get_errno(shmget(first, second, third));
4698 break;
4700 /* IPC_* and SHM_* command values are the same on all linux platforms */
4701 case IPCOP_shmctl:
4702 ret = do_shmctl(first, second, ptr);
4703 break;
4704 default:
4705 qemu_log_mask(LOG_UNIMP, "Unsupported ipc call: %d (version %d)\n",
4706 call, version);
4707 ret = -TARGET_ENOSYS;
4708 break;
4710 return ret;
4712 #endif
4714 /* kernel structure types definitions */
4716 #define STRUCT(name, ...) STRUCT_ ## name,
4717 #define STRUCT_SPECIAL(name) STRUCT_ ## name,
4718 enum {
4719 #include "syscall_types.h"
4720 STRUCT_MAX
4722 #undef STRUCT
4723 #undef STRUCT_SPECIAL
4725 #define STRUCT(name, ...) static const argtype struct_ ## name ## _def[] = { __VA_ARGS__, TYPE_NULL };
4726 #define STRUCT_SPECIAL(name)
4727 #include "syscall_types.h"
4728 #undef STRUCT
4729 #undef STRUCT_SPECIAL
4731 #define MAX_STRUCT_SIZE 4096
4733 #ifdef CONFIG_FIEMAP
4734 /* So fiemap access checks don't overflow on 32 bit systems.
4735 * This is very slightly smaller than the limit imposed by
4736 * the underlying kernel.
4738 #define FIEMAP_MAX_EXTENTS ((UINT_MAX - sizeof(struct fiemap)) \
4739 / sizeof(struct fiemap_extent))
4741 static abi_long do_ioctl_fs_ioc_fiemap(const IOCTLEntry *ie, uint8_t *buf_temp,
4742 int fd, int cmd, abi_long arg)
4744 /* The parameter for this ioctl is a struct fiemap followed
4745 * by an array of struct fiemap_extent whose size is set
4746 * in fiemap->fm_extent_count. The array is filled in by the
4747 * ioctl.
4749 int target_size_in, target_size_out;
4750 struct fiemap *fm;
4751 const argtype *arg_type = ie->arg_type;
4752 const argtype extent_arg_type[] = { MK_STRUCT(STRUCT_fiemap_extent) };
4753 void *argptr, *p;
4754 abi_long ret;
4755 int i, extent_size = thunk_type_size(extent_arg_type, 0);
4756 uint32_t outbufsz;
4757 int free_fm = 0;
4759 assert(arg_type[0] == TYPE_PTR);
4760 assert(ie->access == IOC_RW);
4761 arg_type++;
4762 target_size_in = thunk_type_size(arg_type, 0);
4763 argptr = lock_user(VERIFY_READ, arg, target_size_in, 1);
4764 if (!argptr) {
4765 return -TARGET_EFAULT;
4767 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
4768 unlock_user(argptr, arg, 0);
4769 fm = (struct fiemap *)buf_temp;
4770 if (fm->fm_extent_count > FIEMAP_MAX_EXTENTS) {
4771 return -TARGET_EINVAL;
4774 outbufsz = sizeof (*fm) +
4775 (sizeof(struct fiemap_extent) * fm->fm_extent_count);
4777 if (outbufsz > MAX_STRUCT_SIZE) {
4778 /* We can't fit all the extents into the fixed size buffer.
4779 * Allocate one that is large enough and use it instead.
4781 fm = g_try_malloc(outbufsz);
4782 if (!fm) {
4783 return -TARGET_ENOMEM;
4785 memcpy(fm, buf_temp, sizeof(struct fiemap));
4786 free_fm = 1;
4788 ret = get_errno(safe_ioctl(fd, ie->host_cmd, fm));
4789 if (!is_error(ret)) {
4790 target_size_out = target_size_in;
4791 /* An extent_count of 0 means we were only counting the extents
4792 * so there are no structs to copy
4794 if (fm->fm_extent_count != 0) {
4795 target_size_out += fm->fm_mapped_extents * extent_size;
4797 argptr = lock_user(VERIFY_WRITE, arg, target_size_out, 0);
4798 if (!argptr) {
4799 ret = -TARGET_EFAULT;
4800 } else {
4801 /* Convert the struct fiemap */
4802 thunk_convert(argptr, fm, arg_type, THUNK_TARGET);
4803 if (fm->fm_extent_count != 0) {
4804 p = argptr + target_size_in;
4805 /* ...and then all the struct fiemap_extents */
4806 for (i = 0; i < fm->fm_mapped_extents; i++) {
4807 thunk_convert(p, &fm->fm_extents[i], extent_arg_type,
4808 THUNK_TARGET);
4809 p += extent_size;
4812 unlock_user(argptr, arg, target_size_out);
4815 if (free_fm) {
4816 g_free(fm);
4818 return ret;
4820 #endif
4822 static abi_long do_ioctl_ifconf(const IOCTLEntry *ie, uint8_t *buf_temp,
4823 int fd, int cmd, abi_long arg)
4825 const argtype *arg_type = ie->arg_type;
4826 int target_size;
4827 void *argptr;
4828 int ret;
4829 struct ifconf *host_ifconf;
4830 uint32_t outbufsz;
4831 const argtype ifreq_arg_type[] = { MK_STRUCT(STRUCT_sockaddr_ifreq) };
4832 const argtype ifreq_max_type[] = { MK_STRUCT(STRUCT_ifmap_ifreq) };
4833 int target_ifreq_size;
4834 int nb_ifreq;
4835 int free_buf = 0;
4836 int i;
4837 int target_ifc_len;
4838 abi_long target_ifc_buf;
4839 int host_ifc_len;
4840 char *host_ifc_buf;
4842 assert(arg_type[0] == TYPE_PTR);
4843 assert(ie->access == IOC_RW);
4845 arg_type++;
4846 target_size = thunk_type_size(arg_type, 0);
4848 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
4849 if (!argptr)
4850 return -TARGET_EFAULT;
4851 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
4852 unlock_user(argptr, arg, 0);
4854 host_ifconf = (struct ifconf *)(unsigned long)buf_temp;
4855 target_ifc_buf = (abi_long)(unsigned long)host_ifconf->ifc_buf;
4856 target_ifreq_size = thunk_type_size(ifreq_max_type, 0);
4858 if (target_ifc_buf != 0) {
4859 target_ifc_len = host_ifconf->ifc_len;
4860 nb_ifreq = target_ifc_len / target_ifreq_size;
4861 host_ifc_len = nb_ifreq * sizeof(struct ifreq);
4863 outbufsz = sizeof(*host_ifconf) + host_ifc_len;
4864 if (outbufsz > MAX_STRUCT_SIZE) {
4866 * We can't fit all the extents into the fixed size buffer.
4867 * Allocate one that is large enough and use it instead.
4869 host_ifconf = g_try_malloc(outbufsz);
4870 if (!host_ifconf) {
4871 return -TARGET_ENOMEM;
4873 memcpy(host_ifconf, buf_temp, sizeof(*host_ifconf));
4874 free_buf = 1;
4876 host_ifc_buf = (char *)host_ifconf + sizeof(*host_ifconf);
4878 host_ifconf->ifc_len = host_ifc_len;
4879 } else {
4880 host_ifc_buf = NULL;
4882 host_ifconf->ifc_buf = host_ifc_buf;
4884 ret = get_errno(safe_ioctl(fd, ie->host_cmd, host_ifconf));
4885 if (!is_error(ret)) {
4886 /* convert host ifc_len to target ifc_len */
4888 nb_ifreq = host_ifconf->ifc_len / sizeof(struct ifreq);
4889 target_ifc_len = nb_ifreq * target_ifreq_size;
4890 host_ifconf->ifc_len = target_ifc_len;
4892 /* restore target ifc_buf */
4894 host_ifconf->ifc_buf = (char *)(unsigned long)target_ifc_buf;
4896 /* copy struct ifconf to target user */
4898 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
4899 if (!argptr)
4900 return -TARGET_EFAULT;
4901 thunk_convert(argptr, host_ifconf, arg_type, THUNK_TARGET);
4902 unlock_user(argptr, arg, target_size);
4904 if (target_ifc_buf != 0) {
4905 /* copy ifreq[] to target user */
4906 argptr = lock_user(VERIFY_WRITE, target_ifc_buf, target_ifc_len, 0);
4907 for (i = 0; i < nb_ifreq ; i++) {
4908 thunk_convert(argptr + i * target_ifreq_size,
4909 host_ifc_buf + i * sizeof(struct ifreq),
4910 ifreq_arg_type, THUNK_TARGET);
4912 unlock_user(argptr, target_ifc_buf, target_ifc_len);
4916 if (free_buf) {
4917 g_free(host_ifconf);
4920 return ret;
4923 #if defined(CONFIG_USBFS)
4924 #if HOST_LONG_BITS > 64
4925 #error USBDEVFS thunks do not support >64 bit hosts yet.
4926 #endif
4927 struct live_urb {
4928 uint64_t target_urb_adr;
4929 uint64_t target_buf_adr;
4930 char *target_buf_ptr;
4931 struct usbdevfs_urb host_urb;
4934 static GHashTable *usbdevfs_urb_hashtable(void)
4936 static GHashTable *urb_hashtable;
4938 if (!urb_hashtable) {
4939 urb_hashtable = g_hash_table_new(g_int64_hash, g_int64_equal);
4941 return urb_hashtable;
4944 static void urb_hashtable_insert(struct live_urb *urb)
4946 GHashTable *urb_hashtable = usbdevfs_urb_hashtable();
4947 g_hash_table_insert(urb_hashtable, urb, urb);
4950 static struct live_urb *urb_hashtable_lookup(uint64_t target_urb_adr)
4952 GHashTable *urb_hashtable = usbdevfs_urb_hashtable();
4953 return g_hash_table_lookup(urb_hashtable, &target_urb_adr);
4956 static void urb_hashtable_remove(struct live_urb *urb)
4958 GHashTable *urb_hashtable = usbdevfs_urb_hashtable();
4959 g_hash_table_remove(urb_hashtable, urb);
4962 static abi_long
4963 do_ioctl_usbdevfs_reapurb(const IOCTLEntry *ie, uint8_t *buf_temp,
4964 int fd, int cmd, abi_long arg)
4966 const argtype usbfsurb_arg_type[] = { MK_STRUCT(STRUCT_usbdevfs_urb) };
4967 const argtype ptrvoid_arg_type[] = { TYPE_PTRVOID, 0, 0 };
4968 struct live_urb *lurb;
4969 void *argptr;
4970 uint64_t hurb;
4971 int target_size;
4972 uintptr_t target_urb_adr;
4973 abi_long ret;
4975 target_size = thunk_type_size(usbfsurb_arg_type, THUNK_TARGET);
4977 memset(buf_temp, 0, sizeof(uint64_t));
4978 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
4979 if (is_error(ret)) {
4980 return ret;
4983 memcpy(&hurb, buf_temp, sizeof(uint64_t));
4984 lurb = (void *)((uintptr_t)hurb - offsetof(struct live_urb, host_urb));
4985 if (!lurb->target_urb_adr) {
4986 return -TARGET_EFAULT;
4988 urb_hashtable_remove(lurb);
4989 unlock_user(lurb->target_buf_ptr, lurb->target_buf_adr,
4990 lurb->host_urb.buffer_length);
4991 lurb->target_buf_ptr = NULL;
4993 /* restore the guest buffer pointer */
4994 lurb->host_urb.buffer = (void *)(uintptr_t)lurb->target_buf_adr;
4996 /* update the guest urb struct */
4997 argptr = lock_user(VERIFY_WRITE, lurb->target_urb_adr, target_size, 0);
4998 if (!argptr) {
4999 g_free(lurb);
5000 return -TARGET_EFAULT;
5002 thunk_convert(argptr, &lurb->host_urb, usbfsurb_arg_type, THUNK_TARGET);
5003 unlock_user(argptr, lurb->target_urb_adr, target_size);
5005 target_size = thunk_type_size(ptrvoid_arg_type, THUNK_TARGET);
5006 /* write back the urb handle */
5007 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
5008 if (!argptr) {
5009 g_free(lurb);
5010 return -TARGET_EFAULT;
5013 /* GHashTable uses 64-bit keys but thunk_convert expects uintptr_t */
5014 target_urb_adr = lurb->target_urb_adr;
5015 thunk_convert(argptr, &target_urb_adr, ptrvoid_arg_type, THUNK_TARGET);
5016 unlock_user(argptr, arg, target_size);
5018 g_free(lurb);
5019 return ret;
5022 static abi_long
5023 do_ioctl_usbdevfs_discardurb(const IOCTLEntry *ie,
5024 uint8_t *buf_temp __attribute__((unused)),
5025 int fd, int cmd, abi_long arg)
5027 struct live_urb *lurb;
5029 /* map target address back to host URB with metadata. */
5030 lurb = urb_hashtable_lookup(arg);
5031 if (!lurb) {
5032 return -TARGET_EFAULT;
5034 return get_errno(safe_ioctl(fd, ie->host_cmd, &lurb->host_urb));
5037 static abi_long
5038 do_ioctl_usbdevfs_submiturb(const IOCTLEntry *ie, uint8_t *buf_temp,
5039 int fd, int cmd, abi_long arg)
5041 const argtype *arg_type = ie->arg_type;
5042 int target_size;
5043 abi_long ret;
5044 void *argptr;
5045 int rw_dir;
5046 struct live_urb *lurb;
5049 * each submitted URB needs to map to a unique ID for the
5050 * kernel, and that unique ID needs to be a pointer to
5051 * host memory. hence, we need to malloc for each URB.
5052 * isochronous transfers have a variable length struct.
5054 arg_type++;
5055 target_size = thunk_type_size(arg_type, THUNK_TARGET);
5057 /* construct host copy of urb and metadata */
5058 lurb = g_try_new0(struct live_urb, 1);
5059 if (!lurb) {
5060 return -TARGET_ENOMEM;
5063 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5064 if (!argptr) {
5065 g_free(lurb);
5066 return -TARGET_EFAULT;
5068 thunk_convert(&lurb->host_urb, argptr, arg_type, THUNK_HOST);
5069 unlock_user(argptr, arg, 0);
5071 lurb->target_urb_adr = arg;
5072 lurb->target_buf_adr = (uintptr_t)lurb->host_urb.buffer;
5074 /* buffer space used depends on endpoint type so lock the entire buffer */
5075 /* control type urbs should check the buffer contents for true direction */
5076 rw_dir = lurb->host_urb.endpoint & USB_DIR_IN ? VERIFY_WRITE : VERIFY_READ;
5077 lurb->target_buf_ptr = lock_user(rw_dir, lurb->target_buf_adr,
5078 lurb->host_urb.buffer_length, 1);
5079 if (lurb->target_buf_ptr == NULL) {
5080 g_free(lurb);
5081 return -TARGET_EFAULT;
5084 /* update buffer pointer in host copy */
5085 lurb->host_urb.buffer = lurb->target_buf_ptr;
5087 ret = get_errno(safe_ioctl(fd, ie->host_cmd, &lurb->host_urb));
5088 if (is_error(ret)) {
5089 unlock_user(lurb->target_buf_ptr, lurb->target_buf_adr, 0);
5090 g_free(lurb);
5091 } else {
5092 urb_hashtable_insert(lurb);
5095 return ret;
5097 #endif /* CONFIG_USBFS */
5099 static abi_long do_ioctl_dm(const IOCTLEntry *ie, uint8_t *buf_temp, int fd,
5100 int cmd, abi_long arg)
5102 void *argptr;
5103 struct dm_ioctl *host_dm;
5104 abi_long guest_data;
5105 uint32_t guest_data_size;
5106 int target_size;
5107 const argtype *arg_type = ie->arg_type;
5108 abi_long ret;
5109 void *big_buf = NULL;
5110 char *host_data;
5112 arg_type++;
5113 target_size = thunk_type_size(arg_type, 0);
5114 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5115 if (!argptr) {
5116 ret = -TARGET_EFAULT;
5117 goto out;
5119 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
5120 unlock_user(argptr, arg, 0);
5122 /* buf_temp is too small, so fetch things into a bigger buffer */
5123 big_buf = g_malloc0(((struct dm_ioctl*)buf_temp)->data_size * 2);
5124 memcpy(big_buf, buf_temp, target_size);
5125 buf_temp = big_buf;
5126 host_dm = big_buf;
5128 guest_data = arg + host_dm->data_start;
5129 if ((guest_data - arg) < 0) {
5130 ret = -TARGET_EINVAL;
5131 goto out;
5133 guest_data_size = host_dm->data_size - host_dm->data_start;
5134 host_data = (char*)host_dm + host_dm->data_start;
5136 argptr = lock_user(VERIFY_READ, guest_data, guest_data_size, 1);
5137 if (!argptr) {
5138 ret = -TARGET_EFAULT;
5139 goto out;
5142 switch (ie->host_cmd) {
5143 case DM_REMOVE_ALL:
5144 case DM_LIST_DEVICES:
5145 case DM_DEV_CREATE:
5146 case DM_DEV_REMOVE:
5147 case DM_DEV_SUSPEND:
5148 case DM_DEV_STATUS:
5149 case DM_DEV_WAIT:
5150 case DM_TABLE_STATUS:
5151 case DM_TABLE_CLEAR:
5152 case DM_TABLE_DEPS:
5153 case DM_LIST_VERSIONS:
5154 /* no input data */
5155 break;
5156 case DM_DEV_RENAME:
5157 case DM_DEV_SET_GEOMETRY:
5158 /* data contains only strings */
5159 memcpy(host_data, argptr, guest_data_size);
5160 break;
5161 case DM_TARGET_MSG:
5162 memcpy(host_data, argptr, guest_data_size);
5163 *(uint64_t*)host_data = tswap64(*(uint64_t*)argptr);
5164 break;
5165 case DM_TABLE_LOAD:
5167 void *gspec = argptr;
5168 void *cur_data = host_data;
5169 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_target_spec) };
5170 int spec_size = thunk_type_size(arg_type, 0);
5171 int i;
5173 for (i = 0; i < host_dm->target_count; i++) {
5174 struct dm_target_spec *spec = cur_data;
5175 uint32_t next;
5176 int slen;
5178 thunk_convert(spec, gspec, arg_type, THUNK_HOST);
5179 slen = strlen((char*)gspec + spec_size) + 1;
5180 next = spec->next;
5181 spec->next = sizeof(*spec) + slen;
5182 strcpy((char*)&spec[1], gspec + spec_size);
5183 gspec += next;
5184 cur_data += spec->next;
5186 break;
5188 default:
5189 ret = -TARGET_EINVAL;
5190 unlock_user(argptr, guest_data, 0);
5191 goto out;
5193 unlock_user(argptr, guest_data, 0);
5195 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
5196 if (!is_error(ret)) {
5197 guest_data = arg + host_dm->data_start;
5198 guest_data_size = host_dm->data_size - host_dm->data_start;
5199 argptr = lock_user(VERIFY_WRITE, guest_data, guest_data_size, 0);
5200 switch (ie->host_cmd) {
5201 case DM_REMOVE_ALL:
5202 case DM_DEV_CREATE:
5203 case DM_DEV_REMOVE:
5204 case DM_DEV_RENAME:
5205 case DM_DEV_SUSPEND:
5206 case DM_DEV_STATUS:
5207 case DM_TABLE_LOAD:
5208 case DM_TABLE_CLEAR:
5209 case DM_TARGET_MSG:
5210 case DM_DEV_SET_GEOMETRY:
5211 /* no return data */
5212 break;
5213 case DM_LIST_DEVICES:
5215 struct dm_name_list *nl = (void*)host_dm + host_dm->data_start;
5216 uint32_t remaining_data = guest_data_size;
5217 void *cur_data = argptr;
5218 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_name_list) };
5219 int nl_size = 12; /* can't use thunk_size due to alignment */
5221 while (1) {
5222 uint32_t next = nl->next;
5223 if (next) {
5224 nl->next = nl_size + (strlen(nl->name) + 1);
5226 if (remaining_data < nl->next) {
5227 host_dm->flags |= DM_BUFFER_FULL_FLAG;
5228 break;
5230 thunk_convert(cur_data, nl, arg_type, THUNK_TARGET);
5231 strcpy(cur_data + nl_size, nl->name);
5232 cur_data += nl->next;
5233 remaining_data -= nl->next;
5234 if (!next) {
5235 break;
5237 nl = (void*)nl + next;
5239 break;
5241 case DM_DEV_WAIT:
5242 case DM_TABLE_STATUS:
5244 struct dm_target_spec *spec = (void*)host_dm + host_dm->data_start;
5245 void *cur_data = argptr;
5246 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_target_spec) };
5247 int spec_size = thunk_type_size(arg_type, 0);
5248 int i;
5250 for (i = 0; i < host_dm->target_count; i++) {
5251 uint32_t next = spec->next;
5252 int slen = strlen((char*)&spec[1]) + 1;
5253 spec->next = (cur_data - argptr) + spec_size + slen;
5254 if (guest_data_size < spec->next) {
5255 host_dm->flags |= DM_BUFFER_FULL_FLAG;
5256 break;
5258 thunk_convert(cur_data, spec, arg_type, THUNK_TARGET);
5259 strcpy(cur_data + spec_size, (char*)&spec[1]);
5260 cur_data = argptr + spec->next;
5261 spec = (void*)host_dm + host_dm->data_start + next;
5263 break;
5265 case DM_TABLE_DEPS:
5267 void *hdata = (void*)host_dm + host_dm->data_start;
5268 int count = *(uint32_t*)hdata;
5269 uint64_t *hdev = hdata + 8;
5270 uint64_t *gdev = argptr + 8;
5271 int i;
5273 *(uint32_t*)argptr = tswap32(count);
5274 for (i = 0; i < count; i++) {
5275 *gdev = tswap64(*hdev);
5276 gdev++;
5277 hdev++;
5279 break;
5281 case DM_LIST_VERSIONS:
5283 struct dm_target_versions *vers = (void*)host_dm + host_dm->data_start;
5284 uint32_t remaining_data = guest_data_size;
5285 void *cur_data = argptr;
5286 const argtype arg_type[] = { MK_STRUCT(STRUCT_dm_target_versions) };
5287 int vers_size = thunk_type_size(arg_type, 0);
5289 while (1) {
5290 uint32_t next = vers->next;
5291 if (next) {
5292 vers->next = vers_size + (strlen(vers->name) + 1);
5294 if (remaining_data < vers->next) {
5295 host_dm->flags |= DM_BUFFER_FULL_FLAG;
5296 break;
5298 thunk_convert(cur_data, vers, arg_type, THUNK_TARGET);
5299 strcpy(cur_data + vers_size, vers->name);
5300 cur_data += vers->next;
5301 remaining_data -= vers->next;
5302 if (!next) {
5303 break;
5305 vers = (void*)vers + next;
5307 break;
5309 default:
5310 unlock_user(argptr, guest_data, 0);
5311 ret = -TARGET_EINVAL;
5312 goto out;
5314 unlock_user(argptr, guest_data, guest_data_size);
5316 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
5317 if (!argptr) {
5318 ret = -TARGET_EFAULT;
5319 goto out;
5321 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET);
5322 unlock_user(argptr, arg, target_size);
5324 out:
5325 g_free(big_buf);
5326 return ret;
5329 static abi_long do_ioctl_blkpg(const IOCTLEntry *ie, uint8_t *buf_temp, int fd,
5330 int cmd, abi_long arg)
5332 void *argptr;
5333 int target_size;
5334 const argtype *arg_type = ie->arg_type;
5335 const argtype part_arg_type[] = { MK_STRUCT(STRUCT_blkpg_partition) };
5336 abi_long ret;
5338 struct blkpg_ioctl_arg *host_blkpg = (void*)buf_temp;
5339 struct blkpg_partition host_part;
5341 /* Read and convert blkpg */
5342 arg_type++;
5343 target_size = thunk_type_size(arg_type, 0);
5344 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5345 if (!argptr) {
5346 ret = -TARGET_EFAULT;
5347 goto out;
5349 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
5350 unlock_user(argptr, arg, 0);
5352 switch (host_blkpg->op) {
5353 case BLKPG_ADD_PARTITION:
5354 case BLKPG_DEL_PARTITION:
5355 /* payload is struct blkpg_partition */
5356 break;
5357 default:
5358 /* Unknown opcode */
5359 ret = -TARGET_EINVAL;
5360 goto out;
5363 /* Read and convert blkpg->data */
5364 arg = (abi_long)(uintptr_t)host_blkpg->data;
5365 target_size = thunk_type_size(part_arg_type, 0);
5366 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5367 if (!argptr) {
5368 ret = -TARGET_EFAULT;
5369 goto out;
5371 thunk_convert(&host_part, argptr, part_arg_type, THUNK_HOST);
5372 unlock_user(argptr, arg, 0);
5374 /* Swizzle the data pointer to our local copy and call! */
5375 host_blkpg->data = &host_part;
5376 ret = get_errno(safe_ioctl(fd, ie->host_cmd, host_blkpg));
5378 out:
5379 return ret;
5382 static abi_long do_ioctl_rt(const IOCTLEntry *ie, uint8_t *buf_temp,
5383 int fd, int cmd, abi_long arg)
5385 const argtype *arg_type = ie->arg_type;
5386 const StructEntry *se;
5387 const argtype *field_types;
5388 const int *dst_offsets, *src_offsets;
5389 int target_size;
5390 void *argptr;
5391 abi_ulong *target_rt_dev_ptr = NULL;
5392 unsigned long *host_rt_dev_ptr = NULL;
5393 abi_long ret;
5394 int i;
5396 assert(ie->access == IOC_W);
5397 assert(*arg_type == TYPE_PTR);
5398 arg_type++;
5399 assert(*arg_type == TYPE_STRUCT);
5400 target_size = thunk_type_size(arg_type, 0);
5401 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5402 if (!argptr) {
5403 return -TARGET_EFAULT;
5405 arg_type++;
5406 assert(*arg_type == (int)STRUCT_rtentry);
5407 se = struct_entries + *arg_type++;
5408 assert(se->convert[0] == NULL);
5409 /* convert struct here to be able to catch rt_dev string */
5410 field_types = se->field_types;
5411 dst_offsets = se->field_offsets[THUNK_HOST];
5412 src_offsets = se->field_offsets[THUNK_TARGET];
5413 for (i = 0; i < se->nb_fields; i++) {
5414 if (dst_offsets[i] == offsetof(struct rtentry, rt_dev)) {
5415 assert(*field_types == TYPE_PTRVOID);
5416 target_rt_dev_ptr = (abi_ulong *)(argptr + src_offsets[i]);
5417 host_rt_dev_ptr = (unsigned long *)(buf_temp + dst_offsets[i]);
5418 if (*target_rt_dev_ptr != 0) {
5419 *host_rt_dev_ptr = (unsigned long)lock_user_string(
5420 tswapal(*target_rt_dev_ptr));
5421 if (!*host_rt_dev_ptr) {
5422 unlock_user(argptr, arg, 0);
5423 return -TARGET_EFAULT;
5425 } else {
5426 *host_rt_dev_ptr = 0;
5428 field_types++;
5429 continue;
5431 field_types = thunk_convert(buf_temp + dst_offsets[i],
5432 argptr + src_offsets[i],
5433 field_types, THUNK_HOST);
5435 unlock_user(argptr, arg, 0);
5437 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
5439 assert(host_rt_dev_ptr != NULL);
5440 assert(target_rt_dev_ptr != NULL);
5441 if (*host_rt_dev_ptr != 0) {
5442 unlock_user((void *)*host_rt_dev_ptr,
5443 *target_rt_dev_ptr, 0);
5445 return ret;
5448 static abi_long do_ioctl_kdsigaccept(const IOCTLEntry *ie, uint8_t *buf_temp,
5449 int fd, int cmd, abi_long arg)
5451 int sig = target_to_host_signal(arg);
5452 return get_errno(safe_ioctl(fd, ie->host_cmd, sig));
5455 static abi_long do_ioctl_SIOCGSTAMP(const IOCTLEntry *ie, uint8_t *buf_temp,
5456 int fd, int cmd, abi_long arg)
5458 struct timeval tv;
5459 abi_long ret;
5461 ret = get_errno(safe_ioctl(fd, SIOCGSTAMP, &tv));
5462 if (is_error(ret)) {
5463 return ret;
5466 if (cmd == (int)TARGET_SIOCGSTAMP_OLD) {
5467 if (copy_to_user_timeval(arg, &tv)) {
5468 return -TARGET_EFAULT;
5470 } else {
5471 if (copy_to_user_timeval64(arg, &tv)) {
5472 return -TARGET_EFAULT;
5476 return ret;
5479 static abi_long do_ioctl_SIOCGSTAMPNS(const IOCTLEntry *ie, uint8_t *buf_temp,
5480 int fd, int cmd, abi_long arg)
5482 struct timespec ts;
5483 abi_long ret;
5485 ret = get_errno(safe_ioctl(fd, SIOCGSTAMPNS, &ts));
5486 if (is_error(ret)) {
5487 return ret;
5490 if (cmd == (int)TARGET_SIOCGSTAMPNS_OLD) {
5491 if (host_to_target_timespec(arg, &ts)) {
5492 return -TARGET_EFAULT;
5494 } else{
5495 if (host_to_target_timespec64(arg, &ts)) {
5496 return -TARGET_EFAULT;
5500 return ret;
5503 #ifdef TIOCGPTPEER
5504 static abi_long do_ioctl_tiocgptpeer(const IOCTLEntry *ie, uint8_t *buf_temp,
5505 int fd, int cmd, abi_long arg)
5507 int flags = target_to_host_bitmask(arg, fcntl_flags_tbl);
5508 return get_errno(safe_ioctl(fd, ie->host_cmd, flags));
5510 #endif
5512 #ifdef HAVE_DRM_H
5514 static void unlock_drm_version(struct drm_version *host_ver,
5515 struct target_drm_version *target_ver,
5516 bool copy)
5518 unlock_user(host_ver->name, target_ver->name,
5519 copy ? host_ver->name_len : 0);
5520 unlock_user(host_ver->date, target_ver->date,
5521 copy ? host_ver->date_len : 0);
5522 unlock_user(host_ver->desc, target_ver->desc,
5523 copy ? host_ver->desc_len : 0);
5526 static inline abi_long target_to_host_drmversion(struct drm_version *host_ver,
5527 struct target_drm_version *target_ver)
5529 memset(host_ver, 0, sizeof(*host_ver));
5531 __get_user(host_ver->name_len, &target_ver->name_len);
5532 if (host_ver->name_len) {
5533 host_ver->name = lock_user(VERIFY_WRITE, target_ver->name,
5534 target_ver->name_len, 0);
5535 if (!host_ver->name) {
5536 return -EFAULT;
5540 __get_user(host_ver->date_len, &target_ver->date_len);
5541 if (host_ver->date_len) {
5542 host_ver->date = lock_user(VERIFY_WRITE, target_ver->date,
5543 target_ver->date_len, 0);
5544 if (!host_ver->date) {
5545 goto err;
5549 __get_user(host_ver->desc_len, &target_ver->desc_len);
5550 if (host_ver->desc_len) {
5551 host_ver->desc = lock_user(VERIFY_WRITE, target_ver->desc,
5552 target_ver->desc_len, 0);
5553 if (!host_ver->desc) {
5554 goto err;
5558 return 0;
5559 err:
5560 unlock_drm_version(host_ver, target_ver, false);
5561 return -EFAULT;
5564 static inline void host_to_target_drmversion(
5565 struct target_drm_version *target_ver,
5566 struct drm_version *host_ver)
5568 __put_user(host_ver->version_major, &target_ver->version_major);
5569 __put_user(host_ver->version_minor, &target_ver->version_minor);
5570 __put_user(host_ver->version_patchlevel, &target_ver->version_patchlevel);
5571 __put_user(host_ver->name_len, &target_ver->name_len);
5572 __put_user(host_ver->date_len, &target_ver->date_len);
5573 __put_user(host_ver->desc_len, &target_ver->desc_len);
5574 unlock_drm_version(host_ver, target_ver, true);
5577 static abi_long do_ioctl_drm(const IOCTLEntry *ie, uint8_t *buf_temp,
5578 int fd, int cmd, abi_long arg)
5580 struct drm_version *ver;
5581 struct target_drm_version *target_ver;
5582 abi_long ret;
5584 switch (ie->host_cmd) {
5585 case DRM_IOCTL_VERSION:
5586 if (!lock_user_struct(VERIFY_WRITE, target_ver, arg, 0)) {
5587 return -TARGET_EFAULT;
5589 ver = (struct drm_version *)buf_temp;
5590 ret = target_to_host_drmversion(ver, target_ver);
5591 if (!is_error(ret)) {
5592 ret = get_errno(safe_ioctl(fd, ie->host_cmd, ver));
5593 if (is_error(ret)) {
5594 unlock_drm_version(ver, target_ver, false);
5595 } else {
5596 host_to_target_drmversion(target_ver, ver);
5599 unlock_user_struct(target_ver, arg, 0);
5600 return ret;
5602 return -TARGET_ENOSYS;
5605 static abi_long do_ioctl_drm_i915_getparam(const IOCTLEntry *ie,
5606 struct drm_i915_getparam *gparam,
5607 int fd, abi_long arg)
5609 abi_long ret;
5610 int value;
5611 struct target_drm_i915_getparam *target_gparam;
5613 if (!lock_user_struct(VERIFY_READ, target_gparam, arg, 0)) {
5614 return -TARGET_EFAULT;
5617 __get_user(gparam->param, &target_gparam->param);
5618 gparam->value = &value;
5619 ret = get_errno(safe_ioctl(fd, ie->host_cmd, gparam));
5620 put_user_s32(value, target_gparam->value);
5622 unlock_user_struct(target_gparam, arg, 0);
5623 return ret;
5626 static abi_long do_ioctl_drm_i915(const IOCTLEntry *ie, uint8_t *buf_temp,
5627 int fd, int cmd, abi_long arg)
5629 switch (ie->host_cmd) {
5630 case DRM_IOCTL_I915_GETPARAM:
5631 return do_ioctl_drm_i915_getparam(ie,
5632 (struct drm_i915_getparam *)buf_temp,
5633 fd, arg);
5634 default:
5635 return -TARGET_ENOSYS;
5639 #endif
5641 static abi_long do_ioctl_TUNSETTXFILTER(const IOCTLEntry *ie, uint8_t *buf_temp,
5642 int fd, int cmd, abi_long arg)
5644 struct tun_filter *filter = (struct tun_filter *)buf_temp;
5645 struct tun_filter *target_filter;
5646 char *target_addr;
5648 assert(ie->access == IOC_W);
5650 target_filter = lock_user(VERIFY_READ, arg, sizeof(*target_filter), 1);
5651 if (!target_filter) {
5652 return -TARGET_EFAULT;
5654 filter->flags = tswap16(target_filter->flags);
5655 filter->count = tswap16(target_filter->count);
5656 unlock_user(target_filter, arg, 0);
5658 if (filter->count) {
5659 if (offsetof(struct tun_filter, addr) + filter->count * ETH_ALEN >
5660 MAX_STRUCT_SIZE) {
5661 return -TARGET_EFAULT;
5664 target_addr = lock_user(VERIFY_READ,
5665 arg + offsetof(struct tun_filter, addr),
5666 filter->count * ETH_ALEN, 1);
5667 if (!target_addr) {
5668 return -TARGET_EFAULT;
5670 memcpy(filter->addr, target_addr, filter->count * ETH_ALEN);
5671 unlock_user(target_addr, arg + offsetof(struct tun_filter, addr), 0);
5674 return get_errno(safe_ioctl(fd, ie->host_cmd, filter));
5677 IOCTLEntry ioctl_entries[] = {
5678 #define IOCTL(cmd, access, ...) \
5679 { TARGET_ ## cmd, cmd, #cmd, access, 0, { __VA_ARGS__ } },
5680 #define IOCTL_SPECIAL(cmd, access, dofn, ...) \
5681 { TARGET_ ## cmd, cmd, #cmd, access, dofn, { __VA_ARGS__ } },
5682 #define IOCTL_IGNORE(cmd) \
5683 { TARGET_ ## cmd, 0, #cmd },
5684 #include "ioctls.h"
5685 { 0, 0, },
5688 /* ??? Implement proper locking for ioctls. */
5689 /* do_ioctl() Must return target values and target errnos. */
5690 static abi_long do_ioctl(int fd, int cmd, abi_long arg)
5692 const IOCTLEntry *ie;
5693 const argtype *arg_type;
5694 abi_long ret;
5695 uint8_t buf_temp[MAX_STRUCT_SIZE];
5696 int target_size;
5697 void *argptr;
5699 ie = ioctl_entries;
5700 for(;;) {
5701 if (ie->target_cmd == 0) {
5702 qemu_log_mask(
5703 LOG_UNIMP, "Unsupported ioctl: cmd=0x%04lx\n", (long)cmd);
5704 return -TARGET_ENOSYS;
5706 if (ie->target_cmd == cmd)
5707 break;
5708 ie++;
5710 arg_type = ie->arg_type;
5711 if (ie->do_ioctl) {
5712 return ie->do_ioctl(ie, buf_temp, fd, cmd, arg);
5713 } else if (!ie->host_cmd) {
5714 /* Some architectures define BSD ioctls in their headers
5715 that are not implemented in Linux. */
5716 return -TARGET_ENOSYS;
5719 switch(arg_type[0]) {
5720 case TYPE_NULL:
5721 /* no argument */
5722 ret = get_errno(safe_ioctl(fd, ie->host_cmd));
5723 break;
5724 case TYPE_PTRVOID:
5725 case TYPE_INT:
5726 case TYPE_LONG:
5727 case TYPE_ULONG:
5728 ret = get_errno(safe_ioctl(fd, ie->host_cmd, arg));
5729 break;
5730 case TYPE_PTR:
5731 arg_type++;
5732 target_size = thunk_type_size(arg_type, 0);
5733 switch(ie->access) {
5734 case IOC_R:
5735 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
5736 if (!is_error(ret)) {
5737 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
5738 if (!argptr)
5739 return -TARGET_EFAULT;
5740 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET);
5741 unlock_user(argptr, arg, target_size);
5743 break;
5744 case IOC_W:
5745 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5746 if (!argptr)
5747 return -TARGET_EFAULT;
5748 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
5749 unlock_user(argptr, arg, 0);
5750 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
5751 break;
5752 default:
5753 case IOC_RW:
5754 argptr = lock_user(VERIFY_READ, arg, target_size, 1);
5755 if (!argptr)
5756 return -TARGET_EFAULT;
5757 thunk_convert(buf_temp, argptr, arg_type, THUNK_HOST);
5758 unlock_user(argptr, arg, 0);
5759 ret = get_errno(safe_ioctl(fd, ie->host_cmd, buf_temp));
5760 if (!is_error(ret)) {
5761 argptr = lock_user(VERIFY_WRITE, arg, target_size, 0);
5762 if (!argptr)
5763 return -TARGET_EFAULT;
5764 thunk_convert(argptr, buf_temp, arg_type, THUNK_TARGET);
5765 unlock_user(argptr, arg, target_size);
5767 break;
5769 break;
5770 default:
5771 qemu_log_mask(LOG_UNIMP,
5772 "Unsupported ioctl type: cmd=0x%04lx type=%d\n",
5773 (long)cmd, arg_type[0]);
5774 ret = -TARGET_ENOSYS;
5775 break;
5777 return ret;
5780 static const bitmask_transtbl iflag_tbl[] = {
5781 { TARGET_IGNBRK, TARGET_IGNBRK, IGNBRK, IGNBRK },
5782 { TARGET_BRKINT, TARGET_BRKINT, BRKINT, BRKINT },
5783 { TARGET_IGNPAR, TARGET_IGNPAR, IGNPAR, IGNPAR },
5784 { TARGET_PARMRK, TARGET_PARMRK, PARMRK, PARMRK },
5785 { TARGET_INPCK, TARGET_INPCK, INPCK, INPCK },
5786 { TARGET_ISTRIP, TARGET_ISTRIP, ISTRIP, ISTRIP },
5787 { TARGET_INLCR, TARGET_INLCR, INLCR, INLCR },
5788 { TARGET_IGNCR, TARGET_IGNCR, IGNCR, IGNCR },
5789 { TARGET_ICRNL, TARGET_ICRNL, ICRNL, ICRNL },
5790 { TARGET_IUCLC, TARGET_IUCLC, IUCLC, IUCLC },
5791 { TARGET_IXON, TARGET_IXON, IXON, IXON },
5792 { TARGET_IXANY, TARGET_IXANY, IXANY, IXANY },
5793 { TARGET_IXOFF, TARGET_IXOFF, IXOFF, IXOFF },
5794 { TARGET_IMAXBEL, TARGET_IMAXBEL, IMAXBEL, IMAXBEL },
5795 { TARGET_IUTF8, TARGET_IUTF8, IUTF8, IUTF8},
5796 { 0, 0, 0, 0 }
5799 static const bitmask_transtbl oflag_tbl[] = {
5800 { TARGET_OPOST, TARGET_OPOST, OPOST, OPOST },
5801 { TARGET_OLCUC, TARGET_OLCUC, OLCUC, OLCUC },
5802 { TARGET_ONLCR, TARGET_ONLCR, ONLCR, ONLCR },
5803 { TARGET_OCRNL, TARGET_OCRNL, OCRNL, OCRNL },
5804 { TARGET_ONOCR, TARGET_ONOCR, ONOCR, ONOCR },
5805 { TARGET_ONLRET, TARGET_ONLRET, ONLRET, ONLRET },
5806 { TARGET_OFILL, TARGET_OFILL, OFILL, OFILL },
5807 { TARGET_OFDEL, TARGET_OFDEL, OFDEL, OFDEL },
5808 { TARGET_NLDLY, TARGET_NL0, NLDLY, NL0 },
5809 { TARGET_NLDLY, TARGET_NL1, NLDLY, NL1 },
5810 { TARGET_CRDLY, TARGET_CR0, CRDLY, CR0 },
5811 { TARGET_CRDLY, TARGET_CR1, CRDLY, CR1 },
5812 { TARGET_CRDLY, TARGET_CR2, CRDLY, CR2 },
5813 { TARGET_CRDLY, TARGET_CR3, CRDLY, CR3 },
5814 { TARGET_TABDLY, TARGET_TAB0, TABDLY, TAB0 },
5815 { TARGET_TABDLY, TARGET_TAB1, TABDLY, TAB1 },
5816 { TARGET_TABDLY, TARGET_TAB2, TABDLY, TAB2 },
5817 { TARGET_TABDLY, TARGET_TAB3, TABDLY, TAB3 },
5818 { TARGET_BSDLY, TARGET_BS0, BSDLY, BS0 },
5819 { TARGET_BSDLY, TARGET_BS1, BSDLY, BS1 },
5820 { TARGET_VTDLY, TARGET_VT0, VTDLY, VT0 },
5821 { TARGET_VTDLY, TARGET_VT1, VTDLY, VT1 },
5822 { TARGET_FFDLY, TARGET_FF0, FFDLY, FF0 },
5823 { TARGET_FFDLY, TARGET_FF1, FFDLY, FF1 },
5824 { 0, 0, 0, 0 }
5827 static const bitmask_transtbl cflag_tbl[] = {
5828 { TARGET_CBAUD, TARGET_B0, CBAUD, B0 },
5829 { TARGET_CBAUD, TARGET_B50, CBAUD, B50 },
5830 { TARGET_CBAUD, TARGET_B75, CBAUD, B75 },
5831 { TARGET_CBAUD, TARGET_B110, CBAUD, B110 },
5832 { TARGET_CBAUD, TARGET_B134, CBAUD, B134 },
5833 { TARGET_CBAUD, TARGET_B150, CBAUD, B150 },
5834 { TARGET_CBAUD, TARGET_B200, CBAUD, B200 },
5835 { TARGET_CBAUD, TARGET_B300, CBAUD, B300 },
5836 { TARGET_CBAUD, TARGET_B600, CBAUD, B600 },
5837 { TARGET_CBAUD, TARGET_B1200, CBAUD, B1200 },
5838 { TARGET_CBAUD, TARGET_B1800, CBAUD, B1800 },
5839 { TARGET_CBAUD, TARGET_B2400, CBAUD, B2400 },
5840 { TARGET_CBAUD, TARGET_B4800, CBAUD, B4800 },
5841 { TARGET_CBAUD, TARGET_B9600, CBAUD, B9600 },
5842 { TARGET_CBAUD, TARGET_B19200, CBAUD, B19200 },
5843 { TARGET_CBAUD, TARGET_B38400, CBAUD, B38400 },
5844 { TARGET_CBAUD, TARGET_B57600, CBAUD, B57600 },
5845 { TARGET_CBAUD, TARGET_B115200, CBAUD, B115200 },
5846 { TARGET_CBAUD, TARGET_B230400, CBAUD, B230400 },
5847 { TARGET_CBAUD, TARGET_B460800, CBAUD, B460800 },
5848 { TARGET_CSIZE, TARGET_CS5, CSIZE, CS5 },
5849 { TARGET_CSIZE, TARGET_CS6, CSIZE, CS6 },
5850 { TARGET_CSIZE, TARGET_CS7, CSIZE, CS7 },
5851 { TARGET_CSIZE, TARGET_CS8, CSIZE, CS8 },
5852 { TARGET_CSTOPB, TARGET_CSTOPB, CSTOPB, CSTOPB },
5853 { TARGET_CREAD, TARGET_CREAD, CREAD, CREAD },
5854 { TARGET_PARENB, TARGET_PARENB, PARENB, PARENB },
5855 { TARGET_PARODD, TARGET_PARODD, PARODD, PARODD },
5856 { TARGET_HUPCL, TARGET_HUPCL, HUPCL, HUPCL },
5857 { TARGET_CLOCAL, TARGET_CLOCAL, CLOCAL, CLOCAL },
5858 { TARGET_CRTSCTS, TARGET_CRTSCTS, CRTSCTS, CRTSCTS },
5859 { 0, 0, 0, 0 }
5862 static const bitmask_transtbl lflag_tbl[] = {
5863 { TARGET_ISIG, TARGET_ISIG, ISIG, ISIG },
5864 { TARGET_ICANON, TARGET_ICANON, ICANON, ICANON },
5865 { TARGET_XCASE, TARGET_XCASE, XCASE, XCASE },
5866 { TARGET_ECHO, TARGET_ECHO, ECHO, ECHO },
5867 { TARGET_ECHOE, TARGET_ECHOE, ECHOE, ECHOE },
5868 { TARGET_ECHOK, TARGET_ECHOK, ECHOK, ECHOK },
5869 { TARGET_ECHONL, TARGET_ECHONL, ECHONL, ECHONL },
5870 { TARGET_NOFLSH, TARGET_NOFLSH, NOFLSH, NOFLSH },
5871 { TARGET_TOSTOP, TARGET_TOSTOP, TOSTOP, TOSTOP },
5872 { TARGET_ECHOCTL, TARGET_ECHOCTL, ECHOCTL, ECHOCTL },
5873 { TARGET_ECHOPRT, TARGET_ECHOPRT, ECHOPRT, ECHOPRT },
5874 { TARGET_ECHOKE, TARGET_ECHOKE, ECHOKE, ECHOKE },
5875 { TARGET_FLUSHO, TARGET_FLUSHO, FLUSHO, FLUSHO },
5876 { TARGET_PENDIN, TARGET_PENDIN, PENDIN, PENDIN },
5877 { TARGET_IEXTEN, TARGET_IEXTEN, IEXTEN, IEXTEN },
5878 { TARGET_EXTPROC, TARGET_EXTPROC, EXTPROC, EXTPROC},
5879 { 0, 0, 0, 0 }
5882 static void target_to_host_termios (void *dst, const void *src)
5884 struct host_termios *host = dst;
5885 const struct target_termios *target = src;
5887 host->c_iflag =
5888 target_to_host_bitmask(tswap32(target->c_iflag), iflag_tbl);
5889 host->c_oflag =
5890 target_to_host_bitmask(tswap32(target->c_oflag), oflag_tbl);
5891 host->c_cflag =
5892 target_to_host_bitmask(tswap32(target->c_cflag), cflag_tbl);
5893 host->c_lflag =
5894 target_to_host_bitmask(tswap32(target->c_lflag), lflag_tbl);
5895 host->c_line = target->c_line;
5897 memset(host->c_cc, 0, sizeof(host->c_cc));
5898 host->c_cc[VINTR] = target->c_cc[TARGET_VINTR];
5899 host->c_cc[VQUIT] = target->c_cc[TARGET_VQUIT];
5900 host->c_cc[VERASE] = target->c_cc[TARGET_VERASE];
5901 host->c_cc[VKILL] = target->c_cc[TARGET_VKILL];
5902 host->c_cc[VEOF] = target->c_cc[TARGET_VEOF];
5903 host->c_cc[VTIME] = target->c_cc[TARGET_VTIME];
5904 host->c_cc[VMIN] = target->c_cc[TARGET_VMIN];
5905 host->c_cc[VSWTC] = target->c_cc[TARGET_VSWTC];
5906 host->c_cc[VSTART] = target->c_cc[TARGET_VSTART];
5907 host->c_cc[VSTOP] = target->c_cc[TARGET_VSTOP];
5908 host->c_cc[VSUSP] = target->c_cc[TARGET_VSUSP];
5909 host->c_cc[VEOL] = target->c_cc[TARGET_VEOL];
5910 host->c_cc[VREPRINT] = target->c_cc[TARGET_VREPRINT];
5911 host->c_cc[VDISCARD] = target->c_cc[TARGET_VDISCARD];
5912 host->c_cc[VWERASE] = target->c_cc[TARGET_VWERASE];
5913 host->c_cc[VLNEXT] = target->c_cc[TARGET_VLNEXT];
5914 host->c_cc[VEOL2] = target->c_cc[TARGET_VEOL2];
5917 static void host_to_target_termios (void *dst, const void *src)
5919 struct target_termios *target = dst;
5920 const struct host_termios *host = src;
5922 target->c_iflag =
5923 tswap32(host_to_target_bitmask(host->c_iflag, iflag_tbl));
5924 target->c_oflag =
5925 tswap32(host_to_target_bitmask(host->c_oflag, oflag_tbl));
5926 target->c_cflag =
5927 tswap32(host_to_target_bitmask(host->c_cflag, cflag_tbl));
5928 target->c_lflag =
5929 tswap32(host_to_target_bitmask(host->c_lflag, lflag_tbl));
5930 target->c_line = host->c_line;
5932 memset(target->c_cc, 0, sizeof(target->c_cc));
5933 target->c_cc[TARGET_VINTR] = host->c_cc[VINTR];
5934 target->c_cc[TARGET_VQUIT] = host->c_cc[VQUIT];
5935 target->c_cc[TARGET_VERASE] = host->c_cc[VERASE];
5936 target->c_cc[TARGET_VKILL] = host->c_cc[VKILL];
5937 target->c_cc[TARGET_VEOF] = host->c_cc[VEOF];
5938 target->c_cc[TARGET_VTIME] = host->c_cc[VTIME];
5939 target->c_cc[TARGET_VMIN] = host->c_cc[VMIN];
5940 target->c_cc[TARGET_VSWTC] = host->c_cc[VSWTC];
5941 target->c_cc[TARGET_VSTART] = host->c_cc[VSTART];
5942 target->c_cc[TARGET_VSTOP] = host->c_cc[VSTOP];
5943 target->c_cc[TARGET_VSUSP] = host->c_cc[VSUSP];
5944 target->c_cc[TARGET_VEOL] = host->c_cc[VEOL];
5945 target->c_cc[TARGET_VREPRINT] = host->c_cc[VREPRINT];
5946 target->c_cc[TARGET_VDISCARD] = host->c_cc[VDISCARD];
5947 target->c_cc[TARGET_VWERASE] = host->c_cc[VWERASE];
5948 target->c_cc[TARGET_VLNEXT] = host->c_cc[VLNEXT];
5949 target->c_cc[TARGET_VEOL2] = host->c_cc[VEOL2];
5952 static const StructEntry struct_termios_def = {
5953 .convert = { host_to_target_termios, target_to_host_termios },
5954 .size = { sizeof(struct target_termios), sizeof(struct host_termios) },
5955 .align = { __alignof__(struct target_termios), __alignof__(struct host_termios) },
5956 .print = print_termios,
5959 static const bitmask_transtbl mmap_flags_tbl[] = {
5960 { TARGET_MAP_SHARED, TARGET_MAP_SHARED, MAP_SHARED, MAP_SHARED },
5961 { TARGET_MAP_PRIVATE, TARGET_MAP_PRIVATE, MAP_PRIVATE, MAP_PRIVATE },
5962 { TARGET_MAP_FIXED, TARGET_MAP_FIXED, MAP_FIXED, MAP_FIXED },
5963 { TARGET_MAP_ANONYMOUS, TARGET_MAP_ANONYMOUS,
5964 MAP_ANONYMOUS, MAP_ANONYMOUS },
5965 { TARGET_MAP_GROWSDOWN, TARGET_MAP_GROWSDOWN,
5966 MAP_GROWSDOWN, MAP_GROWSDOWN },
5967 { TARGET_MAP_DENYWRITE, TARGET_MAP_DENYWRITE,
5968 MAP_DENYWRITE, MAP_DENYWRITE },
5969 { TARGET_MAP_EXECUTABLE, TARGET_MAP_EXECUTABLE,
5970 MAP_EXECUTABLE, MAP_EXECUTABLE },
5971 { TARGET_MAP_LOCKED, TARGET_MAP_LOCKED, MAP_LOCKED, MAP_LOCKED },
5972 { TARGET_MAP_NORESERVE, TARGET_MAP_NORESERVE,
5973 MAP_NORESERVE, MAP_NORESERVE },
5974 { TARGET_MAP_HUGETLB, TARGET_MAP_HUGETLB, MAP_HUGETLB, MAP_HUGETLB },
5975 /* MAP_STACK had been ignored by the kernel for quite some time.
5976 Recognize it for the target insofar as we do not want to pass
5977 it through to the host. */
5978 { TARGET_MAP_STACK, TARGET_MAP_STACK, 0, 0 },
5979 { 0, 0, 0, 0 }
5983 * NOTE: TARGET_ABI32 is defined for TARGET_I386 (but not for TARGET_X86_64)
5984 * TARGET_I386 is defined if TARGET_X86_64 is defined
5986 #if defined(TARGET_I386)
5988 /* NOTE: there is really one LDT for all the threads */
5989 static uint8_t *ldt_table;
5991 static abi_long read_ldt(abi_ulong ptr, unsigned long bytecount)
5993 int size;
5994 void *p;
5996 if (!ldt_table)
5997 return 0;
5998 size = TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE;
5999 if (size > bytecount)
6000 size = bytecount;
6001 p = lock_user(VERIFY_WRITE, ptr, size, 0);
6002 if (!p)
6003 return -TARGET_EFAULT;
6004 /* ??? Should this by byteswapped? */
6005 memcpy(p, ldt_table, size);
6006 unlock_user(p, ptr, size);
6007 return size;
6010 /* XXX: add locking support */
6011 static abi_long write_ldt(CPUX86State *env,
6012 abi_ulong ptr, unsigned long bytecount, int oldmode)
6014 struct target_modify_ldt_ldt_s ldt_info;
6015 struct target_modify_ldt_ldt_s *target_ldt_info;
6016 int seg_32bit, contents, read_exec_only, limit_in_pages;
6017 int seg_not_present, useable, lm;
6018 uint32_t *lp, entry_1, entry_2;
6020 if (bytecount != sizeof(ldt_info))
6021 return -TARGET_EINVAL;
6022 if (!lock_user_struct(VERIFY_READ, target_ldt_info, ptr, 1))
6023 return -TARGET_EFAULT;
6024 ldt_info.entry_number = tswap32(target_ldt_info->entry_number);
6025 ldt_info.base_addr = tswapal(target_ldt_info->base_addr);
6026 ldt_info.limit = tswap32(target_ldt_info->limit);
6027 ldt_info.flags = tswap32(target_ldt_info->flags);
6028 unlock_user_struct(target_ldt_info, ptr, 0);
6030 if (ldt_info.entry_number >= TARGET_LDT_ENTRIES)
6031 return -TARGET_EINVAL;
6032 seg_32bit = ldt_info.flags & 1;
6033 contents = (ldt_info.flags >> 1) & 3;
6034 read_exec_only = (ldt_info.flags >> 3) & 1;
6035 limit_in_pages = (ldt_info.flags >> 4) & 1;
6036 seg_not_present = (ldt_info.flags >> 5) & 1;
6037 useable = (ldt_info.flags >> 6) & 1;
6038 #ifdef TARGET_ABI32
6039 lm = 0;
6040 #else
6041 lm = (ldt_info.flags >> 7) & 1;
6042 #endif
6043 if (contents == 3) {
6044 if (oldmode)
6045 return -TARGET_EINVAL;
6046 if (seg_not_present == 0)
6047 return -TARGET_EINVAL;
6049 /* allocate the LDT */
6050 if (!ldt_table) {
6051 env->ldt.base = target_mmap(0,
6052 TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE,
6053 PROT_READ|PROT_WRITE,
6054 MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
6055 if (env->ldt.base == -1)
6056 return -TARGET_ENOMEM;
6057 memset(g2h_untagged(env->ldt.base), 0,
6058 TARGET_LDT_ENTRIES * TARGET_LDT_ENTRY_SIZE);
6059 env->ldt.limit = 0xffff;
6060 ldt_table = g2h_untagged(env->ldt.base);
6063 /* NOTE: same code as Linux kernel */
6064 /* Allow LDTs to be cleared by the user. */
6065 if (ldt_info.base_addr == 0 && ldt_info.limit == 0) {
6066 if (oldmode ||
6067 (contents == 0 &&
6068 read_exec_only == 1 &&
6069 seg_32bit == 0 &&
6070 limit_in_pages == 0 &&
6071 seg_not_present == 1 &&
6072 useable == 0 )) {
6073 entry_1 = 0;
6074 entry_2 = 0;
6075 goto install;
6079 entry_1 = ((ldt_info.base_addr & 0x0000ffff) << 16) |
6080 (ldt_info.limit & 0x0ffff);
6081 entry_2 = (ldt_info.base_addr & 0xff000000) |
6082 ((ldt_info.base_addr & 0x00ff0000) >> 16) |
6083 (ldt_info.limit & 0xf0000) |
6084 ((read_exec_only ^ 1) << 9) |
6085 (contents << 10) |
6086 ((seg_not_present ^ 1) << 15) |
6087 (seg_32bit << 22) |
6088 (limit_in_pages << 23) |
6089 (lm << 21) |
6090 0x7000;
6091 if (!oldmode)
6092 entry_2 |= (useable << 20);
6094 /* Install the new entry ... */
6095 install:
6096 lp = (uint32_t *)(ldt_table + (ldt_info.entry_number << 3));
6097 lp[0] = tswap32(entry_1);
6098 lp[1] = tswap32(entry_2);
6099 return 0;
6102 /* specific and weird i386 syscalls */
6103 static abi_long do_modify_ldt(CPUX86State *env, int func, abi_ulong ptr,
6104 unsigned long bytecount)
6106 abi_long ret;
6108 switch (func) {
6109 case 0:
6110 ret = read_ldt(ptr, bytecount);
6111 break;
6112 case 1:
6113 ret = write_ldt(env, ptr, bytecount, 1);
6114 break;
6115 case 0x11:
6116 ret = write_ldt(env, ptr, bytecount, 0);
6117 break;
6118 default:
6119 ret = -TARGET_ENOSYS;
6120 break;
6122 return ret;
6125 #if defined(TARGET_ABI32)
6126 abi_long do_set_thread_area(CPUX86State *env, abi_ulong ptr)
6128 uint64_t *gdt_table = g2h_untagged(env->gdt.base);
6129 struct target_modify_ldt_ldt_s ldt_info;
6130 struct target_modify_ldt_ldt_s *target_ldt_info;
6131 int seg_32bit, contents, read_exec_only, limit_in_pages;
6132 int seg_not_present, useable, lm;
6133 uint32_t *lp, entry_1, entry_2;
6134 int i;
6136 lock_user_struct(VERIFY_WRITE, target_ldt_info, ptr, 1);
6137 if (!target_ldt_info)
6138 return -TARGET_EFAULT;
6139 ldt_info.entry_number = tswap32(target_ldt_info->entry_number);
6140 ldt_info.base_addr = tswapal(target_ldt_info->base_addr);
6141 ldt_info.limit = tswap32(target_ldt_info->limit);
6142 ldt_info.flags = tswap32(target_ldt_info->flags);
6143 if (ldt_info.entry_number == -1) {
6144 for (i=TARGET_GDT_ENTRY_TLS_MIN; i<=TARGET_GDT_ENTRY_TLS_MAX; i++) {
6145 if (gdt_table[i] == 0) {
6146 ldt_info.entry_number = i;
6147 target_ldt_info->entry_number = tswap32(i);
6148 break;
6152 unlock_user_struct(target_ldt_info, ptr, 1);
6154 if (ldt_info.entry_number < TARGET_GDT_ENTRY_TLS_MIN ||
6155 ldt_info.entry_number > TARGET_GDT_ENTRY_TLS_MAX)
6156 return -TARGET_EINVAL;
6157 seg_32bit = ldt_info.flags & 1;
6158 contents = (ldt_info.flags >> 1) & 3;
6159 read_exec_only = (ldt_info.flags >> 3) & 1;
6160 limit_in_pages = (ldt_info.flags >> 4) & 1;
6161 seg_not_present = (ldt_info.flags >> 5) & 1;
6162 useable = (ldt_info.flags >> 6) & 1;
6163 #ifdef TARGET_ABI32
6164 lm = 0;
6165 #else
6166 lm = (ldt_info.flags >> 7) & 1;
6167 #endif
6169 if (contents == 3) {
6170 if (seg_not_present == 0)
6171 return -TARGET_EINVAL;
6174 /* NOTE: same code as Linux kernel */
6175 /* Allow LDTs to be cleared by the user. */
6176 if (ldt_info.base_addr == 0 && ldt_info.limit == 0) {
6177 if ((contents == 0 &&
6178 read_exec_only == 1 &&
6179 seg_32bit == 0 &&
6180 limit_in_pages == 0 &&
6181 seg_not_present == 1 &&
6182 useable == 0 )) {
6183 entry_1 = 0;
6184 entry_2 = 0;
6185 goto install;
6189 entry_1 = ((ldt_info.base_addr & 0x0000ffff) << 16) |
6190 (ldt_info.limit & 0x0ffff);
6191 entry_2 = (ldt_info.base_addr & 0xff000000) |
6192 ((ldt_info.base_addr & 0x00ff0000) >> 16) |
6193 (ldt_info.limit & 0xf0000) |
6194 ((read_exec_only ^ 1) << 9) |
6195 (contents << 10) |
6196 ((seg_not_present ^ 1) << 15) |
6197 (seg_32bit << 22) |
6198 (limit_in_pages << 23) |
6199 (useable << 20) |
6200 (lm << 21) |
6201 0x7000;
6203 /* Install the new entry ... */
6204 install:
6205 lp = (uint32_t *)(gdt_table + ldt_info.entry_number);
6206 lp[0] = tswap32(entry_1);
6207 lp[1] = tswap32(entry_2);
6208 return 0;
6211 static abi_long do_get_thread_area(CPUX86State *env, abi_ulong ptr)
6213 struct target_modify_ldt_ldt_s *target_ldt_info;
6214 uint64_t *gdt_table = g2h_untagged(env->gdt.base);
6215 uint32_t base_addr, limit, flags;
6216 int seg_32bit, contents, read_exec_only, limit_in_pages, idx;
6217 int seg_not_present, useable, lm;
6218 uint32_t *lp, entry_1, entry_2;
6220 lock_user_struct(VERIFY_WRITE, target_ldt_info, ptr, 1);
6221 if (!target_ldt_info)
6222 return -TARGET_EFAULT;
6223 idx = tswap32(target_ldt_info->entry_number);
6224 if (idx < TARGET_GDT_ENTRY_TLS_MIN ||
6225 idx > TARGET_GDT_ENTRY_TLS_MAX) {
6226 unlock_user_struct(target_ldt_info, ptr, 1);
6227 return -TARGET_EINVAL;
6229 lp = (uint32_t *)(gdt_table + idx);
6230 entry_1 = tswap32(lp[0]);
6231 entry_2 = tswap32(lp[1]);
6233 read_exec_only = ((entry_2 >> 9) & 1) ^ 1;
6234 contents = (entry_2 >> 10) & 3;
6235 seg_not_present = ((entry_2 >> 15) & 1) ^ 1;
6236 seg_32bit = (entry_2 >> 22) & 1;
6237 limit_in_pages = (entry_2 >> 23) & 1;
6238 useable = (entry_2 >> 20) & 1;
6239 #ifdef TARGET_ABI32
6240 lm = 0;
6241 #else
6242 lm = (entry_2 >> 21) & 1;
6243 #endif
6244 flags = (seg_32bit << 0) | (contents << 1) |
6245 (read_exec_only << 3) | (limit_in_pages << 4) |
6246 (seg_not_present << 5) | (useable << 6) | (lm << 7);
6247 limit = (entry_1 & 0xffff) | (entry_2 & 0xf0000);
6248 base_addr = (entry_1 >> 16) |
6249 (entry_2 & 0xff000000) |
6250 ((entry_2 & 0xff) << 16);
6251 target_ldt_info->base_addr = tswapal(base_addr);
6252 target_ldt_info->limit = tswap32(limit);
6253 target_ldt_info->flags = tswap32(flags);
6254 unlock_user_struct(target_ldt_info, ptr, 1);
6255 return 0;
6258 abi_long do_arch_prctl(CPUX86State *env, int code, abi_ulong addr)
6260 return -TARGET_ENOSYS;
6262 #else
6263 abi_long do_arch_prctl(CPUX86State *env, int code, abi_ulong addr)
6265 abi_long ret = 0;
6266 abi_ulong val;
6267 int idx;
6269 switch(code) {
6270 case TARGET_ARCH_SET_GS:
6271 case TARGET_ARCH_SET_FS:
6272 if (code == TARGET_ARCH_SET_GS)
6273 idx = R_GS;
6274 else
6275 idx = R_FS;
6276 cpu_x86_load_seg(env, idx, 0);
6277 env->segs[idx].base = addr;
6278 break;
6279 case TARGET_ARCH_GET_GS:
6280 case TARGET_ARCH_GET_FS:
6281 if (code == TARGET_ARCH_GET_GS)
6282 idx = R_GS;
6283 else
6284 idx = R_FS;
6285 val = env->segs[idx].base;
6286 if (put_user(val, addr, abi_ulong))
6287 ret = -TARGET_EFAULT;
6288 break;
6289 default:
6290 ret = -TARGET_EINVAL;
6291 break;
6293 return ret;
6295 #endif /* defined(TARGET_ABI32 */
6296 #endif /* defined(TARGET_I386) */
6299 * These constants are generic. Supply any that are missing from the host.
6301 #ifndef PR_SET_NAME
6302 # define PR_SET_NAME 15
6303 # define PR_GET_NAME 16
6304 #endif
6305 #ifndef PR_SET_FP_MODE
6306 # define PR_SET_FP_MODE 45
6307 # define PR_GET_FP_MODE 46
6308 # define PR_FP_MODE_FR (1 << 0)
6309 # define PR_FP_MODE_FRE (1 << 1)
6310 #endif
6311 #ifndef PR_SVE_SET_VL
6312 # define PR_SVE_SET_VL 50
6313 # define PR_SVE_GET_VL 51
6314 # define PR_SVE_VL_LEN_MASK 0xffff
6315 # define PR_SVE_VL_INHERIT (1 << 17)
6316 #endif
6317 #ifndef PR_PAC_RESET_KEYS
6318 # define PR_PAC_RESET_KEYS 54
6319 # define PR_PAC_APIAKEY (1 << 0)
6320 # define PR_PAC_APIBKEY (1 << 1)
6321 # define PR_PAC_APDAKEY (1 << 2)
6322 # define PR_PAC_APDBKEY (1 << 3)
6323 # define PR_PAC_APGAKEY (1 << 4)
6324 #endif
6325 #ifndef PR_SET_TAGGED_ADDR_CTRL
6326 # define PR_SET_TAGGED_ADDR_CTRL 55
6327 # define PR_GET_TAGGED_ADDR_CTRL 56
6328 # define PR_TAGGED_ADDR_ENABLE (1UL << 0)
6329 #endif
6330 #ifndef PR_MTE_TCF_SHIFT
6331 # define PR_MTE_TCF_SHIFT 1
6332 # define PR_MTE_TCF_NONE (0UL << PR_MTE_TCF_SHIFT)
6333 # define PR_MTE_TCF_SYNC (1UL << PR_MTE_TCF_SHIFT)
6334 # define PR_MTE_TCF_ASYNC (2UL << PR_MTE_TCF_SHIFT)
6335 # define PR_MTE_TCF_MASK (3UL << PR_MTE_TCF_SHIFT)
6336 # define PR_MTE_TAG_SHIFT 3
6337 # define PR_MTE_TAG_MASK (0xffffUL << PR_MTE_TAG_SHIFT)
6338 #endif
6339 #ifndef PR_SET_IO_FLUSHER
6340 # define PR_SET_IO_FLUSHER 57
6341 # define PR_GET_IO_FLUSHER 58
6342 #endif
6343 #ifndef PR_SET_SYSCALL_USER_DISPATCH
6344 # define PR_SET_SYSCALL_USER_DISPATCH 59
6345 #endif
6347 #include "target_prctl.h"
6349 static abi_long do_prctl_inval0(CPUArchState *env)
6351 return -TARGET_EINVAL;
6354 static abi_long do_prctl_inval1(CPUArchState *env, abi_long arg2)
6356 return -TARGET_EINVAL;
6359 #ifndef do_prctl_get_fp_mode
6360 #define do_prctl_get_fp_mode do_prctl_inval0
6361 #endif
6362 #ifndef do_prctl_set_fp_mode
6363 #define do_prctl_set_fp_mode do_prctl_inval1
6364 #endif
6365 #ifndef do_prctl_get_vl
6366 #define do_prctl_get_vl do_prctl_inval0
6367 #endif
6368 #ifndef do_prctl_set_vl
6369 #define do_prctl_set_vl do_prctl_inval1
6370 #endif
6371 #ifndef do_prctl_reset_keys
6372 #define do_prctl_reset_keys do_prctl_inval1
6373 #endif
6374 #ifndef do_prctl_set_tagged_addr_ctrl
6375 #define do_prctl_set_tagged_addr_ctrl do_prctl_inval1
6376 #endif
6377 #ifndef do_prctl_get_tagged_addr_ctrl
6378 #define do_prctl_get_tagged_addr_ctrl do_prctl_inval0
6379 #endif
6380 #ifndef do_prctl_get_unalign
6381 #define do_prctl_get_unalign do_prctl_inval1
6382 #endif
6383 #ifndef do_prctl_set_unalign
6384 #define do_prctl_set_unalign do_prctl_inval1
6385 #endif
6387 static abi_long do_prctl(CPUArchState *env, abi_long option, abi_long arg2,
6388 abi_long arg3, abi_long arg4, abi_long arg5)
6390 abi_long ret;
6392 switch (option) {
6393 case PR_GET_PDEATHSIG:
6395 int deathsig;
6396 ret = get_errno(prctl(PR_GET_PDEATHSIG, &deathsig,
6397 arg3, arg4, arg5));
6398 if (!is_error(ret) &&
6399 put_user_s32(host_to_target_signal(deathsig), arg2)) {
6400 return -TARGET_EFAULT;
6402 return ret;
6404 case PR_SET_PDEATHSIG:
6405 return get_errno(prctl(PR_SET_PDEATHSIG, target_to_host_signal(arg2),
6406 arg3, arg4, arg5));
6407 case PR_GET_NAME:
6409 void *name = lock_user(VERIFY_WRITE, arg2, 16, 1);
6410 if (!name) {
6411 return -TARGET_EFAULT;
6413 ret = get_errno(prctl(PR_GET_NAME, (uintptr_t)name,
6414 arg3, arg4, arg5));
6415 unlock_user(name, arg2, 16);
6416 return ret;
6418 case PR_SET_NAME:
6420 void *name = lock_user(VERIFY_READ, arg2, 16, 1);
6421 if (!name) {
6422 return -TARGET_EFAULT;
6424 ret = get_errno(prctl(PR_SET_NAME, (uintptr_t)name,
6425 arg3, arg4, arg5));
6426 unlock_user(name, arg2, 0);
6427 return ret;
6429 case PR_GET_FP_MODE:
6430 return do_prctl_get_fp_mode(env);
6431 case PR_SET_FP_MODE:
6432 return do_prctl_set_fp_mode(env, arg2);
6433 case PR_SVE_GET_VL:
6434 return do_prctl_get_vl(env);
6435 case PR_SVE_SET_VL:
6436 return do_prctl_set_vl(env, arg2);
6437 case PR_PAC_RESET_KEYS:
6438 if (arg3 || arg4 || arg5) {
6439 return -TARGET_EINVAL;
6441 return do_prctl_reset_keys(env, arg2);
6442 case PR_SET_TAGGED_ADDR_CTRL:
6443 if (arg3 || arg4 || arg5) {
6444 return -TARGET_EINVAL;
6446 return do_prctl_set_tagged_addr_ctrl(env, arg2);
6447 case PR_GET_TAGGED_ADDR_CTRL:
6448 if (arg2 || arg3 || arg4 || arg5) {
6449 return -TARGET_EINVAL;
6451 return do_prctl_get_tagged_addr_ctrl(env);
6453 case PR_GET_UNALIGN:
6454 return do_prctl_get_unalign(env, arg2);
6455 case PR_SET_UNALIGN:
6456 return do_prctl_set_unalign(env, arg2);
6458 case PR_CAP_AMBIENT:
6459 case PR_CAPBSET_READ:
6460 case PR_CAPBSET_DROP:
6461 case PR_GET_DUMPABLE:
6462 case PR_SET_DUMPABLE:
6463 case PR_GET_KEEPCAPS:
6464 case PR_SET_KEEPCAPS:
6465 case PR_GET_SECUREBITS:
6466 case PR_SET_SECUREBITS:
6467 case PR_GET_TIMING:
6468 case PR_SET_TIMING:
6469 case PR_GET_TIMERSLACK:
6470 case PR_SET_TIMERSLACK:
6471 case PR_MCE_KILL:
6472 case PR_MCE_KILL_GET:
6473 case PR_GET_NO_NEW_PRIVS:
6474 case PR_SET_NO_NEW_PRIVS:
6475 case PR_GET_IO_FLUSHER:
6476 case PR_SET_IO_FLUSHER:
6477 /* Some prctl options have no pointer arguments and we can pass on. */
6478 return get_errno(prctl(option, arg2, arg3, arg4, arg5));
6480 case PR_GET_CHILD_SUBREAPER:
6481 case PR_SET_CHILD_SUBREAPER:
6482 case PR_GET_SPECULATION_CTRL:
6483 case PR_SET_SPECULATION_CTRL:
6484 case PR_GET_TID_ADDRESS:
6485 /* TODO */
6486 return -TARGET_EINVAL;
6488 case PR_GET_FPEXC:
6489 case PR_SET_FPEXC:
6490 /* Was used for SPE on PowerPC. */
6491 return -TARGET_EINVAL;
6493 case PR_GET_ENDIAN:
6494 case PR_SET_ENDIAN:
6495 case PR_GET_FPEMU:
6496 case PR_SET_FPEMU:
6497 case PR_SET_MM:
6498 case PR_GET_SECCOMP:
6499 case PR_SET_SECCOMP:
6500 case PR_SET_SYSCALL_USER_DISPATCH:
6501 case PR_GET_THP_DISABLE:
6502 case PR_SET_THP_DISABLE:
6503 case PR_GET_TSC:
6504 case PR_SET_TSC:
6505 /* Disable to prevent the target disabling stuff we need. */
6506 return -TARGET_EINVAL;
6508 default:
6509 qemu_log_mask(LOG_UNIMP, "Unsupported prctl: " TARGET_ABI_FMT_ld "\n",
6510 option);
6511 return -TARGET_EINVAL;
6515 #define NEW_STACK_SIZE 0x40000
6518 static pthread_mutex_t clone_lock = PTHREAD_MUTEX_INITIALIZER;
6519 typedef struct {
6520 CPUArchState *env;
6521 pthread_mutex_t mutex;
6522 pthread_cond_t cond;
6523 pthread_t thread;
6524 uint32_t tid;
6525 abi_ulong child_tidptr;
6526 abi_ulong parent_tidptr;
6527 sigset_t sigmask;
6528 } new_thread_info;
6530 static void *clone_func(void *arg)
6532 new_thread_info *info = arg;
6533 CPUArchState *env;
6534 CPUState *cpu;
6535 TaskState *ts;
6537 rcu_register_thread();
6538 tcg_register_thread();
6539 env = info->env;
6540 cpu = env_cpu(env);
6541 thread_cpu = cpu;
6542 ts = (TaskState *)cpu->opaque;
6543 info->tid = sys_gettid();
6544 task_settid(ts);
6545 if (info->child_tidptr)
6546 put_user_u32(info->tid, info->child_tidptr);
6547 if (info->parent_tidptr)
6548 put_user_u32(info->tid, info->parent_tidptr);
6549 qemu_guest_random_seed_thread_part2(cpu->random_seed);
6550 /* Enable signals. */
6551 sigprocmask(SIG_SETMASK, &info->sigmask, NULL);
6552 /* Signal to the parent that we're ready. */
6553 pthread_mutex_lock(&info->mutex);
6554 pthread_cond_broadcast(&info->cond);
6555 pthread_mutex_unlock(&info->mutex);
6556 /* Wait until the parent has finished initializing the tls state. */
6557 pthread_mutex_lock(&clone_lock);
6558 pthread_mutex_unlock(&clone_lock);
6559 cpu_loop(env);
6560 /* never exits */
6561 return NULL;
6564 /* do_fork() Must return host values and target errnos (unlike most
6565 do_*() functions). */
6566 static int do_fork(CPUArchState *env, unsigned int flags, abi_ulong newsp,
6567 abi_ulong parent_tidptr, target_ulong newtls,
6568 abi_ulong child_tidptr)
6570 CPUState *cpu = env_cpu(env);
6571 int ret;
6572 TaskState *ts;
6573 CPUState *new_cpu;
6574 CPUArchState *new_env;
6575 sigset_t sigmask;
6577 flags &= ~CLONE_IGNORED_FLAGS;
6579 /* Emulate vfork() with fork() */
6580 if (flags & CLONE_VFORK)
6581 flags &= ~(CLONE_VFORK | CLONE_VM);
6583 if (flags & CLONE_VM) {
6584 TaskState *parent_ts = (TaskState *)cpu->opaque;
6585 new_thread_info info;
6586 pthread_attr_t attr;
6588 if (((flags & CLONE_THREAD_FLAGS) != CLONE_THREAD_FLAGS) ||
6589 (flags & CLONE_INVALID_THREAD_FLAGS)) {
6590 return -TARGET_EINVAL;
6593 ts = g_new0(TaskState, 1);
6594 init_task_state(ts);
6596 /* Grab a mutex so that thread setup appears atomic. */
6597 pthread_mutex_lock(&clone_lock);
6600 * If this is our first additional thread, we need to ensure we
6601 * generate code for parallel execution and flush old translations.
6602 * Do this now so that the copy gets CF_PARALLEL too.
6604 if (!(cpu->tcg_cflags & CF_PARALLEL)) {
6605 cpu->tcg_cflags |= CF_PARALLEL;
6606 tb_flush(cpu);
6609 /* we create a new CPU instance. */
6610 new_env = cpu_copy(env);
6611 /* Init regs that differ from the parent. */
6612 cpu_clone_regs_child(new_env, newsp, flags);
6613 cpu_clone_regs_parent(env, flags);
6614 new_cpu = env_cpu(new_env);
6615 new_cpu->opaque = ts;
6616 ts->bprm = parent_ts->bprm;
6617 ts->info = parent_ts->info;
6618 ts->signal_mask = parent_ts->signal_mask;
6620 if (flags & CLONE_CHILD_CLEARTID) {
6621 ts->child_tidptr = child_tidptr;
6624 if (flags & CLONE_SETTLS) {
6625 cpu_set_tls (new_env, newtls);
6628 memset(&info, 0, sizeof(info));
6629 pthread_mutex_init(&info.mutex, NULL);
6630 pthread_mutex_lock(&info.mutex);
6631 pthread_cond_init(&info.cond, NULL);
6632 info.env = new_env;
6633 if (flags & CLONE_CHILD_SETTID) {
6634 info.child_tidptr = child_tidptr;
6636 if (flags & CLONE_PARENT_SETTID) {
6637 info.parent_tidptr = parent_tidptr;
6640 ret = pthread_attr_init(&attr);
6641 ret = pthread_attr_setstacksize(&attr, NEW_STACK_SIZE);
6642 ret = pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED);
6643 /* It is not safe to deliver signals until the child has finished
6644 initializing, so temporarily block all signals. */
6645 sigfillset(&sigmask);
6646 sigprocmask(SIG_BLOCK, &sigmask, &info.sigmask);
6647 cpu->random_seed = qemu_guest_random_seed_thread_part1();
6649 ret = pthread_create(&info.thread, &attr, clone_func, &info);
6650 /* TODO: Free new CPU state if thread creation failed. */
6652 sigprocmask(SIG_SETMASK, &info.sigmask, NULL);
6653 pthread_attr_destroy(&attr);
6654 if (ret == 0) {
6655 /* Wait for the child to initialize. */
6656 pthread_cond_wait(&info.cond, &info.mutex);
6657 ret = info.tid;
6658 } else {
6659 ret = -1;
6661 pthread_mutex_unlock(&info.mutex);
6662 pthread_cond_destroy(&info.cond);
6663 pthread_mutex_destroy(&info.mutex);
6664 pthread_mutex_unlock(&clone_lock);
6665 } else {
6666 /* if no CLONE_VM, we consider it is a fork */
6667 if (flags & CLONE_INVALID_FORK_FLAGS) {
6668 return -TARGET_EINVAL;
6671 /* We can't support custom termination signals */
6672 if ((flags & CSIGNAL) != TARGET_SIGCHLD) {
6673 return -TARGET_EINVAL;
6676 if (block_signals()) {
6677 return -QEMU_ERESTARTSYS;
6680 fork_start();
6681 ret = fork();
6682 if (ret == 0) {
6683 /* Child Process. */
6684 cpu_clone_regs_child(env, newsp, flags);
6685 fork_end(1);
6686 /* There is a race condition here. The parent process could
6687 theoretically read the TID in the child process before the child
6688 tid is set. This would require using either ptrace
6689 (not implemented) or having *_tidptr to point at a shared memory
6690 mapping. We can't repeat the spinlock hack used above because
6691 the child process gets its own copy of the lock. */
6692 if (flags & CLONE_CHILD_SETTID)
6693 put_user_u32(sys_gettid(), child_tidptr);
6694 if (flags & CLONE_PARENT_SETTID)
6695 put_user_u32(sys_gettid(), parent_tidptr);
6696 ts = (TaskState *)cpu->opaque;
6697 if (flags & CLONE_SETTLS)
6698 cpu_set_tls (env, newtls);
6699 if (flags & CLONE_CHILD_CLEARTID)
6700 ts->child_tidptr = child_tidptr;
6701 } else {
6702 cpu_clone_regs_parent(env, flags);
6703 fork_end(0);
6706 return ret;
6709 /* warning : doesn't handle linux specific flags... */
6710 static int target_to_host_fcntl_cmd(int cmd)
6712 int ret;
6714 switch(cmd) {
6715 case TARGET_F_DUPFD:
6716 case TARGET_F_GETFD:
6717 case TARGET_F_SETFD:
6718 case TARGET_F_GETFL:
6719 case TARGET_F_SETFL:
6720 case TARGET_F_OFD_GETLK:
6721 case TARGET_F_OFD_SETLK:
6722 case TARGET_F_OFD_SETLKW:
6723 ret = cmd;
6724 break;
6725 case TARGET_F_GETLK:
6726 ret = F_GETLK64;
6727 break;
6728 case TARGET_F_SETLK:
6729 ret = F_SETLK64;
6730 break;
6731 case TARGET_F_SETLKW:
6732 ret = F_SETLKW64;
6733 break;
6734 case TARGET_F_GETOWN:
6735 ret = F_GETOWN;
6736 break;
6737 case TARGET_F_SETOWN:
6738 ret = F_SETOWN;
6739 break;
6740 case TARGET_F_GETSIG:
6741 ret = F_GETSIG;
6742 break;
6743 case TARGET_F_SETSIG:
6744 ret = F_SETSIG;
6745 break;
6746 #if TARGET_ABI_BITS == 32
6747 case TARGET_F_GETLK64:
6748 ret = F_GETLK64;
6749 break;
6750 case TARGET_F_SETLK64:
6751 ret = F_SETLK64;
6752 break;
6753 case TARGET_F_SETLKW64:
6754 ret = F_SETLKW64;
6755 break;
6756 #endif
6757 case TARGET_F_SETLEASE:
6758 ret = F_SETLEASE;
6759 break;
6760 case TARGET_F_GETLEASE:
6761 ret = F_GETLEASE;
6762 break;
6763 #ifdef F_DUPFD_CLOEXEC
6764 case TARGET_F_DUPFD_CLOEXEC:
6765 ret = F_DUPFD_CLOEXEC;
6766 break;
6767 #endif
6768 case TARGET_F_NOTIFY:
6769 ret = F_NOTIFY;
6770 break;
6771 #ifdef F_GETOWN_EX
6772 case TARGET_F_GETOWN_EX:
6773 ret = F_GETOWN_EX;
6774 break;
6775 #endif
6776 #ifdef F_SETOWN_EX
6777 case TARGET_F_SETOWN_EX:
6778 ret = F_SETOWN_EX;
6779 break;
6780 #endif
6781 #ifdef F_SETPIPE_SZ
6782 case TARGET_F_SETPIPE_SZ:
6783 ret = F_SETPIPE_SZ;
6784 break;
6785 case TARGET_F_GETPIPE_SZ:
6786 ret = F_GETPIPE_SZ;
6787 break;
6788 #endif
6789 #ifdef F_ADD_SEALS
6790 case TARGET_F_ADD_SEALS:
6791 ret = F_ADD_SEALS;
6792 break;
6793 case TARGET_F_GET_SEALS:
6794 ret = F_GET_SEALS;
6795 break;
6796 #endif
6797 default:
6798 ret = -TARGET_EINVAL;
6799 break;
6802 #if defined(__powerpc64__)
6803 /* On PPC64, glibc headers has the F_*LK* defined to 12, 13 and 14 and
6804 * is not supported by kernel. The glibc fcntl call actually adjusts
6805 * them to 5, 6 and 7 before making the syscall(). Since we make the
6806 * syscall directly, adjust to what is supported by the kernel.
6808 if (ret >= F_GETLK64 && ret <= F_SETLKW64) {
6809 ret -= F_GETLK64 - 5;
6811 #endif
6813 return ret;
6816 #define FLOCK_TRANSTBL \
6817 switch (type) { \
6818 TRANSTBL_CONVERT(F_RDLCK); \
6819 TRANSTBL_CONVERT(F_WRLCK); \
6820 TRANSTBL_CONVERT(F_UNLCK); \
6823 static int target_to_host_flock(int type)
6825 #define TRANSTBL_CONVERT(a) case TARGET_##a: return a
6826 FLOCK_TRANSTBL
6827 #undef TRANSTBL_CONVERT
6828 return -TARGET_EINVAL;
6831 static int host_to_target_flock(int type)
6833 #define TRANSTBL_CONVERT(a) case a: return TARGET_##a
6834 FLOCK_TRANSTBL
6835 #undef TRANSTBL_CONVERT
6836 /* if we don't know how to convert the value coming
6837 * from the host we copy to the target field as-is
6839 return type;
6842 static inline abi_long copy_from_user_flock(struct flock64 *fl,
6843 abi_ulong target_flock_addr)
6845 struct target_flock *target_fl;
6846 int l_type;
6848 if (!lock_user_struct(VERIFY_READ, target_fl, target_flock_addr, 1)) {
6849 return -TARGET_EFAULT;
6852 __get_user(l_type, &target_fl->l_type);
6853 l_type = target_to_host_flock(l_type);
6854 if (l_type < 0) {
6855 return l_type;
6857 fl->l_type = l_type;
6858 __get_user(fl->l_whence, &target_fl->l_whence);
6859 __get_user(fl->l_start, &target_fl->l_start);
6860 __get_user(fl->l_len, &target_fl->l_len);
6861 __get_user(fl->l_pid, &target_fl->l_pid);
6862 unlock_user_struct(target_fl, target_flock_addr, 0);
6863 return 0;
6866 static inline abi_long copy_to_user_flock(abi_ulong target_flock_addr,
6867 const struct flock64 *fl)
6869 struct target_flock *target_fl;
6870 short l_type;
6872 if (!lock_user_struct(VERIFY_WRITE, target_fl, target_flock_addr, 0)) {
6873 return -TARGET_EFAULT;
6876 l_type = host_to_target_flock(fl->l_type);
6877 __put_user(l_type, &target_fl->l_type);
6878 __put_user(fl->l_whence, &target_fl->l_whence);
6879 __put_user(fl->l_start, &target_fl->l_start);
6880 __put_user(fl->l_len, &target_fl->l_len);
6881 __put_user(fl->l_pid, &target_fl->l_pid);
6882 unlock_user_struct(target_fl, target_flock_addr, 1);
6883 return 0;
6886 typedef abi_long from_flock64_fn(struct flock64 *fl, abi_ulong target_addr);
6887 typedef abi_long to_flock64_fn(abi_ulong target_addr, const struct flock64 *fl);
6889 #if defined(TARGET_ARM) && TARGET_ABI_BITS == 32
6890 struct target_oabi_flock64 {
6891 abi_short l_type;
6892 abi_short l_whence;
6893 abi_llong l_start;
6894 abi_llong l_len;
6895 abi_int l_pid;
6896 } QEMU_PACKED;
6898 static inline abi_long copy_from_user_oabi_flock64(struct flock64 *fl,
6899 abi_ulong target_flock_addr)
6901 struct target_oabi_flock64 *target_fl;
6902 int l_type;
6904 if (!lock_user_struct(VERIFY_READ, target_fl, target_flock_addr, 1)) {
6905 return -TARGET_EFAULT;
6908 __get_user(l_type, &target_fl->l_type);
6909 l_type = target_to_host_flock(l_type);
6910 if (l_type < 0) {
6911 return l_type;
6913 fl->l_type = l_type;
6914 __get_user(fl->l_whence, &target_fl->l_whence);
6915 __get_user(fl->l_start, &target_fl->l_start);
6916 __get_user(fl->l_len, &target_fl->l_len);
6917 __get_user(fl->l_pid, &target_fl->l_pid);
6918 unlock_user_struct(target_fl, target_flock_addr, 0);
6919 return 0;
6922 static inline abi_long copy_to_user_oabi_flock64(abi_ulong target_flock_addr,
6923 const struct flock64 *fl)
6925 struct target_oabi_flock64 *target_fl;
6926 short l_type;
6928 if (!lock_user_struct(VERIFY_WRITE, target_fl, target_flock_addr, 0)) {
6929 return -TARGET_EFAULT;
6932 l_type = host_to_target_flock(fl->l_type);
6933 __put_user(l_type, &target_fl->l_type);
6934 __put_user(fl->l_whence, &target_fl->l_whence);
6935 __put_user(fl->l_start, &target_fl->l_start);
6936 __put_user(fl->l_len, &target_fl->l_len);
6937 __put_user(fl->l_pid, &target_fl->l_pid);
6938 unlock_user_struct(target_fl, target_flock_addr, 1);
6939 return 0;
6941 #endif
6943 static inline abi_long copy_from_user_flock64(struct flock64 *fl,
6944 abi_ulong target_flock_addr)
6946 struct target_flock64 *target_fl;
6947 int l_type;
6949 if (!lock_user_struct(VERIFY_READ, target_fl, target_flock_addr, 1)) {
6950 return -TARGET_EFAULT;
6953 __get_user(l_type, &target_fl->l_type);
6954 l_type = target_to_host_flock(l_type);
6955 if (l_type < 0) {
6956 return l_type;
6958 fl->l_type = l_type;
6959 __get_user(fl->l_whence, &target_fl->l_whence);
6960 __get_user(fl->l_start, &target_fl->l_start);
6961 __get_user(fl->l_len, &target_fl->l_len);
6962 __get_user(fl->l_pid, &target_fl->l_pid);
6963 unlock_user_struct(target_fl, target_flock_addr, 0);
6964 return 0;
6967 static inline abi_long copy_to_user_flock64(abi_ulong target_flock_addr,
6968 const struct flock64 *fl)
6970 struct target_flock64 *target_fl;
6971 short l_type;
6973 if (!lock_user_struct(VERIFY_WRITE, target_fl, target_flock_addr, 0)) {
6974 return -TARGET_EFAULT;
6977 l_type = host_to_target_flock(fl->l_type);
6978 __put_user(l_type, &target_fl->l_type);
6979 __put_user(fl->l_whence, &target_fl->l_whence);
6980 __put_user(fl->l_start, &target_fl->l_start);
6981 __put_user(fl->l_len, &target_fl->l_len);
6982 __put_user(fl->l_pid, &target_fl->l_pid);
6983 unlock_user_struct(target_fl, target_flock_addr, 1);
6984 return 0;
6987 static abi_long do_fcntl(int fd, int cmd, abi_ulong arg)
6989 struct flock64 fl64;
6990 #ifdef F_GETOWN_EX
6991 struct f_owner_ex fox;
6992 struct target_f_owner_ex *target_fox;
6993 #endif
6994 abi_long ret;
6995 int host_cmd = target_to_host_fcntl_cmd(cmd);
6997 if (host_cmd == -TARGET_EINVAL)
6998 return host_cmd;
7000 switch(cmd) {
7001 case TARGET_F_GETLK:
7002 ret = copy_from_user_flock(&fl64, arg);
7003 if (ret) {
7004 return ret;
7006 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64));
7007 if (ret == 0) {
7008 ret = copy_to_user_flock(arg, &fl64);
7010 break;
7012 case TARGET_F_SETLK:
7013 case TARGET_F_SETLKW:
7014 ret = copy_from_user_flock(&fl64, arg);
7015 if (ret) {
7016 return ret;
7018 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64));
7019 break;
7021 case TARGET_F_GETLK64:
7022 case TARGET_F_OFD_GETLK:
7023 ret = copy_from_user_flock64(&fl64, arg);
7024 if (ret) {
7025 return ret;
7027 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64));
7028 if (ret == 0) {
7029 ret = copy_to_user_flock64(arg, &fl64);
7031 break;
7032 case TARGET_F_SETLK64:
7033 case TARGET_F_SETLKW64:
7034 case TARGET_F_OFD_SETLK:
7035 case TARGET_F_OFD_SETLKW:
7036 ret = copy_from_user_flock64(&fl64, arg);
7037 if (ret) {
7038 return ret;
7040 ret = get_errno(safe_fcntl(fd, host_cmd, &fl64));
7041 break;
7043 case TARGET_F_GETFL:
7044 ret = get_errno(safe_fcntl(fd, host_cmd, arg));
7045 if (ret >= 0) {
7046 ret = host_to_target_bitmask(ret, fcntl_flags_tbl);
7048 break;
7050 case TARGET_F_SETFL:
7051 ret = get_errno(safe_fcntl(fd, host_cmd,
7052 target_to_host_bitmask(arg,
7053 fcntl_flags_tbl)));
7054 break;
7056 #ifdef F_GETOWN_EX
7057 case TARGET_F_GETOWN_EX:
7058 ret = get_errno(safe_fcntl(fd, host_cmd, &fox));
7059 if (ret >= 0) {
7060 if (!lock_user_struct(VERIFY_WRITE, target_fox, arg, 0))
7061 return -TARGET_EFAULT;
7062 target_fox->type = tswap32(fox.type);
7063 target_fox->pid = tswap32(fox.pid);
7064 unlock_user_struct(target_fox, arg, 1);
7066 break;
7067 #endif
7069 #ifdef F_SETOWN_EX
7070 case TARGET_F_SETOWN_EX:
7071 if (!lock_user_struct(VERIFY_READ, target_fox, arg, 1))
7072 return -TARGET_EFAULT;
7073 fox.type = tswap32(target_fox->type);
7074 fox.pid = tswap32(target_fox->pid);
7075 unlock_user_struct(target_fox, arg, 0);
7076 ret = get_errno(safe_fcntl(fd, host_cmd, &fox));
7077 break;
7078 #endif
7080 case TARGET_F_SETSIG:
7081 ret = get_errno(safe_fcntl(fd, host_cmd, target_to_host_signal(arg)));
7082 break;
7084 case TARGET_F_GETSIG:
7085 ret = host_to_target_signal(get_errno(safe_fcntl(fd, host_cmd, arg)));
7086 break;
7088 case TARGET_F_SETOWN:
7089 case TARGET_F_GETOWN:
7090 case TARGET_F_SETLEASE:
7091 case TARGET_F_GETLEASE:
7092 case TARGET_F_SETPIPE_SZ:
7093 case TARGET_F_GETPIPE_SZ:
7094 case TARGET_F_ADD_SEALS:
7095 case TARGET_F_GET_SEALS:
7096 ret = get_errno(safe_fcntl(fd, host_cmd, arg));
7097 break;
7099 default:
7100 ret = get_errno(safe_fcntl(fd, cmd, arg));
7101 break;
7103 return ret;
7106 #ifdef USE_UID16
7108 static inline int high2lowuid(int uid)
7110 if (uid > 65535)
7111 return 65534;
7112 else
7113 return uid;
7116 static inline int high2lowgid(int gid)
7118 if (gid > 65535)
7119 return 65534;
7120 else
7121 return gid;
7124 static inline int low2highuid(int uid)
7126 if ((int16_t)uid == -1)
7127 return -1;
7128 else
7129 return uid;
7132 static inline int low2highgid(int gid)
7134 if ((int16_t)gid == -1)
7135 return -1;
7136 else
7137 return gid;
7139 static inline int tswapid(int id)
7141 return tswap16(id);
7144 #define put_user_id(x, gaddr) put_user_u16(x, gaddr)
7146 #else /* !USE_UID16 */
7147 static inline int high2lowuid(int uid)
7149 return uid;
7151 static inline int high2lowgid(int gid)
7153 return gid;
7155 static inline int low2highuid(int uid)
7157 return uid;
7159 static inline int low2highgid(int gid)
7161 return gid;
7163 static inline int tswapid(int id)
7165 return tswap32(id);
7168 #define put_user_id(x, gaddr) put_user_u32(x, gaddr)
7170 #endif /* USE_UID16 */
7172 /* We must do direct syscalls for setting UID/GID, because we want to
7173 * implement the Linux system call semantics of "change only for this thread",
7174 * not the libc/POSIX semantics of "change for all threads in process".
7175 * (See http://ewontfix.com/17/ for more details.)
7176 * We use the 32-bit version of the syscalls if present; if it is not
7177 * then either the host architecture supports 32-bit UIDs natively with
7178 * the standard syscall, or the 16-bit UID is the best we can do.
7180 #ifdef __NR_setuid32
7181 #define __NR_sys_setuid __NR_setuid32
7182 #else
7183 #define __NR_sys_setuid __NR_setuid
7184 #endif
7185 #ifdef __NR_setgid32
7186 #define __NR_sys_setgid __NR_setgid32
7187 #else
7188 #define __NR_sys_setgid __NR_setgid
7189 #endif
7190 #ifdef __NR_setresuid32
7191 #define __NR_sys_setresuid __NR_setresuid32
7192 #else
7193 #define __NR_sys_setresuid __NR_setresuid
7194 #endif
7195 #ifdef __NR_setresgid32
7196 #define __NR_sys_setresgid __NR_setresgid32
7197 #else
7198 #define __NR_sys_setresgid __NR_setresgid
7199 #endif
7201 _syscall1(int, sys_setuid, uid_t, uid)
7202 _syscall1(int, sys_setgid, gid_t, gid)
7203 _syscall3(int, sys_setresuid, uid_t, ruid, uid_t, euid, uid_t, suid)
7204 _syscall3(int, sys_setresgid, gid_t, rgid, gid_t, egid, gid_t, sgid)
7206 void syscall_init(void)
7208 IOCTLEntry *ie;
7209 const argtype *arg_type;
7210 int size;
7212 thunk_init(STRUCT_MAX);
7214 #define STRUCT(name, ...) thunk_register_struct(STRUCT_ ## name, #name, struct_ ## name ## _def);
7215 #define STRUCT_SPECIAL(name) thunk_register_struct_direct(STRUCT_ ## name, #name, &struct_ ## name ## _def);
7216 #include "syscall_types.h"
7217 #undef STRUCT
7218 #undef STRUCT_SPECIAL
7220 /* we patch the ioctl size if necessary. We rely on the fact that
7221 no ioctl has all the bits at '1' in the size field */
7222 ie = ioctl_entries;
7223 while (ie->target_cmd != 0) {
7224 if (((ie->target_cmd >> TARGET_IOC_SIZESHIFT) & TARGET_IOC_SIZEMASK) ==
7225 TARGET_IOC_SIZEMASK) {
7226 arg_type = ie->arg_type;
7227 if (arg_type[0] != TYPE_PTR) {
7228 fprintf(stderr, "cannot patch size for ioctl 0x%x\n",
7229 ie->target_cmd);
7230 exit(1);
7232 arg_type++;
7233 size = thunk_type_size(arg_type, 0);
7234 ie->target_cmd = (ie->target_cmd &
7235 ~(TARGET_IOC_SIZEMASK << TARGET_IOC_SIZESHIFT)) |
7236 (size << TARGET_IOC_SIZESHIFT);
7239 /* automatic consistency check if same arch */
7240 #if (defined(__i386__) && defined(TARGET_I386) && defined(TARGET_ABI32)) || \
7241 (defined(__x86_64__) && defined(TARGET_X86_64))
7242 if (unlikely(ie->target_cmd != ie->host_cmd)) {
7243 fprintf(stderr, "ERROR: ioctl(%s): target=0x%x host=0x%x\n",
7244 ie->name, ie->target_cmd, ie->host_cmd);
7246 #endif
7247 ie++;
7251 #ifdef TARGET_NR_truncate64
7252 static inline abi_long target_truncate64(CPUArchState *cpu_env, const char *arg1,
7253 abi_long arg2,
7254 abi_long arg3,
7255 abi_long arg4)
7257 if (regpairs_aligned(cpu_env, TARGET_NR_truncate64)) {
7258 arg2 = arg3;
7259 arg3 = arg4;
7261 return get_errno(truncate64(arg1, target_offset64(arg2, arg3)));
7263 #endif
7265 #ifdef TARGET_NR_ftruncate64
7266 static inline abi_long target_ftruncate64(CPUArchState *cpu_env, abi_long arg1,
7267 abi_long arg2,
7268 abi_long arg3,
7269 abi_long arg4)
7271 if (regpairs_aligned(cpu_env, TARGET_NR_ftruncate64)) {
7272 arg2 = arg3;
7273 arg3 = arg4;
7275 return get_errno(ftruncate64(arg1, target_offset64(arg2, arg3)));
7277 #endif
7279 #if defined(TARGET_NR_timer_settime) || \
7280 (defined(TARGET_NR_timerfd_settime) && defined(CONFIG_TIMERFD))
7281 static inline abi_long target_to_host_itimerspec(struct itimerspec *host_its,
7282 abi_ulong target_addr)
7284 if (target_to_host_timespec(&host_its->it_interval, target_addr +
7285 offsetof(struct target_itimerspec,
7286 it_interval)) ||
7287 target_to_host_timespec(&host_its->it_value, target_addr +
7288 offsetof(struct target_itimerspec,
7289 it_value))) {
7290 return -TARGET_EFAULT;
7293 return 0;
7295 #endif
7297 #if defined(TARGET_NR_timer_settime64) || \
7298 (defined(TARGET_NR_timerfd_settime64) && defined(CONFIG_TIMERFD))
7299 static inline abi_long target_to_host_itimerspec64(struct itimerspec *host_its,
7300 abi_ulong target_addr)
7302 if (target_to_host_timespec64(&host_its->it_interval, target_addr +
7303 offsetof(struct target__kernel_itimerspec,
7304 it_interval)) ||
7305 target_to_host_timespec64(&host_its->it_value, target_addr +
7306 offsetof(struct target__kernel_itimerspec,
7307 it_value))) {
7308 return -TARGET_EFAULT;
7311 return 0;
7313 #endif
7315 #if ((defined(TARGET_NR_timerfd_gettime) || \
7316 defined(TARGET_NR_timerfd_settime)) && defined(CONFIG_TIMERFD)) || \
7317 defined(TARGET_NR_timer_gettime) || defined(TARGET_NR_timer_settime)
7318 static inline abi_long host_to_target_itimerspec(abi_ulong target_addr,
7319 struct itimerspec *host_its)
7321 if (host_to_target_timespec(target_addr + offsetof(struct target_itimerspec,
7322 it_interval),
7323 &host_its->it_interval) ||
7324 host_to_target_timespec(target_addr + offsetof(struct target_itimerspec,
7325 it_value),
7326 &host_its->it_value)) {
7327 return -TARGET_EFAULT;
7329 return 0;
7331 #endif
7333 #if ((defined(TARGET_NR_timerfd_gettime64) || \
7334 defined(TARGET_NR_timerfd_settime64)) && defined(CONFIG_TIMERFD)) || \
7335 defined(TARGET_NR_timer_gettime64) || defined(TARGET_NR_timer_settime64)
7336 static inline abi_long host_to_target_itimerspec64(abi_ulong target_addr,
7337 struct itimerspec *host_its)
7339 if (host_to_target_timespec64(target_addr +
7340 offsetof(struct target__kernel_itimerspec,
7341 it_interval),
7342 &host_its->it_interval) ||
7343 host_to_target_timespec64(target_addr +
7344 offsetof(struct target__kernel_itimerspec,
7345 it_value),
7346 &host_its->it_value)) {
7347 return -TARGET_EFAULT;
7349 return 0;
7351 #endif
7353 #if defined(TARGET_NR_adjtimex) || \
7354 (defined(TARGET_NR_clock_adjtime) && defined(CONFIG_CLOCK_ADJTIME))
7355 static inline abi_long target_to_host_timex(struct timex *host_tx,
7356 abi_long target_addr)
7358 struct target_timex *target_tx;
7360 if (!lock_user_struct(VERIFY_READ, target_tx, target_addr, 1)) {
7361 return -TARGET_EFAULT;
7364 __get_user(host_tx->modes, &target_tx->modes);
7365 __get_user(host_tx->offset, &target_tx->offset);
7366 __get_user(host_tx->freq, &target_tx->freq);
7367 __get_user(host_tx->maxerror, &target_tx->maxerror);
7368 __get_user(host_tx->esterror, &target_tx->esterror);
7369 __get_user(host_tx->status, &target_tx->status);
7370 __get_user(host_tx->constant, &target_tx->constant);
7371 __get_user(host_tx->precision, &target_tx->precision);
7372 __get_user(host_tx->tolerance, &target_tx->tolerance);
7373 __get_user(host_tx->time.tv_sec, &target_tx->time.tv_sec);
7374 __get_user(host_tx->time.tv_usec, &target_tx->time.tv_usec);
7375 __get_user(host_tx->tick, &target_tx->tick);
7376 __get_user(host_tx->ppsfreq, &target_tx->ppsfreq);
7377 __get_user(host_tx->jitter, &target_tx->jitter);
7378 __get_user(host_tx->shift, &target_tx->shift);
7379 __get_user(host_tx->stabil, &target_tx->stabil);
7380 __get_user(host_tx->jitcnt, &target_tx->jitcnt);
7381 __get_user(host_tx->calcnt, &target_tx->calcnt);
7382 __get_user(host_tx->errcnt, &target_tx->errcnt);
7383 __get_user(host_tx->stbcnt, &target_tx->stbcnt);
7384 __get_user(host_tx->tai, &target_tx->tai);
7386 unlock_user_struct(target_tx, target_addr, 0);
7387 return 0;
7390 static inline abi_long host_to_target_timex(abi_long target_addr,
7391 struct timex *host_tx)
7393 struct target_timex *target_tx;
7395 if (!lock_user_struct(VERIFY_WRITE, target_tx, target_addr, 0)) {
7396 return -TARGET_EFAULT;
7399 __put_user(host_tx->modes, &target_tx->modes);
7400 __put_user(host_tx->offset, &target_tx->offset);
7401 __put_user(host_tx->freq, &target_tx->freq);
7402 __put_user(host_tx->maxerror, &target_tx->maxerror);
7403 __put_user(host_tx->esterror, &target_tx->esterror);
7404 __put_user(host_tx->status, &target_tx->status);
7405 __put_user(host_tx->constant, &target_tx->constant);
7406 __put_user(host_tx->precision, &target_tx->precision);
7407 __put_user(host_tx->tolerance, &target_tx->tolerance);
7408 __put_user(host_tx->time.tv_sec, &target_tx->time.tv_sec);
7409 __put_user(host_tx->time.tv_usec, &target_tx->time.tv_usec);
7410 __put_user(host_tx->tick, &target_tx->tick);
7411 __put_user(host_tx->ppsfreq, &target_tx->ppsfreq);
7412 __put_user(host_tx->jitter, &target_tx->jitter);
7413 __put_user(host_tx->shift, &target_tx->shift);
7414 __put_user(host_tx->stabil, &target_tx->stabil);
7415 __put_user(host_tx->jitcnt, &target_tx->jitcnt);
7416 __put_user(host_tx->calcnt, &target_tx->calcnt);
7417 __put_user(host_tx->errcnt, &target_tx->errcnt);
7418 __put_user(host_tx->stbcnt, &target_tx->stbcnt);
7419 __put_user(host_tx->tai, &target_tx->tai);
7421 unlock_user_struct(target_tx, target_addr, 1);
7422 return 0;
7424 #endif
7427 #if defined(TARGET_NR_clock_adjtime64) && defined(CONFIG_CLOCK_ADJTIME)
7428 static inline abi_long target_to_host_timex64(struct timex *host_tx,
7429 abi_long target_addr)
7431 struct target__kernel_timex *target_tx;
7433 if (copy_from_user_timeval64(&host_tx->time, target_addr +
7434 offsetof(struct target__kernel_timex,
7435 time))) {
7436 return -TARGET_EFAULT;
7439 if (!lock_user_struct(VERIFY_READ, target_tx, target_addr, 1)) {
7440 return -TARGET_EFAULT;
7443 __get_user(host_tx->modes, &target_tx->modes);
7444 __get_user(host_tx->offset, &target_tx->offset);
7445 __get_user(host_tx->freq, &target_tx->freq);
7446 __get_user(host_tx->maxerror, &target_tx->maxerror);
7447 __get_user(host_tx->esterror, &target_tx->esterror);
7448 __get_user(host_tx->status, &target_tx->status);
7449 __get_user(host_tx->constant, &target_tx->constant);
7450 __get_user(host_tx->precision, &target_tx->precision);
7451 __get_user(host_tx->tolerance, &target_tx->tolerance);
7452 __get_user(host_tx->tick, &target_tx->tick);
7453 __get_user(host_tx->ppsfreq, &target_tx->ppsfreq);
7454 __get_user(host_tx->jitter, &target_tx->jitter);
7455 __get_user(host_tx->shift, &target_tx->shift);
7456 __get_user(host_tx->stabil, &target_tx->stabil);
7457 __get_user(host_tx->jitcnt, &target_tx->jitcnt);
7458 __get_user(host_tx->calcnt, &target_tx->calcnt);
7459 __get_user(host_tx->errcnt, &target_tx->errcnt);
7460 __get_user(host_tx->stbcnt, &target_tx->stbcnt);
7461 __get_user(host_tx->tai, &target_tx->tai);
7463 unlock_user_struct(target_tx, target_addr, 0);
7464 return 0;
7467 static inline abi_long host_to_target_timex64(abi_long target_addr,
7468 struct timex *host_tx)
7470 struct target__kernel_timex *target_tx;
7472 if (copy_to_user_timeval64(target_addr +
7473 offsetof(struct target__kernel_timex, time),
7474 &host_tx->time)) {
7475 return -TARGET_EFAULT;
7478 if (!lock_user_struct(VERIFY_WRITE, target_tx, target_addr, 0)) {
7479 return -TARGET_EFAULT;
7482 __put_user(host_tx->modes, &target_tx->modes);
7483 __put_user(host_tx->offset, &target_tx->offset);
7484 __put_user(host_tx->freq, &target_tx->freq);
7485 __put_user(host_tx->maxerror, &target_tx->maxerror);
7486 __put_user(host_tx->esterror, &target_tx->esterror);
7487 __put_user(host_tx->status, &target_tx->status);
7488 __put_user(host_tx->constant, &target_tx->constant);
7489 __put_user(host_tx->precision, &target_tx->precision);
7490 __put_user(host_tx->tolerance, &target_tx->tolerance);
7491 __put_user(host_tx->tick, &target_tx->tick);
7492 __put_user(host_tx->ppsfreq, &target_tx->ppsfreq);
7493 __put_user(host_tx->jitter, &target_tx->jitter);
7494 __put_user(host_tx->shift, &target_tx->shift);
7495 __put_user(host_tx->stabil, &target_tx->stabil);
7496 __put_user(host_tx->jitcnt, &target_tx->jitcnt);
7497 __put_user(host_tx->calcnt, &target_tx->calcnt);
7498 __put_user(host_tx->errcnt, &target_tx->errcnt);
7499 __put_user(host_tx->stbcnt, &target_tx->stbcnt);
7500 __put_user(host_tx->tai, &target_tx->tai);
7502 unlock_user_struct(target_tx, target_addr, 1);
7503 return 0;
7505 #endif
7507 #ifndef HAVE_SIGEV_NOTIFY_THREAD_ID
7508 #define sigev_notify_thread_id _sigev_un._tid
7509 #endif
7511 static inline abi_long target_to_host_sigevent(struct sigevent *host_sevp,
7512 abi_ulong target_addr)
7514 struct target_sigevent *target_sevp;
7516 if (!lock_user_struct(VERIFY_READ, target_sevp, target_addr, 1)) {
7517 return -TARGET_EFAULT;
7520 /* This union is awkward on 64 bit systems because it has a 32 bit
7521 * integer and a pointer in it; we follow the conversion approach
7522 * used for handling sigval types in signal.c so the guest should get
7523 * the correct value back even if we did a 64 bit byteswap and it's
7524 * using the 32 bit integer.
7526 host_sevp->sigev_value.sival_ptr =
7527 (void *)(uintptr_t)tswapal(target_sevp->sigev_value.sival_ptr);
7528 host_sevp->sigev_signo =
7529 target_to_host_signal(tswap32(target_sevp->sigev_signo));
7530 host_sevp->sigev_notify = tswap32(target_sevp->sigev_notify);
7531 host_sevp->sigev_notify_thread_id = tswap32(target_sevp->_sigev_un._tid);
7533 unlock_user_struct(target_sevp, target_addr, 1);
7534 return 0;
7537 #if defined(TARGET_NR_mlockall)
7538 static inline int target_to_host_mlockall_arg(int arg)
7540 int result = 0;
7542 if (arg & TARGET_MCL_CURRENT) {
7543 result |= MCL_CURRENT;
7545 if (arg & TARGET_MCL_FUTURE) {
7546 result |= MCL_FUTURE;
7548 #ifdef MCL_ONFAULT
7549 if (arg & TARGET_MCL_ONFAULT) {
7550 result |= MCL_ONFAULT;
7552 #endif
7554 return result;
7556 #endif
7558 #if (defined(TARGET_NR_stat64) || defined(TARGET_NR_lstat64) || \
7559 defined(TARGET_NR_fstat64) || defined(TARGET_NR_fstatat64) || \
7560 defined(TARGET_NR_newfstatat))
7561 static inline abi_long host_to_target_stat64(CPUArchState *cpu_env,
7562 abi_ulong target_addr,
7563 struct stat *host_st)
7565 #if defined(TARGET_ARM) && defined(TARGET_ABI32)
7566 if (cpu_env->eabi) {
7567 struct target_eabi_stat64 *target_st;
7569 if (!lock_user_struct(VERIFY_WRITE, target_st, target_addr, 0))
7570 return -TARGET_EFAULT;
7571 memset(target_st, 0, sizeof(struct target_eabi_stat64));
7572 __put_user(host_st->st_dev, &target_st->st_dev);
7573 __put_user(host_st->st_ino, &target_st->st_ino);
7574 #ifdef TARGET_STAT64_HAS_BROKEN_ST_INO
7575 __put_user(host_st->st_ino, &target_st->__st_ino);
7576 #endif
7577 __put_user(host_st->st_mode, &target_st->st_mode);
7578 __put_user(host_st->st_nlink, &target_st->st_nlink);
7579 __put_user(host_st->st_uid, &target_st->st_uid);
7580 __put_user(host_st->st_gid, &target_st->st_gid);
7581 __put_user(host_st->st_rdev, &target_st->st_rdev);
7582 __put_user(host_st->st_size, &target_st->st_size);
7583 __put_user(host_st->st_blksize, &target_st->st_blksize);
7584 __put_user(host_st->st_blocks, &target_st->st_blocks);
7585 __put_user(host_st->st_atime, &target_st->target_st_atime);
7586 __put_user(host_st->st_mtime, &target_st->target_st_mtime);
7587 __put_user(host_st->st_ctime, &target_st->target_st_ctime);
7588 #ifdef HAVE_STRUCT_STAT_ST_ATIM
7589 __put_user(host_st->st_atim.tv_nsec, &target_st->target_st_atime_nsec);
7590 __put_user(host_st->st_mtim.tv_nsec, &target_st->target_st_mtime_nsec);
7591 __put_user(host_st->st_ctim.tv_nsec, &target_st->target_st_ctime_nsec);
7592 #endif
7593 unlock_user_struct(target_st, target_addr, 1);
7594 } else
7595 #endif
7597 #if defined(TARGET_HAS_STRUCT_STAT64)
7598 struct target_stat64 *target_st;
7599 #else
7600 struct target_stat *target_st;
7601 #endif
7603 if (!lock_user_struct(VERIFY_WRITE, target_st, target_addr, 0))
7604 return -TARGET_EFAULT;
7605 memset(target_st, 0, sizeof(*target_st));
7606 __put_user(host_st->st_dev, &target_st->st_dev);
7607 __put_user(host_st->st_ino, &target_st->st_ino);
7608 #ifdef TARGET_STAT64_HAS_BROKEN_ST_INO
7609 __put_user(host_st->st_ino, &target_st->__st_ino);
7610 #endif
7611 __put_user(host_st->st_mode, &target_st->st_mode);
7612 __put_user(host_st->st_nlink, &target_st->st_nlink);
7613 __put_user(host_st->st_uid, &target_st->st_uid);
7614 __put_user(host_st->st_gid, &target_st->st_gid);
7615 __put_user(host_st->st_rdev, &target_st->st_rdev);
7616 /* XXX: better use of kernel struct */
7617 __put_user(host_st->st_size, &target_st->st_size);
7618 __put_user(host_st->st_blksize, &target_st->st_blksize);
7619 __put_user(host_st->st_blocks, &target_st->st_blocks);
7620 __put_user(host_st->st_atime, &target_st->target_st_atime);
7621 __put_user(host_st->st_mtime, &target_st->target_st_mtime);
7622 __put_user(host_st->st_ctime, &target_st->target_st_ctime);
7623 #ifdef HAVE_STRUCT_STAT_ST_ATIM
7624 __put_user(host_st->st_atim.tv_nsec, &target_st->target_st_atime_nsec);
7625 __put_user(host_st->st_mtim.tv_nsec, &target_st->target_st_mtime_nsec);
7626 __put_user(host_st->st_ctim.tv_nsec, &target_st->target_st_ctime_nsec);
7627 #endif
7628 unlock_user_struct(target_st, target_addr, 1);
7631 return 0;
7633 #endif
7635 #if defined(TARGET_NR_statx) && defined(__NR_statx)
7636 static inline abi_long host_to_target_statx(struct target_statx *host_stx,
7637 abi_ulong target_addr)
7639 struct target_statx *target_stx;
7641 if (!lock_user_struct(VERIFY_WRITE, target_stx, target_addr, 0)) {
7642 return -TARGET_EFAULT;
7644 memset(target_stx, 0, sizeof(*target_stx));
7646 __put_user(host_stx->stx_mask, &target_stx->stx_mask);
7647 __put_user(host_stx->stx_blksize, &target_stx->stx_blksize);
7648 __put_user(host_stx->stx_attributes, &target_stx->stx_attributes);
7649 __put_user(host_stx->stx_nlink, &target_stx->stx_nlink);
7650 __put_user(host_stx->stx_uid, &target_stx->stx_uid);
7651 __put_user(host_stx->stx_gid, &target_stx->stx_gid);
7652 __put_user(host_stx->stx_mode, &target_stx->stx_mode);
7653 __put_user(host_stx->stx_ino, &target_stx->stx_ino);
7654 __put_user(host_stx->stx_size, &target_stx->stx_size);
7655 __put_user(host_stx->stx_blocks, &target_stx->stx_blocks);
7656 __put_user(host_stx->stx_attributes_mask, &target_stx->stx_attributes_mask);
7657 __put_user(host_stx->stx_atime.tv_sec, &target_stx->stx_atime.tv_sec);
7658 __put_user(host_stx->stx_atime.tv_nsec, &target_stx->stx_atime.tv_nsec);
7659 __put_user(host_stx->stx_btime.tv_sec, &target_stx->stx_btime.tv_sec);
7660 __put_user(host_stx->stx_btime.tv_nsec, &target_stx->stx_btime.tv_nsec);
7661 __put_user(host_stx->stx_ctime.tv_sec, &target_stx->stx_ctime.tv_sec);
7662 __put_user(host_stx->stx_ctime.tv_nsec, &target_stx->stx_ctime.tv_nsec);
7663 __put_user(host_stx->stx_mtime.tv_sec, &target_stx->stx_mtime.tv_sec);
7664 __put_user(host_stx->stx_mtime.tv_nsec, &target_stx->stx_mtime.tv_nsec);
7665 __put_user(host_stx->stx_rdev_major, &target_stx->stx_rdev_major);
7666 __put_user(host_stx->stx_rdev_minor, &target_stx->stx_rdev_minor);
7667 __put_user(host_stx->stx_dev_major, &target_stx->stx_dev_major);
7668 __put_user(host_stx->stx_dev_minor, &target_stx->stx_dev_minor);
7670 unlock_user_struct(target_stx, target_addr, 1);
7672 return 0;
7674 #endif
7676 static int do_sys_futex(int *uaddr, int op, int val,
7677 const struct timespec *timeout, int *uaddr2,
7678 int val3)
7680 #if HOST_LONG_BITS == 64
7681 #if defined(__NR_futex)
7682 /* always a 64-bit time_t, it doesn't define _time64 version */
7683 return sys_futex(uaddr, op, val, timeout, uaddr2, val3);
7685 #endif
7686 #else /* HOST_LONG_BITS == 64 */
7687 #if defined(__NR_futex_time64)
7688 if (sizeof(timeout->tv_sec) == 8) {
7689 /* _time64 function on 32bit arch */
7690 return sys_futex_time64(uaddr, op, val, timeout, uaddr2, val3);
7692 #endif
7693 #if defined(__NR_futex)
7694 /* old function on 32bit arch */
7695 return sys_futex(uaddr, op, val, timeout, uaddr2, val3);
7696 #endif
7697 #endif /* HOST_LONG_BITS == 64 */
7698 g_assert_not_reached();
7701 static int do_safe_futex(int *uaddr, int op, int val,
7702 const struct timespec *timeout, int *uaddr2,
7703 int val3)
7705 #if HOST_LONG_BITS == 64
7706 #if defined(__NR_futex)
7707 /* always a 64-bit time_t, it doesn't define _time64 version */
7708 return get_errno(safe_futex(uaddr, op, val, timeout, uaddr2, val3));
7709 #endif
7710 #else /* HOST_LONG_BITS == 64 */
7711 #if defined(__NR_futex_time64)
7712 if (sizeof(timeout->tv_sec) == 8) {
7713 /* _time64 function on 32bit arch */
7714 return get_errno(safe_futex_time64(uaddr, op, val, timeout, uaddr2,
7715 val3));
7717 #endif
7718 #if defined(__NR_futex)
7719 /* old function on 32bit arch */
7720 return get_errno(safe_futex(uaddr, op, val, timeout, uaddr2, val3));
7721 #endif
7722 #endif /* HOST_LONG_BITS == 64 */
7723 return -TARGET_ENOSYS;
7726 /* ??? Using host futex calls even when target atomic operations
7727 are not really atomic probably breaks things. However implementing
7728 futexes locally would make futexes shared between multiple processes
7729 tricky. However they're probably useless because guest atomic
7730 operations won't work either. */
7731 #if defined(TARGET_NR_futex)
7732 static int do_futex(CPUState *cpu, target_ulong uaddr, int op, int val,
7733 target_ulong timeout, target_ulong uaddr2, int val3)
7735 struct timespec ts, *pts;
7736 int base_op;
7738 /* ??? We assume FUTEX_* constants are the same on both host
7739 and target. */
7740 #ifdef FUTEX_CMD_MASK
7741 base_op = op & FUTEX_CMD_MASK;
7742 #else
7743 base_op = op;
7744 #endif
7745 switch (base_op) {
7746 case FUTEX_WAIT:
7747 case FUTEX_WAIT_BITSET:
7748 if (timeout) {
7749 pts = &ts;
7750 target_to_host_timespec(pts, timeout);
7751 } else {
7752 pts = NULL;
7754 return do_safe_futex(g2h(cpu, uaddr),
7755 op, tswap32(val), pts, NULL, val3);
7756 case FUTEX_WAKE:
7757 return do_safe_futex(g2h(cpu, uaddr),
7758 op, val, NULL, NULL, 0);
7759 case FUTEX_FD:
7760 return do_safe_futex(g2h(cpu, uaddr),
7761 op, val, NULL, NULL, 0);
7762 case FUTEX_REQUEUE:
7763 case FUTEX_CMP_REQUEUE:
7764 case FUTEX_WAKE_OP:
7765 /* For FUTEX_REQUEUE, FUTEX_CMP_REQUEUE, and FUTEX_WAKE_OP, the
7766 TIMEOUT parameter is interpreted as a uint32_t by the kernel.
7767 But the prototype takes a `struct timespec *'; insert casts
7768 to satisfy the compiler. We do not need to tswap TIMEOUT
7769 since it's not compared to guest memory. */
7770 pts = (struct timespec *)(uintptr_t) timeout;
7771 return do_safe_futex(g2h(cpu, uaddr), op, val, pts, g2h(cpu, uaddr2),
7772 (base_op == FUTEX_CMP_REQUEUE
7773 ? tswap32(val3) : val3));
7774 default:
7775 return -TARGET_ENOSYS;
7778 #endif
7780 #if defined(TARGET_NR_futex_time64)
7781 static int do_futex_time64(CPUState *cpu, target_ulong uaddr, int op,
7782 int val, target_ulong timeout,
7783 target_ulong uaddr2, int val3)
7785 struct timespec ts, *pts;
7786 int base_op;
7788 /* ??? We assume FUTEX_* constants are the same on both host
7789 and target. */
7790 #ifdef FUTEX_CMD_MASK
7791 base_op = op & FUTEX_CMD_MASK;
7792 #else
7793 base_op = op;
7794 #endif
7795 switch (base_op) {
7796 case FUTEX_WAIT:
7797 case FUTEX_WAIT_BITSET:
7798 if (timeout) {
7799 pts = &ts;
7800 if (target_to_host_timespec64(pts, timeout)) {
7801 return -TARGET_EFAULT;
7803 } else {
7804 pts = NULL;
7806 return do_safe_futex(g2h(cpu, uaddr), op,
7807 tswap32(val), pts, NULL, val3);
7808 case FUTEX_WAKE:
7809 return do_safe_futex(g2h(cpu, uaddr), op, val, NULL, NULL, 0);
7810 case FUTEX_FD:
7811 return do_safe_futex(g2h(cpu, uaddr), op, val, NULL, NULL, 0);
7812 case FUTEX_REQUEUE:
7813 case FUTEX_CMP_REQUEUE:
7814 case FUTEX_WAKE_OP:
7815 /* For FUTEX_REQUEUE, FUTEX_CMP_REQUEUE, and FUTEX_WAKE_OP, the
7816 TIMEOUT parameter is interpreted as a uint32_t by the kernel.
7817 But the prototype takes a `struct timespec *'; insert casts
7818 to satisfy the compiler. We do not need to tswap TIMEOUT
7819 since it's not compared to guest memory. */
7820 pts = (struct timespec *)(uintptr_t) timeout;
7821 return do_safe_futex(g2h(cpu, uaddr), op, val, pts, g2h(cpu, uaddr2),
7822 (base_op == FUTEX_CMP_REQUEUE
7823 ? tswap32(val3) : val3));
7824 default:
7825 return -TARGET_ENOSYS;
7828 #endif
7830 #if defined(TARGET_NR_name_to_handle_at) && defined(CONFIG_OPEN_BY_HANDLE)
7831 static abi_long do_name_to_handle_at(abi_long dirfd, abi_long pathname,
7832 abi_long handle, abi_long mount_id,
7833 abi_long flags)
7835 struct file_handle *target_fh;
7836 struct file_handle *fh;
7837 int mid = 0;
7838 abi_long ret;
7839 char *name;
7840 unsigned int size, total_size;
7842 if (get_user_s32(size, handle)) {
7843 return -TARGET_EFAULT;
7846 name = lock_user_string(pathname);
7847 if (!name) {
7848 return -TARGET_EFAULT;
7851 total_size = sizeof(struct file_handle) + size;
7852 target_fh = lock_user(VERIFY_WRITE, handle, total_size, 0);
7853 if (!target_fh) {
7854 unlock_user(name, pathname, 0);
7855 return -TARGET_EFAULT;
7858 fh = g_malloc0(total_size);
7859 fh->handle_bytes = size;
7861 ret = get_errno(name_to_handle_at(dirfd, path(name), fh, &mid, flags));
7862 unlock_user(name, pathname, 0);
7864 /* man name_to_handle_at(2):
7865 * Other than the use of the handle_bytes field, the caller should treat
7866 * the file_handle structure as an opaque data type
7869 memcpy(target_fh, fh, total_size);
7870 target_fh->handle_bytes = tswap32(fh->handle_bytes);
7871 target_fh->handle_type = tswap32(fh->handle_type);
7872 g_free(fh);
7873 unlock_user(target_fh, handle, total_size);
7875 if (put_user_s32(mid, mount_id)) {
7876 return -TARGET_EFAULT;
7879 return ret;
7882 #endif
7884 #if defined(TARGET_NR_open_by_handle_at) && defined(CONFIG_OPEN_BY_HANDLE)
7885 static abi_long do_open_by_handle_at(abi_long mount_fd, abi_long handle,
7886 abi_long flags)
7888 struct file_handle *target_fh;
7889 struct file_handle *fh;
7890 unsigned int size, total_size;
7891 abi_long ret;
7893 if (get_user_s32(size, handle)) {
7894 return -TARGET_EFAULT;
7897 total_size = sizeof(struct file_handle) + size;
7898 target_fh = lock_user(VERIFY_READ, handle, total_size, 1);
7899 if (!target_fh) {
7900 return -TARGET_EFAULT;
7903 fh = g_memdup(target_fh, total_size);
7904 fh->handle_bytes = size;
7905 fh->handle_type = tswap32(target_fh->handle_type);
7907 ret = get_errno(open_by_handle_at(mount_fd, fh,
7908 target_to_host_bitmask(flags, fcntl_flags_tbl)));
7910 g_free(fh);
7912 unlock_user(target_fh, handle, total_size);
7914 return ret;
7916 #endif
7918 #if defined(TARGET_NR_signalfd) || defined(TARGET_NR_signalfd4)
7920 static abi_long do_signalfd4(int fd, abi_long mask, int flags)
7922 int host_flags;
7923 target_sigset_t *target_mask;
7924 sigset_t host_mask;
7925 abi_long ret;
7927 if (flags & ~(TARGET_O_NONBLOCK_MASK | TARGET_O_CLOEXEC)) {
7928 return -TARGET_EINVAL;
7930 if (!lock_user_struct(VERIFY_READ, target_mask, mask, 1)) {
7931 return -TARGET_EFAULT;
7934 target_to_host_sigset(&host_mask, target_mask);
7936 host_flags = target_to_host_bitmask(flags, fcntl_flags_tbl);
7938 ret = get_errno(signalfd(fd, &host_mask, host_flags));
7939 if (ret >= 0) {
7940 fd_trans_register(ret, &target_signalfd_trans);
7943 unlock_user_struct(target_mask, mask, 0);
7945 return ret;
7947 #endif
7949 /* Map host to target signal numbers for the wait family of syscalls.
7950 Assume all other status bits are the same. */
7951 int host_to_target_waitstatus(int status)
7953 if (WIFSIGNALED(status)) {
7954 return host_to_target_signal(WTERMSIG(status)) | (status & ~0x7f);
7956 if (WIFSTOPPED(status)) {
7957 return (host_to_target_signal(WSTOPSIG(status)) << 8)
7958 | (status & 0xff);
7960 return status;
7963 static int open_self_cmdline(CPUArchState *cpu_env, int fd)
7965 CPUState *cpu = env_cpu(cpu_env);
7966 struct linux_binprm *bprm = ((TaskState *)cpu->opaque)->bprm;
7967 int i;
7969 for (i = 0; i < bprm->argc; i++) {
7970 size_t len = strlen(bprm->argv[i]) + 1;
7972 if (write(fd, bprm->argv[i], len) != len) {
7973 return -1;
7977 return 0;
7980 static int open_self_maps(CPUArchState *cpu_env, int fd)
7982 CPUState *cpu = env_cpu(cpu_env);
7983 TaskState *ts = cpu->opaque;
7984 GSList *map_info = read_self_maps();
7985 GSList *s;
7986 int count;
7988 for (s = map_info; s; s = g_slist_next(s)) {
7989 MapInfo *e = (MapInfo *) s->data;
7991 if (h2g_valid(e->start)) {
7992 unsigned long min = e->start;
7993 unsigned long max = e->end;
7994 int flags = page_get_flags(h2g(min));
7995 const char *path;
7997 max = h2g_valid(max - 1) ?
7998 max : (uintptr_t) g2h_untagged(GUEST_ADDR_MAX) + 1;
8000 if (page_check_range(h2g(min), max - min, flags) == -1) {
8001 continue;
8004 if (h2g(min) == ts->info->stack_limit) {
8005 path = "[stack]";
8006 } else {
8007 path = e->path;
8010 count = dprintf(fd, TARGET_ABI_FMT_ptr "-" TARGET_ABI_FMT_ptr
8011 " %c%c%c%c %08" PRIx64 " %s %"PRId64,
8012 h2g(min), h2g(max - 1) + 1,
8013 (flags & PAGE_READ) ? 'r' : '-',
8014 (flags & PAGE_WRITE_ORG) ? 'w' : '-',
8015 (flags & PAGE_EXEC) ? 'x' : '-',
8016 e->is_priv ? 'p' : 's',
8017 (uint64_t) e->offset, e->dev, e->inode);
8018 if (path) {
8019 dprintf(fd, "%*s%s\n", 73 - count, "", path);
8020 } else {
8021 dprintf(fd, "\n");
8026 free_self_maps(map_info);
8028 #ifdef TARGET_VSYSCALL_PAGE
8030 * We only support execution from the vsyscall page.
8031 * This is as if CONFIG_LEGACY_VSYSCALL_XONLY=y from v5.3.
8033 count = dprintf(fd, TARGET_FMT_lx "-" TARGET_FMT_lx
8034 " --xp 00000000 00:00 0",
8035 TARGET_VSYSCALL_PAGE, TARGET_VSYSCALL_PAGE + TARGET_PAGE_SIZE);
8036 dprintf(fd, "%*s%s\n", 73 - count, "", "[vsyscall]");
8037 #endif
8039 return 0;
8042 static int open_self_stat(CPUArchState *cpu_env, int fd)
8044 CPUState *cpu = env_cpu(cpu_env);
8045 TaskState *ts = cpu->opaque;
8046 g_autoptr(GString) buf = g_string_new(NULL);
8047 int i;
8049 for (i = 0; i < 44; i++) {
8050 if (i == 0) {
8051 /* pid */
8052 g_string_printf(buf, FMT_pid " ", getpid());
8053 } else if (i == 1) {
8054 /* app name */
8055 gchar *bin = g_strrstr(ts->bprm->argv[0], "/");
8056 bin = bin ? bin + 1 : ts->bprm->argv[0];
8057 g_string_printf(buf, "(%.15s) ", bin);
8058 } else if (i == 3) {
8059 /* ppid */
8060 g_string_printf(buf, FMT_pid " ", getppid());
8061 } else if (i == 21) {
8062 /* starttime */
8063 g_string_printf(buf, "%" PRIu64 " ", ts->start_boottime);
8064 } else if (i == 27) {
8065 /* stack bottom */
8066 g_string_printf(buf, TARGET_ABI_FMT_ld " ", ts->info->start_stack);
8067 } else {
8068 /* for the rest, there is MasterCard */
8069 g_string_printf(buf, "0%c", i == 43 ? '\n' : ' ');
8072 if (write(fd, buf->str, buf->len) != buf->len) {
8073 return -1;
8077 return 0;
8080 static int open_self_auxv(CPUArchState *cpu_env, int fd)
8082 CPUState *cpu = env_cpu(cpu_env);
8083 TaskState *ts = cpu->opaque;
8084 abi_ulong auxv = ts->info->saved_auxv;
8085 abi_ulong len = ts->info->auxv_len;
8086 char *ptr;
8089 * Auxiliary vector is stored in target process stack.
8090 * read in whole auxv vector and copy it to file
8092 ptr = lock_user(VERIFY_READ, auxv, len, 0);
8093 if (ptr != NULL) {
8094 while (len > 0) {
8095 ssize_t r;
8096 r = write(fd, ptr, len);
8097 if (r <= 0) {
8098 break;
8100 len -= r;
8101 ptr += r;
8103 lseek(fd, 0, SEEK_SET);
8104 unlock_user(ptr, auxv, len);
8107 return 0;
8110 static int is_proc_myself(const char *filename, const char *entry)
8112 if (!strncmp(filename, "/proc/", strlen("/proc/"))) {
8113 filename += strlen("/proc/");
8114 if (!strncmp(filename, "self/", strlen("self/"))) {
8115 filename += strlen("self/");
8116 } else if (*filename >= '1' && *filename <= '9') {
8117 char myself[80];
8118 snprintf(myself, sizeof(myself), "%d/", getpid());
8119 if (!strncmp(filename, myself, strlen(myself))) {
8120 filename += strlen(myself);
8121 } else {
8122 return 0;
8124 } else {
8125 return 0;
8127 if (!strcmp(filename, entry)) {
8128 return 1;
8131 return 0;
8134 #if HOST_BIG_ENDIAN != TARGET_BIG_ENDIAN || \
8135 defined(TARGET_SPARC) || defined(TARGET_M68K) || defined(TARGET_HPPA)
8136 static int is_proc(const char *filename, const char *entry)
8138 return strcmp(filename, entry) == 0;
8140 #endif
8142 #if HOST_BIG_ENDIAN != TARGET_BIG_ENDIAN
8143 static int open_net_route(CPUArchState *cpu_env, int fd)
8145 FILE *fp;
8146 char *line = NULL;
8147 size_t len = 0;
8148 ssize_t read;
8150 fp = fopen("/proc/net/route", "r");
8151 if (fp == NULL) {
8152 return -1;
8155 /* read header */
8157 read = getline(&line, &len, fp);
8158 dprintf(fd, "%s", line);
8160 /* read routes */
8162 while ((read = getline(&line, &len, fp)) != -1) {
8163 char iface[16];
8164 uint32_t dest, gw, mask;
8165 unsigned int flags, refcnt, use, metric, mtu, window, irtt;
8166 int fields;
8168 fields = sscanf(line,
8169 "%s\t%08x\t%08x\t%04x\t%d\t%d\t%d\t%08x\t%d\t%u\t%u\n",
8170 iface, &dest, &gw, &flags, &refcnt, &use, &metric,
8171 &mask, &mtu, &window, &irtt);
8172 if (fields != 11) {
8173 continue;
8175 dprintf(fd, "%s\t%08x\t%08x\t%04x\t%d\t%d\t%d\t%08x\t%d\t%u\t%u\n",
8176 iface, tswap32(dest), tswap32(gw), flags, refcnt, use,
8177 metric, tswap32(mask), mtu, window, irtt);
8180 free(line);
8181 fclose(fp);
8183 return 0;
8185 #endif
8187 #if defined(TARGET_SPARC)
8188 static int open_cpuinfo(CPUArchState *cpu_env, int fd)
8190 dprintf(fd, "type\t\t: sun4u\n");
8191 return 0;
8193 #endif
8195 #if defined(TARGET_HPPA)
8196 static int open_cpuinfo(CPUArchState *cpu_env, int fd)
8198 dprintf(fd, "cpu family\t: PA-RISC 1.1e\n");
8199 dprintf(fd, "cpu\t\t: PA7300LC (PCX-L2)\n");
8200 dprintf(fd, "capabilities\t: os32\n");
8201 dprintf(fd, "model\t\t: 9000/778/B160L\n");
8202 dprintf(fd, "model name\t: Merlin L2 160 QEMU (9000/778/B160L)\n");
8203 return 0;
8205 #endif
8207 #if defined(TARGET_M68K)
8208 static int open_hardware(CPUArchState *cpu_env, int fd)
8210 dprintf(fd, "Model:\t\tqemu-m68k\n");
8211 return 0;
8213 #endif
8215 static int do_openat(CPUArchState *cpu_env, int dirfd, const char *pathname, int flags, mode_t mode)
8217 struct fake_open {
8218 const char *filename;
8219 int (*fill)(CPUArchState *cpu_env, int fd);
8220 int (*cmp)(const char *s1, const char *s2);
8222 const struct fake_open *fake_open;
8223 static const struct fake_open fakes[] = {
8224 { "maps", open_self_maps, is_proc_myself },
8225 { "stat", open_self_stat, is_proc_myself },
8226 { "auxv", open_self_auxv, is_proc_myself },
8227 { "cmdline", open_self_cmdline, is_proc_myself },
8228 #if HOST_BIG_ENDIAN != TARGET_BIG_ENDIAN
8229 { "/proc/net/route", open_net_route, is_proc },
8230 #endif
8231 #if defined(TARGET_SPARC) || defined(TARGET_HPPA)
8232 { "/proc/cpuinfo", open_cpuinfo, is_proc },
8233 #endif
8234 #if defined(TARGET_M68K)
8235 { "/proc/hardware", open_hardware, is_proc },
8236 #endif
8237 { NULL, NULL, NULL }
8240 if (is_proc_myself(pathname, "exe")) {
8241 int execfd = qemu_getauxval(AT_EXECFD);
8242 return execfd ? execfd : safe_openat(dirfd, exec_path, flags, mode);
8245 for (fake_open = fakes; fake_open->filename; fake_open++) {
8246 if (fake_open->cmp(pathname, fake_open->filename)) {
8247 break;
8251 if (fake_open->filename) {
8252 const char *tmpdir;
8253 char filename[PATH_MAX];
8254 int fd, r;
8256 /* create temporary file to map stat to */
8257 tmpdir = getenv("TMPDIR");
8258 if (!tmpdir)
8259 tmpdir = "/tmp";
8260 snprintf(filename, sizeof(filename), "%s/qemu-open.XXXXXX", tmpdir);
8261 fd = mkstemp(filename);
8262 if (fd < 0) {
8263 return fd;
8265 unlink(filename);
8267 if ((r = fake_open->fill(cpu_env, fd))) {
8268 int e = errno;
8269 close(fd);
8270 errno = e;
8271 return r;
8273 lseek(fd, 0, SEEK_SET);
8275 return fd;
8278 return safe_openat(dirfd, path(pathname), flags, mode);
8281 #define TIMER_MAGIC 0x0caf0000
8282 #define TIMER_MAGIC_MASK 0xffff0000
8284 /* Convert QEMU provided timer ID back to internal 16bit index format */
8285 static target_timer_t get_timer_id(abi_long arg)
8287 target_timer_t timerid = arg;
8289 if ((timerid & TIMER_MAGIC_MASK) != TIMER_MAGIC) {
8290 return -TARGET_EINVAL;
8293 timerid &= 0xffff;
8295 if (timerid >= ARRAY_SIZE(g_posix_timers)) {
8296 return -TARGET_EINVAL;
8299 return timerid;
8302 static int target_to_host_cpu_mask(unsigned long *host_mask,
8303 size_t host_size,
8304 abi_ulong target_addr,
8305 size_t target_size)
8307 unsigned target_bits = sizeof(abi_ulong) * 8;
8308 unsigned host_bits = sizeof(*host_mask) * 8;
8309 abi_ulong *target_mask;
8310 unsigned i, j;
8312 assert(host_size >= target_size);
8314 target_mask = lock_user(VERIFY_READ, target_addr, target_size, 1);
8315 if (!target_mask) {
8316 return -TARGET_EFAULT;
8318 memset(host_mask, 0, host_size);
8320 for (i = 0 ; i < target_size / sizeof(abi_ulong); i++) {
8321 unsigned bit = i * target_bits;
8322 abi_ulong val;
8324 __get_user(val, &target_mask[i]);
8325 for (j = 0; j < target_bits; j++, bit++) {
8326 if (val & (1UL << j)) {
8327 host_mask[bit / host_bits] |= 1UL << (bit % host_bits);
8332 unlock_user(target_mask, target_addr, 0);
8333 return 0;
8336 static int host_to_target_cpu_mask(const unsigned long *host_mask,
8337 size_t host_size,
8338 abi_ulong target_addr,
8339 size_t target_size)
8341 unsigned target_bits = sizeof(abi_ulong) * 8;
8342 unsigned host_bits = sizeof(*host_mask) * 8;
8343 abi_ulong *target_mask;
8344 unsigned i, j;
8346 assert(host_size >= target_size);
8348 target_mask = lock_user(VERIFY_WRITE, target_addr, target_size, 0);
8349 if (!target_mask) {
8350 return -TARGET_EFAULT;
8353 for (i = 0 ; i < target_size / sizeof(abi_ulong); i++) {
8354 unsigned bit = i * target_bits;
8355 abi_ulong val = 0;
8357 for (j = 0; j < target_bits; j++, bit++) {
8358 if (host_mask[bit / host_bits] & (1UL << (bit % host_bits))) {
8359 val |= 1UL << j;
8362 __put_user(val, &target_mask[i]);
8365 unlock_user(target_mask, target_addr, target_size);
8366 return 0;
8369 #ifdef TARGET_NR_getdents
8370 static int do_getdents(abi_long dirfd, abi_long arg2, abi_long count)
8372 g_autofree void *hdirp = NULL;
8373 void *tdirp;
8374 int hlen, hoff, toff;
8375 int hreclen, treclen;
8376 off64_t prev_diroff = 0;
8378 hdirp = g_try_malloc(count);
8379 if (!hdirp) {
8380 return -TARGET_ENOMEM;
8383 #ifdef EMULATE_GETDENTS_WITH_GETDENTS
8384 hlen = sys_getdents(dirfd, hdirp, count);
8385 #else
8386 hlen = sys_getdents64(dirfd, hdirp, count);
8387 #endif
8389 hlen = get_errno(hlen);
8390 if (is_error(hlen)) {
8391 return hlen;
8394 tdirp = lock_user(VERIFY_WRITE, arg2, count, 0);
8395 if (!tdirp) {
8396 return -TARGET_EFAULT;
8399 for (hoff = toff = 0; hoff < hlen; hoff += hreclen, toff += treclen) {
8400 #ifdef EMULATE_GETDENTS_WITH_GETDENTS
8401 struct linux_dirent *hde = hdirp + hoff;
8402 #else
8403 struct linux_dirent64 *hde = hdirp + hoff;
8404 #endif
8405 struct target_dirent *tde = tdirp + toff;
8406 int namelen;
8407 uint8_t type;
8409 namelen = strlen(hde->d_name);
8410 hreclen = hde->d_reclen;
8411 treclen = offsetof(struct target_dirent, d_name) + namelen + 2;
8412 treclen = QEMU_ALIGN_UP(treclen, __alignof(struct target_dirent));
8414 if (toff + treclen > count) {
8416 * If the host struct is smaller than the target struct, or
8417 * requires less alignment and thus packs into less space,
8418 * then the host can return more entries than we can pass
8419 * on to the guest.
8421 if (toff == 0) {
8422 toff = -TARGET_EINVAL; /* result buffer is too small */
8423 break;
8426 * Return what we have, resetting the file pointer to the
8427 * location of the first record not returned.
8429 lseek64(dirfd, prev_diroff, SEEK_SET);
8430 break;
8433 prev_diroff = hde->d_off;
8434 tde->d_ino = tswapal(hde->d_ino);
8435 tde->d_off = tswapal(hde->d_off);
8436 tde->d_reclen = tswap16(treclen);
8437 memcpy(tde->d_name, hde->d_name, namelen + 1);
8440 * The getdents type is in what was formerly a padding byte at the
8441 * end of the structure.
8443 #ifdef EMULATE_GETDENTS_WITH_GETDENTS
8444 type = *((uint8_t *)hde + hreclen - 1);
8445 #else
8446 type = hde->d_type;
8447 #endif
8448 *((uint8_t *)tde + treclen - 1) = type;
8451 unlock_user(tdirp, arg2, toff);
8452 return toff;
8454 #endif /* TARGET_NR_getdents */
8456 #if defined(TARGET_NR_getdents64) && defined(__NR_getdents64)
8457 static int do_getdents64(abi_long dirfd, abi_long arg2, abi_long count)
8459 g_autofree void *hdirp = NULL;
8460 void *tdirp;
8461 int hlen, hoff, toff;
8462 int hreclen, treclen;
8463 off64_t prev_diroff = 0;
8465 hdirp = g_try_malloc(count);
8466 if (!hdirp) {
8467 return -TARGET_ENOMEM;
8470 hlen = get_errno(sys_getdents64(dirfd, hdirp, count));
8471 if (is_error(hlen)) {
8472 return hlen;
8475 tdirp = lock_user(VERIFY_WRITE, arg2, count, 0);
8476 if (!tdirp) {
8477 return -TARGET_EFAULT;
8480 for (hoff = toff = 0; hoff < hlen; hoff += hreclen, toff += treclen) {
8481 struct linux_dirent64 *hde = hdirp + hoff;
8482 struct target_dirent64 *tde = tdirp + toff;
8483 int namelen;
8485 namelen = strlen(hde->d_name) + 1;
8486 hreclen = hde->d_reclen;
8487 treclen = offsetof(struct target_dirent64, d_name) + namelen;
8488 treclen = QEMU_ALIGN_UP(treclen, __alignof(struct target_dirent64));
8490 if (toff + treclen > count) {
8492 * If the host struct is smaller than the target struct, or
8493 * requires less alignment and thus packs into less space,
8494 * then the host can return more entries than we can pass
8495 * on to the guest.
8497 if (toff == 0) {
8498 toff = -TARGET_EINVAL; /* result buffer is too small */
8499 break;
8502 * Return what we have, resetting the file pointer to the
8503 * location of the first record not returned.
8505 lseek64(dirfd, prev_diroff, SEEK_SET);
8506 break;
8509 prev_diroff = hde->d_off;
8510 tde->d_ino = tswap64(hde->d_ino);
8511 tde->d_off = tswap64(hde->d_off);
8512 tde->d_reclen = tswap16(treclen);
8513 tde->d_type = hde->d_type;
8514 memcpy(tde->d_name, hde->d_name, namelen);
8517 unlock_user(tdirp, arg2, toff);
8518 return toff;
8520 #endif /* TARGET_NR_getdents64 */
8522 #if defined(TARGET_NR_pivot_root) && defined(__NR_pivot_root)
8523 _syscall2(int, pivot_root, const char *, new_root, const char *, put_old)
8524 #endif
8526 /* This is an internal helper for do_syscall so that it is easier
8527 * to have a single return point, so that actions, such as logging
8528 * of syscall results, can be performed.
8529 * All errnos that do_syscall() returns must be -TARGET_<errcode>.
8531 static abi_long do_syscall1(CPUArchState *cpu_env, int num, abi_long arg1,
8532 abi_long arg2, abi_long arg3, abi_long arg4,
8533 abi_long arg5, abi_long arg6, abi_long arg7,
8534 abi_long arg8)
8536 CPUState *cpu = env_cpu(cpu_env);
8537 abi_long ret;
8538 #if defined(TARGET_NR_stat) || defined(TARGET_NR_stat64) \
8539 || defined(TARGET_NR_lstat) || defined(TARGET_NR_lstat64) \
8540 || defined(TARGET_NR_fstat) || defined(TARGET_NR_fstat64) \
8541 || defined(TARGET_NR_statx)
8542 struct stat st;
8543 #endif
8544 #if defined(TARGET_NR_statfs) || defined(TARGET_NR_statfs64) \
8545 || defined(TARGET_NR_fstatfs)
8546 struct statfs stfs;
8547 #endif
8548 void *p;
8550 switch(num) {
8551 case TARGET_NR_exit:
8552 /* In old applications this may be used to implement _exit(2).
8553 However in threaded applications it is used for thread termination,
8554 and _exit_group is used for application termination.
8555 Do thread termination if we have more then one thread. */
8557 if (block_signals()) {
8558 return -QEMU_ERESTARTSYS;
8561 pthread_mutex_lock(&clone_lock);
8563 if (CPU_NEXT(first_cpu)) {
8564 TaskState *ts = cpu->opaque;
8566 object_property_set_bool(OBJECT(cpu), "realized", false, NULL);
8567 object_unref(OBJECT(cpu));
8569 * At this point the CPU should be unrealized and removed
8570 * from cpu lists. We can clean-up the rest of the thread
8571 * data without the lock held.
8574 pthread_mutex_unlock(&clone_lock);
8576 if (ts->child_tidptr) {
8577 put_user_u32(0, ts->child_tidptr);
8578 do_sys_futex(g2h(cpu, ts->child_tidptr),
8579 FUTEX_WAKE, INT_MAX, NULL, NULL, 0);
8581 thread_cpu = NULL;
8582 g_free(ts);
8583 rcu_unregister_thread();
8584 pthread_exit(NULL);
8587 pthread_mutex_unlock(&clone_lock);
8588 preexit_cleanup(cpu_env, arg1);
8589 _exit(arg1);
8590 return 0; /* avoid warning */
8591 case TARGET_NR_read:
8592 if (arg2 == 0 && arg3 == 0) {
8593 return get_errno(safe_read(arg1, 0, 0));
8594 } else {
8595 if (!(p = lock_user(VERIFY_WRITE, arg2, arg3, 0)))
8596 return -TARGET_EFAULT;
8597 ret = get_errno(safe_read(arg1, p, arg3));
8598 if (ret >= 0 &&
8599 fd_trans_host_to_target_data(arg1)) {
8600 ret = fd_trans_host_to_target_data(arg1)(p, ret);
8602 unlock_user(p, arg2, ret);
8604 return ret;
8605 case TARGET_NR_write:
8606 if (arg2 == 0 && arg3 == 0) {
8607 return get_errno(safe_write(arg1, 0, 0));
8609 if (!(p = lock_user(VERIFY_READ, arg2, arg3, 1)))
8610 return -TARGET_EFAULT;
8611 if (fd_trans_target_to_host_data(arg1)) {
8612 void *copy = g_malloc(arg3);
8613 memcpy(copy, p, arg3);
8614 ret = fd_trans_target_to_host_data(arg1)(copy, arg3);
8615 if (ret >= 0) {
8616 ret = get_errno(safe_write(arg1, copy, ret));
8618 g_free(copy);
8619 } else {
8620 ret = get_errno(safe_write(arg1, p, arg3));
8622 unlock_user(p, arg2, 0);
8623 return ret;
8625 #ifdef TARGET_NR_open
8626 case TARGET_NR_open:
8627 if (!(p = lock_user_string(arg1)))
8628 return -TARGET_EFAULT;
8629 ret = get_errno(do_openat(cpu_env, AT_FDCWD, p,
8630 target_to_host_bitmask(arg2, fcntl_flags_tbl),
8631 arg3));
8632 fd_trans_unregister(ret);
8633 unlock_user(p, arg1, 0);
8634 return ret;
8635 #endif
8636 case TARGET_NR_openat:
8637 if (!(p = lock_user_string(arg2)))
8638 return -TARGET_EFAULT;
8639 ret = get_errno(do_openat(cpu_env, arg1, p,
8640 target_to_host_bitmask(arg3, fcntl_flags_tbl),
8641 arg4));
8642 fd_trans_unregister(ret);
8643 unlock_user(p, arg2, 0);
8644 return ret;
8645 #if defined(TARGET_NR_name_to_handle_at) && defined(CONFIG_OPEN_BY_HANDLE)
8646 case TARGET_NR_name_to_handle_at:
8647 ret = do_name_to_handle_at(arg1, arg2, arg3, arg4, arg5);
8648 return ret;
8649 #endif
8650 #if defined(TARGET_NR_open_by_handle_at) && defined(CONFIG_OPEN_BY_HANDLE)
8651 case TARGET_NR_open_by_handle_at:
8652 ret = do_open_by_handle_at(arg1, arg2, arg3);
8653 fd_trans_unregister(ret);
8654 return ret;
8655 #endif
8656 case TARGET_NR_close:
8657 fd_trans_unregister(arg1);
8658 return get_errno(close(arg1));
8660 case TARGET_NR_brk:
8661 return do_brk(arg1);
8662 #ifdef TARGET_NR_fork
8663 case TARGET_NR_fork:
8664 return get_errno(do_fork(cpu_env, TARGET_SIGCHLD, 0, 0, 0, 0));
8665 #endif
8666 #ifdef TARGET_NR_waitpid
8667 case TARGET_NR_waitpid:
8669 int status;
8670 ret = get_errno(safe_wait4(arg1, &status, arg3, 0));
8671 if (!is_error(ret) && arg2 && ret
8672 && put_user_s32(host_to_target_waitstatus(status), arg2))
8673 return -TARGET_EFAULT;
8675 return ret;
8676 #endif
8677 #ifdef TARGET_NR_waitid
8678 case TARGET_NR_waitid:
8680 siginfo_t info;
8681 info.si_pid = 0;
8682 ret = get_errno(safe_waitid(arg1, arg2, &info, arg4, NULL));
8683 if (!is_error(ret) && arg3 && info.si_pid != 0) {
8684 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_siginfo_t), 0)))
8685 return -TARGET_EFAULT;
8686 host_to_target_siginfo(p, &info);
8687 unlock_user(p, arg3, sizeof(target_siginfo_t));
8690 return ret;
8691 #endif
8692 #ifdef TARGET_NR_creat /* not on alpha */
8693 case TARGET_NR_creat:
8694 if (!(p = lock_user_string(arg1)))
8695 return -TARGET_EFAULT;
8696 ret = get_errno(creat(p, arg2));
8697 fd_trans_unregister(ret);
8698 unlock_user(p, arg1, 0);
8699 return ret;
8700 #endif
8701 #ifdef TARGET_NR_link
8702 case TARGET_NR_link:
8704 void * p2;
8705 p = lock_user_string(arg1);
8706 p2 = lock_user_string(arg2);
8707 if (!p || !p2)
8708 ret = -TARGET_EFAULT;
8709 else
8710 ret = get_errno(link(p, p2));
8711 unlock_user(p2, arg2, 0);
8712 unlock_user(p, arg1, 0);
8714 return ret;
8715 #endif
8716 #if defined(TARGET_NR_linkat)
8717 case TARGET_NR_linkat:
8719 void * p2 = NULL;
8720 if (!arg2 || !arg4)
8721 return -TARGET_EFAULT;
8722 p = lock_user_string(arg2);
8723 p2 = lock_user_string(arg4);
8724 if (!p || !p2)
8725 ret = -TARGET_EFAULT;
8726 else
8727 ret = get_errno(linkat(arg1, p, arg3, p2, arg5));
8728 unlock_user(p, arg2, 0);
8729 unlock_user(p2, arg4, 0);
8731 return ret;
8732 #endif
8733 #ifdef TARGET_NR_unlink
8734 case TARGET_NR_unlink:
8735 if (!(p = lock_user_string(arg1)))
8736 return -TARGET_EFAULT;
8737 ret = get_errno(unlink(p));
8738 unlock_user(p, arg1, 0);
8739 return ret;
8740 #endif
8741 #if defined(TARGET_NR_unlinkat)
8742 case TARGET_NR_unlinkat:
8743 if (!(p = lock_user_string(arg2)))
8744 return -TARGET_EFAULT;
8745 ret = get_errno(unlinkat(arg1, p, arg3));
8746 unlock_user(p, arg2, 0);
8747 return ret;
8748 #endif
8749 case TARGET_NR_execve:
8751 char **argp, **envp;
8752 int argc, envc;
8753 abi_ulong gp;
8754 abi_ulong guest_argp;
8755 abi_ulong guest_envp;
8756 abi_ulong addr;
8757 char **q;
8759 argc = 0;
8760 guest_argp = arg2;
8761 for (gp = guest_argp; gp; gp += sizeof(abi_ulong)) {
8762 if (get_user_ual(addr, gp))
8763 return -TARGET_EFAULT;
8764 if (!addr)
8765 break;
8766 argc++;
8768 envc = 0;
8769 guest_envp = arg3;
8770 for (gp = guest_envp; gp; gp += sizeof(abi_ulong)) {
8771 if (get_user_ual(addr, gp))
8772 return -TARGET_EFAULT;
8773 if (!addr)
8774 break;
8775 envc++;
8778 argp = g_new0(char *, argc + 1);
8779 envp = g_new0(char *, envc + 1);
8781 for (gp = guest_argp, q = argp; gp;
8782 gp += sizeof(abi_ulong), q++) {
8783 if (get_user_ual(addr, gp))
8784 goto execve_efault;
8785 if (!addr)
8786 break;
8787 if (!(*q = lock_user_string(addr)))
8788 goto execve_efault;
8790 *q = NULL;
8792 for (gp = guest_envp, q = envp; gp;
8793 gp += sizeof(abi_ulong), q++) {
8794 if (get_user_ual(addr, gp))
8795 goto execve_efault;
8796 if (!addr)
8797 break;
8798 if (!(*q = lock_user_string(addr)))
8799 goto execve_efault;
8801 *q = NULL;
8803 if (!(p = lock_user_string(arg1)))
8804 goto execve_efault;
8805 /* Although execve() is not an interruptible syscall it is
8806 * a special case where we must use the safe_syscall wrapper:
8807 * if we allow a signal to happen before we make the host
8808 * syscall then we will 'lose' it, because at the point of
8809 * execve the process leaves QEMU's control. So we use the
8810 * safe syscall wrapper to ensure that we either take the
8811 * signal as a guest signal, or else it does not happen
8812 * before the execve completes and makes it the other
8813 * program's problem.
8815 ret = get_errno(safe_execve(p, argp, envp));
8816 unlock_user(p, arg1, 0);
8818 goto execve_end;
8820 execve_efault:
8821 ret = -TARGET_EFAULT;
8823 execve_end:
8824 for (gp = guest_argp, q = argp; *q;
8825 gp += sizeof(abi_ulong), q++) {
8826 if (get_user_ual(addr, gp)
8827 || !addr)
8828 break;
8829 unlock_user(*q, addr, 0);
8831 for (gp = guest_envp, q = envp; *q;
8832 gp += sizeof(abi_ulong), q++) {
8833 if (get_user_ual(addr, gp)
8834 || !addr)
8835 break;
8836 unlock_user(*q, addr, 0);
8839 g_free(argp);
8840 g_free(envp);
8842 return ret;
8843 case TARGET_NR_chdir:
8844 if (!(p = lock_user_string(arg1)))
8845 return -TARGET_EFAULT;
8846 ret = get_errno(chdir(p));
8847 unlock_user(p, arg1, 0);
8848 return ret;
8849 #ifdef TARGET_NR_time
8850 case TARGET_NR_time:
8852 time_t host_time;
8853 ret = get_errno(time(&host_time));
8854 if (!is_error(ret)
8855 && arg1
8856 && put_user_sal(host_time, arg1))
8857 return -TARGET_EFAULT;
8859 return ret;
8860 #endif
8861 #ifdef TARGET_NR_mknod
8862 case TARGET_NR_mknod:
8863 if (!(p = lock_user_string(arg1)))
8864 return -TARGET_EFAULT;
8865 ret = get_errno(mknod(p, arg2, arg3));
8866 unlock_user(p, arg1, 0);
8867 return ret;
8868 #endif
8869 #if defined(TARGET_NR_mknodat)
8870 case TARGET_NR_mknodat:
8871 if (!(p = lock_user_string(arg2)))
8872 return -TARGET_EFAULT;
8873 ret = get_errno(mknodat(arg1, p, arg3, arg4));
8874 unlock_user(p, arg2, 0);
8875 return ret;
8876 #endif
8877 #ifdef TARGET_NR_chmod
8878 case TARGET_NR_chmod:
8879 if (!(p = lock_user_string(arg1)))
8880 return -TARGET_EFAULT;
8881 ret = get_errno(chmod(p, arg2));
8882 unlock_user(p, arg1, 0);
8883 return ret;
8884 #endif
8885 #ifdef TARGET_NR_lseek
8886 case TARGET_NR_lseek:
8887 return get_errno(lseek(arg1, arg2, arg3));
8888 #endif
8889 #if defined(TARGET_NR_getxpid) && defined(TARGET_ALPHA)
8890 /* Alpha specific */
8891 case TARGET_NR_getxpid:
8892 cpu_env->ir[IR_A4] = getppid();
8893 return get_errno(getpid());
8894 #endif
8895 #ifdef TARGET_NR_getpid
8896 case TARGET_NR_getpid:
8897 return get_errno(getpid());
8898 #endif
8899 case TARGET_NR_mount:
8901 /* need to look at the data field */
8902 void *p2, *p3;
8904 if (arg1) {
8905 p = lock_user_string(arg1);
8906 if (!p) {
8907 return -TARGET_EFAULT;
8909 } else {
8910 p = NULL;
8913 p2 = lock_user_string(arg2);
8914 if (!p2) {
8915 if (arg1) {
8916 unlock_user(p, arg1, 0);
8918 return -TARGET_EFAULT;
8921 if (arg3) {
8922 p3 = lock_user_string(arg3);
8923 if (!p3) {
8924 if (arg1) {
8925 unlock_user(p, arg1, 0);
8927 unlock_user(p2, arg2, 0);
8928 return -TARGET_EFAULT;
8930 } else {
8931 p3 = NULL;
8934 /* FIXME - arg5 should be locked, but it isn't clear how to
8935 * do that since it's not guaranteed to be a NULL-terminated
8936 * string.
8938 if (!arg5) {
8939 ret = mount(p, p2, p3, (unsigned long)arg4, NULL);
8940 } else {
8941 ret = mount(p, p2, p3, (unsigned long)arg4, g2h(cpu, arg5));
8943 ret = get_errno(ret);
8945 if (arg1) {
8946 unlock_user(p, arg1, 0);
8948 unlock_user(p2, arg2, 0);
8949 if (arg3) {
8950 unlock_user(p3, arg3, 0);
8953 return ret;
8954 #if defined(TARGET_NR_umount) || defined(TARGET_NR_oldumount)
8955 #if defined(TARGET_NR_umount)
8956 case TARGET_NR_umount:
8957 #endif
8958 #if defined(TARGET_NR_oldumount)
8959 case TARGET_NR_oldumount:
8960 #endif
8961 if (!(p = lock_user_string(arg1)))
8962 return -TARGET_EFAULT;
8963 ret = get_errno(umount(p));
8964 unlock_user(p, arg1, 0);
8965 return ret;
8966 #endif
8967 #ifdef TARGET_NR_stime /* not on alpha */
8968 case TARGET_NR_stime:
8970 struct timespec ts;
8971 ts.tv_nsec = 0;
8972 if (get_user_sal(ts.tv_sec, arg1)) {
8973 return -TARGET_EFAULT;
8975 return get_errno(clock_settime(CLOCK_REALTIME, &ts));
8977 #endif
8978 #ifdef TARGET_NR_alarm /* not on alpha */
8979 case TARGET_NR_alarm:
8980 return alarm(arg1);
8981 #endif
8982 #ifdef TARGET_NR_pause /* not on alpha */
8983 case TARGET_NR_pause:
8984 if (!block_signals()) {
8985 sigsuspend(&((TaskState *)cpu->opaque)->signal_mask);
8987 return -TARGET_EINTR;
8988 #endif
8989 #ifdef TARGET_NR_utime
8990 case TARGET_NR_utime:
8992 struct utimbuf tbuf, *host_tbuf;
8993 struct target_utimbuf *target_tbuf;
8994 if (arg2) {
8995 if (!lock_user_struct(VERIFY_READ, target_tbuf, arg2, 1))
8996 return -TARGET_EFAULT;
8997 tbuf.actime = tswapal(target_tbuf->actime);
8998 tbuf.modtime = tswapal(target_tbuf->modtime);
8999 unlock_user_struct(target_tbuf, arg2, 0);
9000 host_tbuf = &tbuf;
9001 } else {
9002 host_tbuf = NULL;
9004 if (!(p = lock_user_string(arg1)))
9005 return -TARGET_EFAULT;
9006 ret = get_errno(utime(p, host_tbuf));
9007 unlock_user(p, arg1, 0);
9009 return ret;
9010 #endif
9011 #ifdef TARGET_NR_utimes
9012 case TARGET_NR_utimes:
9014 struct timeval *tvp, tv[2];
9015 if (arg2) {
9016 if (copy_from_user_timeval(&tv[0], arg2)
9017 || copy_from_user_timeval(&tv[1],
9018 arg2 + sizeof(struct target_timeval)))
9019 return -TARGET_EFAULT;
9020 tvp = tv;
9021 } else {
9022 tvp = NULL;
9024 if (!(p = lock_user_string(arg1)))
9025 return -TARGET_EFAULT;
9026 ret = get_errno(utimes(p, tvp));
9027 unlock_user(p, arg1, 0);
9029 return ret;
9030 #endif
9031 #if defined(TARGET_NR_futimesat)
9032 case TARGET_NR_futimesat:
9034 struct timeval *tvp, tv[2];
9035 if (arg3) {
9036 if (copy_from_user_timeval(&tv[0], arg3)
9037 || copy_from_user_timeval(&tv[1],
9038 arg3 + sizeof(struct target_timeval)))
9039 return -TARGET_EFAULT;
9040 tvp = tv;
9041 } else {
9042 tvp = NULL;
9044 if (!(p = lock_user_string(arg2))) {
9045 return -TARGET_EFAULT;
9047 ret = get_errno(futimesat(arg1, path(p), tvp));
9048 unlock_user(p, arg2, 0);
9050 return ret;
9051 #endif
9052 #ifdef TARGET_NR_access
9053 case TARGET_NR_access:
9054 if (!(p = lock_user_string(arg1))) {
9055 return -TARGET_EFAULT;
9057 ret = get_errno(access(path(p), arg2));
9058 unlock_user(p, arg1, 0);
9059 return ret;
9060 #endif
9061 #if defined(TARGET_NR_faccessat) && defined(__NR_faccessat)
9062 case TARGET_NR_faccessat:
9063 if (!(p = lock_user_string(arg2))) {
9064 return -TARGET_EFAULT;
9066 ret = get_errno(faccessat(arg1, p, arg3, 0));
9067 unlock_user(p, arg2, 0);
9068 return ret;
9069 #endif
9070 #ifdef TARGET_NR_nice /* not on alpha */
9071 case TARGET_NR_nice:
9072 return get_errno(nice(arg1));
9073 #endif
9074 case TARGET_NR_sync:
9075 sync();
9076 return 0;
9077 #if defined(TARGET_NR_syncfs) && defined(CONFIG_SYNCFS)
9078 case TARGET_NR_syncfs:
9079 return get_errno(syncfs(arg1));
9080 #endif
9081 case TARGET_NR_kill:
9082 return get_errno(safe_kill(arg1, target_to_host_signal(arg2)));
9083 #ifdef TARGET_NR_rename
9084 case TARGET_NR_rename:
9086 void *p2;
9087 p = lock_user_string(arg1);
9088 p2 = lock_user_string(arg2);
9089 if (!p || !p2)
9090 ret = -TARGET_EFAULT;
9091 else
9092 ret = get_errno(rename(p, p2));
9093 unlock_user(p2, arg2, 0);
9094 unlock_user(p, arg1, 0);
9096 return ret;
9097 #endif
9098 #if defined(TARGET_NR_renameat)
9099 case TARGET_NR_renameat:
9101 void *p2;
9102 p = lock_user_string(arg2);
9103 p2 = lock_user_string(arg4);
9104 if (!p || !p2)
9105 ret = -TARGET_EFAULT;
9106 else
9107 ret = get_errno(renameat(arg1, p, arg3, p2));
9108 unlock_user(p2, arg4, 0);
9109 unlock_user(p, arg2, 0);
9111 return ret;
9112 #endif
9113 #if defined(TARGET_NR_renameat2)
9114 case TARGET_NR_renameat2:
9116 void *p2;
9117 p = lock_user_string(arg2);
9118 p2 = lock_user_string(arg4);
9119 if (!p || !p2) {
9120 ret = -TARGET_EFAULT;
9121 } else {
9122 ret = get_errno(sys_renameat2(arg1, p, arg3, p2, arg5));
9124 unlock_user(p2, arg4, 0);
9125 unlock_user(p, arg2, 0);
9127 return ret;
9128 #endif
9129 #ifdef TARGET_NR_mkdir
9130 case TARGET_NR_mkdir:
9131 if (!(p = lock_user_string(arg1)))
9132 return -TARGET_EFAULT;
9133 ret = get_errno(mkdir(p, arg2));
9134 unlock_user(p, arg1, 0);
9135 return ret;
9136 #endif
9137 #if defined(TARGET_NR_mkdirat)
9138 case TARGET_NR_mkdirat:
9139 if (!(p = lock_user_string(arg2)))
9140 return -TARGET_EFAULT;
9141 ret = get_errno(mkdirat(arg1, p, arg3));
9142 unlock_user(p, arg2, 0);
9143 return ret;
9144 #endif
9145 #ifdef TARGET_NR_rmdir
9146 case TARGET_NR_rmdir:
9147 if (!(p = lock_user_string(arg1)))
9148 return -TARGET_EFAULT;
9149 ret = get_errno(rmdir(p));
9150 unlock_user(p, arg1, 0);
9151 return ret;
9152 #endif
9153 case TARGET_NR_dup:
9154 ret = get_errno(dup(arg1));
9155 if (ret >= 0) {
9156 fd_trans_dup(arg1, ret);
9158 return ret;
9159 #ifdef TARGET_NR_pipe
9160 case TARGET_NR_pipe:
9161 return do_pipe(cpu_env, arg1, 0, 0);
9162 #endif
9163 #ifdef TARGET_NR_pipe2
9164 case TARGET_NR_pipe2:
9165 return do_pipe(cpu_env, arg1,
9166 target_to_host_bitmask(arg2, fcntl_flags_tbl), 1);
9167 #endif
9168 case TARGET_NR_times:
9170 struct target_tms *tmsp;
9171 struct tms tms;
9172 ret = get_errno(times(&tms));
9173 if (arg1) {
9174 tmsp = lock_user(VERIFY_WRITE, arg1, sizeof(struct target_tms), 0);
9175 if (!tmsp)
9176 return -TARGET_EFAULT;
9177 tmsp->tms_utime = tswapal(host_to_target_clock_t(tms.tms_utime));
9178 tmsp->tms_stime = tswapal(host_to_target_clock_t(tms.tms_stime));
9179 tmsp->tms_cutime = tswapal(host_to_target_clock_t(tms.tms_cutime));
9180 tmsp->tms_cstime = tswapal(host_to_target_clock_t(tms.tms_cstime));
9182 if (!is_error(ret))
9183 ret = host_to_target_clock_t(ret);
9185 return ret;
9186 case TARGET_NR_acct:
9187 if (arg1 == 0) {
9188 ret = get_errno(acct(NULL));
9189 } else {
9190 if (!(p = lock_user_string(arg1))) {
9191 return -TARGET_EFAULT;
9193 ret = get_errno(acct(path(p)));
9194 unlock_user(p, arg1, 0);
9196 return ret;
9197 #ifdef TARGET_NR_umount2
9198 case TARGET_NR_umount2:
9199 if (!(p = lock_user_string(arg1)))
9200 return -TARGET_EFAULT;
9201 ret = get_errno(umount2(p, arg2));
9202 unlock_user(p, arg1, 0);
9203 return ret;
9204 #endif
9205 case TARGET_NR_ioctl:
9206 return do_ioctl(arg1, arg2, arg3);
9207 #ifdef TARGET_NR_fcntl
9208 case TARGET_NR_fcntl:
9209 return do_fcntl(arg1, arg2, arg3);
9210 #endif
9211 case TARGET_NR_setpgid:
9212 return get_errno(setpgid(arg1, arg2));
9213 case TARGET_NR_umask:
9214 return get_errno(umask(arg1));
9215 case TARGET_NR_chroot:
9216 if (!(p = lock_user_string(arg1)))
9217 return -TARGET_EFAULT;
9218 ret = get_errno(chroot(p));
9219 unlock_user(p, arg1, 0);
9220 return ret;
9221 #ifdef TARGET_NR_dup2
9222 case TARGET_NR_dup2:
9223 ret = get_errno(dup2(arg1, arg2));
9224 if (ret >= 0) {
9225 fd_trans_dup(arg1, arg2);
9227 return ret;
9228 #endif
9229 #if defined(CONFIG_DUP3) && defined(TARGET_NR_dup3)
9230 case TARGET_NR_dup3:
9232 int host_flags;
9234 if ((arg3 & ~TARGET_O_CLOEXEC) != 0) {
9235 return -EINVAL;
9237 host_flags = target_to_host_bitmask(arg3, fcntl_flags_tbl);
9238 ret = get_errno(dup3(arg1, arg2, host_flags));
9239 if (ret >= 0) {
9240 fd_trans_dup(arg1, arg2);
9242 return ret;
9244 #endif
9245 #ifdef TARGET_NR_getppid /* not on alpha */
9246 case TARGET_NR_getppid:
9247 return get_errno(getppid());
9248 #endif
9249 #ifdef TARGET_NR_getpgrp
9250 case TARGET_NR_getpgrp:
9251 return get_errno(getpgrp());
9252 #endif
9253 case TARGET_NR_setsid:
9254 return get_errno(setsid());
9255 #ifdef TARGET_NR_sigaction
9256 case TARGET_NR_sigaction:
9258 #if defined(TARGET_MIPS)
9259 struct target_sigaction act, oact, *pact, *old_act;
9261 if (arg2) {
9262 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1))
9263 return -TARGET_EFAULT;
9264 act._sa_handler = old_act->_sa_handler;
9265 target_siginitset(&act.sa_mask, old_act->sa_mask.sig[0]);
9266 act.sa_flags = old_act->sa_flags;
9267 unlock_user_struct(old_act, arg2, 0);
9268 pact = &act;
9269 } else {
9270 pact = NULL;
9273 ret = get_errno(do_sigaction(arg1, pact, &oact, 0));
9275 if (!is_error(ret) && arg3) {
9276 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0))
9277 return -TARGET_EFAULT;
9278 old_act->_sa_handler = oact._sa_handler;
9279 old_act->sa_flags = oact.sa_flags;
9280 old_act->sa_mask.sig[0] = oact.sa_mask.sig[0];
9281 old_act->sa_mask.sig[1] = 0;
9282 old_act->sa_mask.sig[2] = 0;
9283 old_act->sa_mask.sig[3] = 0;
9284 unlock_user_struct(old_act, arg3, 1);
9286 #else
9287 struct target_old_sigaction *old_act;
9288 struct target_sigaction act, oact, *pact;
9289 if (arg2) {
9290 if (!lock_user_struct(VERIFY_READ, old_act, arg2, 1))
9291 return -TARGET_EFAULT;
9292 act._sa_handler = old_act->_sa_handler;
9293 target_siginitset(&act.sa_mask, old_act->sa_mask);
9294 act.sa_flags = old_act->sa_flags;
9295 #ifdef TARGET_ARCH_HAS_SA_RESTORER
9296 act.sa_restorer = old_act->sa_restorer;
9297 #endif
9298 unlock_user_struct(old_act, arg2, 0);
9299 pact = &act;
9300 } else {
9301 pact = NULL;
9303 ret = get_errno(do_sigaction(arg1, pact, &oact, 0));
9304 if (!is_error(ret) && arg3) {
9305 if (!lock_user_struct(VERIFY_WRITE, old_act, arg3, 0))
9306 return -TARGET_EFAULT;
9307 old_act->_sa_handler = oact._sa_handler;
9308 old_act->sa_mask = oact.sa_mask.sig[0];
9309 old_act->sa_flags = oact.sa_flags;
9310 #ifdef TARGET_ARCH_HAS_SA_RESTORER
9311 old_act->sa_restorer = oact.sa_restorer;
9312 #endif
9313 unlock_user_struct(old_act, arg3, 1);
9315 #endif
9317 return ret;
9318 #endif
9319 case TARGET_NR_rt_sigaction:
9322 * For Alpha and SPARC this is a 5 argument syscall, with
9323 * a 'restorer' parameter which must be copied into the
9324 * sa_restorer field of the sigaction struct.
9325 * For Alpha that 'restorer' is arg5; for SPARC it is arg4,
9326 * and arg5 is the sigsetsize.
9328 #if defined(TARGET_ALPHA)
9329 target_ulong sigsetsize = arg4;
9330 target_ulong restorer = arg5;
9331 #elif defined(TARGET_SPARC)
9332 target_ulong restorer = arg4;
9333 target_ulong sigsetsize = arg5;
9334 #else
9335 target_ulong sigsetsize = arg4;
9336 target_ulong restorer = 0;
9337 #endif
9338 struct target_sigaction *act = NULL;
9339 struct target_sigaction *oact = NULL;
9341 if (sigsetsize != sizeof(target_sigset_t)) {
9342 return -TARGET_EINVAL;
9344 if (arg2 && !lock_user_struct(VERIFY_READ, act, arg2, 1)) {
9345 return -TARGET_EFAULT;
9347 if (arg3 && !lock_user_struct(VERIFY_WRITE, oact, arg3, 0)) {
9348 ret = -TARGET_EFAULT;
9349 } else {
9350 ret = get_errno(do_sigaction(arg1, act, oact, restorer));
9351 if (oact) {
9352 unlock_user_struct(oact, arg3, 1);
9355 if (act) {
9356 unlock_user_struct(act, arg2, 0);
9359 return ret;
9360 #ifdef TARGET_NR_sgetmask /* not on alpha */
9361 case TARGET_NR_sgetmask:
9363 sigset_t cur_set;
9364 abi_ulong target_set;
9365 ret = do_sigprocmask(0, NULL, &cur_set);
9366 if (!ret) {
9367 host_to_target_old_sigset(&target_set, &cur_set);
9368 ret = target_set;
9371 return ret;
9372 #endif
9373 #ifdef TARGET_NR_ssetmask /* not on alpha */
9374 case TARGET_NR_ssetmask:
9376 sigset_t set, oset;
9377 abi_ulong target_set = arg1;
9378 target_to_host_old_sigset(&set, &target_set);
9379 ret = do_sigprocmask(SIG_SETMASK, &set, &oset);
9380 if (!ret) {
9381 host_to_target_old_sigset(&target_set, &oset);
9382 ret = target_set;
9385 return ret;
9386 #endif
9387 #ifdef TARGET_NR_sigprocmask
9388 case TARGET_NR_sigprocmask:
9390 #if defined(TARGET_ALPHA)
9391 sigset_t set, oldset;
9392 abi_ulong mask;
9393 int how;
9395 switch (arg1) {
9396 case TARGET_SIG_BLOCK:
9397 how = SIG_BLOCK;
9398 break;
9399 case TARGET_SIG_UNBLOCK:
9400 how = SIG_UNBLOCK;
9401 break;
9402 case TARGET_SIG_SETMASK:
9403 how = SIG_SETMASK;
9404 break;
9405 default:
9406 return -TARGET_EINVAL;
9408 mask = arg2;
9409 target_to_host_old_sigset(&set, &mask);
9411 ret = do_sigprocmask(how, &set, &oldset);
9412 if (!is_error(ret)) {
9413 host_to_target_old_sigset(&mask, &oldset);
9414 ret = mask;
9415 cpu_env->ir[IR_V0] = 0; /* force no error */
9417 #else
9418 sigset_t set, oldset, *set_ptr;
9419 int how;
9421 if (arg2) {
9422 p = lock_user(VERIFY_READ, arg2, sizeof(target_sigset_t), 1);
9423 if (!p) {
9424 return -TARGET_EFAULT;
9426 target_to_host_old_sigset(&set, p);
9427 unlock_user(p, arg2, 0);
9428 set_ptr = &set;
9429 switch (arg1) {
9430 case TARGET_SIG_BLOCK:
9431 how = SIG_BLOCK;
9432 break;
9433 case TARGET_SIG_UNBLOCK:
9434 how = SIG_UNBLOCK;
9435 break;
9436 case TARGET_SIG_SETMASK:
9437 how = SIG_SETMASK;
9438 break;
9439 default:
9440 return -TARGET_EINVAL;
9442 } else {
9443 how = 0;
9444 set_ptr = NULL;
9446 ret = do_sigprocmask(how, set_ptr, &oldset);
9447 if (!is_error(ret) && arg3) {
9448 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_sigset_t), 0)))
9449 return -TARGET_EFAULT;
9450 host_to_target_old_sigset(p, &oldset);
9451 unlock_user(p, arg3, sizeof(target_sigset_t));
9453 #endif
9455 return ret;
9456 #endif
9457 case TARGET_NR_rt_sigprocmask:
9459 int how = arg1;
9460 sigset_t set, oldset, *set_ptr;
9462 if (arg4 != sizeof(target_sigset_t)) {
9463 return -TARGET_EINVAL;
9466 if (arg2) {
9467 p = lock_user(VERIFY_READ, arg2, sizeof(target_sigset_t), 1);
9468 if (!p) {
9469 return -TARGET_EFAULT;
9471 target_to_host_sigset(&set, p);
9472 unlock_user(p, arg2, 0);
9473 set_ptr = &set;
9474 switch(how) {
9475 case TARGET_SIG_BLOCK:
9476 how = SIG_BLOCK;
9477 break;
9478 case TARGET_SIG_UNBLOCK:
9479 how = SIG_UNBLOCK;
9480 break;
9481 case TARGET_SIG_SETMASK:
9482 how = SIG_SETMASK;
9483 break;
9484 default:
9485 return -TARGET_EINVAL;
9487 } else {
9488 how = 0;
9489 set_ptr = NULL;
9491 ret = do_sigprocmask(how, set_ptr, &oldset);
9492 if (!is_error(ret) && arg3) {
9493 if (!(p = lock_user(VERIFY_WRITE, arg3, sizeof(target_sigset_t), 0)))
9494 return -TARGET_EFAULT;
9495 host_to_target_sigset(p, &oldset);
9496 unlock_user(p, arg3, sizeof(target_sigset_t));
9499 return ret;
9500 #ifdef TARGET_NR_sigpending
9501 case TARGET_NR_sigpending:
9503 sigset_t set;
9504 ret = get_errno(sigpending(&set));
9505 if (!is_error(ret)) {
9506 if (!(p = lock_user(VERIFY_WRITE, arg1, sizeof(target_sigset_t), 0)))
9507 return -TARGET_EFAULT;
9508 host_to_target_old_sigset(p, &set);
9509 unlock_user(p, arg1, sizeof(target_sigset_t));
9512 return ret;
9513 #endif
9514 case TARGET_NR_rt_sigpending:
9516 sigset_t set;
9518 /* Yes, this check is >, not != like most. We follow the kernel's
9519 * logic and it does it like this because it implements
9520 * NR_sigpending through the same code path, and in that case
9521 * the old_sigset_t is smaller in size.
9523 if (arg2 > sizeof(target_sigset_t)) {
9524 return -TARGET_EINVAL;
9527 ret = get_errno(sigpending(&set));
9528 if (!is_error(ret)) {
9529 if (!(p = lock_user(VERIFY_WRITE, arg1, sizeof(target_sigset_t), 0)))
9530 return -TARGET_EFAULT;
9531 host_to_target_sigset(p, &set);
9532 unlock_user(p, arg1, sizeof(target_sigset_t));
9535 return ret;
9536 #ifdef TARGET_NR_sigsuspend
9537 case TARGET_NR_sigsuspend:
9539 sigset_t *set;
9541 #if defined(TARGET_ALPHA)
9542 TaskState *ts = cpu->opaque;
9543 /* target_to_host_old_sigset will bswap back */
9544 abi_ulong mask = tswapal(arg1);
9545 set = &ts->sigsuspend_mask;
9546 target_to_host_old_sigset(set, &mask);
9547 #else
9548 ret = process_sigsuspend_mask(&set, arg1, sizeof(target_sigset_t));
9549 if (ret != 0) {
9550 return ret;
9552 #endif
9553 ret = get_errno(safe_rt_sigsuspend(set, SIGSET_T_SIZE));
9554 finish_sigsuspend_mask(ret);
9556 return ret;
9557 #endif
9558 case TARGET_NR_rt_sigsuspend:
9560 sigset_t *set;
9562 ret = process_sigsuspend_mask(&set, arg1, arg2);
9563 if (ret != 0) {
9564 return ret;
9566 ret = get_errno(safe_rt_sigsuspend(set, SIGSET_T_SIZE));
9567 finish_sigsuspend_mask(ret);
9569 return ret;
9570 #ifdef TARGET_NR_rt_sigtimedwait
9571 case TARGET_NR_rt_sigtimedwait:
9573 sigset_t set;
9574 struct timespec uts, *puts;
9575 siginfo_t uinfo;
9577 if (arg4 != sizeof(target_sigset_t)) {
9578 return -TARGET_EINVAL;
9581 if (!(p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1)))
9582 return -TARGET_EFAULT;
9583 target_to_host_sigset(&set, p);
9584 unlock_user(p, arg1, 0);
9585 if (arg3) {
9586 puts = &uts;
9587 if (target_to_host_timespec(puts, arg3)) {
9588 return -TARGET_EFAULT;
9590 } else {
9591 puts = NULL;
9593 ret = get_errno(safe_rt_sigtimedwait(&set, &uinfo, puts,
9594 SIGSET_T_SIZE));
9595 if (!is_error(ret)) {
9596 if (arg2) {
9597 p = lock_user(VERIFY_WRITE, arg2, sizeof(target_siginfo_t),
9599 if (!p) {
9600 return -TARGET_EFAULT;
9602 host_to_target_siginfo(p, &uinfo);
9603 unlock_user(p, arg2, sizeof(target_siginfo_t));
9605 ret = host_to_target_signal(ret);
9608 return ret;
9609 #endif
9610 #ifdef TARGET_NR_rt_sigtimedwait_time64
9611 case TARGET_NR_rt_sigtimedwait_time64:
9613 sigset_t set;
9614 struct timespec uts, *puts;
9615 siginfo_t uinfo;
9617 if (arg4 != sizeof(target_sigset_t)) {
9618 return -TARGET_EINVAL;
9621 p = lock_user(VERIFY_READ, arg1, sizeof(target_sigset_t), 1);
9622 if (!p) {
9623 return -TARGET_EFAULT;
9625 target_to_host_sigset(&set, p);
9626 unlock_user(p, arg1, 0);
9627 if (arg3) {
9628 puts = &uts;
9629 if (target_to_host_timespec64(puts, arg3)) {
9630 return -TARGET_EFAULT;
9632 } else {
9633 puts = NULL;
9635 ret = get_errno(safe_rt_sigtimedwait(&set, &uinfo, puts,
9636 SIGSET_T_SIZE));
9637 if (!is_error(ret)) {
9638 if (arg2) {
9639 p = lock_user(VERIFY_WRITE, arg2,
9640 sizeof(target_siginfo_t), 0);
9641 if (!p) {
9642 return -TARGET_EFAULT;
9644 host_to_target_siginfo(p, &uinfo);
9645 unlock_user(p, arg2, sizeof(target_siginfo_t));
9647 ret = host_to_target_signal(ret);
9650 return ret;
9651 #endif
9652 case TARGET_NR_rt_sigqueueinfo:
9654 siginfo_t uinfo;
9656 p = lock_user(VERIFY_READ, arg3, sizeof(target_siginfo_t), 1);
9657 if (!p) {
9658 return -TARGET_EFAULT;
9660 target_to_host_siginfo(&uinfo, p);
9661 unlock_user(p, arg3, 0);
9662 ret = get_errno(sys_rt_sigqueueinfo(arg1, arg2, &uinfo));
9664 return ret;
9665 case TARGET_NR_rt_tgsigqueueinfo:
9667 siginfo_t uinfo;
9669 p = lock_user(VERIFY_READ, arg4, sizeof(target_siginfo_t), 1);
9670 if (!p) {
9671 return -TARGET_EFAULT;
9673 target_to_host_siginfo(&uinfo, p);
9674 unlock_user(p, arg4, 0);
9675 ret = get_errno(sys_rt_tgsigqueueinfo(arg1, arg2, arg3, &uinfo));
9677 return ret;
9678 #ifdef TARGET_NR_sigreturn
9679 case TARGET_NR_sigreturn:
9680 if (block_signals()) {
9681 return -QEMU_ERESTARTSYS;
9683 return do_sigreturn(cpu_env);
9684 #endif
9685 case TARGET_NR_rt_sigreturn:
9686 if (block_signals()) {
9687 return -QEMU_ERESTARTSYS;
9689 return do_rt_sigreturn(cpu_env);
9690 case TARGET_NR_sethostname:
9691 if (!(p = lock_user_string(arg1)))
9692 return -TARGET_EFAULT;
9693 ret = get_errno(sethostname(p, arg2));
9694 unlock_user(p, arg1, 0);
9695 return ret;
9696 #ifdef TARGET_NR_setrlimit
9697 case TARGET_NR_setrlimit:
9699 int resource = target_to_host_resource(arg1);
9700 struct target_rlimit *target_rlim;
9701 struct rlimit rlim;
9702 if (!lock_user_struct(VERIFY_READ, target_rlim, arg2, 1))
9703 return -TARGET_EFAULT;
9704 rlim.rlim_cur = target_to_host_rlim(target_rlim->rlim_cur);
9705 rlim.rlim_max = target_to_host_rlim(target_rlim->rlim_max);
9706 unlock_user_struct(target_rlim, arg2, 0);
9708 * If we just passed through resource limit settings for memory then
9709 * they would also apply to QEMU's own allocations, and QEMU will
9710 * crash or hang or die if its allocations fail. Ideally we would
9711 * track the guest allocations in QEMU and apply the limits ourselves.
9712 * For now, just tell the guest the call succeeded but don't actually
9713 * limit anything.
9715 if (resource != RLIMIT_AS &&
9716 resource != RLIMIT_DATA &&
9717 resource != RLIMIT_STACK) {
9718 return get_errno(setrlimit(resource, &rlim));
9719 } else {
9720 return 0;
9723 #endif
9724 #ifdef TARGET_NR_getrlimit
9725 case TARGET_NR_getrlimit:
9727 int resource = target_to_host_resource(arg1);
9728 struct target_rlimit *target_rlim;
9729 struct rlimit rlim;
9731 ret = get_errno(getrlimit(resource, &rlim));
9732 if (!is_error(ret)) {
9733 if (!lock_user_struct(VERIFY_WRITE, target_rlim, arg2, 0))
9734 return -TARGET_EFAULT;
9735 target_rlim->rlim_cur = host_to_target_rlim(rlim.rlim_cur);
9736 target_rlim->rlim_max = host_to_target_rlim(rlim.rlim_max);
9737 unlock_user_struct(target_rlim, arg2, 1);
9740 return ret;
9741 #endif
9742 case TARGET_NR_getrusage:
9744 struct rusage rusage;
9745 ret = get_errno(getrusage(arg1, &rusage));
9746 if (!is_error(ret)) {
9747 ret = host_to_target_rusage(arg2, &rusage);
9750 return ret;
9751 #if defined(TARGET_NR_gettimeofday)
9752 case TARGET_NR_gettimeofday:
9754 struct timeval tv;
9755 struct timezone tz;
9757 ret = get_errno(gettimeofday(&tv, &tz));
9758 if (!is_error(ret)) {
9759 if (arg1 && copy_to_user_timeval(arg1, &tv)) {
9760 return -TARGET_EFAULT;
9762 if (arg2 && copy_to_user_timezone(arg2, &tz)) {
9763 return -TARGET_EFAULT;
9767 return ret;
9768 #endif
9769 #if defined(TARGET_NR_settimeofday)
9770 case TARGET_NR_settimeofday:
9772 struct timeval tv, *ptv = NULL;
9773 struct timezone tz, *ptz = NULL;
9775 if (arg1) {
9776 if (copy_from_user_timeval(&tv, arg1)) {
9777 return -TARGET_EFAULT;
9779 ptv = &tv;
9782 if (arg2) {
9783 if (copy_from_user_timezone(&tz, arg2)) {
9784 return -TARGET_EFAULT;
9786 ptz = &tz;
9789 return get_errno(settimeofday(ptv, ptz));
9791 #endif
9792 #if defined(TARGET_NR_select)
9793 case TARGET_NR_select:
9794 #if defined(TARGET_WANT_NI_OLD_SELECT)
9795 /* some architectures used to have old_select here
9796 * but now ENOSYS it.
9798 ret = -TARGET_ENOSYS;
9799 #elif defined(TARGET_WANT_OLD_SYS_SELECT)
9800 ret = do_old_select(arg1);
9801 #else
9802 ret = do_select(arg1, arg2, arg3, arg4, arg5);
9803 #endif
9804 return ret;
9805 #endif
9806 #ifdef TARGET_NR_pselect6
9807 case TARGET_NR_pselect6:
9808 return do_pselect6(arg1, arg2, arg3, arg4, arg5, arg6, false);
9809 #endif
9810 #ifdef TARGET_NR_pselect6_time64
9811 case TARGET_NR_pselect6_time64:
9812 return do_pselect6(arg1, arg2, arg3, arg4, arg5, arg6, true);
9813 #endif
9814 #ifdef TARGET_NR_symlink
9815 case TARGET_NR_symlink:
9817 void *p2;
9818 p = lock_user_string(arg1);
9819 p2 = lock_user_string(arg2);
9820 if (!p || !p2)
9821 ret = -TARGET_EFAULT;
9822 else
9823 ret = get_errno(symlink(p, p2));
9824 unlock_user(p2, arg2, 0);
9825 unlock_user(p, arg1, 0);
9827 return ret;
9828 #endif
9829 #if defined(TARGET_NR_symlinkat)
9830 case TARGET_NR_symlinkat:
9832 void *p2;
9833 p = lock_user_string(arg1);
9834 p2 = lock_user_string(arg3);
9835 if (!p || !p2)
9836 ret = -TARGET_EFAULT;
9837 else
9838 ret = get_errno(symlinkat(p, arg2, p2));
9839 unlock_user(p2, arg3, 0);
9840 unlock_user(p, arg1, 0);
9842 return ret;
9843 #endif
9844 #ifdef TARGET_NR_readlink
9845 case TARGET_NR_readlink:
9847 void *p2;
9848 p = lock_user_string(arg1);
9849 p2 = lock_user(VERIFY_WRITE, arg2, arg3, 0);
9850 if (!p || !p2) {
9851 ret = -TARGET_EFAULT;
9852 } else if (!arg3) {
9853 /* Short circuit this for the magic exe check. */
9854 ret = -TARGET_EINVAL;
9855 } else if (is_proc_myself((const char *)p, "exe")) {
9856 char real[PATH_MAX], *temp;
9857 temp = realpath(exec_path, real);
9858 /* Return value is # of bytes that we wrote to the buffer. */
9859 if (temp == NULL) {
9860 ret = get_errno(-1);
9861 } else {
9862 /* Don't worry about sign mismatch as earlier mapping
9863 * logic would have thrown a bad address error. */
9864 ret = MIN(strlen(real), arg3);
9865 /* We cannot NUL terminate the string. */
9866 memcpy(p2, real, ret);
9868 } else {
9869 ret = get_errno(readlink(path(p), p2, arg3));
9871 unlock_user(p2, arg2, ret);
9872 unlock_user(p, arg1, 0);
9874 return ret;
9875 #endif
9876 #if defined(TARGET_NR_readlinkat)
9877 case TARGET_NR_readlinkat:
9879 void *p2;
9880 p = lock_user_string(arg2);
9881 p2 = lock_user(VERIFY_WRITE, arg3, arg4, 0);
9882 if (!p || !p2) {
9883 ret = -TARGET_EFAULT;
9884 } else if (is_proc_myself((const char *)p, "exe")) {
9885 char real[PATH_MAX], *temp;
9886 temp = realpath(exec_path, real);
9887 ret = temp == NULL ? get_errno(-1) : strlen(real) ;
9888 snprintf((char *)p2, arg4, "%s", real);
9889 } else {
9890 ret = get_errno(readlinkat(arg1, path(p), p2, arg4));
9892 unlock_user(p2, arg3, ret);
9893 unlock_user(p, arg2, 0);
9895 return ret;
9896 #endif
9897 #ifdef TARGET_NR_swapon
9898 case TARGET_NR_swapon:
9899 if (!(p = lock_user_string(arg1)))
9900 return -TARGET_EFAULT;
9901 ret = get_errno(swapon(p, arg2));
9902 unlock_user(p, arg1, 0);
9903 return ret;
9904 #endif
9905 case TARGET_NR_reboot:
9906 if (arg3 == LINUX_REBOOT_CMD_RESTART2) {
9907 /* arg4 must be ignored in all other cases */
9908 p = lock_user_string(arg4);
9909 if (!p) {
9910 return -TARGET_EFAULT;
9912 ret = get_errno(reboot(arg1, arg2, arg3, p));
9913 unlock_user(p, arg4, 0);
9914 } else {
9915 ret = get_errno(reboot(arg1, arg2, arg3, NULL));
9917 return ret;
9918 #ifdef TARGET_NR_mmap
9919 case TARGET_NR_mmap:
9920 #if (defined(TARGET_I386) && defined(TARGET_ABI32)) || \
9921 (defined(TARGET_ARM) && defined(TARGET_ABI32)) || \
9922 defined(TARGET_M68K) || defined(TARGET_CRIS) || defined(TARGET_MICROBLAZE) \
9923 || defined(TARGET_S390X)
9925 abi_ulong *v;
9926 abi_ulong v1, v2, v3, v4, v5, v6;
9927 if (!(v = lock_user(VERIFY_READ, arg1, 6 * sizeof(abi_ulong), 1)))
9928 return -TARGET_EFAULT;
9929 v1 = tswapal(v[0]);
9930 v2 = tswapal(v[1]);
9931 v3 = tswapal(v[2]);
9932 v4 = tswapal(v[3]);
9933 v5 = tswapal(v[4]);
9934 v6 = tswapal(v[5]);
9935 unlock_user(v, arg1, 0);
9936 ret = get_errno(target_mmap(v1, v2, v3,
9937 target_to_host_bitmask(v4, mmap_flags_tbl),
9938 v5, v6));
9940 #else
9941 /* mmap pointers are always untagged */
9942 ret = get_errno(target_mmap(arg1, arg2, arg3,
9943 target_to_host_bitmask(arg4, mmap_flags_tbl),
9944 arg5,
9945 arg6));
9946 #endif
9947 return ret;
9948 #endif
9949 #ifdef TARGET_NR_mmap2
9950 case TARGET_NR_mmap2:
9951 #ifndef MMAP_SHIFT
9952 #define MMAP_SHIFT 12
9953 #endif
9954 ret = target_mmap(arg1, arg2, arg3,
9955 target_to_host_bitmask(arg4, mmap_flags_tbl),
9956 arg5, arg6 << MMAP_SHIFT);
9957 return get_errno(ret);
9958 #endif
9959 case TARGET_NR_munmap:
9960 arg1 = cpu_untagged_addr(cpu, arg1);
9961 return get_errno(target_munmap(arg1, arg2));
9962 case TARGET_NR_mprotect:
9963 arg1 = cpu_untagged_addr(cpu, arg1);
9965 TaskState *ts = cpu->opaque;
9966 /* Special hack to detect libc making the stack executable. */
9967 if ((arg3 & PROT_GROWSDOWN)
9968 && arg1 >= ts->info->stack_limit
9969 && arg1 <= ts->info->start_stack) {
9970 arg3 &= ~PROT_GROWSDOWN;
9971 arg2 = arg2 + arg1 - ts->info->stack_limit;
9972 arg1 = ts->info->stack_limit;
9975 return get_errno(target_mprotect(arg1, arg2, arg3));
9976 #ifdef TARGET_NR_mremap
9977 case TARGET_NR_mremap:
9978 arg1 = cpu_untagged_addr(cpu, arg1);
9979 /* mremap new_addr (arg5) is always untagged */
9980 return get_errno(target_mremap(arg1, arg2, arg3, arg4, arg5));
9981 #endif
9982 /* ??? msync/mlock/munlock are broken for softmmu. */
9983 #ifdef TARGET_NR_msync
9984 case TARGET_NR_msync:
9985 return get_errno(msync(g2h(cpu, arg1), arg2, arg3));
9986 #endif
9987 #ifdef TARGET_NR_mlock
9988 case TARGET_NR_mlock:
9989 return get_errno(mlock(g2h(cpu, arg1), arg2));
9990 #endif
9991 #ifdef TARGET_NR_munlock
9992 case TARGET_NR_munlock:
9993 return get_errno(munlock(g2h(cpu, arg1), arg2));
9994 #endif
9995 #ifdef TARGET_NR_mlockall
9996 case TARGET_NR_mlockall:
9997 return get_errno(mlockall(target_to_host_mlockall_arg(arg1)));
9998 #endif
9999 #ifdef TARGET_NR_munlockall
10000 case TARGET_NR_munlockall:
10001 return get_errno(munlockall());
10002 #endif
10003 #ifdef TARGET_NR_truncate
10004 case TARGET_NR_truncate:
10005 if (!(p = lock_user_string(arg1)))
10006 return -TARGET_EFAULT;
10007 ret = get_errno(truncate(p, arg2));
10008 unlock_user(p, arg1, 0);
10009 return ret;
10010 #endif
10011 #ifdef TARGET_NR_ftruncate
10012 case TARGET_NR_ftruncate:
10013 return get_errno(ftruncate(arg1, arg2));
10014 #endif
10015 case TARGET_NR_fchmod:
10016 return get_errno(fchmod(arg1, arg2));
10017 #if defined(TARGET_NR_fchmodat)
10018 case TARGET_NR_fchmodat:
10019 if (!(p = lock_user_string(arg2)))
10020 return -TARGET_EFAULT;
10021 ret = get_errno(fchmodat(arg1, p, arg3, 0));
10022 unlock_user(p, arg2, 0);
10023 return ret;
10024 #endif
10025 case TARGET_NR_getpriority:
10026 /* Note that negative values are valid for getpriority, so we must
10027 differentiate based on errno settings. */
10028 errno = 0;
10029 ret = getpriority(arg1, arg2);
10030 if (ret == -1 && errno != 0) {
10031 return -host_to_target_errno(errno);
10033 #ifdef TARGET_ALPHA
10034 /* Return value is the unbiased priority. Signal no error. */
10035 cpu_env->ir[IR_V0] = 0;
10036 #else
10037 /* Return value is a biased priority to avoid negative numbers. */
10038 ret = 20 - ret;
10039 #endif
10040 return ret;
10041 case TARGET_NR_setpriority:
10042 return get_errno(setpriority(arg1, arg2, arg3));
10043 #ifdef TARGET_NR_statfs
10044 case TARGET_NR_statfs:
10045 if (!(p = lock_user_string(arg1))) {
10046 return -TARGET_EFAULT;
10048 ret = get_errno(statfs(path(p), &stfs));
10049 unlock_user(p, arg1, 0);
10050 convert_statfs:
10051 if (!is_error(ret)) {
10052 struct target_statfs *target_stfs;
10054 if (!lock_user_struct(VERIFY_WRITE, target_stfs, arg2, 0))
10055 return -TARGET_EFAULT;
10056 __put_user(stfs.f_type, &target_stfs->f_type);
10057 __put_user(stfs.f_bsize, &target_stfs->f_bsize);
10058 __put_user(stfs.f_blocks, &target_stfs->f_blocks);
10059 __put_user(stfs.f_bfree, &target_stfs->f_bfree);
10060 __put_user(stfs.f_bavail, &target_stfs->f_bavail);
10061 __put_user(stfs.f_files, &target_stfs->f_files);
10062 __put_user(stfs.f_ffree, &target_stfs->f_ffree);
10063 __put_user(stfs.f_fsid.__val[0], &target_stfs->f_fsid.val[0]);
10064 __put_user(stfs.f_fsid.__val[1], &target_stfs->f_fsid.val[1]);
10065 __put_user(stfs.f_namelen, &target_stfs->f_namelen);
10066 __put_user(stfs.f_frsize, &target_stfs->f_frsize);
10067 #ifdef _STATFS_F_FLAGS
10068 __put_user(stfs.f_flags, &target_stfs->f_flags);
10069 #else
10070 __put_user(0, &target_stfs->f_flags);
10071 #endif
10072 memset(target_stfs->f_spare, 0, sizeof(target_stfs->f_spare));
10073 unlock_user_struct(target_stfs, arg2, 1);
10075 return ret;
10076 #endif
10077 #ifdef TARGET_NR_fstatfs
10078 case TARGET_NR_fstatfs:
10079 ret = get_errno(fstatfs(arg1, &stfs));
10080 goto convert_statfs;
10081 #endif
10082 #ifdef TARGET_NR_statfs64
10083 case TARGET_NR_statfs64:
10084 if (!(p = lock_user_string(arg1))) {
10085 return -TARGET_EFAULT;
10087 ret = get_errno(statfs(path(p), &stfs));
10088 unlock_user(p, arg1, 0);
10089 convert_statfs64:
10090 if (!is_error(ret)) {
10091 struct target_statfs64 *target_stfs;
10093 if (!lock_user_struct(VERIFY_WRITE, target_stfs, arg3, 0))
10094 return -TARGET_EFAULT;
10095 __put_user(stfs.f_type, &target_stfs->f_type);
10096 __put_user(stfs.f_bsize, &target_stfs->f_bsize);
10097 __put_user(stfs.f_blocks, &target_stfs->f_blocks);
10098 __put_user(stfs.f_bfree, &target_stfs->f_bfree);
10099 __put_user(stfs.f_bavail, &target_stfs->f_bavail);
10100 __put_user(stfs.f_files, &target_stfs->f_files);
10101 __put_user(stfs.f_ffree, &target_stfs->f_ffree);
10102 __put_user(stfs.f_fsid.__val[0], &target_stfs->f_fsid.val[0]);
10103 __put_user(stfs.f_fsid.__val[1], &target_stfs->f_fsid.val[1]);
10104 __put_user(stfs.f_namelen, &target_stfs->f_namelen);
10105 __put_user(stfs.f_frsize, &target_stfs->f_frsize);
10106 #ifdef _STATFS_F_FLAGS
10107 __put_user(stfs.f_flags, &target_stfs->f_flags);
10108 #else
10109 __put_user(0, &target_stfs->f_flags);
10110 #endif
10111 memset(target_stfs->f_spare, 0, sizeof(target_stfs->f_spare));
10112 unlock_user_struct(target_stfs, arg3, 1);
10114 return ret;
10115 case TARGET_NR_fstatfs64:
10116 ret = get_errno(fstatfs(arg1, &stfs));
10117 goto convert_statfs64;
10118 #endif
10119 #ifdef TARGET_NR_socketcall
10120 case TARGET_NR_socketcall:
10121 return do_socketcall(arg1, arg2);
10122 #endif
10123 #ifdef TARGET_NR_accept
10124 case TARGET_NR_accept:
10125 return do_accept4(arg1, arg2, arg3, 0);
10126 #endif
10127 #ifdef TARGET_NR_accept4
10128 case TARGET_NR_accept4:
10129 return do_accept4(arg1, arg2, arg3, arg4);
10130 #endif
10131 #ifdef TARGET_NR_bind
10132 case TARGET_NR_bind:
10133 return do_bind(arg1, arg2, arg3);
10134 #endif
10135 #ifdef TARGET_NR_connect
10136 case TARGET_NR_connect:
10137 return do_connect(arg1, arg2, arg3);
10138 #endif
10139 #ifdef TARGET_NR_getpeername
10140 case TARGET_NR_getpeername:
10141 return do_getpeername(arg1, arg2, arg3);
10142 #endif
10143 #ifdef TARGET_NR_getsockname
10144 case TARGET_NR_getsockname:
10145 return do_getsockname(arg1, arg2, arg3);
10146 #endif
10147 #ifdef TARGET_NR_getsockopt
10148 case TARGET_NR_getsockopt:
10149 return do_getsockopt(arg1, arg2, arg3, arg4, arg5);
10150 #endif
10151 #ifdef TARGET_NR_listen
10152 case TARGET_NR_listen:
10153 return get_errno(listen(arg1, arg2));
10154 #endif
10155 #ifdef TARGET_NR_recv
10156 case TARGET_NR_recv:
10157 return do_recvfrom(arg1, arg2, arg3, arg4, 0, 0);
10158 #endif
10159 #ifdef TARGET_NR_recvfrom
10160 case TARGET_NR_recvfrom:
10161 return do_recvfrom(arg1, arg2, arg3, arg4, arg5, arg6);
10162 #endif
10163 #ifdef TARGET_NR_recvmsg
10164 case TARGET_NR_recvmsg:
10165 return do_sendrecvmsg(arg1, arg2, arg3, 0);
10166 #endif
10167 #ifdef TARGET_NR_send
10168 case TARGET_NR_send:
10169 return do_sendto(arg1, arg2, arg3, arg4, 0, 0);
10170 #endif
10171 #ifdef TARGET_NR_sendmsg
10172 case TARGET_NR_sendmsg:
10173 return do_sendrecvmsg(arg1, arg2, arg3, 1);
10174 #endif
10175 #ifdef TARGET_NR_sendmmsg
10176 case TARGET_NR_sendmmsg:
10177 return do_sendrecvmmsg(arg1, arg2, arg3, arg4, 1);
10178 #endif
10179 #ifdef TARGET_NR_recvmmsg
10180 case TARGET_NR_recvmmsg:
10181 return do_sendrecvmmsg(arg1, arg2, arg3, arg4, 0);
10182 #endif
10183 #ifdef TARGET_NR_sendto
10184 case TARGET_NR_sendto:
10185 return do_sendto(arg1, arg2, arg3, arg4, arg5, arg6);
10186 #endif
10187 #ifdef TARGET_NR_shutdown
10188 case TARGET_NR_shutdown:
10189 return get_errno(shutdown(arg1, arg2));
10190 #endif
10191 #if defined(TARGET_NR_getrandom) && defined(__NR_getrandom)
10192 case TARGET_NR_getrandom:
10193 p = lock_user(VERIFY_WRITE, arg1, arg2, 0);
10194 if (!p) {
10195 return -TARGET_EFAULT;
10197 ret = get_errno(getrandom(p, arg2, arg3));
10198 unlock_user(p, arg1, ret);
10199 return ret;
10200 #endif
10201 #ifdef TARGET_NR_socket
10202 case TARGET_NR_socket:
10203 return do_socket(arg1, arg2, arg3);
10204 #endif
10205 #ifdef TARGET_NR_socketpair
10206 case TARGET_NR_socketpair:
10207 return do_socketpair(arg1, arg2, arg3, arg4);
10208 #endif
10209 #ifdef TARGET_NR_setsockopt
10210 case TARGET_NR_setsockopt:
10211 return do_setsockopt(arg1, arg2, arg3, arg4, (socklen_t) arg5);
10212 #endif
10213 #if defined(TARGET_NR_syslog)
10214 case TARGET_NR_syslog:
10216 int len = arg2;
10218 switch (arg1) {
10219 case TARGET_SYSLOG_ACTION_CLOSE: /* Close log */
10220 case TARGET_SYSLOG_ACTION_OPEN: /* Open log */
10221 case TARGET_SYSLOG_ACTION_CLEAR: /* Clear ring buffer */
10222 case TARGET_SYSLOG_ACTION_CONSOLE_OFF: /* Disable logging */
10223 case TARGET_SYSLOG_ACTION_CONSOLE_ON: /* Enable logging */
10224 case TARGET_SYSLOG_ACTION_CONSOLE_LEVEL: /* Set messages level */
10225 case TARGET_SYSLOG_ACTION_SIZE_UNREAD: /* Number of chars */
10226 case TARGET_SYSLOG_ACTION_SIZE_BUFFER: /* Size of the buffer */
10227 return get_errno(sys_syslog((int)arg1, NULL, (int)arg3));
10228 case TARGET_SYSLOG_ACTION_READ: /* Read from log */
10229 case TARGET_SYSLOG_ACTION_READ_CLEAR: /* Read/clear msgs */
10230 case TARGET_SYSLOG_ACTION_READ_ALL: /* Read last messages */
10232 if (len < 0) {
10233 return -TARGET_EINVAL;
10235 if (len == 0) {
10236 return 0;
10238 p = lock_user(VERIFY_WRITE, arg2, arg3, 0);
10239 if (!p) {
10240 return -TARGET_EFAULT;
10242 ret = get_errno(sys_syslog((int)arg1, p, (int)arg3));
10243 unlock_user(p, arg2, arg3);
10245 return ret;
10246 default:
10247 return -TARGET_EINVAL;
10250 break;
10251 #endif
10252 case TARGET_NR_setitimer:
10254 struct itimerval value, ovalue, *pvalue;
10256 if (arg2) {
10257 pvalue = &value;
10258 if (copy_from_user_timeval(&pvalue->it_interval, arg2)
10259 || copy_from_user_timeval(&pvalue->it_value,
10260 arg2 + sizeof(struct target_timeval)))
10261 return -TARGET_EFAULT;
10262 } else {
10263 pvalue = NULL;
10265 ret = get_errno(setitimer(arg1, pvalue, &ovalue));
10266 if (!is_error(ret) && arg3) {
10267 if (copy_to_user_timeval(arg3,
10268 &ovalue.it_interval)
10269 || copy_to_user_timeval(arg3 + sizeof(struct target_timeval),
10270 &ovalue.it_value))
10271 return -TARGET_EFAULT;
10274 return ret;
10275 case TARGET_NR_getitimer:
10277 struct itimerval value;
10279 ret = get_errno(getitimer(arg1, &value));
10280 if (!is_error(ret) && arg2) {
10281 if (copy_to_user_timeval(arg2,
10282 &value.it_interval)
10283 || copy_to_user_timeval(arg2 + sizeof(struct target_timeval),
10284 &value.it_value))
10285 return -TARGET_EFAULT;
10288 return ret;
10289 #ifdef TARGET_NR_stat
10290 case TARGET_NR_stat:
10291 if (!(p = lock_user_string(arg1))) {
10292 return -TARGET_EFAULT;
10294 ret = get_errno(stat(path(p), &st));
10295 unlock_user(p, arg1, 0);
10296 goto do_stat;
10297 #endif
10298 #ifdef TARGET_NR_lstat
10299 case TARGET_NR_lstat:
10300 if (!(p = lock_user_string(arg1))) {
10301 return -TARGET_EFAULT;
10303 ret = get_errno(lstat(path(p), &st));
10304 unlock_user(p, arg1, 0);
10305 goto do_stat;
10306 #endif
10307 #ifdef TARGET_NR_fstat
10308 case TARGET_NR_fstat:
10310 ret = get_errno(fstat(arg1, &st));
10311 #if defined(TARGET_NR_stat) || defined(TARGET_NR_lstat)
10312 do_stat:
10313 #endif
10314 if (!is_error(ret)) {
10315 struct target_stat *target_st;
10317 if (!lock_user_struct(VERIFY_WRITE, target_st, arg2, 0))
10318 return -TARGET_EFAULT;
10319 memset(target_st, 0, sizeof(*target_st));
10320 __put_user(st.st_dev, &target_st->st_dev);
10321 __put_user(st.st_ino, &target_st->st_ino);
10322 __put_user(st.st_mode, &target_st->st_mode);
10323 __put_user(st.st_uid, &target_st->st_uid);
10324 __put_user(st.st_gid, &target_st->st_gid);
10325 __put_user(st.st_nlink, &target_st->st_nlink);
10326 __put_user(st.st_rdev, &target_st->st_rdev);
10327 __put_user(st.st_size, &target_st->st_size);
10328 __put_user(st.st_blksize, &target_st->st_blksize);
10329 __put_user(st.st_blocks, &target_st->st_blocks);
10330 __put_user(st.st_atime, &target_st->target_st_atime);
10331 __put_user(st.st_mtime, &target_st->target_st_mtime);
10332 __put_user(st.st_ctime, &target_st->target_st_ctime);
10333 #if defined(HAVE_STRUCT_STAT_ST_ATIM) && defined(TARGET_STAT_HAVE_NSEC)
10334 __put_user(st.st_atim.tv_nsec,
10335 &target_st->target_st_atime_nsec);
10336 __put_user(st.st_mtim.tv_nsec,
10337 &target_st->target_st_mtime_nsec);
10338 __put_user(st.st_ctim.tv_nsec,
10339 &target_st->target_st_ctime_nsec);
10340 #endif
10341 unlock_user_struct(target_st, arg2, 1);
10344 return ret;
10345 #endif
10346 case TARGET_NR_vhangup:
10347 return get_errno(vhangup());
10348 #ifdef TARGET_NR_syscall
10349 case TARGET_NR_syscall:
10350 return do_syscall(cpu_env, arg1 & 0xffff, arg2, arg3, arg4, arg5,
10351 arg6, arg7, arg8, 0);
10352 #endif
10353 #if defined(TARGET_NR_wait4)
10354 case TARGET_NR_wait4:
10356 int status;
10357 abi_long status_ptr = arg2;
10358 struct rusage rusage, *rusage_ptr;
10359 abi_ulong target_rusage = arg4;
10360 abi_long rusage_err;
10361 if (target_rusage)
10362 rusage_ptr = &rusage;
10363 else
10364 rusage_ptr = NULL;
10365 ret = get_errno(safe_wait4(arg1, &status, arg3, rusage_ptr));
10366 if (!is_error(ret)) {
10367 if (status_ptr && ret) {
10368 status = host_to_target_waitstatus(status);
10369 if (put_user_s32(status, status_ptr))
10370 return -TARGET_EFAULT;
10372 if (target_rusage) {
10373 rusage_err = host_to_target_rusage(target_rusage, &rusage);
10374 if (rusage_err) {
10375 ret = rusage_err;
10380 return ret;
10381 #endif
10382 #ifdef TARGET_NR_swapoff
10383 case TARGET_NR_swapoff:
10384 if (!(p = lock_user_string(arg1)))
10385 return -TARGET_EFAULT;
10386 ret = get_errno(swapoff(p));
10387 unlock_user(p, arg1, 0);
10388 return ret;
10389 #endif
10390 case TARGET_NR_sysinfo:
10392 struct target_sysinfo *target_value;
10393 struct sysinfo value;
10394 ret = get_errno(sysinfo(&value));
10395 if (!is_error(ret) && arg1)
10397 if (!lock_user_struct(VERIFY_WRITE, target_value, arg1, 0))
10398 return -TARGET_EFAULT;
10399 __put_user(value.uptime, &target_value->uptime);
10400 __put_user(value.loads[0], &target_value->loads[0]);
10401 __put_user(value.loads[1], &target_value->loads[1]);
10402 __put_user(value.loads[2], &target_value->loads[2]);
10403 __put_user(value.totalram, &target_value->totalram);
10404 __put_user(value.freeram, &target_value->freeram);
10405 __put_user(value.sharedram, &target_value->sharedram);
10406 __put_user(value.bufferram, &target_value->bufferram);
10407 __put_user(value.totalswap, &target_value->totalswap);
10408 __put_user(value.freeswap, &target_value->freeswap);
10409 __put_user(value.procs, &target_value->procs);
10410 __put_user(value.totalhigh, &target_value->totalhigh);
10411 __put_user(value.freehigh, &target_value->freehigh);
10412 __put_user(value.mem_unit, &target_value->mem_unit);
10413 unlock_user_struct(target_value, arg1, 1);
10416 return ret;
10417 #ifdef TARGET_NR_ipc
10418 case TARGET_NR_ipc:
10419 return do_ipc(cpu_env, arg1, arg2, arg3, arg4, arg5, arg6);
10420 #endif
10421 #ifdef TARGET_NR_semget
10422 case TARGET_NR_semget:
10423 return get_errno(semget(arg1, arg2, arg3));
10424 #endif
10425 #ifdef TARGET_NR_semop
10426 case TARGET_NR_semop:
10427 return do_semtimedop(arg1, arg2, arg3, 0, false);
10428 #endif
10429 #ifdef TARGET_NR_semtimedop
10430 case TARGET_NR_semtimedop:
10431 return do_semtimedop(arg1, arg2, arg3, arg4, false);
10432 #endif
10433 #ifdef TARGET_NR_semtimedop_time64
10434 case TARGET_NR_semtimedop_time64:
10435 return do_semtimedop(arg1, arg2, arg3, arg4, true);
10436 #endif
10437 #ifdef TARGET_NR_semctl
10438 case TARGET_NR_semctl:
10439 return do_semctl(arg1, arg2, arg3, arg4);
10440 #endif
10441 #ifdef TARGET_NR_msgctl
10442 case TARGET_NR_msgctl:
10443 return do_msgctl(arg1, arg2, arg3);
10444 #endif
10445 #ifdef TARGET_NR_msgget
10446 case TARGET_NR_msgget:
10447 return get_errno(msgget(arg1, arg2));
10448 #endif
10449 #ifdef TARGET_NR_msgrcv
10450 case TARGET_NR_msgrcv:
10451 return do_msgrcv(arg1, arg2, arg3, arg4, arg5);
10452 #endif
10453 #ifdef TARGET_NR_msgsnd
10454 case TARGET_NR_msgsnd:
10455 return do_msgsnd(arg1, arg2, arg3, arg4);
10456 #endif
10457 #ifdef TARGET_NR_shmget
10458 case TARGET_NR_shmget:
10459 return get_errno(shmget(arg1, arg2, arg3));
10460 #endif
10461 #ifdef TARGET_NR_shmctl
10462 case TARGET_NR_shmctl:
10463 return do_shmctl(arg1, arg2, arg3);
10464 #endif
10465 #ifdef TARGET_NR_shmat
10466 case TARGET_NR_shmat:
10467 return do_shmat(cpu_env, arg1, arg2, arg3);
10468 #endif
10469 #ifdef TARGET_NR_shmdt
10470 case TARGET_NR_shmdt:
10471 return do_shmdt(arg1);
10472 #endif
10473 case TARGET_NR_fsync:
10474 return get_errno(fsync(arg1));
10475 case TARGET_NR_clone:
10476 /* Linux manages to have three different orderings for its
10477 * arguments to clone(); the BACKWARDS and BACKWARDS2 defines
10478 * match the kernel's CONFIG_CLONE_* settings.
10479 * Microblaze is further special in that it uses a sixth
10480 * implicit argument to clone for the TLS pointer.
10482 #if defined(TARGET_MICROBLAZE)
10483 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg4, arg6, arg5));
10484 #elif defined(TARGET_CLONE_BACKWARDS)
10485 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg3, arg4, arg5));
10486 #elif defined(TARGET_CLONE_BACKWARDS2)
10487 ret = get_errno(do_fork(cpu_env, arg2, arg1, arg3, arg5, arg4));
10488 #else
10489 ret = get_errno(do_fork(cpu_env, arg1, arg2, arg3, arg5, arg4));
10490 #endif
10491 return ret;
10492 #ifdef __NR_exit_group
10493 /* new thread calls */
10494 case TARGET_NR_exit_group:
10495 preexit_cleanup(cpu_env, arg1);
10496 return get_errno(exit_group(arg1));
10497 #endif
10498 case TARGET_NR_setdomainname:
10499 if (!(p = lock_user_string(arg1)))
10500 return -TARGET_EFAULT;
10501 ret = get_errno(setdomainname(p, arg2));
10502 unlock_user(p, arg1, 0);
10503 return ret;
10504 case TARGET_NR_uname:
10505 /* no need to transcode because we use the linux syscall */
10507 struct new_utsname * buf;
10509 if (!lock_user_struct(VERIFY_WRITE, buf, arg1, 0))
10510 return -TARGET_EFAULT;
10511 ret = get_errno(sys_uname(buf));
10512 if (!is_error(ret)) {
10513 /* Overwrite the native machine name with whatever is being
10514 emulated. */
10515 g_strlcpy(buf->machine, cpu_to_uname_machine(cpu_env),
10516 sizeof(buf->machine));
10517 /* Allow the user to override the reported release. */
10518 if (qemu_uname_release && *qemu_uname_release) {
10519 g_strlcpy(buf->release, qemu_uname_release,
10520 sizeof(buf->release));
10523 unlock_user_struct(buf, arg1, 1);
10525 return ret;
10526 #ifdef TARGET_I386
10527 case TARGET_NR_modify_ldt:
10528 return do_modify_ldt(cpu_env, arg1, arg2, arg3);
10529 #if !defined(TARGET_X86_64)
10530 case TARGET_NR_vm86:
10531 return do_vm86(cpu_env, arg1, arg2);
10532 #endif
10533 #endif
10534 #if defined(TARGET_NR_adjtimex)
10535 case TARGET_NR_adjtimex:
10537 struct timex host_buf;
10539 if (target_to_host_timex(&host_buf, arg1) != 0) {
10540 return -TARGET_EFAULT;
10542 ret = get_errno(adjtimex(&host_buf));
10543 if (!is_error(ret)) {
10544 if (host_to_target_timex(arg1, &host_buf) != 0) {
10545 return -TARGET_EFAULT;
10549 return ret;
10550 #endif
10551 #if defined(TARGET_NR_clock_adjtime) && defined(CONFIG_CLOCK_ADJTIME)
10552 case TARGET_NR_clock_adjtime:
10554 struct timex htx, *phtx = &htx;
10556 if (target_to_host_timex(phtx, arg2) != 0) {
10557 return -TARGET_EFAULT;
10559 ret = get_errno(clock_adjtime(arg1, phtx));
10560 if (!is_error(ret) && phtx) {
10561 if (host_to_target_timex(arg2, phtx) != 0) {
10562 return -TARGET_EFAULT;
10566 return ret;
10567 #endif
10568 #if defined(TARGET_NR_clock_adjtime64) && defined(CONFIG_CLOCK_ADJTIME)
10569 case TARGET_NR_clock_adjtime64:
10571 struct timex htx;
10573 if (target_to_host_timex64(&htx, arg2) != 0) {
10574 return -TARGET_EFAULT;
10576 ret = get_errno(clock_adjtime(arg1, &htx));
10577 if (!is_error(ret) && host_to_target_timex64(arg2, &htx)) {
10578 return -TARGET_EFAULT;
10581 return ret;
10582 #endif
10583 case TARGET_NR_getpgid:
10584 return get_errno(getpgid(arg1));
10585 case TARGET_NR_fchdir:
10586 return get_errno(fchdir(arg1));
10587 case TARGET_NR_personality:
10588 return get_errno(personality(arg1));
10589 #ifdef TARGET_NR__llseek /* Not on alpha */
10590 case TARGET_NR__llseek:
10592 int64_t res;
10593 #if !defined(__NR_llseek)
10594 res = lseek(arg1, ((uint64_t)arg2 << 32) | (abi_ulong)arg3, arg5);
10595 if (res == -1) {
10596 ret = get_errno(res);
10597 } else {
10598 ret = 0;
10600 #else
10601 ret = get_errno(_llseek(arg1, arg2, arg3, &res, arg5));
10602 #endif
10603 if ((ret == 0) && put_user_s64(res, arg4)) {
10604 return -TARGET_EFAULT;
10607 return ret;
10608 #endif
10609 #ifdef TARGET_NR_getdents
10610 case TARGET_NR_getdents:
10611 return do_getdents(arg1, arg2, arg3);
10612 #endif /* TARGET_NR_getdents */
10613 #if defined(TARGET_NR_getdents64) && defined(__NR_getdents64)
10614 case TARGET_NR_getdents64:
10615 return do_getdents64(arg1, arg2, arg3);
10616 #endif /* TARGET_NR_getdents64 */
10617 #if defined(TARGET_NR__newselect)
10618 case TARGET_NR__newselect:
10619 return do_select(arg1, arg2, arg3, arg4, arg5);
10620 #endif
10621 #ifdef TARGET_NR_poll
10622 case TARGET_NR_poll:
10623 return do_ppoll(arg1, arg2, arg3, arg4, arg5, false, false);
10624 #endif
10625 #ifdef TARGET_NR_ppoll
10626 case TARGET_NR_ppoll:
10627 return do_ppoll(arg1, arg2, arg3, arg4, arg5, true, false);
10628 #endif
10629 #ifdef TARGET_NR_ppoll_time64
10630 case TARGET_NR_ppoll_time64:
10631 return do_ppoll(arg1, arg2, arg3, arg4, arg5, true, true);
10632 #endif
10633 case TARGET_NR_flock:
10634 /* NOTE: the flock constant seems to be the same for every
10635 Linux platform */
10636 return get_errno(safe_flock(arg1, arg2));
10637 case TARGET_NR_readv:
10639 struct iovec *vec = lock_iovec(VERIFY_WRITE, arg2, arg3, 0);
10640 if (vec != NULL) {
10641 ret = get_errno(safe_readv(arg1, vec, arg3));
10642 unlock_iovec(vec, arg2, arg3, 1);
10643 } else {
10644 ret = -host_to_target_errno(errno);
10647 return ret;
10648 case TARGET_NR_writev:
10650 struct iovec *vec = lock_iovec(VERIFY_READ, arg2, arg3, 1);
10651 if (vec != NULL) {
10652 ret = get_errno(safe_writev(arg1, vec, arg3));
10653 unlock_iovec(vec, arg2, arg3, 0);
10654 } else {
10655 ret = -host_to_target_errno(errno);
10658 return ret;
10659 #if defined(TARGET_NR_preadv)
10660 case TARGET_NR_preadv:
10662 struct iovec *vec = lock_iovec(VERIFY_WRITE, arg2, arg3, 0);
10663 if (vec != NULL) {
10664 unsigned long low, high;
10666 target_to_host_low_high(arg4, arg5, &low, &high);
10667 ret = get_errno(safe_preadv(arg1, vec, arg3, low, high));
10668 unlock_iovec(vec, arg2, arg3, 1);
10669 } else {
10670 ret = -host_to_target_errno(errno);
10673 return ret;
10674 #endif
10675 #if defined(TARGET_NR_pwritev)
10676 case TARGET_NR_pwritev:
10678 struct iovec *vec = lock_iovec(VERIFY_READ, arg2, arg3, 1);
10679 if (vec != NULL) {
10680 unsigned long low, high;
10682 target_to_host_low_high(arg4, arg5, &low, &high);
10683 ret = get_errno(safe_pwritev(arg1, vec, arg3, low, high));
10684 unlock_iovec(vec, arg2, arg3, 0);
10685 } else {
10686 ret = -host_to_target_errno(errno);
10689 return ret;
10690 #endif
10691 case TARGET_NR_getsid:
10692 return get_errno(getsid(arg1));
10693 #if defined(TARGET_NR_fdatasync) /* Not on alpha (osf_datasync ?) */
10694 case TARGET_NR_fdatasync:
10695 return get_errno(fdatasync(arg1));
10696 #endif
10697 case TARGET_NR_sched_getaffinity:
10699 unsigned int mask_size;
10700 unsigned long *mask;
10703 * sched_getaffinity needs multiples of ulong, so need to take
10704 * care of mismatches between target ulong and host ulong sizes.
10706 if (arg2 & (sizeof(abi_ulong) - 1)) {
10707 return -TARGET_EINVAL;
10709 mask_size = (arg2 + (sizeof(*mask) - 1)) & ~(sizeof(*mask) - 1);
10711 mask = alloca(mask_size);
10712 memset(mask, 0, mask_size);
10713 ret = get_errno(sys_sched_getaffinity(arg1, mask_size, mask));
10715 if (!is_error(ret)) {
10716 if (ret > arg2) {
10717 /* More data returned than the caller's buffer will fit.
10718 * This only happens if sizeof(abi_long) < sizeof(long)
10719 * and the caller passed us a buffer holding an odd number
10720 * of abi_longs. If the host kernel is actually using the
10721 * extra 4 bytes then fail EINVAL; otherwise we can just
10722 * ignore them and only copy the interesting part.
10724 int numcpus = sysconf(_SC_NPROCESSORS_CONF);
10725 if (numcpus > arg2 * 8) {
10726 return -TARGET_EINVAL;
10728 ret = arg2;
10731 if (host_to_target_cpu_mask(mask, mask_size, arg3, ret)) {
10732 return -TARGET_EFAULT;
10736 return ret;
10737 case TARGET_NR_sched_setaffinity:
10739 unsigned int mask_size;
10740 unsigned long *mask;
10743 * sched_setaffinity needs multiples of ulong, so need to take
10744 * care of mismatches between target ulong and host ulong sizes.
10746 if (arg2 & (sizeof(abi_ulong) - 1)) {
10747 return -TARGET_EINVAL;
10749 mask_size = (arg2 + (sizeof(*mask) - 1)) & ~(sizeof(*mask) - 1);
10750 mask = alloca(mask_size);
10752 ret = target_to_host_cpu_mask(mask, mask_size, arg3, arg2);
10753 if (ret) {
10754 return ret;
10757 return get_errno(sys_sched_setaffinity(arg1, mask_size, mask));
10759 case TARGET_NR_getcpu:
10761 unsigned cpu, node;
10762 ret = get_errno(sys_getcpu(arg1 ? &cpu : NULL,
10763 arg2 ? &node : NULL,
10764 NULL));
10765 if (is_error(ret)) {
10766 return ret;
10768 if (arg1 && put_user_u32(cpu, arg1)) {
10769 return -TARGET_EFAULT;
10771 if (arg2 && put_user_u32(node, arg2)) {
10772 return -TARGET_EFAULT;
10775 return ret;
10776 case TARGET_NR_sched_setparam:
10778 struct target_sched_param *target_schp;
10779 struct sched_param schp;
10781 if (arg2 == 0) {
10782 return -TARGET_EINVAL;
10784 if (!lock_user_struct(VERIFY_READ, target_schp, arg2, 1)) {
10785 return -TARGET_EFAULT;
10787 schp.sched_priority = tswap32(target_schp->sched_priority);
10788 unlock_user_struct(target_schp, arg2, 0);
10789 return get_errno(sys_sched_setparam(arg1, &schp));
10791 case TARGET_NR_sched_getparam:
10793 struct target_sched_param *target_schp;
10794 struct sched_param schp;
10796 if (arg2 == 0) {
10797 return -TARGET_EINVAL;
10799 ret = get_errno(sys_sched_getparam(arg1, &schp));
10800 if (!is_error(ret)) {
10801 if (!lock_user_struct(VERIFY_WRITE, target_schp, arg2, 0)) {
10802 return -TARGET_EFAULT;
10804 target_schp->sched_priority = tswap32(schp.sched_priority);
10805 unlock_user_struct(target_schp, arg2, 1);
10808 return ret;
10809 case TARGET_NR_sched_setscheduler:
10811 struct target_sched_param *target_schp;
10812 struct sched_param schp;
10813 if (arg3 == 0) {
10814 return -TARGET_EINVAL;
10816 if (!lock_user_struct(VERIFY_READ, target_schp, arg3, 1)) {
10817 return -TARGET_EFAULT;
10819 schp.sched_priority = tswap32(target_schp->sched_priority);
10820 unlock_user_struct(target_schp, arg3, 0);
10821 return get_errno(sys_sched_setscheduler(arg1, arg2, &schp));
10823 case TARGET_NR_sched_getscheduler:
10824 return get_errno(sys_sched_getscheduler(arg1));
10825 case TARGET_NR_sched_getattr:
10827 struct target_sched_attr *target_scha;
10828 struct sched_attr scha;
10829 if (arg2 == 0) {
10830 return -TARGET_EINVAL;
10832 if (arg3 > sizeof(scha)) {
10833 arg3 = sizeof(scha);
10835 ret = get_errno(sys_sched_getattr(arg1, &scha, arg3, arg4));
10836 if (!is_error(ret)) {
10837 target_scha = lock_user(VERIFY_WRITE, arg2, arg3, 0);
10838 if (!target_scha) {
10839 return -TARGET_EFAULT;
10841 target_scha->size = tswap32(scha.size);
10842 target_scha->sched_policy = tswap32(scha.sched_policy);
10843 target_scha->sched_flags = tswap64(scha.sched_flags);
10844 target_scha->sched_nice = tswap32(scha.sched_nice);
10845 target_scha->sched_priority = tswap32(scha.sched_priority);
10846 target_scha->sched_runtime = tswap64(scha.sched_runtime);
10847 target_scha->sched_deadline = tswap64(scha.sched_deadline);
10848 target_scha->sched_period = tswap64(scha.sched_period);
10849 if (scha.size > offsetof(struct sched_attr, sched_util_min)) {
10850 target_scha->sched_util_min = tswap32(scha.sched_util_min);
10851 target_scha->sched_util_max = tswap32(scha.sched_util_max);
10853 unlock_user(target_scha, arg2, arg3);
10855 return ret;
10857 case TARGET_NR_sched_setattr:
10859 struct target_sched_attr *target_scha;
10860 struct sched_attr scha;
10861 uint32_t size;
10862 int zeroed;
10863 if (arg2 == 0) {
10864 return -TARGET_EINVAL;
10866 if (get_user_u32(size, arg2)) {
10867 return -TARGET_EFAULT;
10869 if (!size) {
10870 size = offsetof(struct target_sched_attr, sched_util_min);
10872 if (size < offsetof(struct target_sched_attr, sched_util_min)) {
10873 if (put_user_u32(sizeof(struct target_sched_attr), arg2)) {
10874 return -TARGET_EFAULT;
10876 return -TARGET_E2BIG;
10879 zeroed = check_zeroed_user(arg2, sizeof(struct target_sched_attr), size);
10880 if (zeroed < 0) {
10881 return zeroed;
10882 } else if (zeroed == 0) {
10883 if (put_user_u32(sizeof(struct target_sched_attr), arg2)) {
10884 return -TARGET_EFAULT;
10886 return -TARGET_E2BIG;
10888 if (size > sizeof(struct target_sched_attr)) {
10889 size = sizeof(struct target_sched_attr);
10892 target_scha = lock_user(VERIFY_READ, arg2, size, 1);
10893 if (!target_scha) {
10894 return -TARGET_EFAULT;
10896 scha.size = size;
10897 scha.sched_policy = tswap32(target_scha->sched_policy);
10898 scha.sched_flags = tswap64(target_scha->sched_flags);
10899 scha.sched_nice = tswap32(target_scha->sched_nice);
10900 scha.sched_priority = tswap32(target_scha->sched_priority);
10901 scha.sched_runtime = tswap64(target_scha->sched_runtime);
10902 scha.sched_deadline = tswap64(target_scha->sched_deadline);
10903 scha.sched_period = tswap64(target_scha->sched_period);
10904 if (size > offsetof(struct target_sched_attr, sched_util_min)) {
10905 scha.sched_util_min = tswap32(target_scha->sched_util_min);
10906 scha.sched_util_max = tswap32(target_scha->sched_util_max);
10908 unlock_user(target_scha, arg2, 0);
10909 return get_errno(sys_sched_setattr(arg1, &scha, arg3));
10911 case TARGET_NR_sched_yield:
10912 return get_errno(sched_yield());
10913 case TARGET_NR_sched_get_priority_max:
10914 return get_errno(sched_get_priority_max(arg1));
10915 case TARGET_NR_sched_get_priority_min:
10916 return get_errno(sched_get_priority_min(arg1));
10917 #ifdef TARGET_NR_sched_rr_get_interval
10918 case TARGET_NR_sched_rr_get_interval:
10920 struct timespec ts;
10921 ret = get_errno(sched_rr_get_interval(arg1, &ts));
10922 if (!is_error(ret)) {
10923 ret = host_to_target_timespec(arg2, &ts);
10926 return ret;
10927 #endif
10928 #ifdef TARGET_NR_sched_rr_get_interval_time64
10929 case TARGET_NR_sched_rr_get_interval_time64:
10931 struct timespec ts;
10932 ret = get_errno(sched_rr_get_interval(arg1, &ts));
10933 if (!is_error(ret)) {
10934 ret = host_to_target_timespec64(arg2, &ts);
10937 return ret;
10938 #endif
10939 #if defined(TARGET_NR_nanosleep)
10940 case TARGET_NR_nanosleep:
10942 struct timespec req, rem;
10943 target_to_host_timespec(&req, arg1);
10944 ret = get_errno(safe_nanosleep(&req, &rem));
10945 if (is_error(ret) && arg2) {
10946 host_to_target_timespec(arg2, &rem);
10949 return ret;
10950 #endif
10951 case TARGET_NR_prctl:
10952 return do_prctl(cpu_env, arg1, arg2, arg3, arg4, arg5);
10953 break;
10954 #ifdef TARGET_NR_arch_prctl
10955 case TARGET_NR_arch_prctl:
10956 return do_arch_prctl(cpu_env, arg1, arg2);
10957 #endif
10958 #ifdef TARGET_NR_pread64
10959 case TARGET_NR_pread64:
10960 if (regpairs_aligned(cpu_env, num)) {
10961 arg4 = arg5;
10962 arg5 = arg6;
10964 if (arg2 == 0 && arg3 == 0) {
10965 /* Special-case NULL buffer and zero length, which should succeed */
10966 p = 0;
10967 } else {
10968 p = lock_user(VERIFY_WRITE, arg2, arg3, 0);
10969 if (!p) {
10970 return -TARGET_EFAULT;
10973 ret = get_errno(pread64(arg1, p, arg3, target_offset64(arg4, arg5)));
10974 unlock_user(p, arg2, ret);
10975 return ret;
10976 case TARGET_NR_pwrite64:
10977 if (regpairs_aligned(cpu_env, num)) {
10978 arg4 = arg5;
10979 arg5 = arg6;
10981 if (arg2 == 0 && arg3 == 0) {
10982 /* Special-case NULL buffer and zero length, which should succeed */
10983 p = 0;
10984 } else {
10985 p = lock_user(VERIFY_READ, arg2, arg3, 1);
10986 if (!p) {
10987 return -TARGET_EFAULT;
10990 ret = get_errno(pwrite64(arg1, p, arg3, target_offset64(arg4, arg5)));
10991 unlock_user(p, arg2, 0);
10992 return ret;
10993 #endif
10994 case TARGET_NR_getcwd:
10995 if (!(p = lock_user(VERIFY_WRITE, arg1, arg2, 0)))
10996 return -TARGET_EFAULT;
10997 ret = get_errno(sys_getcwd1(p, arg2));
10998 unlock_user(p, arg1, ret);
10999 return ret;
11000 case TARGET_NR_capget:
11001 case TARGET_NR_capset:
11003 struct target_user_cap_header *target_header;
11004 struct target_user_cap_data *target_data = NULL;
11005 struct __user_cap_header_struct header;
11006 struct __user_cap_data_struct data[2];
11007 struct __user_cap_data_struct *dataptr = NULL;
11008 int i, target_datalen;
11009 int data_items = 1;
11011 if (!lock_user_struct(VERIFY_WRITE, target_header, arg1, 1)) {
11012 return -TARGET_EFAULT;
11014 header.version = tswap32(target_header->version);
11015 header.pid = tswap32(target_header->pid);
11017 if (header.version != _LINUX_CAPABILITY_VERSION) {
11018 /* Version 2 and up takes pointer to two user_data structs */
11019 data_items = 2;
11022 target_datalen = sizeof(*target_data) * data_items;
11024 if (arg2) {
11025 if (num == TARGET_NR_capget) {
11026 target_data = lock_user(VERIFY_WRITE, arg2, target_datalen, 0);
11027 } else {
11028 target_data = lock_user(VERIFY_READ, arg2, target_datalen, 1);
11030 if (!target_data) {
11031 unlock_user_struct(target_header, arg1, 0);
11032 return -TARGET_EFAULT;
11035 if (num == TARGET_NR_capset) {
11036 for (i = 0; i < data_items; i++) {
11037 data[i].effective = tswap32(target_data[i].effective);
11038 data[i].permitted = tswap32(target_data[i].permitted);
11039 data[i].inheritable = tswap32(target_data[i].inheritable);
11043 dataptr = data;
11046 if (num == TARGET_NR_capget) {
11047 ret = get_errno(capget(&header, dataptr));
11048 } else {
11049 ret = get_errno(capset(&header, dataptr));
11052 /* The kernel always updates version for both capget and capset */
11053 target_header->version = tswap32(header.version);
11054 unlock_user_struct(target_header, arg1, 1);
11056 if (arg2) {
11057 if (num == TARGET_NR_capget) {
11058 for (i = 0; i < data_items; i++) {
11059 target_data[i].effective = tswap32(data[i].effective);
11060 target_data[i].permitted = tswap32(data[i].permitted);
11061 target_data[i].inheritable = tswap32(data[i].inheritable);
11063 unlock_user(target_data, arg2, target_datalen);
11064 } else {
11065 unlock_user(target_data, arg2, 0);
11068 return ret;
11070 case TARGET_NR_sigaltstack:
11071 return do_sigaltstack(arg1, arg2, cpu_env);
11073 #ifdef CONFIG_SENDFILE
11074 #ifdef TARGET_NR_sendfile
11075 case TARGET_NR_sendfile:
11077 off_t *offp = NULL;
11078 off_t off;
11079 if (arg3) {
11080 ret = get_user_sal(off, arg3);
11081 if (is_error(ret)) {
11082 return ret;
11084 offp = &off;
11086 ret = get_errno(sendfile(arg1, arg2, offp, arg4));
11087 if (!is_error(ret) && arg3) {
11088 abi_long ret2 = put_user_sal(off, arg3);
11089 if (is_error(ret2)) {
11090 ret = ret2;
11093 return ret;
11095 #endif
11096 #ifdef TARGET_NR_sendfile64
11097 case TARGET_NR_sendfile64:
11099 off_t *offp = NULL;
11100 off_t off;
11101 if (arg3) {
11102 ret = get_user_s64(off, arg3);
11103 if (is_error(ret)) {
11104 return ret;
11106 offp = &off;
11108 ret = get_errno(sendfile(arg1, arg2, offp, arg4));
11109 if (!is_error(ret) && arg3) {
11110 abi_long ret2 = put_user_s64(off, arg3);
11111 if (is_error(ret2)) {
11112 ret = ret2;
11115 return ret;
11117 #endif
11118 #endif
11119 #ifdef TARGET_NR_vfork
11120 case TARGET_NR_vfork:
11121 return get_errno(do_fork(cpu_env,
11122 CLONE_VFORK | CLONE_VM | TARGET_SIGCHLD,
11123 0, 0, 0, 0));
11124 #endif
11125 #ifdef TARGET_NR_ugetrlimit
11126 case TARGET_NR_ugetrlimit:
11128 struct rlimit rlim;
11129 int resource = target_to_host_resource(arg1);
11130 ret = get_errno(getrlimit(resource, &rlim));
11131 if (!is_error(ret)) {
11132 struct target_rlimit *target_rlim;
11133 if (!lock_user_struct(VERIFY_WRITE, target_rlim, arg2, 0))
11134 return -TARGET_EFAULT;
11135 target_rlim->rlim_cur = host_to_target_rlim(rlim.rlim_cur);
11136 target_rlim->rlim_max = host_to_target_rlim(rlim.rlim_max);
11137 unlock_user_struct(target_rlim, arg2, 1);
11139 return ret;
11141 #endif
11142 #ifdef TARGET_NR_truncate64
11143 case TARGET_NR_truncate64:
11144 if (!(p = lock_user_string(arg1)))
11145 return -TARGET_EFAULT;
11146 ret = target_truncate64(cpu_env, p, arg2, arg3, arg4);
11147 unlock_user(p, arg1, 0);
11148 return ret;
11149 #endif
11150 #ifdef TARGET_NR_ftruncate64
11151 case TARGET_NR_ftruncate64:
11152 return target_ftruncate64(cpu_env, arg1, arg2, arg3, arg4);
11153 #endif
11154 #ifdef TARGET_NR_stat64
11155 case TARGET_NR_stat64:
11156 if (!(p = lock_user_string(arg1))) {
11157 return -TARGET_EFAULT;
11159 ret = get_errno(stat(path(p), &st));
11160 unlock_user(p, arg1, 0);
11161 if (!is_error(ret))
11162 ret = host_to_target_stat64(cpu_env, arg2, &st);
11163 return ret;
11164 #endif
11165 #ifdef TARGET_NR_lstat64
11166 case TARGET_NR_lstat64:
11167 if (!(p = lock_user_string(arg1))) {
11168 return -TARGET_EFAULT;
11170 ret = get_errno(lstat(path(p), &st));
11171 unlock_user(p, arg1, 0);
11172 if (!is_error(ret))
11173 ret = host_to_target_stat64(cpu_env, arg2, &st);
11174 return ret;
11175 #endif
11176 #ifdef TARGET_NR_fstat64
11177 case TARGET_NR_fstat64:
11178 ret = get_errno(fstat(arg1, &st));
11179 if (!is_error(ret))
11180 ret = host_to_target_stat64(cpu_env, arg2, &st);
11181 return ret;
11182 #endif
11183 #if (defined(TARGET_NR_fstatat64) || defined(TARGET_NR_newfstatat))
11184 #ifdef TARGET_NR_fstatat64
11185 case TARGET_NR_fstatat64:
11186 #endif
11187 #ifdef TARGET_NR_newfstatat
11188 case TARGET_NR_newfstatat:
11189 #endif
11190 if (!(p = lock_user_string(arg2))) {
11191 return -TARGET_EFAULT;
11193 ret = get_errno(fstatat(arg1, path(p), &st, arg4));
11194 unlock_user(p, arg2, 0);
11195 if (!is_error(ret))
11196 ret = host_to_target_stat64(cpu_env, arg3, &st);
11197 return ret;
11198 #endif
11199 #if defined(TARGET_NR_statx)
11200 case TARGET_NR_statx:
11202 struct target_statx *target_stx;
11203 int dirfd = arg1;
11204 int flags = arg3;
11206 p = lock_user_string(arg2);
11207 if (p == NULL) {
11208 return -TARGET_EFAULT;
11210 #if defined(__NR_statx)
11213 * It is assumed that struct statx is architecture independent.
11215 struct target_statx host_stx;
11216 int mask = arg4;
11218 ret = get_errno(sys_statx(dirfd, p, flags, mask, &host_stx));
11219 if (!is_error(ret)) {
11220 if (host_to_target_statx(&host_stx, arg5) != 0) {
11221 unlock_user(p, arg2, 0);
11222 return -TARGET_EFAULT;
11226 if (ret != -TARGET_ENOSYS) {
11227 unlock_user(p, arg2, 0);
11228 return ret;
11231 #endif
11232 ret = get_errno(fstatat(dirfd, path(p), &st, flags));
11233 unlock_user(p, arg2, 0);
11235 if (!is_error(ret)) {
11236 if (!lock_user_struct(VERIFY_WRITE, target_stx, arg5, 0)) {
11237 return -TARGET_EFAULT;
11239 memset(target_stx, 0, sizeof(*target_stx));
11240 __put_user(major(st.st_dev), &target_stx->stx_dev_major);
11241 __put_user(minor(st.st_dev), &target_stx->stx_dev_minor);
11242 __put_user(st.st_ino, &target_stx->stx_ino);
11243 __put_user(st.st_mode, &target_stx->stx_mode);
11244 __put_user(st.st_uid, &target_stx->stx_uid);
11245 __put_user(st.st_gid, &target_stx->stx_gid);
11246 __put_user(st.st_nlink, &target_stx->stx_nlink);
11247 __put_user(major(st.st_rdev), &target_stx->stx_rdev_major);
11248 __put_user(minor(st.st_rdev), &target_stx->stx_rdev_minor);
11249 __put_user(st.st_size, &target_stx->stx_size);
11250 __put_user(st.st_blksize, &target_stx->stx_blksize);
11251 __put_user(st.st_blocks, &target_stx->stx_blocks);
11252 __put_user(st.st_atime, &target_stx->stx_atime.tv_sec);
11253 __put_user(st.st_mtime, &target_stx->stx_mtime.tv_sec);
11254 __put_user(st.st_ctime, &target_stx->stx_ctime.tv_sec);
11255 unlock_user_struct(target_stx, arg5, 1);
11258 return ret;
11259 #endif
11260 #ifdef TARGET_NR_lchown
11261 case TARGET_NR_lchown:
11262 if (!(p = lock_user_string(arg1)))
11263 return -TARGET_EFAULT;
11264 ret = get_errno(lchown(p, low2highuid(arg2), low2highgid(arg3)));
11265 unlock_user(p, arg1, 0);
11266 return ret;
11267 #endif
11268 #ifdef TARGET_NR_getuid
11269 case TARGET_NR_getuid:
11270 return get_errno(high2lowuid(getuid()));
11271 #endif
11272 #ifdef TARGET_NR_getgid
11273 case TARGET_NR_getgid:
11274 return get_errno(high2lowgid(getgid()));
11275 #endif
11276 #ifdef TARGET_NR_geteuid
11277 case TARGET_NR_geteuid:
11278 return get_errno(high2lowuid(geteuid()));
11279 #endif
11280 #ifdef TARGET_NR_getegid
11281 case TARGET_NR_getegid:
11282 return get_errno(high2lowgid(getegid()));
11283 #endif
11284 case TARGET_NR_setreuid:
11285 return get_errno(setreuid(low2highuid(arg1), low2highuid(arg2)));
11286 case TARGET_NR_setregid:
11287 return get_errno(setregid(low2highgid(arg1), low2highgid(arg2)));
11288 case TARGET_NR_getgroups:
11290 int gidsetsize = arg1;
11291 target_id *target_grouplist;
11292 gid_t *grouplist;
11293 int i;
11295 grouplist = alloca(gidsetsize * sizeof(gid_t));
11296 ret = get_errno(getgroups(gidsetsize, grouplist));
11297 if (gidsetsize == 0)
11298 return ret;
11299 if (!is_error(ret)) {
11300 target_grouplist = lock_user(VERIFY_WRITE, arg2, gidsetsize * sizeof(target_id), 0);
11301 if (!target_grouplist)
11302 return -TARGET_EFAULT;
11303 for(i = 0;i < ret; i++)
11304 target_grouplist[i] = tswapid(high2lowgid(grouplist[i]));
11305 unlock_user(target_grouplist, arg2, gidsetsize * sizeof(target_id));
11308 return ret;
11309 case TARGET_NR_setgroups:
11311 int gidsetsize = arg1;
11312 target_id *target_grouplist;
11313 gid_t *grouplist = NULL;
11314 int i;
11315 if (gidsetsize) {
11316 grouplist = alloca(gidsetsize * sizeof(gid_t));
11317 target_grouplist = lock_user(VERIFY_READ, arg2, gidsetsize * sizeof(target_id), 1);
11318 if (!target_grouplist) {
11319 return -TARGET_EFAULT;
11321 for (i = 0; i < gidsetsize; i++) {
11322 grouplist[i] = low2highgid(tswapid(target_grouplist[i]));
11324 unlock_user(target_grouplist, arg2, 0);
11326 return get_errno(setgroups(gidsetsize, grouplist));
11328 case TARGET_NR_fchown:
11329 return get_errno(fchown(arg1, low2highuid(arg2), low2highgid(arg3)));
11330 #if defined(TARGET_NR_fchownat)
11331 case TARGET_NR_fchownat:
11332 if (!(p = lock_user_string(arg2)))
11333 return -TARGET_EFAULT;
11334 ret = get_errno(fchownat(arg1, p, low2highuid(arg3),
11335 low2highgid(arg4), arg5));
11336 unlock_user(p, arg2, 0);
11337 return ret;
11338 #endif
11339 #ifdef TARGET_NR_setresuid
11340 case TARGET_NR_setresuid:
11341 return get_errno(sys_setresuid(low2highuid(arg1),
11342 low2highuid(arg2),
11343 low2highuid(arg3)));
11344 #endif
11345 #ifdef TARGET_NR_getresuid
11346 case TARGET_NR_getresuid:
11348 uid_t ruid, euid, suid;
11349 ret = get_errno(getresuid(&ruid, &euid, &suid));
11350 if (!is_error(ret)) {
11351 if (put_user_id(high2lowuid(ruid), arg1)
11352 || put_user_id(high2lowuid(euid), arg2)
11353 || put_user_id(high2lowuid(suid), arg3))
11354 return -TARGET_EFAULT;
11357 return ret;
11358 #endif
11359 #ifdef TARGET_NR_getresgid
11360 case TARGET_NR_setresgid:
11361 return get_errno(sys_setresgid(low2highgid(arg1),
11362 low2highgid(arg2),
11363 low2highgid(arg3)));
11364 #endif
11365 #ifdef TARGET_NR_getresgid
11366 case TARGET_NR_getresgid:
11368 gid_t rgid, egid, sgid;
11369 ret = get_errno(getresgid(&rgid, &egid, &sgid));
11370 if (!is_error(ret)) {
11371 if (put_user_id(high2lowgid(rgid), arg1)
11372 || put_user_id(high2lowgid(egid), arg2)
11373 || put_user_id(high2lowgid(sgid), arg3))
11374 return -TARGET_EFAULT;
11377 return ret;
11378 #endif
11379 #ifdef TARGET_NR_chown
11380 case TARGET_NR_chown:
11381 if (!(p = lock_user_string(arg1)))
11382 return -TARGET_EFAULT;
11383 ret = get_errno(chown(p, low2highuid(arg2), low2highgid(arg3)));
11384 unlock_user(p, arg1, 0);
11385 return ret;
11386 #endif
11387 case TARGET_NR_setuid:
11388 return get_errno(sys_setuid(low2highuid(arg1)));
11389 case TARGET_NR_setgid:
11390 return get_errno(sys_setgid(low2highgid(arg1)));
11391 case TARGET_NR_setfsuid:
11392 return get_errno(setfsuid(arg1));
11393 case TARGET_NR_setfsgid:
11394 return get_errno(setfsgid(arg1));
11396 #ifdef TARGET_NR_lchown32
11397 case TARGET_NR_lchown32:
11398 if (!(p = lock_user_string(arg1)))
11399 return -TARGET_EFAULT;
11400 ret = get_errno(lchown(p, arg2, arg3));
11401 unlock_user(p, arg1, 0);
11402 return ret;
11403 #endif
11404 #ifdef TARGET_NR_getuid32
11405 case TARGET_NR_getuid32:
11406 return get_errno(getuid());
11407 #endif
11409 #if defined(TARGET_NR_getxuid) && defined(TARGET_ALPHA)
11410 /* Alpha specific */
11411 case TARGET_NR_getxuid:
11413 uid_t euid;
11414 euid=geteuid();
11415 cpu_env->ir[IR_A4]=euid;
11417 return get_errno(getuid());
11418 #endif
11419 #if defined(TARGET_NR_getxgid) && defined(TARGET_ALPHA)
11420 /* Alpha specific */
11421 case TARGET_NR_getxgid:
11423 uid_t egid;
11424 egid=getegid();
11425 cpu_env->ir[IR_A4]=egid;
11427 return get_errno(getgid());
11428 #endif
11429 #if defined(TARGET_NR_osf_getsysinfo) && defined(TARGET_ALPHA)
11430 /* Alpha specific */
11431 case TARGET_NR_osf_getsysinfo:
11432 ret = -TARGET_EOPNOTSUPP;
11433 switch (arg1) {
11434 case TARGET_GSI_IEEE_FP_CONTROL:
11436 uint64_t fpcr = cpu_alpha_load_fpcr(cpu_env);
11437 uint64_t swcr = cpu_env->swcr;
11439 swcr &= ~SWCR_STATUS_MASK;
11440 swcr |= (fpcr >> 35) & SWCR_STATUS_MASK;
11442 if (put_user_u64 (swcr, arg2))
11443 return -TARGET_EFAULT;
11444 ret = 0;
11446 break;
11448 /* case GSI_IEEE_STATE_AT_SIGNAL:
11449 -- Not implemented in linux kernel.
11450 case GSI_UACPROC:
11451 -- Retrieves current unaligned access state; not much used.
11452 case GSI_PROC_TYPE:
11453 -- Retrieves implver information; surely not used.
11454 case GSI_GET_HWRPB:
11455 -- Grabs a copy of the HWRPB; surely not used.
11458 return ret;
11459 #endif
11460 #if defined(TARGET_NR_osf_setsysinfo) && defined(TARGET_ALPHA)
11461 /* Alpha specific */
11462 case TARGET_NR_osf_setsysinfo:
11463 ret = -TARGET_EOPNOTSUPP;
11464 switch (arg1) {
11465 case TARGET_SSI_IEEE_FP_CONTROL:
11467 uint64_t swcr, fpcr;
11469 if (get_user_u64 (swcr, arg2)) {
11470 return -TARGET_EFAULT;
11474 * The kernel calls swcr_update_status to update the
11475 * status bits from the fpcr at every point that it
11476 * could be queried. Therefore, we store the status
11477 * bits only in FPCR.
11479 cpu_env->swcr = swcr & (SWCR_TRAP_ENABLE_MASK | SWCR_MAP_MASK);
11481 fpcr = cpu_alpha_load_fpcr(cpu_env);
11482 fpcr &= ((uint64_t)FPCR_DYN_MASK << 32);
11483 fpcr |= alpha_ieee_swcr_to_fpcr(swcr);
11484 cpu_alpha_store_fpcr(cpu_env, fpcr);
11485 ret = 0;
11487 break;
11489 case TARGET_SSI_IEEE_RAISE_EXCEPTION:
11491 uint64_t exc, fpcr, fex;
11493 if (get_user_u64(exc, arg2)) {
11494 return -TARGET_EFAULT;
11496 exc &= SWCR_STATUS_MASK;
11497 fpcr = cpu_alpha_load_fpcr(cpu_env);
11499 /* Old exceptions are not signaled. */
11500 fex = alpha_ieee_fpcr_to_swcr(fpcr);
11501 fex = exc & ~fex;
11502 fex >>= SWCR_STATUS_TO_EXCSUM_SHIFT;
11503 fex &= (cpu_env)->swcr;
11505 /* Update the hardware fpcr. */
11506 fpcr |= alpha_ieee_swcr_to_fpcr(exc);
11507 cpu_alpha_store_fpcr(cpu_env, fpcr);
11509 if (fex) {
11510 int si_code = TARGET_FPE_FLTUNK;
11511 target_siginfo_t info;
11513 if (fex & SWCR_TRAP_ENABLE_DNO) {
11514 si_code = TARGET_FPE_FLTUND;
11516 if (fex & SWCR_TRAP_ENABLE_INE) {
11517 si_code = TARGET_FPE_FLTRES;
11519 if (fex & SWCR_TRAP_ENABLE_UNF) {
11520 si_code = TARGET_FPE_FLTUND;
11522 if (fex & SWCR_TRAP_ENABLE_OVF) {
11523 si_code = TARGET_FPE_FLTOVF;
11525 if (fex & SWCR_TRAP_ENABLE_DZE) {
11526 si_code = TARGET_FPE_FLTDIV;
11528 if (fex & SWCR_TRAP_ENABLE_INV) {
11529 si_code = TARGET_FPE_FLTINV;
11532 info.si_signo = SIGFPE;
11533 info.si_errno = 0;
11534 info.si_code = si_code;
11535 info._sifields._sigfault._addr = (cpu_env)->pc;
11536 queue_signal(cpu_env, info.si_signo,
11537 QEMU_SI_FAULT, &info);
11539 ret = 0;
11541 break;
11543 /* case SSI_NVPAIRS:
11544 -- Used with SSIN_UACPROC to enable unaligned accesses.
11545 case SSI_IEEE_STATE_AT_SIGNAL:
11546 case SSI_IEEE_IGNORE_STATE_AT_SIGNAL:
11547 -- Not implemented in linux kernel
11550 return ret;
11551 #endif
11552 #ifdef TARGET_NR_osf_sigprocmask
11553 /* Alpha specific. */
11554 case TARGET_NR_osf_sigprocmask:
11556 abi_ulong mask;
11557 int how;
11558 sigset_t set, oldset;
11560 switch(arg1) {
11561 case TARGET_SIG_BLOCK:
11562 how = SIG_BLOCK;
11563 break;
11564 case TARGET_SIG_UNBLOCK:
11565 how = SIG_UNBLOCK;
11566 break;
11567 case TARGET_SIG_SETMASK:
11568 how = SIG_SETMASK;
11569 break;
11570 default:
11571 return -TARGET_EINVAL;
11573 mask = arg2;
11574 target_to_host_old_sigset(&set, &mask);
11575 ret = do_sigprocmask(how, &set, &oldset);
11576 if (!ret) {
11577 host_to_target_old_sigset(&mask, &oldset);
11578 ret = mask;
11581 return ret;
11582 #endif
11584 #ifdef TARGET_NR_getgid32
11585 case TARGET_NR_getgid32:
11586 return get_errno(getgid());
11587 #endif
11588 #ifdef TARGET_NR_geteuid32
11589 case TARGET_NR_geteuid32:
11590 return get_errno(geteuid());
11591 #endif
11592 #ifdef TARGET_NR_getegid32
11593 case TARGET_NR_getegid32:
11594 return get_errno(getegid());
11595 #endif
11596 #ifdef TARGET_NR_setreuid32
11597 case TARGET_NR_setreuid32:
11598 return get_errno(setreuid(arg1, arg2));
11599 #endif
11600 #ifdef TARGET_NR_setregid32
11601 case TARGET_NR_setregid32:
11602 return get_errno(setregid(arg1, arg2));
11603 #endif
11604 #ifdef TARGET_NR_getgroups32
11605 case TARGET_NR_getgroups32:
11607 int gidsetsize = arg1;
11608 uint32_t *target_grouplist;
11609 gid_t *grouplist;
11610 int i;
11612 grouplist = alloca(gidsetsize * sizeof(gid_t));
11613 ret = get_errno(getgroups(gidsetsize, grouplist));
11614 if (gidsetsize == 0)
11615 return ret;
11616 if (!is_error(ret)) {
11617 target_grouplist = lock_user(VERIFY_WRITE, arg2, gidsetsize * 4, 0);
11618 if (!target_grouplist) {
11619 return -TARGET_EFAULT;
11621 for(i = 0;i < ret; i++)
11622 target_grouplist[i] = tswap32(grouplist[i]);
11623 unlock_user(target_grouplist, arg2, gidsetsize * 4);
11626 return ret;
11627 #endif
11628 #ifdef TARGET_NR_setgroups32
11629 case TARGET_NR_setgroups32:
11631 int gidsetsize = arg1;
11632 uint32_t *target_grouplist;
11633 gid_t *grouplist;
11634 int i;
11636 grouplist = alloca(gidsetsize * sizeof(gid_t));
11637 target_grouplist = lock_user(VERIFY_READ, arg2, gidsetsize * 4, 1);
11638 if (!target_grouplist) {
11639 return -TARGET_EFAULT;
11641 for(i = 0;i < gidsetsize; i++)
11642 grouplist[i] = tswap32(target_grouplist[i]);
11643 unlock_user(target_grouplist, arg2, 0);
11644 return get_errno(setgroups(gidsetsize, grouplist));
11646 #endif
11647 #ifdef TARGET_NR_fchown32
11648 case TARGET_NR_fchown32:
11649 return get_errno(fchown(arg1, arg2, arg3));
11650 #endif
11651 #ifdef TARGET_NR_setresuid32
11652 case TARGET_NR_setresuid32:
11653 return get_errno(sys_setresuid(arg1, arg2, arg3));
11654 #endif
11655 #ifdef TARGET_NR_getresuid32
11656 case TARGET_NR_getresuid32:
11658 uid_t ruid, euid, suid;
11659 ret = get_errno(getresuid(&ruid, &euid, &suid));
11660 if (!is_error(ret)) {
11661 if (put_user_u32(ruid, arg1)
11662 || put_user_u32(euid, arg2)
11663 || put_user_u32(suid, arg3))
11664 return -TARGET_EFAULT;
11667 return ret;
11668 #endif
11669 #ifdef TARGET_NR_setresgid32
11670 case TARGET_NR_setresgid32:
11671 return get_errno(sys_setresgid(arg1, arg2, arg3));
11672 #endif
11673 #ifdef TARGET_NR_getresgid32
11674 case TARGET_NR_getresgid32:
11676 gid_t rgid, egid, sgid;
11677 ret = get_errno(getresgid(&rgid, &egid, &sgid));
11678 if (!is_error(ret)) {
11679 if (put_user_u32(rgid, arg1)
11680 || put_user_u32(egid, arg2)
11681 || put_user_u32(sgid, arg3))
11682 return -TARGET_EFAULT;
11685 return ret;
11686 #endif
11687 #ifdef TARGET_NR_chown32
11688 case TARGET_NR_chown32:
11689 if (!(p = lock_user_string(arg1)))
11690 return -TARGET_EFAULT;
11691 ret = get_errno(chown(p, arg2, arg3));
11692 unlock_user(p, arg1, 0);
11693 return ret;
11694 #endif
11695 #ifdef TARGET_NR_setuid32
11696 case TARGET_NR_setuid32:
11697 return get_errno(sys_setuid(arg1));
11698 #endif
11699 #ifdef TARGET_NR_setgid32
11700 case TARGET_NR_setgid32:
11701 return get_errno(sys_setgid(arg1));
11702 #endif
11703 #ifdef TARGET_NR_setfsuid32
11704 case TARGET_NR_setfsuid32:
11705 return get_errno(setfsuid(arg1));
11706 #endif
11707 #ifdef TARGET_NR_setfsgid32
11708 case TARGET_NR_setfsgid32:
11709 return get_errno(setfsgid(arg1));
11710 #endif
11711 #ifdef TARGET_NR_mincore
11712 case TARGET_NR_mincore:
11714 void *a = lock_user(VERIFY_READ, arg1, arg2, 0);
11715 if (!a) {
11716 return -TARGET_ENOMEM;
11718 p = lock_user_string(arg3);
11719 if (!p) {
11720 ret = -TARGET_EFAULT;
11721 } else {
11722 ret = get_errno(mincore(a, arg2, p));
11723 unlock_user(p, arg3, ret);
11725 unlock_user(a, arg1, 0);
11727 return ret;
11728 #endif
11729 #ifdef TARGET_NR_arm_fadvise64_64
11730 case TARGET_NR_arm_fadvise64_64:
11731 /* arm_fadvise64_64 looks like fadvise64_64 but
11732 * with different argument order: fd, advice, offset, len
11733 * rather than the usual fd, offset, len, advice.
11734 * Note that offset and len are both 64-bit so appear as
11735 * pairs of 32-bit registers.
11737 ret = posix_fadvise(arg1, target_offset64(arg3, arg4),
11738 target_offset64(arg5, arg6), arg2);
11739 return -host_to_target_errno(ret);
11740 #endif
11742 #if TARGET_ABI_BITS == 32
11744 #ifdef TARGET_NR_fadvise64_64
11745 case TARGET_NR_fadvise64_64:
11746 #if defined(TARGET_PPC) || defined(TARGET_XTENSA)
11747 /* 6 args: fd, advice, offset (high, low), len (high, low) */
11748 ret = arg2;
11749 arg2 = arg3;
11750 arg3 = arg4;
11751 arg4 = arg5;
11752 arg5 = arg6;
11753 arg6 = ret;
11754 #else
11755 /* 6 args: fd, offset (high, low), len (high, low), advice */
11756 if (regpairs_aligned(cpu_env, num)) {
11757 /* offset is in (3,4), len in (5,6) and advice in 7 */
11758 arg2 = arg3;
11759 arg3 = arg4;
11760 arg4 = arg5;
11761 arg5 = arg6;
11762 arg6 = arg7;
11764 #endif
11765 ret = posix_fadvise(arg1, target_offset64(arg2, arg3),
11766 target_offset64(arg4, arg5), arg6);
11767 return -host_to_target_errno(ret);
11768 #endif
11770 #ifdef TARGET_NR_fadvise64
11771 case TARGET_NR_fadvise64:
11772 /* 5 args: fd, offset (high, low), len, advice */
11773 if (regpairs_aligned(cpu_env, num)) {
11774 /* offset is in (3,4), len in 5 and advice in 6 */
11775 arg2 = arg3;
11776 arg3 = arg4;
11777 arg4 = arg5;
11778 arg5 = arg6;
11780 ret = posix_fadvise(arg1, target_offset64(arg2, arg3), arg4, arg5);
11781 return -host_to_target_errno(ret);
11782 #endif
11784 #else /* not a 32-bit ABI */
11785 #if defined(TARGET_NR_fadvise64_64) || defined(TARGET_NR_fadvise64)
11786 #ifdef TARGET_NR_fadvise64_64
11787 case TARGET_NR_fadvise64_64:
11788 #endif
11789 #ifdef TARGET_NR_fadvise64
11790 case TARGET_NR_fadvise64:
11791 #endif
11792 #ifdef TARGET_S390X
11793 switch (arg4) {
11794 case 4: arg4 = POSIX_FADV_NOREUSE + 1; break; /* make sure it's an invalid value */
11795 case 5: arg4 = POSIX_FADV_NOREUSE + 2; break; /* ditto */
11796 case 6: arg4 = POSIX_FADV_DONTNEED; break;
11797 case 7: arg4 = POSIX_FADV_NOREUSE; break;
11798 default: break;
11800 #endif
11801 return -host_to_target_errno(posix_fadvise(arg1, arg2, arg3, arg4));
11802 #endif
11803 #endif /* end of 64-bit ABI fadvise handling */
11805 #ifdef TARGET_NR_madvise
11806 case TARGET_NR_madvise:
11807 return target_madvise(arg1, arg2, arg3);
11808 #endif
11809 #ifdef TARGET_NR_fcntl64
11810 case TARGET_NR_fcntl64:
11812 int cmd;
11813 struct flock64 fl;
11814 from_flock64_fn *copyfrom = copy_from_user_flock64;
11815 to_flock64_fn *copyto = copy_to_user_flock64;
11817 #ifdef TARGET_ARM
11818 if (!cpu_env->eabi) {
11819 copyfrom = copy_from_user_oabi_flock64;
11820 copyto = copy_to_user_oabi_flock64;
11822 #endif
11824 cmd = target_to_host_fcntl_cmd(arg2);
11825 if (cmd == -TARGET_EINVAL) {
11826 return cmd;
11829 switch(arg2) {
11830 case TARGET_F_GETLK64:
11831 ret = copyfrom(&fl, arg3);
11832 if (ret) {
11833 break;
11835 ret = get_errno(safe_fcntl(arg1, cmd, &fl));
11836 if (ret == 0) {
11837 ret = copyto(arg3, &fl);
11839 break;
11841 case TARGET_F_SETLK64:
11842 case TARGET_F_SETLKW64:
11843 ret = copyfrom(&fl, arg3);
11844 if (ret) {
11845 break;
11847 ret = get_errno(safe_fcntl(arg1, cmd, &fl));
11848 break;
11849 default:
11850 ret = do_fcntl(arg1, arg2, arg3);
11851 break;
11853 return ret;
11855 #endif
11856 #ifdef TARGET_NR_cacheflush
11857 case TARGET_NR_cacheflush:
11858 /* self-modifying code is handled automatically, so nothing needed */
11859 return 0;
11860 #endif
11861 #ifdef TARGET_NR_getpagesize
11862 case TARGET_NR_getpagesize:
11863 return TARGET_PAGE_SIZE;
11864 #endif
11865 case TARGET_NR_gettid:
11866 return get_errno(sys_gettid());
11867 #ifdef TARGET_NR_readahead
11868 case TARGET_NR_readahead:
11869 #if TARGET_ABI_BITS == 32
11870 if (regpairs_aligned(cpu_env, num)) {
11871 arg2 = arg3;
11872 arg3 = arg4;
11873 arg4 = arg5;
11875 ret = get_errno(readahead(arg1, target_offset64(arg2, arg3) , arg4));
11876 #else
11877 ret = get_errno(readahead(arg1, arg2, arg3));
11878 #endif
11879 return ret;
11880 #endif
11881 #ifdef CONFIG_ATTR
11882 #ifdef TARGET_NR_setxattr
11883 case TARGET_NR_listxattr:
11884 case TARGET_NR_llistxattr:
11886 void *p, *b = 0;
11887 if (arg2) {
11888 b = lock_user(VERIFY_WRITE, arg2, arg3, 0);
11889 if (!b) {
11890 return -TARGET_EFAULT;
11893 p = lock_user_string(arg1);
11894 if (p) {
11895 if (num == TARGET_NR_listxattr) {
11896 ret = get_errno(listxattr(p, b, arg3));
11897 } else {
11898 ret = get_errno(llistxattr(p, b, arg3));
11900 } else {
11901 ret = -TARGET_EFAULT;
11903 unlock_user(p, arg1, 0);
11904 unlock_user(b, arg2, arg3);
11905 return ret;
11907 case TARGET_NR_flistxattr:
11909 void *b = 0;
11910 if (arg2) {
11911 b = lock_user(VERIFY_WRITE, arg2, arg3, 0);
11912 if (!b) {
11913 return -TARGET_EFAULT;
11916 ret = get_errno(flistxattr(arg1, b, arg3));
11917 unlock_user(b, arg2, arg3);
11918 return ret;
11920 case TARGET_NR_setxattr:
11921 case TARGET_NR_lsetxattr:
11923 void *p, *n, *v = 0;
11924 if (arg3) {
11925 v = lock_user(VERIFY_READ, arg3, arg4, 1);
11926 if (!v) {
11927 return -TARGET_EFAULT;
11930 p = lock_user_string(arg1);
11931 n = lock_user_string(arg2);
11932 if (p && n) {
11933 if (num == TARGET_NR_setxattr) {
11934 ret = get_errno(setxattr(p, n, v, arg4, arg5));
11935 } else {
11936 ret = get_errno(lsetxattr(p, n, v, arg4, arg5));
11938 } else {
11939 ret = -TARGET_EFAULT;
11941 unlock_user(p, arg1, 0);
11942 unlock_user(n, arg2, 0);
11943 unlock_user(v, arg3, 0);
11945 return ret;
11946 case TARGET_NR_fsetxattr:
11948 void *n, *v = 0;
11949 if (arg3) {
11950 v = lock_user(VERIFY_READ, arg3, arg4, 1);
11951 if (!v) {
11952 return -TARGET_EFAULT;
11955 n = lock_user_string(arg2);
11956 if (n) {
11957 ret = get_errno(fsetxattr(arg1, n, v, arg4, arg5));
11958 } else {
11959 ret = -TARGET_EFAULT;
11961 unlock_user(n, arg2, 0);
11962 unlock_user(v, arg3, 0);
11964 return ret;
11965 case TARGET_NR_getxattr:
11966 case TARGET_NR_lgetxattr:
11968 void *p, *n, *v = 0;
11969 if (arg3) {
11970 v = lock_user(VERIFY_WRITE, arg3, arg4, 0);
11971 if (!v) {
11972 return -TARGET_EFAULT;
11975 p = lock_user_string(arg1);
11976 n = lock_user_string(arg2);
11977 if (p && n) {
11978 if (num == TARGET_NR_getxattr) {
11979 ret = get_errno(getxattr(p, n, v, arg4));
11980 } else {
11981 ret = get_errno(lgetxattr(p, n, v, arg4));
11983 } else {
11984 ret = -TARGET_EFAULT;
11986 unlock_user(p, arg1, 0);
11987 unlock_user(n, arg2, 0);
11988 unlock_user(v, arg3, arg4);
11990 return ret;
11991 case TARGET_NR_fgetxattr:
11993 void *n, *v = 0;
11994 if (arg3) {
11995 v = lock_user(VERIFY_WRITE, arg3, arg4, 0);
11996 if (!v) {
11997 return -TARGET_EFAULT;
12000 n = lock_user_string(arg2);
12001 if (n) {
12002 ret = get_errno(fgetxattr(arg1, n, v, arg4));
12003 } else {
12004 ret = -TARGET_EFAULT;
12006 unlock_user(n, arg2, 0);
12007 unlock_user(v, arg3, arg4);
12009 return ret;
12010 case TARGET_NR_removexattr:
12011 case TARGET_NR_lremovexattr:
12013 void *p, *n;
12014 p = lock_user_string(arg1);
12015 n = lock_user_string(arg2);
12016 if (p && n) {
12017 if (num == TARGET_NR_removexattr) {
12018 ret = get_errno(removexattr(p, n));
12019 } else {
12020 ret = get_errno(lremovexattr(p, n));
12022 } else {
12023 ret = -TARGET_EFAULT;
12025 unlock_user(p, arg1, 0);
12026 unlock_user(n, arg2, 0);
12028 return ret;
12029 case TARGET_NR_fremovexattr:
12031 void *n;
12032 n = lock_user_string(arg2);
12033 if (n) {
12034 ret = get_errno(fremovexattr(arg1, n));
12035 } else {
12036 ret = -TARGET_EFAULT;
12038 unlock_user(n, arg2, 0);
12040 return ret;
12041 #endif
12042 #endif /* CONFIG_ATTR */
12043 #ifdef TARGET_NR_set_thread_area
12044 case TARGET_NR_set_thread_area:
12045 #if defined(TARGET_MIPS)
12046 cpu_env->active_tc.CP0_UserLocal = arg1;
12047 return 0;
12048 #elif defined(TARGET_CRIS)
12049 if (arg1 & 0xff)
12050 ret = -TARGET_EINVAL;
12051 else {
12052 cpu_env->pregs[PR_PID] = arg1;
12053 ret = 0;
12055 return ret;
12056 #elif defined(TARGET_I386) && defined(TARGET_ABI32)
12057 return do_set_thread_area(cpu_env, arg1);
12058 #elif defined(TARGET_M68K)
12060 TaskState *ts = cpu->opaque;
12061 ts->tp_value = arg1;
12062 return 0;
12064 #else
12065 return -TARGET_ENOSYS;
12066 #endif
12067 #endif
12068 #ifdef TARGET_NR_get_thread_area
12069 case TARGET_NR_get_thread_area:
12070 #if defined(TARGET_I386) && defined(TARGET_ABI32)
12071 return do_get_thread_area(cpu_env, arg1);
12072 #elif defined(TARGET_M68K)
12074 TaskState *ts = cpu->opaque;
12075 return ts->tp_value;
12077 #else
12078 return -TARGET_ENOSYS;
12079 #endif
12080 #endif
12081 #ifdef TARGET_NR_getdomainname
12082 case TARGET_NR_getdomainname:
12083 return -TARGET_ENOSYS;
12084 #endif
12086 #ifdef TARGET_NR_clock_settime
12087 case TARGET_NR_clock_settime:
12089 struct timespec ts;
12091 ret = target_to_host_timespec(&ts, arg2);
12092 if (!is_error(ret)) {
12093 ret = get_errno(clock_settime(arg1, &ts));
12095 return ret;
12097 #endif
12098 #ifdef TARGET_NR_clock_settime64
12099 case TARGET_NR_clock_settime64:
12101 struct timespec ts;
12103 ret = target_to_host_timespec64(&ts, arg2);
12104 if (!is_error(ret)) {
12105 ret = get_errno(clock_settime(arg1, &ts));
12107 return ret;
12109 #endif
12110 #ifdef TARGET_NR_clock_gettime
12111 case TARGET_NR_clock_gettime:
12113 struct timespec ts;
12114 ret = get_errno(clock_gettime(arg1, &ts));
12115 if (!is_error(ret)) {
12116 ret = host_to_target_timespec(arg2, &ts);
12118 return ret;
12120 #endif
12121 #ifdef TARGET_NR_clock_gettime64
12122 case TARGET_NR_clock_gettime64:
12124 struct timespec ts;
12125 ret = get_errno(clock_gettime(arg1, &ts));
12126 if (!is_error(ret)) {
12127 ret = host_to_target_timespec64(arg2, &ts);
12129 return ret;
12131 #endif
12132 #ifdef TARGET_NR_clock_getres
12133 case TARGET_NR_clock_getres:
12135 struct timespec ts;
12136 ret = get_errno(clock_getres(arg1, &ts));
12137 if (!is_error(ret)) {
12138 host_to_target_timespec(arg2, &ts);
12140 return ret;
12142 #endif
12143 #ifdef TARGET_NR_clock_getres_time64
12144 case TARGET_NR_clock_getres_time64:
12146 struct timespec ts;
12147 ret = get_errno(clock_getres(arg1, &ts));
12148 if (!is_error(ret)) {
12149 host_to_target_timespec64(arg2, &ts);
12151 return ret;
12153 #endif
12154 #ifdef TARGET_NR_clock_nanosleep
12155 case TARGET_NR_clock_nanosleep:
12157 struct timespec ts;
12158 if (target_to_host_timespec(&ts, arg3)) {
12159 return -TARGET_EFAULT;
12161 ret = get_errno(safe_clock_nanosleep(arg1, arg2,
12162 &ts, arg4 ? &ts : NULL));
12164 * if the call is interrupted by a signal handler, it fails
12165 * with error -TARGET_EINTR and if arg4 is not NULL and arg2 is not
12166 * TIMER_ABSTIME, it returns the remaining unslept time in arg4.
12168 if (ret == -TARGET_EINTR && arg4 && arg2 != TIMER_ABSTIME &&
12169 host_to_target_timespec(arg4, &ts)) {
12170 return -TARGET_EFAULT;
12173 return ret;
12175 #endif
12176 #ifdef TARGET_NR_clock_nanosleep_time64
12177 case TARGET_NR_clock_nanosleep_time64:
12179 struct timespec ts;
12181 if (target_to_host_timespec64(&ts, arg3)) {
12182 return -TARGET_EFAULT;
12185 ret = get_errno(safe_clock_nanosleep(arg1, arg2,
12186 &ts, arg4 ? &ts : NULL));
12188 if (ret == -TARGET_EINTR && arg4 && arg2 != TIMER_ABSTIME &&
12189 host_to_target_timespec64(arg4, &ts)) {
12190 return -TARGET_EFAULT;
12192 return ret;
12194 #endif
12196 #if defined(TARGET_NR_set_tid_address)
12197 case TARGET_NR_set_tid_address:
12199 TaskState *ts = cpu->opaque;
12200 ts->child_tidptr = arg1;
12201 /* do not call host set_tid_address() syscall, instead return tid() */
12202 return get_errno(sys_gettid());
12204 #endif
12206 case TARGET_NR_tkill:
12207 return get_errno(safe_tkill((int)arg1, target_to_host_signal(arg2)));
12209 case TARGET_NR_tgkill:
12210 return get_errno(safe_tgkill((int)arg1, (int)arg2,
12211 target_to_host_signal(arg3)));
12213 #ifdef TARGET_NR_set_robust_list
12214 case TARGET_NR_set_robust_list:
12215 case TARGET_NR_get_robust_list:
12216 /* The ABI for supporting robust futexes has userspace pass
12217 * the kernel a pointer to a linked list which is updated by
12218 * userspace after the syscall; the list is walked by the kernel
12219 * when the thread exits. Since the linked list in QEMU guest
12220 * memory isn't a valid linked list for the host and we have
12221 * no way to reliably intercept the thread-death event, we can't
12222 * support these. Silently return ENOSYS so that guest userspace
12223 * falls back to a non-robust futex implementation (which should
12224 * be OK except in the corner case of the guest crashing while
12225 * holding a mutex that is shared with another process via
12226 * shared memory).
12228 return -TARGET_ENOSYS;
12229 #endif
12231 #if defined(TARGET_NR_utimensat)
12232 case TARGET_NR_utimensat:
12234 struct timespec *tsp, ts[2];
12235 if (!arg3) {
12236 tsp = NULL;
12237 } else {
12238 if (target_to_host_timespec(ts, arg3)) {
12239 return -TARGET_EFAULT;
12241 if (target_to_host_timespec(ts + 1, arg3 +
12242 sizeof(struct target_timespec))) {
12243 return -TARGET_EFAULT;
12245 tsp = ts;
12247 if (!arg2)
12248 ret = get_errno(sys_utimensat(arg1, NULL, tsp, arg4));
12249 else {
12250 if (!(p = lock_user_string(arg2))) {
12251 return -TARGET_EFAULT;
12253 ret = get_errno(sys_utimensat(arg1, path(p), tsp, arg4));
12254 unlock_user(p, arg2, 0);
12257 return ret;
12258 #endif
12259 #ifdef TARGET_NR_utimensat_time64
12260 case TARGET_NR_utimensat_time64:
12262 struct timespec *tsp, ts[2];
12263 if (!arg3) {
12264 tsp = NULL;
12265 } else {
12266 if (target_to_host_timespec64(ts, arg3)) {
12267 return -TARGET_EFAULT;
12269 if (target_to_host_timespec64(ts + 1, arg3 +
12270 sizeof(struct target__kernel_timespec))) {
12271 return -TARGET_EFAULT;
12273 tsp = ts;
12275 if (!arg2)
12276 ret = get_errno(sys_utimensat(arg1, NULL, tsp, arg4));
12277 else {
12278 p = lock_user_string(arg2);
12279 if (!p) {
12280 return -TARGET_EFAULT;
12282 ret = get_errno(sys_utimensat(arg1, path(p), tsp, arg4));
12283 unlock_user(p, arg2, 0);
12286 return ret;
12287 #endif
12288 #ifdef TARGET_NR_futex
12289 case TARGET_NR_futex:
12290 return do_futex(cpu, arg1, arg2, arg3, arg4, arg5, arg6);
12291 #endif
12292 #ifdef TARGET_NR_futex_time64
12293 case TARGET_NR_futex_time64:
12294 return do_futex_time64(cpu, arg1, arg2, arg3, arg4, arg5, arg6);
12295 #endif
12296 #ifdef CONFIG_INOTIFY
12297 #if defined(TARGET_NR_inotify_init)
12298 case TARGET_NR_inotify_init:
12299 ret = get_errno(inotify_init());
12300 if (ret >= 0) {
12301 fd_trans_register(ret, &target_inotify_trans);
12303 return ret;
12304 #endif
12305 #if defined(TARGET_NR_inotify_init1) && defined(CONFIG_INOTIFY1)
12306 case TARGET_NR_inotify_init1:
12307 ret = get_errno(inotify_init1(target_to_host_bitmask(arg1,
12308 fcntl_flags_tbl)));
12309 if (ret >= 0) {
12310 fd_trans_register(ret, &target_inotify_trans);
12312 return ret;
12313 #endif
12314 #if defined(TARGET_NR_inotify_add_watch)
12315 case TARGET_NR_inotify_add_watch:
12316 p = lock_user_string(arg2);
12317 ret = get_errno(inotify_add_watch(arg1, path(p), arg3));
12318 unlock_user(p, arg2, 0);
12319 return ret;
12320 #endif
12321 #if defined(TARGET_NR_inotify_rm_watch)
12322 case TARGET_NR_inotify_rm_watch:
12323 return get_errno(inotify_rm_watch(arg1, arg2));
12324 #endif
12325 #endif
12327 #if defined(TARGET_NR_mq_open) && defined(__NR_mq_open)
12328 case TARGET_NR_mq_open:
12330 struct mq_attr posix_mq_attr;
12331 struct mq_attr *pposix_mq_attr;
12332 int host_flags;
12334 host_flags = target_to_host_bitmask(arg2, fcntl_flags_tbl);
12335 pposix_mq_attr = NULL;
12336 if (arg4) {
12337 if (copy_from_user_mq_attr(&posix_mq_attr, arg4) != 0) {
12338 return -TARGET_EFAULT;
12340 pposix_mq_attr = &posix_mq_attr;
12342 p = lock_user_string(arg1 - 1);
12343 if (!p) {
12344 return -TARGET_EFAULT;
12346 ret = get_errno(mq_open(p, host_flags, arg3, pposix_mq_attr));
12347 unlock_user (p, arg1, 0);
12349 return ret;
12351 case TARGET_NR_mq_unlink:
12352 p = lock_user_string(arg1 - 1);
12353 if (!p) {
12354 return -TARGET_EFAULT;
12356 ret = get_errno(mq_unlink(p));
12357 unlock_user (p, arg1, 0);
12358 return ret;
12360 #ifdef TARGET_NR_mq_timedsend
12361 case TARGET_NR_mq_timedsend:
12363 struct timespec ts;
12365 p = lock_user (VERIFY_READ, arg2, arg3, 1);
12366 if (arg5 != 0) {
12367 if (target_to_host_timespec(&ts, arg5)) {
12368 return -TARGET_EFAULT;
12370 ret = get_errno(safe_mq_timedsend(arg1, p, arg3, arg4, &ts));
12371 if (!is_error(ret) && host_to_target_timespec(arg5, &ts)) {
12372 return -TARGET_EFAULT;
12374 } else {
12375 ret = get_errno(safe_mq_timedsend(arg1, p, arg3, arg4, NULL));
12377 unlock_user (p, arg2, arg3);
12379 return ret;
12380 #endif
12381 #ifdef TARGET_NR_mq_timedsend_time64
12382 case TARGET_NR_mq_timedsend_time64:
12384 struct timespec ts;
12386 p = lock_user(VERIFY_READ, arg2, arg3, 1);
12387 if (arg5 != 0) {
12388 if (target_to_host_timespec64(&ts, arg5)) {
12389 return -TARGET_EFAULT;
12391 ret = get_errno(safe_mq_timedsend(arg1, p, arg3, arg4, &ts));
12392 if (!is_error(ret) && host_to_target_timespec64(arg5, &ts)) {
12393 return -TARGET_EFAULT;
12395 } else {
12396 ret = get_errno(safe_mq_timedsend(arg1, p, arg3, arg4, NULL));
12398 unlock_user(p, arg2, arg3);
12400 return ret;
12401 #endif
12403 #ifdef TARGET_NR_mq_timedreceive
12404 case TARGET_NR_mq_timedreceive:
12406 struct timespec ts;
12407 unsigned int prio;
12409 p = lock_user (VERIFY_READ, arg2, arg3, 1);
12410 if (arg5 != 0) {
12411 if (target_to_host_timespec(&ts, arg5)) {
12412 return -TARGET_EFAULT;
12414 ret = get_errno(safe_mq_timedreceive(arg1, p, arg3,
12415 &prio, &ts));
12416 if (!is_error(ret) && host_to_target_timespec(arg5, &ts)) {
12417 return -TARGET_EFAULT;
12419 } else {
12420 ret = get_errno(safe_mq_timedreceive(arg1, p, arg3,
12421 &prio, NULL));
12423 unlock_user (p, arg2, arg3);
12424 if (arg4 != 0)
12425 put_user_u32(prio, arg4);
12427 return ret;
12428 #endif
12429 #ifdef TARGET_NR_mq_timedreceive_time64
12430 case TARGET_NR_mq_timedreceive_time64:
12432 struct timespec ts;
12433 unsigned int prio;
12435 p = lock_user(VERIFY_READ, arg2, arg3, 1);
12436 if (arg5 != 0) {
12437 if (target_to_host_timespec64(&ts, arg5)) {
12438 return -TARGET_EFAULT;
12440 ret = get_errno(safe_mq_timedreceive(arg1, p, arg3,
12441 &prio, &ts));
12442 if (!is_error(ret) && host_to_target_timespec64(arg5, &ts)) {
12443 return -TARGET_EFAULT;
12445 } else {
12446 ret = get_errno(safe_mq_timedreceive(arg1, p, arg3,
12447 &prio, NULL));
12449 unlock_user(p, arg2, arg3);
12450 if (arg4 != 0) {
12451 put_user_u32(prio, arg4);
12454 return ret;
12455 #endif
12457 /* Not implemented for now... */
12458 /* case TARGET_NR_mq_notify: */
12459 /* break; */
12461 case TARGET_NR_mq_getsetattr:
12463 struct mq_attr posix_mq_attr_in, posix_mq_attr_out;
12464 ret = 0;
12465 if (arg2 != 0) {
12466 copy_from_user_mq_attr(&posix_mq_attr_in, arg2);
12467 ret = get_errno(mq_setattr(arg1, &posix_mq_attr_in,
12468 &posix_mq_attr_out));
12469 } else if (arg3 != 0) {
12470 ret = get_errno(mq_getattr(arg1, &posix_mq_attr_out));
12472 if (ret == 0 && arg3 != 0) {
12473 copy_to_user_mq_attr(arg3, &posix_mq_attr_out);
12476 return ret;
12477 #endif
12479 #ifdef CONFIG_SPLICE
12480 #ifdef TARGET_NR_tee
12481 case TARGET_NR_tee:
12483 ret = get_errno(tee(arg1,arg2,arg3,arg4));
12485 return ret;
12486 #endif
12487 #ifdef TARGET_NR_splice
12488 case TARGET_NR_splice:
12490 loff_t loff_in, loff_out;
12491 loff_t *ploff_in = NULL, *ploff_out = NULL;
12492 if (arg2) {
12493 if (get_user_u64(loff_in, arg2)) {
12494 return -TARGET_EFAULT;
12496 ploff_in = &loff_in;
12498 if (arg4) {
12499 if (get_user_u64(loff_out, arg4)) {
12500 return -TARGET_EFAULT;
12502 ploff_out = &loff_out;
12504 ret = get_errno(splice(arg1, ploff_in, arg3, ploff_out, arg5, arg6));
12505 if (arg2) {
12506 if (put_user_u64(loff_in, arg2)) {
12507 return -TARGET_EFAULT;
12510 if (arg4) {
12511 if (put_user_u64(loff_out, arg4)) {
12512 return -TARGET_EFAULT;
12516 return ret;
12517 #endif
12518 #ifdef TARGET_NR_vmsplice
12519 case TARGET_NR_vmsplice:
12521 struct iovec *vec = lock_iovec(VERIFY_READ, arg2, arg3, 1);
12522 if (vec != NULL) {
12523 ret = get_errno(vmsplice(arg1, vec, arg3, arg4));
12524 unlock_iovec(vec, arg2, arg3, 0);
12525 } else {
12526 ret = -host_to_target_errno(errno);
12529 return ret;
12530 #endif
12531 #endif /* CONFIG_SPLICE */
12532 #ifdef CONFIG_EVENTFD
12533 #if defined(TARGET_NR_eventfd)
12534 case TARGET_NR_eventfd:
12535 ret = get_errno(eventfd(arg1, 0));
12536 if (ret >= 0) {
12537 fd_trans_register(ret, &target_eventfd_trans);
12539 return ret;
12540 #endif
12541 #if defined(TARGET_NR_eventfd2)
12542 case TARGET_NR_eventfd2:
12544 int host_flags = arg2 & (~(TARGET_O_NONBLOCK_MASK | TARGET_O_CLOEXEC));
12545 if (arg2 & TARGET_O_NONBLOCK) {
12546 host_flags |= O_NONBLOCK;
12548 if (arg2 & TARGET_O_CLOEXEC) {
12549 host_flags |= O_CLOEXEC;
12551 ret = get_errno(eventfd(arg1, host_flags));
12552 if (ret >= 0) {
12553 fd_trans_register(ret, &target_eventfd_trans);
12555 return ret;
12557 #endif
12558 #endif /* CONFIG_EVENTFD */
12559 #if defined(CONFIG_FALLOCATE) && defined(TARGET_NR_fallocate)
12560 case TARGET_NR_fallocate:
12561 #if TARGET_ABI_BITS == 32
12562 ret = get_errno(fallocate(arg1, arg2, target_offset64(arg3, arg4),
12563 target_offset64(arg5, arg6)));
12564 #else
12565 ret = get_errno(fallocate(arg1, arg2, arg3, arg4));
12566 #endif
12567 return ret;
12568 #endif
12569 #if defined(CONFIG_SYNC_FILE_RANGE)
12570 #if defined(TARGET_NR_sync_file_range)
12571 case TARGET_NR_sync_file_range:
12572 #if TARGET_ABI_BITS == 32
12573 #if defined(TARGET_MIPS)
12574 ret = get_errno(sync_file_range(arg1, target_offset64(arg3, arg4),
12575 target_offset64(arg5, arg6), arg7));
12576 #else
12577 ret = get_errno(sync_file_range(arg1, target_offset64(arg2, arg3),
12578 target_offset64(arg4, arg5), arg6));
12579 #endif /* !TARGET_MIPS */
12580 #else
12581 ret = get_errno(sync_file_range(arg1, arg2, arg3, arg4));
12582 #endif
12583 return ret;
12584 #endif
12585 #if defined(TARGET_NR_sync_file_range2) || \
12586 defined(TARGET_NR_arm_sync_file_range)
12587 #if defined(TARGET_NR_sync_file_range2)
12588 case TARGET_NR_sync_file_range2:
12589 #endif
12590 #if defined(TARGET_NR_arm_sync_file_range)
12591 case TARGET_NR_arm_sync_file_range:
12592 #endif
12593 /* This is like sync_file_range but the arguments are reordered */
12594 #if TARGET_ABI_BITS == 32
12595 ret = get_errno(sync_file_range(arg1, target_offset64(arg3, arg4),
12596 target_offset64(arg5, arg6), arg2));
12597 #else
12598 ret = get_errno(sync_file_range(arg1, arg3, arg4, arg2));
12599 #endif
12600 return ret;
12601 #endif
12602 #endif
12603 #if defined(TARGET_NR_signalfd4)
12604 case TARGET_NR_signalfd4:
12605 return do_signalfd4(arg1, arg2, arg4);
12606 #endif
12607 #if defined(TARGET_NR_signalfd)
12608 case TARGET_NR_signalfd:
12609 return do_signalfd4(arg1, arg2, 0);
12610 #endif
12611 #if defined(CONFIG_EPOLL)
12612 #if defined(TARGET_NR_epoll_create)
12613 case TARGET_NR_epoll_create:
12614 return get_errno(epoll_create(arg1));
12615 #endif
12616 #if defined(TARGET_NR_epoll_create1) && defined(CONFIG_EPOLL_CREATE1)
12617 case TARGET_NR_epoll_create1:
12618 return get_errno(epoll_create1(target_to_host_bitmask(arg1, fcntl_flags_tbl)));
12619 #endif
12620 #if defined(TARGET_NR_epoll_ctl)
12621 case TARGET_NR_epoll_ctl:
12623 struct epoll_event ep;
12624 struct epoll_event *epp = 0;
12625 if (arg4) {
12626 if (arg2 != EPOLL_CTL_DEL) {
12627 struct target_epoll_event *target_ep;
12628 if (!lock_user_struct(VERIFY_READ, target_ep, arg4, 1)) {
12629 return -TARGET_EFAULT;
12631 ep.events = tswap32(target_ep->events);
12633 * The epoll_data_t union is just opaque data to the kernel,
12634 * so we transfer all 64 bits across and need not worry what
12635 * actual data type it is.
12637 ep.data.u64 = tswap64(target_ep->data.u64);
12638 unlock_user_struct(target_ep, arg4, 0);
12641 * before kernel 2.6.9, EPOLL_CTL_DEL operation required a
12642 * non-null pointer, even though this argument is ignored.
12645 epp = &ep;
12647 return get_errno(epoll_ctl(arg1, arg2, arg3, epp));
12649 #endif
12651 #if defined(TARGET_NR_epoll_wait) || defined(TARGET_NR_epoll_pwait)
12652 #if defined(TARGET_NR_epoll_wait)
12653 case TARGET_NR_epoll_wait:
12654 #endif
12655 #if defined(TARGET_NR_epoll_pwait)
12656 case TARGET_NR_epoll_pwait:
12657 #endif
12659 struct target_epoll_event *target_ep;
12660 struct epoll_event *ep;
12661 int epfd = arg1;
12662 int maxevents = arg3;
12663 int timeout = arg4;
12665 if (maxevents <= 0 || maxevents > TARGET_EP_MAX_EVENTS) {
12666 return -TARGET_EINVAL;
12669 target_ep = lock_user(VERIFY_WRITE, arg2,
12670 maxevents * sizeof(struct target_epoll_event), 1);
12671 if (!target_ep) {
12672 return -TARGET_EFAULT;
12675 ep = g_try_new(struct epoll_event, maxevents);
12676 if (!ep) {
12677 unlock_user(target_ep, arg2, 0);
12678 return -TARGET_ENOMEM;
12681 switch (num) {
12682 #if defined(TARGET_NR_epoll_pwait)
12683 case TARGET_NR_epoll_pwait:
12685 sigset_t *set = NULL;
12687 if (arg5) {
12688 ret = process_sigsuspend_mask(&set, arg5, arg6);
12689 if (ret != 0) {
12690 break;
12694 ret = get_errno(safe_epoll_pwait(epfd, ep, maxevents, timeout,
12695 set, SIGSET_T_SIZE));
12697 if (set) {
12698 finish_sigsuspend_mask(ret);
12700 break;
12702 #endif
12703 #if defined(TARGET_NR_epoll_wait)
12704 case TARGET_NR_epoll_wait:
12705 ret = get_errno(safe_epoll_pwait(epfd, ep, maxevents, timeout,
12706 NULL, 0));
12707 break;
12708 #endif
12709 default:
12710 ret = -TARGET_ENOSYS;
12712 if (!is_error(ret)) {
12713 int i;
12714 for (i = 0; i < ret; i++) {
12715 target_ep[i].events = tswap32(ep[i].events);
12716 target_ep[i].data.u64 = tswap64(ep[i].data.u64);
12718 unlock_user(target_ep, arg2,
12719 ret * sizeof(struct target_epoll_event));
12720 } else {
12721 unlock_user(target_ep, arg2, 0);
12723 g_free(ep);
12724 return ret;
12726 #endif
12727 #endif
12728 #ifdef TARGET_NR_prlimit64
12729 case TARGET_NR_prlimit64:
12731 /* args: pid, resource number, ptr to new rlimit, ptr to old rlimit */
12732 struct target_rlimit64 *target_rnew, *target_rold;
12733 struct host_rlimit64 rnew, rold, *rnewp = 0;
12734 int resource = target_to_host_resource(arg2);
12736 if (arg3 && (resource != RLIMIT_AS &&
12737 resource != RLIMIT_DATA &&
12738 resource != RLIMIT_STACK)) {
12739 if (!lock_user_struct(VERIFY_READ, target_rnew, arg3, 1)) {
12740 return -TARGET_EFAULT;
12742 rnew.rlim_cur = tswap64(target_rnew->rlim_cur);
12743 rnew.rlim_max = tswap64(target_rnew->rlim_max);
12744 unlock_user_struct(target_rnew, arg3, 0);
12745 rnewp = &rnew;
12748 ret = get_errno(sys_prlimit64(arg1, resource, rnewp, arg4 ? &rold : 0));
12749 if (!is_error(ret) && arg4) {
12750 if (!lock_user_struct(VERIFY_WRITE, target_rold, arg4, 1)) {
12751 return -TARGET_EFAULT;
12753 target_rold->rlim_cur = tswap64(rold.rlim_cur);
12754 target_rold->rlim_max = tswap64(rold.rlim_max);
12755 unlock_user_struct(target_rold, arg4, 1);
12757 return ret;
12759 #endif
12760 #ifdef TARGET_NR_gethostname
12761 case TARGET_NR_gethostname:
12763 char *name = lock_user(VERIFY_WRITE, arg1, arg2, 0);
12764 if (name) {
12765 ret = get_errno(gethostname(name, arg2));
12766 unlock_user(name, arg1, arg2);
12767 } else {
12768 ret = -TARGET_EFAULT;
12770 return ret;
12772 #endif
12773 #ifdef TARGET_NR_atomic_cmpxchg_32
12774 case TARGET_NR_atomic_cmpxchg_32:
12776 /* should use start_exclusive from main.c */
12777 abi_ulong mem_value;
12778 if (get_user_u32(mem_value, arg6)) {
12779 target_siginfo_t info;
12780 info.si_signo = SIGSEGV;
12781 info.si_errno = 0;
12782 info.si_code = TARGET_SEGV_MAPERR;
12783 info._sifields._sigfault._addr = arg6;
12784 queue_signal(cpu_env, info.si_signo, QEMU_SI_FAULT, &info);
12785 ret = 0xdeadbeef;
12788 if (mem_value == arg2)
12789 put_user_u32(arg1, arg6);
12790 return mem_value;
12792 #endif
12793 #ifdef TARGET_NR_atomic_barrier
12794 case TARGET_NR_atomic_barrier:
12795 /* Like the kernel implementation and the
12796 qemu arm barrier, no-op this? */
12797 return 0;
12798 #endif
12800 #ifdef TARGET_NR_timer_create
12801 case TARGET_NR_timer_create:
12803 /* args: clockid_t clockid, struct sigevent *sevp, timer_t *timerid */
12805 struct sigevent host_sevp = { {0}, }, *phost_sevp = NULL;
12807 int clkid = arg1;
12808 int timer_index = next_free_host_timer();
12810 if (timer_index < 0) {
12811 ret = -TARGET_EAGAIN;
12812 } else {
12813 timer_t *phtimer = g_posix_timers + timer_index;
12815 if (arg2) {
12816 phost_sevp = &host_sevp;
12817 ret = target_to_host_sigevent(phost_sevp, arg2);
12818 if (ret != 0) {
12819 return ret;
12823 ret = get_errno(timer_create(clkid, phost_sevp, phtimer));
12824 if (ret) {
12825 phtimer = NULL;
12826 } else {
12827 if (put_user(TIMER_MAGIC | timer_index, arg3, target_timer_t)) {
12828 return -TARGET_EFAULT;
12832 return ret;
12834 #endif
12836 #ifdef TARGET_NR_timer_settime
12837 case TARGET_NR_timer_settime:
12839 /* args: timer_t timerid, int flags, const struct itimerspec *new_value,
12840 * struct itimerspec * old_value */
12841 target_timer_t timerid = get_timer_id(arg1);
12843 if (timerid < 0) {
12844 ret = timerid;
12845 } else if (arg3 == 0) {
12846 ret = -TARGET_EINVAL;
12847 } else {
12848 timer_t htimer = g_posix_timers[timerid];
12849 struct itimerspec hspec_new = {{0},}, hspec_old = {{0},};
12851 if (target_to_host_itimerspec(&hspec_new, arg3)) {
12852 return -TARGET_EFAULT;
12854 ret = get_errno(
12855 timer_settime(htimer, arg2, &hspec_new, &hspec_old));
12856 if (arg4 && host_to_target_itimerspec(arg4, &hspec_old)) {
12857 return -TARGET_EFAULT;
12860 return ret;
12862 #endif
12864 #ifdef TARGET_NR_timer_settime64
12865 case TARGET_NR_timer_settime64:
12867 target_timer_t timerid = get_timer_id(arg1);
12869 if (timerid < 0) {
12870 ret = timerid;
12871 } else if (arg3 == 0) {
12872 ret = -TARGET_EINVAL;
12873 } else {
12874 timer_t htimer = g_posix_timers[timerid];
12875 struct itimerspec hspec_new = {{0},}, hspec_old = {{0},};
12877 if (target_to_host_itimerspec64(&hspec_new, arg3)) {
12878 return -TARGET_EFAULT;
12880 ret = get_errno(
12881 timer_settime(htimer, arg2, &hspec_new, &hspec_old));
12882 if (arg4 && host_to_target_itimerspec64(arg4, &hspec_old)) {
12883 return -TARGET_EFAULT;
12886 return ret;
12888 #endif
12890 #ifdef TARGET_NR_timer_gettime
12891 case TARGET_NR_timer_gettime:
12893 /* args: timer_t timerid, struct itimerspec *curr_value */
12894 target_timer_t timerid = get_timer_id(arg1);
12896 if (timerid < 0) {
12897 ret = timerid;
12898 } else if (!arg2) {
12899 ret = -TARGET_EFAULT;
12900 } else {
12901 timer_t htimer = g_posix_timers[timerid];
12902 struct itimerspec hspec;
12903 ret = get_errno(timer_gettime(htimer, &hspec));
12905 if (host_to_target_itimerspec(arg2, &hspec)) {
12906 ret = -TARGET_EFAULT;
12909 return ret;
12911 #endif
12913 #ifdef TARGET_NR_timer_gettime64
12914 case TARGET_NR_timer_gettime64:
12916 /* args: timer_t timerid, struct itimerspec64 *curr_value */
12917 target_timer_t timerid = get_timer_id(arg1);
12919 if (timerid < 0) {
12920 ret = timerid;
12921 } else if (!arg2) {
12922 ret = -TARGET_EFAULT;
12923 } else {
12924 timer_t htimer = g_posix_timers[timerid];
12925 struct itimerspec hspec;
12926 ret = get_errno(timer_gettime(htimer, &hspec));
12928 if (host_to_target_itimerspec64(arg2, &hspec)) {
12929 ret = -TARGET_EFAULT;
12932 return ret;
12934 #endif
12936 #ifdef TARGET_NR_timer_getoverrun
12937 case TARGET_NR_timer_getoverrun:
12939 /* args: timer_t timerid */
12940 target_timer_t timerid = get_timer_id(arg1);
12942 if (timerid < 0) {
12943 ret = timerid;
12944 } else {
12945 timer_t htimer = g_posix_timers[timerid];
12946 ret = get_errno(timer_getoverrun(htimer));
12948 return ret;
12950 #endif
12952 #ifdef TARGET_NR_timer_delete
12953 case TARGET_NR_timer_delete:
12955 /* args: timer_t timerid */
12956 target_timer_t timerid = get_timer_id(arg1);
12958 if (timerid < 0) {
12959 ret = timerid;
12960 } else {
12961 timer_t htimer = g_posix_timers[timerid];
12962 ret = get_errno(timer_delete(htimer));
12963 g_posix_timers[timerid] = 0;
12965 return ret;
12967 #endif
12969 #if defined(TARGET_NR_timerfd_create) && defined(CONFIG_TIMERFD)
12970 case TARGET_NR_timerfd_create:
12971 return get_errno(timerfd_create(arg1,
12972 target_to_host_bitmask(arg2, fcntl_flags_tbl)));
12973 #endif
12975 #if defined(TARGET_NR_timerfd_gettime) && defined(CONFIG_TIMERFD)
12976 case TARGET_NR_timerfd_gettime:
12978 struct itimerspec its_curr;
12980 ret = get_errno(timerfd_gettime(arg1, &its_curr));
12982 if (arg2 && host_to_target_itimerspec(arg2, &its_curr)) {
12983 return -TARGET_EFAULT;
12986 return ret;
12987 #endif
12989 #if defined(TARGET_NR_timerfd_gettime64) && defined(CONFIG_TIMERFD)
12990 case TARGET_NR_timerfd_gettime64:
12992 struct itimerspec its_curr;
12994 ret = get_errno(timerfd_gettime(arg1, &its_curr));
12996 if (arg2 && host_to_target_itimerspec64(arg2, &its_curr)) {
12997 return -TARGET_EFAULT;
13000 return ret;
13001 #endif
13003 #if defined(TARGET_NR_timerfd_settime) && defined(CONFIG_TIMERFD)
13004 case TARGET_NR_timerfd_settime:
13006 struct itimerspec its_new, its_old, *p_new;
13008 if (arg3) {
13009 if (target_to_host_itimerspec(&its_new, arg3)) {
13010 return -TARGET_EFAULT;
13012 p_new = &its_new;
13013 } else {
13014 p_new = NULL;
13017 ret = get_errno(timerfd_settime(arg1, arg2, p_new, &its_old));
13019 if (arg4 && host_to_target_itimerspec(arg4, &its_old)) {
13020 return -TARGET_EFAULT;
13023 return ret;
13024 #endif
13026 #if defined(TARGET_NR_timerfd_settime64) && defined(CONFIG_TIMERFD)
13027 case TARGET_NR_timerfd_settime64:
13029 struct itimerspec its_new, its_old, *p_new;
13031 if (arg3) {
13032 if (target_to_host_itimerspec64(&its_new, arg3)) {
13033 return -TARGET_EFAULT;
13035 p_new = &its_new;
13036 } else {
13037 p_new = NULL;
13040 ret = get_errno(timerfd_settime(arg1, arg2, p_new, &its_old));
13042 if (arg4 && host_to_target_itimerspec64(arg4, &its_old)) {
13043 return -TARGET_EFAULT;
13046 return ret;
13047 #endif
13049 #if defined(TARGET_NR_ioprio_get) && defined(__NR_ioprio_get)
13050 case TARGET_NR_ioprio_get:
13051 return get_errno(ioprio_get(arg1, arg2));
13052 #endif
13054 #if defined(TARGET_NR_ioprio_set) && defined(__NR_ioprio_set)
13055 case TARGET_NR_ioprio_set:
13056 return get_errno(ioprio_set(arg1, arg2, arg3));
13057 #endif
13059 #if defined(TARGET_NR_setns) && defined(CONFIG_SETNS)
13060 case TARGET_NR_setns:
13061 return get_errno(setns(arg1, arg2));
13062 #endif
13063 #if defined(TARGET_NR_unshare) && defined(CONFIG_SETNS)
13064 case TARGET_NR_unshare:
13065 return get_errno(unshare(arg1));
13066 #endif
13067 #if defined(TARGET_NR_kcmp) && defined(__NR_kcmp)
13068 case TARGET_NR_kcmp:
13069 return get_errno(kcmp(arg1, arg2, arg3, arg4, arg5));
13070 #endif
13071 #ifdef TARGET_NR_swapcontext
13072 case TARGET_NR_swapcontext:
13073 /* PowerPC specific. */
13074 return do_swapcontext(cpu_env, arg1, arg2, arg3);
13075 #endif
13076 #ifdef TARGET_NR_memfd_create
13077 case TARGET_NR_memfd_create:
13078 p = lock_user_string(arg1);
13079 if (!p) {
13080 return -TARGET_EFAULT;
13082 ret = get_errno(memfd_create(p, arg2));
13083 fd_trans_unregister(ret);
13084 unlock_user(p, arg1, 0);
13085 return ret;
13086 #endif
13087 #if defined TARGET_NR_membarrier && defined __NR_membarrier
13088 case TARGET_NR_membarrier:
13089 return get_errno(membarrier(arg1, arg2));
13090 #endif
13092 #if defined(TARGET_NR_copy_file_range) && defined(__NR_copy_file_range)
13093 case TARGET_NR_copy_file_range:
13095 loff_t inoff, outoff;
13096 loff_t *pinoff = NULL, *poutoff = NULL;
13098 if (arg2) {
13099 if (get_user_u64(inoff, arg2)) {
13100 return -TARGET_EFAULT;
13102 pinoff = &inoff;
13104 if (arg4) {
13105 if (get_user_u64(outoff, arg4)) {
13106 return -TARGET_EFAULT;
13108 poutoff = &outoff;
13110 /* Do not sign-extend the count parameter. */
13111 ret = get_errno(safe_copy_file_range(arg1, pinoff, arg3, poutoff,
13112 (abi_ulong)arg5, arg6));
13113 if (!is_error(ret) && ret > 0) {
13114 if (arg2) {
13115 if (put_user_u64(inoff, arg2)) {
13116 return -TARGET_EFAULT;
13119 if (arg4) {
13120 if (put_user_u64(outoff, arg4)) {
13121 return -TARGET_EFAULT;
13126 return ret;
13127 #endif
13129 #if defined(TARGET_NR_pivot_root)
13130 case TARGET_NR_pivot_root:
13132 void *p2;
13133 p = lock_user_string(arg1); /* new_root */
13134 p2 = lock_user_string(arg2); /* put_old */
13135 if (!p || !p2) {
13136 ret = -TARGET_EFAULT;
13137 } else {
13138 ret = get_errno(pivot_root(p, p2));
13140 unlock_user(p2, arg2, 0);
13141 unlock_user(p, arg1, 0);
13143 return ret;
13144 #endif
13146 default:
13147 qemu_log_mask(LOG_UNIMP, "Unsupported syscall: %d\n", num);
13148 return -TARGET_ENOSYS;
13150 return ret;
13153 abi_long do_syscall(CPUArchState *cpu_env, int num, abi_long arg1,
13154 abi_long arg2, abi_long arg3, abi_long arg4,
13155 abi_long arg5, abi_long arg6, abi_long arg7,
13156 abi_long arg8)
13158 CPUState *cpu = env_cpu(cpu_env);
13159 abi_long ret;
13161 #ifdef DEBUG_ERESTARTSYS
13162 /* Debug-only code for exercising the syscall-restart code paths
13163 * in the per-architecture cpu main loops: restart every syscall
13164 * the guest makes once before letting it through.
13167 static bool flag;
13168 flag = !flag;
13169 if (flag) {
13170 return -QEMU_ERESTARTSYS;
13173 #endif
13175 record_syscall_start(cpu, num, arg1,
13176 arg2, arg3, arg4, arg5, arg6, arg7, arg8);
13178 if (unlikely(qemu_loglevel_mask(LOG_STRACE))) {
13179 print_syscall(cpu_env, num, arg1, arg2, arg3, arg4, arg5, arg6);
13182 ret = do_syscall1(cpu_env, num, arg1, arg2, arg3, arg4,
13183 arg5, arg6, arg7, arg8);
13185 if (unlikely(qemu_loglevel_mask(LOG_STRACE))) {
13186 print_syscall_ret(cpu_env, num, ret, arg1, arg2,
13187 arg3, arg4, arg5, arg6);
13190 record_syscall_return(cpu, num, ret);
13191 return ret;