rust: Implement TARGET_RUST_OS_INFO for *-*-darwin*
[official-gcc.git] / libsanitizer / sanitizer_common / sanitizer_stoptheworld_netbsd_libcdep.cpp
blob701db72619a3d47c4f44b7096c6a8775be460f86
1 //===-- sanitizer_stoptheworld_netbsd_libcdep.cpp -------------------------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // See sanitizer_stoptheworld.h for details.
10 // This implementation was inspired by Markus Gutschke's linuxthreads.cc.
12 // This is a NetBSD variation of Linux stoptheworld implementation
13 // See sanitizer_stoptheworld_linux_libcdep.cpp for code comments.
15 //===----------------------------------------------------------------------===//
17 #include "sanitizer_platform.h"
19 #if SANITIZER_NETBSD
21 #include "sanitizer_stoptheworld.h"
23 #include "sanitizer_atomic.h"
24 #include "sanitizer_platform_limits_posix.h"
26 #include <sys/types.h>
28 #include <sys/ptrace.h>
29 #include <sys/uio.h>
30 #include <sys/wait.h>
32 #include <machine/reg.h>
34 #include <elf.h>
35 #include <errno.h>
36 #include <sched.h>
37 #include <signal.h>
38 #include <stddef.h>
40 #define internal_sigaction_norestorer internal_sigaction
42 #include "sanitizer_common.h"
43 #include "sanitizer_flags.h"
44 #include "sanitizer_libc.h"
45 #include "sanitizer_linux.h"
46 #include "sanitizer_mutex.h"
47 #include "sanitizer_placement_new.h"
49 namespace __sanitizer {
51 class SuspendedThreadsListNetBSD final : public SuspendedThreadsList {
52 public:
53 SuspendedThreadsListNetBSD() { thread_ids_.reserve(1024); }
55 tid_t GetThreadID(uptr index) const;
56 uptr ThreadCount() const;
57 bool ContainsTid(tid_t thread_id) const;
58 void Append(tid_t tid);
60 PtraceRegistersStatus GetRegistersAndSP(uptr index,
61 InternalMmapVector<uptr> *buffer,
62 uptr *sp) const;
64 private:
65 InternalMmapVector<tid_t> thread_ids_;
68 struct TracerThreadArgument {
69 StopTheWorldCallback callback;
70 void *callback_argument;
71 Mutex mutex;
72 atomic_uintptr_t done;
73 uptr parent_pid;
76 class ThreadSuspender {
77 public:
78 explicit ThreadSuspender(pid_t pid, TracerThreadArgument *arg)
79 : arg(arg), pid_(pid) {
80 CHECK_GE(pid, 0);
82 bool SuspendAllThreads();
83 void ResumeAllThreads();
84 void KillAllThreads();
85 SuspendedThreadsListNetBSD &suspended_threads_list() {
86 return suspended_threads_list_;
88 TracerThreadArgument *arg;
90 private:
91 SuspendedThreadsListNetBSD suspended_threads_list_;
92 pid_t pid_;
95 void ThreadSuspender::ResumeAllThreads() {
96 int pterrno;
97 if (!internal_iserror(internal_ptrace(PT_DETACH, pid_, (void *)(uptr)1, 0),
98 &pterrno)) {
99 VReport(2, "Detached from process %d.\n", pid_);
100 } else {
101 VReport(1, "Could not detach from process %d (errno %d).\n", pid_, pterrno);
105 void ThreadSuspender::KillAllThreads() {
106 internal_ptrace(PT_KILL, pid_, nullptr, 0);
109 bool ThreadSuspender::SuspendAllThreads() {
110 int pterrno;
111 if (internal_iserror(internal_ptrace(PT_ATTACH, pid_, nullptr, 0),
112 &pterrno)) {
113 Printf("Could not attach to process %d (errno %d).\n", pid_, pterrno);
114 return false;
117 int status;
118 uptr waitpid_status;
119 HANDLE_EINTR(waitpid_status, internal_waitpid(pid_, &status, 0));
121 VReport(2, "Attached to process %d.\n", pid_);
123 #ifdef PT_LWPNEXT
124 struct ptrace_lwpstatus pl;
125 int op = PT_LWPNEXT;
126 #else
127 struct ptrace_lwpinfo pl;
128 int op = PT_LWPINFO;
129 #endif
131 pl.pl_lwpid = 0;
133 int val;
134 while ((val = internal_ptrace(op, pid_, (void *)&pl, sizeof(pl))) != -1 &&
135 pl.pl_lwpid != 0) {
136 suspended_threads_list_.Append(pl.pl_lwpid);
137 VReport(2, "Appended thread %d in process %d.\n", pl.pl_lwpid, pid_);
139 return true;
142 // Pointer to the ThreadSuspender instance for use in signal handler.
143 static ThreadSuspender *thread_suspender_instance = nullptr;
145 // Synchronous signals that should not be blocked.
146 static const int kSyncSignals[] = {SIGABRT, SIGILL, SIGFPE, SIGSEGV,
147 SIGBUS, SIGXCPU, SIGXFSZ};
149 static void TracerThreadDieCallback() {
150 ThreadSuspender *inst = thread_suspender_instance;
151 if (inst && stoptheworld_tracer_pid == internal_getpid()) {
152 inst->KillAllThreads();
153 thread_suspender_instance = nullptr;
157 // Signal handler to wake up suspended threads when the tracer thread dies.
158 static void TracerThreadSignalHandler(int signum, __sanitizer_siginfo *siginfo,
159 void *uctx) {
160 SignalContext ctx(siginfo, uctx);
161 Printf("Tracer caught signal %d: addr=0x%zx pc=0x%zx sp=0x%zx\n", signum,
162 ctx.addr, ctx.pc, ctx.sp);
163 ThreadSuspender *inst = thread_suspender_instance;
164 if (inst) {
165 if (signum == SIGABRT)
166 inst->KillAllThreads();
167 else
168 inst->ResumeAllThreads();
169 RAW_CHECK(RemoveDieCallback(TracerThreadDieCallback));
170 thread_suspender_instance = nullptr;
171 atomic_store(&inst->arg->done, 1, memory_order_relaxed);
173 internal__exit((signum == SIGABRT) ? 1 : 2);
176 // Size of alternative stack for signal handlers in the tracer thread.
177 static const int kHandlerStackSize = 8192;
179 // This function will be run as a cloned task.
180 static int TracerThread(void *argument) {
181 TracerThreadArgument *tracer_thread_argument =
182 (TracerThreadArgument *)argument;
184 // Check if parent is already dead.
185 if (internal_getppid() != tracer_thread_argument->parent_pid)
186 internal__exit(4);
188 // Wait for the parent thread to finish preparations.
189 tracer_thread_argument->mutex.Lock();
190 tracer_thread_argument->mutex.Unlock();
192 RAW_CHECK(AddDieCallback(TracerThreadDieCallback));
194 ThreadSuspender thread_suspender(internal_getppid(), tracer_thread_argument);
195 // Global pointer for the signal handler.
196 thread_suspender_instance = &thread_suspender;
198 // Alternate stack for signal handling.
199 InternalMmapVector<char> handler_stack_memory(kHandlerStackSize);
200 stack_t handler_stack;
201 internal_memset(&handler_stack, 0, sizeof(handler_stack));
202 handler_stack.ss_sp = handler_stack_memory.data();
203 handler_stack.ss_size = kHandlerStackSize;
204 internal_sigaltstack(&handler_stack, nullptr);
206 // Install our handler for synchronous signals. Other signals should be
207 // blocked by the mask we inherited from the parent thread.
208 for (uptr i = 0; i < ARRAY_SIZE(kSyncSignals); i++) {
209 __sanitizer_sigaction act;
210 internal_memset(&act, 0, sizeof(act));
211 act.sigaction = TracerThreadSignalHandler;
212 act.sa_flags = SA_ONSTACK | SA_SIGINFO;
213 internal_sigaction_norestorer(kSyncSignals[i], &act, 0);
216 int exit_code = 0;
217 if (!thread_suspender.SuspendAllThreads()) {
218 VReport(1, "Failed suspending threads.\n");
219 exit_code = 3;
220 } else {
221 tracer_thread_argument->callback(thread_suspender.suspended_threads_list(),
222 tracer_thread_argument->callback_argument);
223 thread_suspender.ResumeAllThreads();
224 exit_code = 0;
226 RAW_CHECK(RemoveDieCallback(TracerThreadDieCallback));
227 thread_suspender_instance = nullptr;
228 atomic_store(&tracer_thread_argument->done, 1, memory_order_relaxed);
229 return exit_code;
232 class ScopedStackSpaceWithGuard {
233 public:
234 explicit ScopedStackSpaceWithGuard(uptr stack_size) {
235 stack_size_ = stack_size;
236 guard_size_ = GetPageSizeCached();
237 // FIXME: Omitting MAP_STACK here works in current kernels but might break
238 // in the future.
239 guard_start_ =
240 (uptr)MmapOrDie(stack_size_ + guard_size_, "ScopedStackWithGuard");
241 CHECK(MprotectNoAccess((uptr)guard_start_, guard_size_));
243 ~ScopedStackSpaceWithGuard() {
244 UnmapOrDie((void *)guard_start_, stack_size_ + guard_size_);
246 void *Bottom() const {
247 return (void *)(guard_start_ + stack_size_ + guard_size_);
250 private:
251 uptr stack_size_;
252 uptr guard_size_;
253 uptr guard_start_;
256 static __sanitizer_sigset_t blocked_sigset;
257 static __sanitizer_sigset_t old_sigset;
259 struct ScopedSetTracerPID {
260 explicit ScopedSetTracerPID(uptr tracer_pid) {
261 stoptheworld_tracer_pid = tracer_pid;
262 stoptheworld_tracer_ppid = internal_getpid();
264 ~ScopedSetTracerPID() {
265 stoptheworld_tracer_pid = 0;
266 stoptheworld_tracer_ppid = 0;
270 void StopTheWorld(StopTheWorldCallback callback, void *argument) {
271 // Prepare the arguments for TracerThread.
272 struct TracerThreadArgument tracer_thread_argument;
273 tracer_thread_argument.callback = callback;
274 tracer_thread_argument.callback_argument = argument;
275 tracer_thread_argument.parent_pid = internal_getpid();
276 atomic_store(&tracer_thread_argument.done, 0, memory_order_relaxed);
277 const uptr kTracerStackSize = 2 * 1024 * 1024;
278 ScopedStackSpaceWithGuard tracer_stack(kTracerStackSize);
280 tracer_thread_argument.mutex.Lock();
282 internal_sigfillset(&blocked_sigset);
283 for (uptr i = 0; i < ARRAY_SIZE(kSyncSignals); i++)
284 internal_sigdelset(&blocked_sigset, kSyncSignals[i]);
285 int rv = internal_sigprocmask(SIG_BLOCK, &blocked_sigset, &old_sigset);
286 CHECK_EQ(rv, 0);
287 uptr tracer_pid = internal_clone(TracerThread, tracer_stack.Bottom(),
288 CLONE_VM | CLONE_FS | CLONE_FILES,
289 &tracer_thread_argument);
290 internal_sigprocmask(SIG_SETMASK, &old_sigset, 0);
291 int local_errno = 0;
292 if (internal_iserror(tracer_pid, &local_errno)) {
293 VReport(1, "Failed spawning a tracer thread (errno %d).\n", local_errno);
294 tracer_thread_argument.mutex.Unlock();
295 } else {
296 ScopedSetTracerPID scoped_set_tracer_pid(tracer_pid);
298 tracer_thread_argument.mutex.Unlock();
300 while (atomic_load(&tracer_thread_argument.done, memory_order_relaxed) == 0)
301 sched_yield();
303 for (;;) {
304 uptr waitpid_status = internal_waitpid(tracer_pid, nullptr, __WALL);
305 if (!internal_iserror(waitpid_status, &local_errno))
306 break;
307 if (local_errno == EINTR)
308 continue;
309 VReport(1, "Waiting on the tracer thread failed (errno %d).\n",
310 local_errno);
311 break;
316 tid_t SuspendedThreadsListNetBSD::GetThreadID(uptr index) const {
317 CHECK_LT(index, thread_ids_.size());
318 return thread_ids_[index];
321 uptr SuspendedThreadsListNetBSD::ThreadCount() const {
322 return thread_ids_.size();
325 bool SuspendedThreadsListNetBSD::ContainsTid(tid_t thread_id) const {
326 for (uptr i = 0; i < thread_ids_.size(); i++) {
327 if (thread_ids_[i] == thread_id)
328 return true;
330 return false;
333 void SuspendedThreadsListNetBSD::Append(tid_t tid) {
334 thread_ids_.push_back(tid);
337 PtraceRegistersStatus SuspendedThreadsListNetBSD::GetRegistersAndSP(
338 uptr index, InternalMmapVector<uptr> *buffer, uptr *sp) const {
339 lwpid_t tid = GetThreadID(index);
340 pid_t ppid = internal_getppid();
341 struct reg regs;
342 int pterrno;
343 bool isErr =
344 internal_iserror(internal_ptrace(PT_GETREGS, ppid, &regs, tid), &pterrno);
345 if (isErr) {
346 VReport(1,
347 "Could not get registers from process %d thread %d (errno %d).\n",
348 ppid, tid, pterrno);
349 return pterrno == ESRCH ? REGISTERS_UNAVAILABLE_FATAL
350 : REGISTERS_UNAVAILABLE;
353 *sp = PTRACE_REG_SP(&regs);
354 buffer->resize(RoundUpTo(sizeof(regs), sizeof(uptr)) / sizeof(uptr));
355 internal_memcpy(buffer->data(), &regs, sizeof(regs));
357 return REGISTERS_AVAILABLE;
360 } // namespace __sanitizer
362 #endif