target-arm: Rename CPU types
[qemu/agraf.git] / trace / simple.c
blobce17d64bd7b9e9c96bbfa2cc07620bc68fe4d6ba
1 /*
2 * Simple trace backend
4 * Copyright IBM, Corp. 2010
6 * This work is licensed under the terms of the GNU GPL, version 2. See
7 * the COPYING file in the top-level directory.
9 */
11 #include <stdlib.h>
12 #include <stdint.h>
13 #include <stdio.h>
14 #include <time.h>
15 #ifndef _WIN32
16 #include <signal.h>
17 #include <pthread.h>
18 #endif
19 #include "qemu/timer.h"
20 #include "trace.h"
21 #include "trace/control.h"
23 /** Trace file header event ID */
24 #define HEADER_EVENT_ID (~(uint64_t)0) /* avoids conflicting with TraceEventIDs */
26 /** Trace file magic number */
27 #define HEADER_MAGIC 0xf2b177cb0aa429b4ULL
29 /** Trace file version number, bump if format changes */
30 #define HEADER_VERSION 2
32 /** Records were dropped event ID */
33 #define DROPPED_EVENT_ID (~(uint64_t)0 - 1)
35 /** Trace record is valid */
36 #define TRACE_RECORD_VALID ((uint64_t)1 << 63)
39 * Trace records are written out by a dedicated thread. The thread waits for
40 * records to become available, writes them out, and then waits again.
42 static GStaticMutex trace_lock = G_STATIC_MUTEX_INIT;
43 static GCond *trace_available_cond;
44 static GCond *trace_empty_cond;
45 static bool trace_available;
46 static bool trace_writeout_enabled;
48 enum {
49 TRACE_BUF_LEN = 4096 * 64,
50 TRACE_BUF_FLUSH_THRESHOLD = TRACE_BUF_LEN / 4,
53 uint8_t trace_buf[TRACE_BUF_LEN];
54 static unsigned int trace_idx;
55 static unsigned int writeout_idx;
56 static uint64_t dropped_events;
57 static FILE *trace_fp;
58 static char *trace_file_name;
60 /* * Trace buffer entry */
61 typedef struct {
62 uint64_t event; /* TraceEventID */
63 uint64_t timestamp_ns;
64 uint32_t length; /* in bytes */
65 uint32_t reserved; /* unused */
66 uint8_t arguments[];
67 } TraceRecord;
69 typedef struct {
70 uint64_t header_event_id; /* HEADER_EVENT_ID */
71 uint64_t header_magic; /* HEADER_MAGIC */
72 uint64_t header_version; /* HEADER_VERSION */
73 } TraceLogHeader;
76 static void read_from_buffer(unsigned int idx, void *dataptr, size_t size);
77 static unsigned int write_to_buffer(unsigned int idx, void *dataptr, size_t size);
79 static void clear_buffer_range(unsigned int idx, size_t len)
81 uint32_t num = 0;
82 while (num < len) {
83 if (idx >= TRACE_BUF_LEN) {
84 idx = idx % TRACE_BUF_LEN;
86 trace_buf[idx++] = 0;
87 num++;
90 /**
91 * Read a trace record from the trace buffer
93 * @idx Trace buffer index
94 * @record Trace record to fill
96 * Returns false if the record is not valid.
98 static bool get_trace_record(unsigned int idx, TraceRecord **recordptr)
100 uint64_t event_flag = 0;
101 TraceRecord record;
102 /* read the event flag to see if its a valid record */
103 read_from_buffer(idx, &record, sizeof(event_flag));
105 if (!(record.event & TRACE_RECORD_VALID)) {
106 return false;
109 smp_rmb(); /* read memory barrier before accessing record */
110 /* read the record header to know record length */
111 read_from_buffer(idx, &record, sizeof(TraceRecord));
112 *recordptr = malloc(record.length); /* dont use g_malloc, can deadlock when traced */
113 /* make a copy of record to avoid being overwritten */
114 read_from_buffer(idx, *recordptr, record.length);
115 smp_rmb(); /* memory barrier before clearing valid flag */
116 (*recordptr)->event &= ~TRACE_RECORD_VALID;
117 /* clear the trace buffer range for consumed record otherwise any byte
118 * with its MSB set may be considered as a valid event id when the writer
119 * thread crosses this range of buffer again.
121 clear_buffer_range(idx, record.length);
122 return true;
126 * Kick writeout thread
128 * @wait Whether to wait for writeout thread to complete
130 static void flush_trace_file(bool wait)
132 g_static_mutex_lock(&trace_lock);
133 trace_available = true;
134 g_cond_signal(trace_available_cond);
136 if (wait) {
137 g_cond_wait(trace_empty_cond, g_static_mutex_get_mutex(&trace_lock));
140 g_static_mutex_unlock(&trace_lock);
143 static void wait_for_trace_records_available(void)
145 g_static_mutex_lock(&trace_lock);
146 while (!(trace_available && trace_writeout_enabled)) {
147 g_cond_signal(trace_empty_cond);
148 g_cond_wait(trace_available_cond,
149 g_static_mutex_get_mutex(&trace_lock));
151 trace_available = false;
152 g_static_mutex_unlock(&trace_lock);
155 static gpointer writeout_thread(gpointer opaque)
157 TraceRecord *recordptr;
158 union {
159 TraceRecord rec;
160 uint8_t bytes[sizeof(TraceRecord) + sizeof(uint64_t)];
161 } dropped;
162 unsigned int idx = 0;
163 uint64_t dropped_count;
164 size_t unused __attribute__ ((unused));
166 for (;;) {
167 wait_for_trace_records_available();
169 if (dropped_events) {
170 dropped.rec.event = DROPPED_EVENT_ID,
171 dropped.rec.timestamp_ns = get_clock();
172 dropped.rec.length = sizeof(TraceRecord) + sizeof(dropped_events),
173 dropped.rec.reserved = 0;
174 while (1) {
175 dropped_count = dropped_events;
176 if (g_atomic_int_compare_and_exchange((gint *)&dropped_events,
177 dropped_count, 0)) {
178 break;
181 memcpy(dropped.rec.arguments, &dropped_count, sizeof(uint64_t));
182 unused = fwrite(&dropped.rec, dropped.rec.length, 1, trace_fp);
185 while (get_trace_record(idx, &recordptr)) {
186 unused = fwrite(recordptr, recordptr->length, 1, trace_fp);
187 writeout_idx += recordptr->length;
188 free(recordptr); /* dont use g_free, can deadlock when traced */
189 idx = writeout_idx % TRACE_BUF_LEN;
192 fflush(trace_fp);
194 return NULL;
197 void trace_record_write_u64(TraceBufferRecord *rec, uint64_t val)
199 rec->rec_off = write_to_buffer(rec->rec_off, &val, sizeof(uint64_t));
202 void trace_record_write_str(TraceBufferRecord *rec, const char *s, uint32_t slen)
204 /* Write string length first */
205 rec->rec_off = write_to_buffer(rec->rec_off, &slen, sizeof(slen));
206 /* Write actual string now */
207 rec->rec_off = write_to_buffer(rec->rec_off, (void*)s, slen);
210 int trace_record_start(TraceBufferRecord *rec, TraceEventID event, size_t datasize)
212 unsigned int idx, rec_off, old_idx, new_idx;
213 uint32_t rec_len = sizeof(TraceRecord) + datasize;
214 uint64_t timestamp_ns = get_clock();
216 while (1) {
217 old_idx = trace_idx;
218 smp_rmb();
219 new_idx = old_idx + rec_len;
221 if (new_idx - writeout_idx > TRACE_BUF_LEN) {
222 /* Trace Buffer Full, Event dropped ! */
223 g_atomic_int_inc((gint *)&dropped_events);
224 return -ENOSPC;
227 if (g_atomic_int_compare_and_exchange((gint *)&trace_idx,
228 old_idx, new_idx)) {
229 break;
233 idx = old_idx % TRACE_BUF_LEN;
235 rec_off = idx;
236 rec_off = write_to_buffer(rec_off, &event, sizeof(event));
237 rec_off = write_to_buffer(rec_off, &timestamp_ns, sizeof(timestamp_ns));
238 rec_off = write_to_buffer(rec_off, &rec_len, sizeof(rec_len));
240 rec->tbuf_idx = idx;
241 rec->rec_off = (idx + sizeof(TraceRecord)) % TRACE_BUF_LEN;
242 return 0;
245 static void read_from_buffer(unsigned int idx, void *dataptr, size_t size)
247 uint8_t *data_ptr = dataptr;
248 uint32_t x = 0;
249 while (x < size) {
250 if (idx >= TRACE_BUF_LEN) {
251 idx = idx % TRACE_BUF_LEN;
253 data_ptr[x++] = trace_buf[idx++];
257 static unsigned int write_to_buffer(unsigned int idx, void *dataptr, size_t size)
259 uint8_t *data_ptr = dataptr;
260 uint32_t x = 0;
261 while (x < size) {
262 if (idx >= TRACE_BUF_LEN) {
263 idx = idx % TRACE_BUF_LEN;
265 trace_buf[idx++] = data_ptr[x++];
267 return idx; /* most callers wants to know where to write next */
270 void trace_record_finish(TraceBufferRecord *rec)
272 TraceRecord record;
273 read_from_buffer(rec->tbuf_idx, &record, sizeof(TraceRecord));
274 smp_wmb(); /* write barrier before marking as valid */
275 record.event |= TRACE_RECORD_VALID;
276 write_to_buffer(rec->tbuf_idx, &record, sizeof(TraceRecord));
278 if ((trace_idx - writeout_idx) > TRACE_BUF_FLUSH_THRESHOLD) {
279 flush_trace_file(false);
283 void st_set_trace_file_enabled(bool enable)
285 if (enable == !!trace_fp) {
286 return; /* no change */
289 /* Halt trace writeout */
290 flush_trace_file(true);
291 trace_writeout_enabled = false;
292 flush_trace_file(true);
294 if (enable) {
295 static const TraceLogHeader header = {
296 .header_event_id = HEADER_EVENT_ID,
297 .header_magic = HEADER_MAGIC,
298 /* Older log readers will check for version at next location */
299 .header_version = HEADER_VERSION,
302 trace_fp = fopen(trace_file_name, "wb");
303 if (!trace_fp) {
304 return;
307 if (fwrite(&header, sizeof header, 1, trace_fp) != 1) {
308 fclose(trace_fp);
309 trace_fp = NULL;
310 return;
313 /* Resume trace writeout */
314 trace_writeout_enabled = true;
315 flush_trace_file(false);
316 } else {
317 fclose(trace_fp);
318 trace_fp = NULL;
323 * Set the name of a trace file
325 * @file The trace file name or NULL for the default name-<pid> set at
326 * config time
328 bool st_set_trace_file(const char *file)
330 st_set_trace_file_enabled(false);
332 g_free(trace_file_name);
334 if (!file) {
335 trace_file_name = g_strdup_printf(CONFIG_TRACE_FILE, getpid());
336 } else {
337 trace_file_name = g_strdup_printf("%s", file);
340 st_set_trace_file_enabled(true);
341 return true;
344 void st_print_trace_file_status(FILE *stream, int (*stream_printf)(FILE *stream, const char *fmt, ...))
346 stream_printf(stream, "Trace file \"%s\" %s.\n",
347 trace_file_name, trace_fp ? "on" : "off");
350 void st_flush_trace_buffer(void)
352 flush_trace_file(true);
355 void trace_print_events(FILE *stream, fprintf_function stream_printf)
357 unsigned int i;
359 for (i = 0; i < NR_TRACE_EVENTS; i++) {
360 stream_printf(stream, "%s [Event ID %u] : state %u\n",
361 trace_list[i].tp_name, i, trace_list[i].state);
365 bool trace_event_set_state(const char *name, bool state)
367 unsigned int i;
368 unsigned int len;
369 bool wildcard = false;
370 bool matched = false;
372 len = strlen(name);
373 if (len > 0 && name[len - 1] == '*') {
374 wildcard = true;
375 len -= 1;
377 for (i = 0; i < NR_TRACE_EVENTS; i++) {
378 if (wildcard) {
379 if (!strncmp(trace_list[i].tp_name, name, len)) {
380 trace_list[i].state = state;
381 matched = true;
383 continue;
385 if (!strcmp(trace_list[i].tp_name, name)) {
386 trace_list[i].state = state;
387 return true;
390 return matched;
393 /* Helper function to create a thread with signals blocked. Use glib's
394 * portable threads since QEMU abstractions cannot be used due to reentrancy in
395 * the tracer. Also note the signal masking on POSIX hosts so that the thread
396 * does not steal signals when the rest of the program wants them blocked.
398 static GThread *trace_thread_create(GThreadFunc fn)
400 GThread *thread;
401 #ifndef _WIN32
402 sigset_t set, oldset;
404 sigfillset(&set);
405 pthread_sigmask(SIG_SETMASK, &set, &oldset);
406 #endif
407 thread = g_thread_create(fn, NULL, FALSE, NULL);
408 #ifndef _WIN32
409 pthread_sigmask(SIG_SETMASK, &oldset, NULL);
410 #endif
412 return thread;
415 bool trace_backend_init(const char *events, const char *file)
417 GThread *thread;
419 if (!g_thread_supported()) {
420 #if !GLIB_CHECK_VERSION(2, 31, 0)
421 g_thread_init(NULL);
422 #else
423 fprintf(stderr, "glib threading failed to initialize.\n");
424 exit(1);
425 #endif
428 trace_available_cond = g_cond_new();
429 trace_empty_cond = g_cond_new();
431 thread = trace_thread_create(writeout_thread);
432 if (!thread) {
433 fprintf(stderr, "warning: unable to initialize simple trace backend\n");
434 return false;
437 atexit(st_flush_trace_buffer);
438 trace_backend_init_events(events);
439 st_set_trace_file(file);
440 return true;