2 * QEMU Plugin Core code
4 * This is the core code that deals with injecting instrumentation into the code
6 * Copyright (C) 2017, Emilio G. Cota <cota@braap.org>
7 * Copyright (C) 2019, Linaro
9 * License: GNU GPL, version 2 or later.
10 * See the COPYING file in the top-level directory.
12 * SPDX-License-Identifier: GPL-2.0-or-later
14 #include "qemu/osdep.h"
15 #include "qemu/error-report.h"
16 #include "qemu/config-file.h"
17 #include "qapi/error.h"
18 #include "qemu/lockable.h"
19 #include "qemu/option.h"
20 #include "qemu/rcu_queue.h"
21 #include "qemu/xxhash.h"
23 #include "hw/core/cpu.h"
24 #include "exec/cpu-common.h"
26 #include "exec/exec-all.h"
27 #include "exec/tb-flush.h"
28 #include "exec/helper-proto.h"
30 #include "tcg/tcg-op.h"
32 #include "qemu/compiler.h"
34 struct qemu_plugin_cb
{
35 struct qemu_plugin_ctx
*ctx
;
36 union qemu_plugin_cb_sig f
;
38 QLIST_ENTRY(qemu_plugin_cb
) entry
;
41 struct qemu_plugin_state plugin
;
43 struct qemu_plugin_ctx
*plugin_id_to_ctx_locked(qemu_plugin_id_t id
)
45 struct qemu_plugin_ctx
*ctx
;
46 qemu_plugin_id_t
*id_p
;
48 id_p
= g_hash_table_lookup(plugin
.id_ht
, &id
);
49 ctx
= container_of(id_p
, struct qemu_plugin_ctx
, id
);
51 error_report("plugin: invalid plugin id %" PRIu64
, id
);
57 static void plugin_cpu_update__async(CPUState
*cpu
, run_on_cpu_data data
)
59 bitmap_copy(cpu
->plugin_mask
, &data
.host_ulong
, QEMU_PLUGIN_EV_MAX
);
60 tcg_flush_jmp_cache(cpu
);
63 static void plugin_cpu_update__locked(gpointer k
, gpointer v
, gpointer udata
)
65 CPUState
*cpu
= container_of(k
, CPUState
, cpu_index
);
66 run_on_cpu_data mask
= RUN_ON_CPU_HOST_ULONG(*plugin
.mask
);
69 async_run_on_cpu(cpu
, plugin_cpu_update__async
, mask
);
71 plugin_cpu_update__async(cpu
, mask
);
75 void plugin_unregister_cb__locked(struct qemu_plugin_ctx
*ctx
,
76 enum qemu_plugin_event ev
)
78 struct qemu_plugin_cb
*cb
= ctx
->callbacks
[ev
];
83 QLIST_REMOVE_RCU(cb
, entry
);
85 ctx
->callbacks
[ev
] = NULL
;
86 if (QLIST_EMPTY_RCU(&plugin
.cb_lists
[ev
])) {
87 clear_bit(ev
, plugin
.mask
);
88 g_hash_table_foreach(plugin
.cpu_ht
, plugin_cpu_update__locked
, NULL
);
94 * The callback function has been loaded from an external library so we do not
95 * have type information
98 static void plugin_vcpu_cb__simple(CPUState
*cpu
, enum qemu_plugin_event ev
)
100 struct qemu_plugin_cb
*cb
, *next
;
103 case QEMU_PLUGIN_EV_VCPU_INIT
:
104 case QEMU_PLUGIN_EV_VCPU_EXIT
:
105 case QEMU_PLUGIN_EV_VCPU_IDLE
:
106 case QEMU_PLUGIN_EV_VCPU_RESUME
:
107 /* iterate safely; plugins might uninstall themselves at any time */
108 QLIST_FOREACH_SAFE_RCU(cb
, &plugin
.cb_lists
[ev
], entry
, next
) {
109 qemu_plugin_vcpu_simple_cb_t func
= cb
->f
.vcpu_simple
;
111 func(cb
->ctx
->id
, cpu
->cpu_index
);
115 g_assert_not_reached();
120 * Disable CFI checks.
121 * The callback function has been loaded from an external library so we do not
122 * have type information
125 static void plugin_cb__simple(enum qemu_plugin_event ev
)
127 struct qemu_plugin_cb
*cb
, *next
;
130 case QEMU_PLUGIN_EV_FLUSH
:
131 QLIST_FOREACH_SAFE_RCU(cb
, &plugin
.cb_lists
[ev
], entry
, next
) {
132 qemu_plugin_simple_cb_t func
= cb
->f
.simple
;
138 g_assert_not_reached();
143 * Disable CFI checks.
144 * The callback function has been loaded from an external library so we do not
145 * have type information
148 static void plugin_cb__udata(enum qemu_plugin_event ev
)
150 struct qemu_plugin_cb
*cb
, *next
;
153 case QEMU_PLUGIN_EV_ATEXIT
:
154 QLIST_FOREACH_SAFE_RCU(cb
, &plugin
.cb_lists
[ev
], entry
, next
) {
155 qemu_plugin_udata_cb_t func
= cb
->f
.udata
;
157 func(cb
->ctx
->id
, cb
->udata
);
161 g_assert_not_reached();
166 do_plugin_register_cb(qemu_plugin_id_t id
, enum qemu_plugin_event ev
,
167 void *func
, void *udata
)
169 struct qemu_plugin_ctx
*ctx
;
171 QEMU_LOCK_GUARD(&plugin
.lock
);
172 ctx
= plugin_id_to_ctx_locked(id
);
173 /* if the plugin is on its way out, ignore this request */
174 if (unlikely(ctx
->uninstalling
)) {
178 struct qemu_plugin_cb
*cb
= ctx
->callbacks
[ev
];
181 cb
->f
.generic
= func
;
184 cb
= g_new(struct qemu_plugin_cb
, 1);
186 cb
->f
.generic
= func
;
188 ctx
->callbacks
[ev
] = cb
;
189 QLIST_INSERT_HEAD_RCU(&plugin
.cb_lists
[ev
], cb
, entry
);
190 if (!test_bit(ev
, plugin
.mask
)) {
191 set_bit(ev
, plugin
.mask
);
192 g_hash_table_foreach(plugin
.cpu_ht
, plugin_cpu_update__locked
,
197 plugin_unregister_cb__locked(ctx
, ev
);
201 void plugin_register_cb(qemu_plugin_id_t id
, enum qemu_plugin_event ev
,
204 do_plugin_register_cb(id
, ev
, func
, NULL
);
208 plugin_register_cb_udata(qemu_plugin_id_t id
, enum qemu_plugin_event ev
,
209 void *func
, void *udata
)
211 do_plugin_register_cb(id
, ev
, func
, udata
);
214 void qemu_plugin_vcpu_init_hook(CPUState
*cpu
)
218 qemu_rec_mutex_lock(&plugin
.lock
);
219 plugin_cpu_update__locked(&cpu
->cpu_index
, NULL
, NULL
);
220 success
= g_hash_table_insert(plugin
.cpu_ht
, &cpu
->cpu_index
,
223 qemu_rec_mutex_unlock(&plugin
.lock
);
225 plugin_vcpu_cb__simple(cpu
, QEMU_PLUGIN_EV_VCPU_INIT
);
228 void qemu_plugin_vcpu_exit_hook(CPUState
*cpu
)
232 plugin_vcpu_cb__simple(cpu
, QEMU_PLUGIN_EV_VCPU_EXIT
);
234 qemu_rec_mutex_lock(&plugin
.lock
);
235 success
= g_hash_table_remove(plugin
.cpu_ht
, &cpu
->cpu_index
);
237 qemu_rec_mutex_unlock(&plugin
.lock
);
240 struct plugin_for_each_args
{
241 struct qemu_plugin_ctx
*ctx
;
242 qemu_plugin_vcpu_simple_cb_t cb
;
245 static void plugin_vcpu_for_each(gpointer k
, gpointer v
, gpointer udata
)
247 struct plugin_for_each_args
*args
= udata
;
248 int cpu_index
= *(int *)k
;
250 args
->cb(args
->ctx
->id
, cpu_index
);
253 void qemu_plugin_vcpu_for_each(qemu_plugin_id_t id
,
254 qemu_plugin_vcpu_simple_cb_t cb
)
256 struct plugin_for_each_args args
;
261 qemu_rec_mutex_lock(&plugin
.lock
);
262 args
.ctx
= plugin_id_to_ctx_locked(id
);
264 g_hash_table_foreach(plugin
.cpu_ht
, plugin_vcpu_for_each
, &args
);
265 qemu_rec_mutex_unlock(&plugin
.lock
);
268 /* Allocate and return a callback record */
269 static struct qemu_plugin_dyn_cb
*plugin_get_dyn_cb(GArray
**arr
)
274 cbs
= g_array_sized_new(false, false,
275 sizeof(struct qemu_plugin_dyn_cb
), 1);
279 g_array_set_size(cbs
, cbs
->len
+ 1);
280 return &g_array_index(cbs
, struct qemu_plugin_dyn_cb
, cbs
->len
- 1);
283 void plugin_register_inline_op(GArray
**arr
,
284 enum qemu_plugin_mem_rw rw
,
285 enum qemu_plugin_op op
, void *ptr
,
288 struct qemu_plugin_dyn_cb
*dyn_cb
;
290 dyn_cb
= plugin_get_dyn_cb(arr
);
292 dyn_cb
->type
= PLUGIN_CB_INLINE
;
294 dyn_cb
->inline_insn
.op
= op
;
295 dyn_cb
->inline_insn
.imm
= imm
;
298 void plugin_register_dyn_cb__udata(GArray
**arr
,
299 qemu_plugin_vcpu_udata_cb_t cb
,
300 enum qemu_plugin_cb_flags flags
,
303 struct qemu_plugin_dyn_cb
*dyn_cb
= plugin_get_dyn_cb(arr
);
305 dyn_cb
->userp
= udata
;
306 /* Note flags are discarded as unused. */
307 dyn_cb
->f
.vcpu_udata
= cb
;
308 dyn_cb
->type
= PLUGIN_CB_REGULAR
;
311 void plugin_register_vcpu_mem_cb(GArray
**arr
,
313 enum qemu_plugin_cb_flags flags
,
314 enum qemu_plugin_mem_rw rw
,
317 struct qemu_plugin_dyn_cb
*dyn_cb
;
319 dyn_cb
= plugin_get_dyn_cb(arr
);
320 dyn_cb
->userp
= udata
;
321 /* Note flags are discarded as unused. */
322 dyn_cb
->type
= PLUGIN_CB_REGULAR
;
324 dyn_cb
->f
.generic
= cb
;
328 * Disable CFI checks.
329 * The callback function has been loaded from an external library so we do not
330 * have type information
333 void qemu_plugin_tb_trans_cb(CPUState
*cpu
, struct qemu_plugin_tb
*tb
)
335 struct qemu_plugin_cb
*cb
, *next
;
336 enum qemu_plugin_event ev
= QEMU_PLUGIN_EV_VCPU_TB_TRANS
;
338 /* no plugin_mask check here; caller should have checked */
340 QLIST_FOREACH_SAFE_RCU(cb
, &plugin
.cb_lists
[ev
], entry
, next
) {
341 qemu_plugin_vcpu_tb_trans_cb_t func
= cb
->f
.vcpu_tb_trans
;
343 func(cb
->ctx
->id
, tb
);
348 * Disable CFI checks.
349 * The callback function has been loaded from an external library so we do not
350 * have type information
354 qemu_plugin_vcpu_syscall(CPUState
*cpu
, int64_t num
, uint64_t a1
, uint64_t a2
,
355 uint64_t a3
, uint64_t a4
, uint64_t a5
,
356 uint64_t a6
, uint64_t a7
, uint64_t a8
)
358 struct qemu_plugin_cb
*cb
, *next
;
359 enum qemu_plugin_event ev
= QEMU_PLUGIN_EV_VCPU_SYSCALL
;
361 if (!test_bit(ev
, cpu
->plugin_mask
)) {
365 QLIST_FOREACH_SAFE_RCU(cb
, &plugin
.cb_lists
[ev
], entry
, next
) {
366 qemu_plugin_vcpu_syscall_cb_t func
= cb
->f
.vcpu_syscall
;
368 func(cb
->ctx
->id
, cpu
->cpu_index
, num
, a1
, a2
, a3
, a4
, a5
, a6
, a7
, a8
);
373 * Disable CFI checks.
374 * The callback function has been loaded from an external library so we do not
375 * have type information
378 void qemu_plugin_vcpu_syscall_ret(CPUState
*cpu
, int64_t num
, int64_t ret
)
380 struct qemu_plugin_cb
*cb
, *next
;
381 enum qemu_plugin_event ev
= QEMU_PLUGIN_EV_VCPU_SYSCALL_RET
;
383 if (!test_bit(ev
, cpu
->plugin_mask
)) {
387 QLIST_FOREACH_SAFE_RCU(cb
, &plugin
.cb_lists
[ev
], entry
, next
) {
388 qemu_plugin_vcpu_syscall_ret_cb_t func
= cb
->f
.vcpu_syscall_ret
;
390 func(cb
->ctx
->id
, cpu
->cpu_index
, num
, ret
);
394 void qemu_plugin_vcpu_idle_cb(CPUState
*cpu
)
396 plugin_vcpu_cb__simple(cpu
, QEMU_PLUGIN_EV_VCPU_IDLE
);
399 void qemu_plugin_vcpu_resume_cb(CPUState
*cpu
)
401 plugin_vcpu_cb__simple(cpu
, QEMU_PLUGIN_EV_VCPU_RESUME
);
404 void qemu_plugin_register_vcpu_idle_cb(qemu_plugin_id_t id
,
405 qemu_plugin_vcpu_simple_cb_t cb
)
407 plugin_register_cb(id
, QEMU_PLUGIN_EV_VCPU_IDLE
, cb
);
410 void qemu_plugin_register_vcpu_resume_cb(qemu_plugin_id_t id
,
411 qemu_plugin_vcpu_simple_cb_t cb
)
413 plugin_register_cb(id
, QEMU_PLUGIN_EV_VCPU_RESUME
, cb
);
416 void qemu_plugin_register_flush_cb(qemu_plugin_id_t id
,
417 qemu_plugin_simple_cb_t cb
)
419 plugin_register_cb(id
, QEMU_PLUGIN_EV_FLUSH
, cb
);
422 static bool free_dyn_cb_arr(void *p
, uint32_t h
, void *userp
)
424 g_array_free((GArray
*) p
, true);
428 void qemu_plugin_flush_cb(void)
430 qht_iter_remove(&plugin
.dyn_cb_arr_ht
, free_dyn_cb_arr
, NULL
);
431 qht_reset(&plugin
.dyn_cb_arr_ht
);
433 plugin_cb__simple(QEMU_PLUGIN_EV_FLUSH
);
436 void exec_inline_op(struct qemu_plugin_dyn_cb
*cb
)
438 uint64_t *val
= cb
->userp
;
440 switch (cb
->inline_insn
.op
) {
441 case QEMU_PLUGIN_INLINE_ADD_U64
:
442 *val
+= cb
->inline_insn
.imm
;
445 g_assert_not_reached();
449 void qemu_plugin_vcpu_mem_cb(CPUState
*cpu
, uint64_t vaddr
,
450 MemOpIdx oi
, enum qemu_plugin_mem_rw rw
)
452 GArray
*arr
= cpu
->plugin_mem_cbs
;
458 for (i
= 0; i
< arr
->len
; i
++) {
459 struct qemu_plugin_dyn_cb
*cb
=
460 &g_array_index(arr
, struct qemu_plugin_dyn_cb
, i
);
462 if (!(rw
& cb
->rw
)) {
466 case PLUGIN_CB_REGULAR
:
467 cb
->f
.vcpu_mem(cpu
->cpu_index
, make_plugin_meminfo(oi
, rw
),
470 case PLUGIN_CB_INLINE
:
474 g_assert_not_reached();
479 void qemu_plugin_atexit_cb(void)
481 plugin_cb__udata(QEMU_PLUGIN_EV_ATEXIT
);
484 void qemu_plugin_register_atexit_cb(qemu_plugin_id_t id
,
485 qemu_plugin_udata_cb_t cb
,
488 plugin_register_cb_udata(id
, QEMU_PLUGIN_EV_ATEXIT
, cb
, udata
);
492 * Handle exit from linux-user. Unlike the normal atexit() mechanism
493 * we need to handle the clean-up manually as it's possible threads
494 * are still running. We need to remove all callbacks from code
495 * generation, flush the current translations and then we can safely
496 * trigger the exit callbacks.
499 void qemu_plugin_user_exit(void)
501 enum qemu_plugin_event ev
;
505 * Locking order: we must acquire locks in an order that is consistent
506 * with the one in fork_start(). That is:
507 * - start_exclusive(), which acquires qemu_cpu_list_lock,
508 * must be called before acquiring plugin.lock.
509 * - tb_flush(), which acquires mmap_lock(), must be called
510 * while plugin.lock is not held.
514 qemu_rec_mutex_lock(&plugin
.lock
);
515 /* un-register all callbacks except the final AT_EXIT one */
516 for (ev
= 0; ev
< QEMU_PLUGIN_EV_MAX
; ev
++) {
517 if (ev
!= QEMU_PLUGIN_EV_ATEXIT
) {
518 struct qemu_plugin_cb
*cb
, *next
;
520 QLIST_FOREACH_SAFE_RCU(cb
, &plugin
.cb_lists
[ev
], entry
, next
) {
521 plugin_unregister_cb__locked(cb
->ctx
, ev
);
526 qemu_plugin_disable_mem_helpers(cpu
);
528 qemu_rec_mutex_unlock(&plugin
.lock
);
530 tb_flush(current_cpu
);
533 /* now it's safe to handle the exit case */
534 qemu_plugin_atexit_cb();
538 * Helpers for *-user to ensure locks are sane across fork() events.
541 void qemu_plugin_user_prefork_lock(void)
543 qemu_rec_mutex_lock(&plugin
.lock
);
546 void qemu_plugin_user_postfork(bool is_child
)
549 /* should we just reset via plugin_init? */
550 qemu_rec_mutex_init(&plugin
.lock
);
552 qemu_rec_mutex_unlock(&plugin
.lock
);
556 static bool plugin_dyn_cb_arr_cmp(const void *ap
, const void *bp
)
561 static void __attribute__((__constructor__
)) plugin_init(void)
565 for (i
= 0; i
< QEMU_PLUGIN_EV_MAX
; i
++) {
566 QLIST_INIT(&plugin
.cb_lists
[i
]);
568 qemu_rec_mutex_init(&plugin
.lock
);
569 plugin
.id_ht
= g_hash_table_new(g_int64_hash
, g_int64_equal
);
570 plugin
.cpu_ht
= g_hash_table_new(g_int_hash
, g_int_equal
);
571 QTAILQ_INIT(&plugin
.ctxs
);
572 qht_init(&plugin
.dyn_cb_arr_ht
, plugin_dyn_cb_arr_cmp
, 16,
573 QHT_MODE_AUTO_RESIZE
);
574 atexit(qemu_plugin_atexit_cb
);