4 * Copyright (c) 2003-2004 Fabrice Bellard
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
25 #include "qemu/osdep.h"
26 #include "monitor-internal.h"
27 #include "qapi/error.h"
28 #include "qapi/qapi-emit-events.h"
29 #include "qapi/qmp/qdict.h"
30 #include "qapi/qmp/qstring.h"
31 #include "qemu/error-report.h"
32 #include "qemu/option.h"
33 #include "sysemu/qtest.h"
34 #include "sysemu/sysemu.h"
38 * To prevent flooding clients, events can be throttled. The
39 * throttling is calculated globally, rather than per-Monitor
42 typedef struct MonitorQAPIEventState
{
43 QAPIEvent event
; /* Throttling state for this event type and... */
44 QDict
*data
; /* ... data, see qapi_event_throttle_equal() */
45 QEMUTimer
*timer
; /* Timer for handling delayed events */
46 QDict
*qdict
; /* Delayed event (if any) */
47 } MonitorQAPIEventState
;
50 int64_t rate
; /* Minimum time (in ns) between two events */
51 } MonitorQAPIEventConf
;
53 /* Shared monitor I/O thread */
54 IOThread
*mon_iothread
;
56 /* Bottom half to dispatch the requests received from I/O thread */
57 QEMUBH
*qmp_dispatcher_bh
;
59 /* Protects mon_list, monitor_qapi_event_state, monitor_destroyed. */
60 QemuMutex monitor_lock
;
61 static GHashTable
*monitor_qapi_event_state
;
65 static bool monitor_destroyed
;
67 __thread Monitor
*cur_mon
;
70 * Is the current monitor, if any, a QMP monitor?
72 bool monitor_cur_is_qmp(void)
74 return cur_mon
&& monitor_is_qmp(cur_mon
);
78 * Is @mon is using readline?
79 * Note: not all HMP monitors use readline, e.g., gdbserver has a
80 * non-interactive HMP monitor, so readline is not used there.
82 static inline bool monitor_uses_readline(const MonitorHMP
*mon
)
84 return mon
->use_readline
;
87 static inline bool monitor_is_hmp_non_interactive(const Monitor
*mon
)
89 if (monitor_is_qmp(mon
)) {
93 return !monitor_uses_readline(container_of(mon
, MonitorHMP
, common
));
96 static void monitor_flush_locked(Monitor
*mon
);
98 static gboolean
monitor_unblocked(GIOChannel
*chan
, GIOCondition cond
,
101 Monitor
*mon
= opaque
;
103 qemu_mutex_lock(&mon
->mon_lock
);
105 monitor_flush_locked(mon
);
106 qemu_mutex_unlock(&mon
->mon_lock
);
110 /* Caller must hold mon->mon_lock */
111 static void monitor_flush_locked(Monitor
*mon
)
117 if (mon
->skip_flush
) {
121 buf
= qstring_get_str(mon
->outbuf
);
122 len
= qstring_get_length(mon
->outbuf
);
124 if (len
&& !mon
->mux_out
) {
125 rc
= qemu_chr_fe_write(&mon
->chr
, (const uint8_t *) buf
, len
);
126 if ((rc
< 0 && errno
!= EAGAIN
) || (rc
== len
)) {
127 /* all flushed or error */
128 qobject_unref(mon
->outbuf
);
129 mon
->outbuf
= qstring_new();
134 QString
*tmp
= qstring_from_str(buf
+ rc
);
135 qobject_unref(mon
->outbuf
);
138 if (mon
->out_watch
== 0) {
140 qemu_chr_fe_add_watch(&mon
->chr
, G_IO_OUT
| G_IO_HUP
,
141 monitor_unblocked
, mon
);
146 void monitor_flush(Monitor
*mon
)
148 qemu_mutex_lock(&mon
->mon_lock
);
149 monitor_flush_locked(mon
);
150 qemu_mutex_unlock(&mon
->mon_lock
);
153 /* flush at every end of line */
154 int monitor_puts(Monitor
*mon
, const char *str
)
159 qemu_mutex_lock(&mon
->mon_lock
);
160 for (i
= 0; str
[i
]; i
++) {
163 qstring_append_chr(mon
->outbuf
, '\r');
165 qstring_append_chr(mon
->outbuf
, c
);
167 monitor_flush_locked(mon
);
170 qemu_mutex_unlock(&mon
->mon_lock
);
175 int monitor_vprintf(Monitor
*mon
, const char *fmt
, va_list ap
)
184 if (monitor_is_qmp(mon
)) {
188 buf
= g_strdup_vprintf(fmt
, ap
);
189 n
= monitor_puts(mon
, buf
);
194 int monitor_printf(Monitor
*mon
, const char *fmt
, ...)
200 ret
= monitor_vprintf(mon
, fmt
, ap
);
206 * Print to current monitor if we have one, else to stderr.
208 int error_vprintf(const char *fmt
, va_list ap
)
210 if (cur_mon
&& !monitor_cur_is_qmp()) {
211 return monitor_vprintf(cur_mon
, fmt
, ap
);
213 return vfprintf(stderr
, fmt
, ap
);
216 int error_vprintf_unless_qmp(const char *fmt
, va_list ap
)
219 return vfprintf(stderr
, fmt
, ap
);
221 if (!monitor_cur_is_qmp()) {
222 return monitor_vprintf(cur_mon
, fmt
, ap
);
228 static MonitorQAPIEventConf monitor_qapi_event_conf
[QAPI_EVENT__MAX
] = {
229 /* Limit guest-triggerable events to 1 per second */
230 [QAPI_EVENT_RTC_CHANGE
] = { 1000 * SCALE_MS
},
231 [QAPI_EVENT_WATCHDOG
] = { 1000 * SCALE_MS
},
232 [QAPI_EVENT_BALLOON_CHANGE
] = { 1000 * SCALE_MS
},
233 [QAPI_EVENT_QUORUM_REPORT_BAD
] = { 1000 * SCALE_MS
},
234 [QAPI_EVENT_QUORUM_FAILURE
] = { 1000 * SCALE_MS
},
235 [QAPI_EVENT_VSERPORT_CHANGE
] = { 1000 * SCALE_MS
},
239 * Return the clock to use for recording an event's time.
240 * It's QEMU_CLOCK_REALTIME, except for qtests it's
241 * QEMU_CLOCK_VIRTUAL, to support testing rate limits.
242 * Beware: result is invalid before configure_accelerator().
244 static inline QEMUClockType
monitor_get_event_clock(void)
246 return qtest_enabled() ? QEMU_CLOCK_VIRTUAL
: QEMU_CLOCK_REALTIME
;
250 * Broadcast an event to all monitors.
251 * @qdict is the event object. Its member "event" must match @event.
252 * Caller must hold monitor_lock.
254 static void monitor_qapi_event_emit(QAPIEvent event
, QDict
*qdict
)
259 trace_monitor_protocol_event_emit(event
, qdict
);
260 QTAILQ_FOREACH(mon
, &mon_list
, entry
) {
261 if (!monitor_is_qmp(mon
)) {
265 qmp_mon
= container_of(mon
, MonitorQMP
, common
);
266 if (qmp_mon
->commands
!= &qmp_cap_negotiation_commands
) {
267 qmp_send_response(qmp_mon
, qdict
);
272 static void monitor_qapi_event_handler(void *opaque
);
275 * Queue a new event for emission to Monitor instances,
276 * applying any rate limiting if required.
279 monitor_qapi_event_queue_no_reenter(QAPIEvent event
, QDict
*qdict
)
281 MonitorQAPIEventConf
*evconf
;
282 MonitorQAPIEventState
*evstate
;
284 assert(event
< QAPI_EVENT__MAX
);
285 evconf
= &monitor_qapi_event_conf
[event
];
286 trace_monitor_protocol_event_queue(event
, qdict
, evconf
->rate
);
288 qemu_mutex_lock(&monitor_lock
);
291 /* Unthrottled event */
292 monitor_qapi_event_emit(event
, qdict
);
294 QDict
*data
= qobject_to(QDict
, qdict_get(qdict
, "data"));
295 MonitorQAPIEventState key
= { .event
= event
, .data
= data
};
297 evstate
= g_hash_table_lookup(monitor_qapi_event_state
, &key
);
298 assert(!evstate
|| timer_pending(evstate
->timer
));
302 * Timer is pending for (at least) evconf->rate ns after
303 * last send. Store event for sending when timer fires,
304 * replacing a prior stored event if any.
306 qobject_unref(evstate
->qdict
);
307 evstate
->qdict
= qobject_ref(qdict
);
310 * Last send was (at least) evconf->rate ns ago.
311 * Send immediately, and arm the timer to call
312 * monitor_qapi_event_handler() in evconf->rate ns. Any
313 * events arriving before then will be delayed until then.
315 int64_t now
= qemu_clock_get_ns(monitor_get_event_clock());
317 monitor_qapi_event_emit(event
, qdict
);
319 evstate
= g_new(MonitorQAPIEventState
, 1);
320 evstate
->event
= event
;
321 evstate
->data
= qobject_ref(data
);
322 evstate
->qdict
= NULL
;
323 evstate
->timer
= timer_new_ns(monitor_get_event_clock(),
324 monitor_qapi_event_handler
,
326 g_hash_table_add(monitor_qapi_event_state
, evstate
);
327 timer_mod_ns(evstate
->timer
, now
+ evconf
->rate
);
331 qemu_mutex_unlock(&monitor_lock
);
334 void qapi_event_emit(QAPIEvent event
, QDict
*qdict
)
337 * monitor_qapi_event_queue_no_reenter() is not reentrant: it
338 * would deadlock on monitor_lock. Work around by queueing
339 * events in thread-local storage.
340 * TODO: remove this, make it re-enter safe.
342 typedef struct MonitorQapiEvent
{
345 QSIMPLEQ_ENTRY(MonitorQapiEvent
) entry
;
347 static __thread
QSIMPLEQ_HEAD(, MonitorQapiEvent
) event_queue
;
348 static __thread
bool reentered
;
349 MonitorQapiEvent
*ev
;
352 QSIMPLEQ_INIT(&event_queue
);
355 ev
= g_new(MonitorQapiEvent
, 1);
356 ev
->qdict
= qobject_ref(qdict
);
358 QSIMPLEQ_INSERT_TAIL(&event_queue
, ev
, entry
);
365 while ((ev
= QSIMPLEQ_FIRST(&event_queue
)) != NULL
) {
366 QSIMPLEQ_REMOVE_HEAD(&event_queue
, entry
);
367 monitor_qapi_event_queue_no_reenter(ev
->event
, ev
->qdict
);
368 qobject_unref(ev
->qdict
);
376 * This function runs evconf->rate ns after sending a throttled
378 * If another event has since been stored, send it.
380 static void monitor_qapi_event_handler(void *opaque
)
382 MonitorQAPIEventState
*evstate
= opaque
;
383 MonitorQAPIEventConf
*evconf
= &monitor_qapi_event_conf
[evstate
->event
];
385 trace_monitor_protocol_event_handler(evstate
->event
, evstate
->qdict
);
386 qemu_mutex_lock(&monitor_lock
);
388 if (evstate
->qdict
) {
389 int64_t now
= qemu_clock_get_ns(monitor_get_event_clock());
391 monitor_qapi_event_emit(evstate
->event
, evstate
->qdict
);
392 qobject_unref(evstate
->qdict
);
393 evstate
->qdict
= NULL
;
394 timer_mod_ns(evstate
->timer
, now
+ evconf
->rate
);
396 g_hash_table_remove(monitor_qapi_event_state
, evstate
);
397 qobject_unref(evstate
->data
);
398 timer_free(evstate
->timer
);
402 qemu_mutex_unlock(&monitor_lock
);
405 static unsigned int qapi_event_throttle_hash(const void *key
)
407 const MonitorQAPIEventState
*evstate
= key
;
408 unsigned int hash
= evstate
->event
* 255;
410 if (evstate
->event
== QAPI_EVENT_VSERPORT_CHANGE
) {
411 hash
+= g_str_hash(qdict_get_str(evstate
->data
, "id"));
414 if (evstate
->event
== QAPI_EVENT_QUORUM_REPORT_BAD
) {
415 hash
+= g_str_hash(qdict_get_str(evstate
->data
, "node-name"));
421 static gboolean
qapi_event_throttle_equal(const void *a
, const void *b
)
423 const MonitorQAPIEventState
*eva
= a
;
424 const MonitorQAPIEventState
*evb
= b
;
426 if (eva
->event
!= evb
->event
) {
430 if (eva
->event
== QAPI_EVENT_VSERPORT_CHANGE
) {
431 return !strcmp(qdict_get_str(eva
->data
, "id"),
432 qdict_get_str(evb
->data
, "id"));
435 if (eva
->event
== QAPI_EVENT_QUORUM_REPORT_BAD
) {
436 return !strcmp(qdict_get_str(eva
->data
, "node-name"),
437 qdict_get_str(evb
->data
, "node-name"));
443 int monitor_suspend(Monitor
*mon
)
445 if (monitor_is_hmp_non_interactive(mon
)) {
449 atomic_inc(&mon
->suspend_cnt
);
451 if (mon
->use_io_thread
) {
453 * Kick I/O thread to make sure this takes effect. It'll be
454 * evaluated again in prepare() of the watch object.
456 aio_notify(iothread_get_aio_context(mon_iothread
));
459 trace_monitor_suspend(mon
, 1);
463 static void monitor_accept_input(void *opaque
)
465 Monitor
*mon
= opaque
;
467 qemu_chr_fe_accept_input(&mon
->chr
);
470 void monitor_resume(Monitor
*mon
)
472 if (monitor_is_hmp_non_interactive(mon
)) {
476 if (atomic_dec_fetch(&mon
->suspend_cnt
) == 0) {
479 if (mon
->use_io_thread
) {
480 ctx
= iothread_get_aio_context(mon_iothread
);
482 ctx
= qemu_get_aio_context();
485 if (!monitor_is_qmp(mon
)) {
486 MonitorHMP
*hmp_mon
= container_of(mon
, MonitorHMP
, common
);
488 readline_show_prompt(hmp_mon
->rs
);
491 aio_bh_schedule_oneshot(ctx
, monitor_accept_input
, mon
);
494 trace_monitor_suspend(mon
, -1);
497 int monitor_can_read(void *opaque
)
499 Monitor
*mon
= opaque
;
501 return !atomic_mb_read(&mon
->suspend_cnt
);
504 void monitor_list_append(Monitor
*mon
)
506 qemu_mutex_lock(&monitor_lock
);
508 * This prevents inserting new monitors during monitor_cleanup().
509 * A cleaner solution would involve the main thread telling other
510 * threads to terminate, waiting for their termination.
512 if (!monitor_destroyed
) {
513 QTAILQ_INSERT_HEAD(&mon_list
, mon
, entry
);
516 qemu_mutex_unlock(&monitor_lock
);
519 monitor_data_destroy(mon
);
524 static void monitor_iothread_init(void)
526 mon_iothread
= iothread_create("mon_iothread", &error_abort
);
529 void monitor_data_init(Monitor
*mon
, bool is_qmp
, bool skip_flush
,
532 if (use_io_thread
&& !mon_iothread
) {
533 monitor_iothread_init();
535 qemu_mutex_init(&mon
->mon_lock
);
536 mon
->is_qmp
= is_qmp
;
537 mon
->outbuf
= qstring_new();
538 mon
->skip_flush
= skip_flush
;
539 mon
->use_io_thread
= use_io_thread
;
542 void monitor_data_destroy(Monitor
*mon
)
544 g_free(mon
->mon_cpu_path
);
545 qemu_chr_fe_deinit(&mon
->chr
, false);
546 if (monitor_is_qmp(mon
)) {
547 monitor_data_destroy_qmp(container_of(mon
, MonitorQMP
, common
));
549 readline_free(container_of(mon
, MonitorHMP
, common
)->rs
);
551 qobject_unref(mon
->outbuf
);
552 qemu_mutex_destroy(&mon
->mon_lock
);
555 void monitor_cleanup(void)
558 * We need to explicitly stop the I/O thread (but not destroy it),
559 * clean up the monitor resources, then destroy the I/O thread since
560 * we need to unregister from chardev below in
561 * monitor_data_destroy(), and chardev is not thread-safe yet
564 iothread_stop(mon_iothread
);
567 /* Flush output buffers and destroy monitors */
568 qemu_mutex_lock(&monitor_lock
);
569 monitor_destroyed
= true;
570 while (!QTAILQ_EMPTY(&mon_list
)) {
571 Monitor
*mon
= QTAILQ_FIRST(&mon_list
);
572 QTAILQ_REMOVE(&mon_list
, mon
, entry
);
573 /* Permit QAPI event emission from character frontend release */
574 qemu_mutex_unlock(&monitor_lock
);
576 monitor_data_destroy(mon
);
577 qemu_mutex_lock(&monitor_lock
);
580 qemu_mutex_unlock(&monitor_lock
);
582 /* QEMUBHs needs to be deleted before destroying the I/O thread */
583 qemu_bh_delete(qmp_dispatcher_bh
);
584 qmp_dispatcher_bh
= NULL
;
586 iothread_destroy(mon_iothread
);
591 static void monitor_qapi_event_init(void)
593 monitor_qapi_event_state
= g_hash_table_new(qapi_event_throttle_hash
,
594 qapi_event_throttle_equal
);
597 void monitor_init_globals_core(void)
599 monitor_qapi_event_init();
600 qemu_mutex_init(&monitor_lock
);
603 * The dispatcher BH must run in the main loop thread, since we
604 * have commands assuming that context. It would be nice to get
605 * rid of those assumptions.
607 qmp_dispatcher_bh
= aio_bh_new(iohandler_get_aio_context(),
608 monitor_qmp_bh_dispatcher
,
612 QemuOptsList qemu_mon_opts
= {
614 .implied_opt_name
= "chardev",
615 .head
= QTAILQ_HEAD_INITIALIZER(qemu_mon_opts
.head
),
619 .type
= QEMU_OPT_STRING
,
622 .type
= QEMU_OPT_STRING
,
625 .type
= QEMU_OPT_BOOL
,
627 { /* end of list */ }