replay: rng-builtin support
[qemu/ar7.git] / monitor / qmp.c
blob43880fa6232be1f0d4ff10830103e0ad8e9fc71d
1 /*
2 * QEMU monitor
4 * Copyright (c) 2003-2004 Fabrice Bellard
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
25 #include "qemu/osdep.h"
27 #include "chardev/char-io.h"
28 #include "monitor-internal.h"
29 #include "qapi/error.h"
30 #include "qapi/qapi-commands-control.h"
31 #include "qapi/qmp/qdict.h"
32 #include "qapi/qmp/qjson.h"
33 #include "qapi/qmp/qlist.h"
34 #include "trace.h"
36 struct QMPRequest {
37 /* Owner of the request */
38 MonitorQMP *mon;
40 * Request object to be handled or Error to be reported
41 * (exactly one of them is non-null)
43 QObject *req;
44 Error *err;
46 typedef struct QMPRequest QMPRequest;
48 QmpCommandList qmp_commands, qmp_cap_negotiation_commands;
50 static bool qmp_oob_enabled(MonitorQMP *mon)
52 return mon->capab[QMP_CAPABILITY_OOB];
55 static void monitor_qmp_caps_reset(MonitorQMP *mon)
57 memset(mon->capab_offered, 0, sizeof(mon->capab_offered));
58 memset(mon->capab, 0, sizeof(mon->capab));
59 mon->capab_offered[QMP_CAPABILITY_OOB] = mon->common.use_io_thread;
62 static void qmp_request_free(QMPRequest *req)
64 qobject_unref(req->req);
65 error_free(req->err);
66 g_free(req);
69 /* Caller must hold mon->qmp.qmp_queue_lock */
70 static void monitor_qmp_cleanup_req_queue_locked(MonitorQMP *mon)
72 while (!g_queue_is_empty(mon->qmp_requests)) {
73 qmp_request_free(g_queue_pop_head(mon->qmp_requests));
77 static void monitor_qmp_cleanup_queue_and_resume(MonitorQMP *mon)
79 qemu_mutex_lock(&mon->qmp_queue_lock);
82 * Same condition as in monitor_qmp_dispatcher_co(), but before
83 * removing an element from the queue (hence no `- 1`).
84 * Also, the queue should not be empty either, otherwise the
85 * monitor hasn't been suspended yet (or was already resumed).
87 bool need_resume = (!qmp_oob_enabled(mon) ||
88 mon->qmp_requests->length == QMP_REQ_QUEUE_LEN_MAX)
89 && !g_queue_is_empty(mon->qmp_requests);
91 monitor_qmp_cleanup_req_queue_locked(mon);
93 if (need_resume) {
95 * handle_qmp_command() suspended the monitor because the
96 * request queue filled up, to be resumed when the queue has
97 * space again. We just emptied it; resume the monitor.
99 * Without this, the monitor would remain suspended forever
100 * when we get here while the monitor is suspended. An
101 * unfortunately timed CHR_EVENT_CLOSED can do the trick.
103 monitor_resume(&mon->common);
106 qemu_mutex_unlock(&mon->qmp_queue_lock);
109 void qmp_send_response(MonitorQMP *mon, const QDict *rsp)
111 const QObject *data = QOBJECT(rsp);
112 GString *json;
114 json = qobject_to_json_pretty(data, mon->pretty);
115 assert(json != NULL);
116 trace_monitor_qmp_respond(mon, json->str);
118 g_string_append_c(json, '\n');
119 monitor_puts(&mon->common, json->str);
121 g_string_free(json, true);
125 * Emit QMP response @rsp to @mon.
126 * Null @rsp can only happen for commands with QCO_NO_SUCCESS_RESP.
127 * Nothing is emitted then.
129 static void monitor_qmp_respond(MonitorQMP *mon, QDict *rsp)
131 if (rsp) {
132 qmp_send_response(mon, rsp);
137 * Runs outside of coroutine context for OOB commands, but in
138 * coroutine context for everything else.
140 static void monitor_qmp_dispatch(MonitorQMP *mon, QObject *req)
142 QDict *rsp;
143 QDict *error;
145 rsp = qmp_dispatch(mon->commands, req, qmp_oob_enabled(mon),
146 &mon->common);
148 if (mon->commands == &qmp_cap_negotiation_commands) {
149 error = qdict_get_qdict(rsp, "error");
150 if (error
151 && !g_strcmp0(qdict_get_try_str(error, "class"),
152 QapiErrorClass_str(ERROR_CLASS_COMMAND_NOT_FOUND))) {
153 /* Provide a more useful error message */
154 qdict_del(error, "desc");
155 qdict_put_str(error, "desc", "Expecting capabilities negotiation"
156 " with 'qmp_capabilities'");
160 monitor_qmp_respond(mon, rsp);
161 qobject_unref(rsp);
165 * Pop a QMP request from a monitor request queue.
166 * Return the request, or NULL all request queues are empty.
167 * We are using round-robin fashion to pop the request, to avoid
168 * processing commands only on a very busy monitor. To achieve that,
169 * when we process one request on a specific monitor, we put that
170 * monitor to the end of mon_list queue.
172 * Note: if the function returned with non-NULL, then the caller will
173 * be with qmp_mon->qmp_queue_lock held, and the caller is responsible
174 * to release it.
176 static QMPRequest *monitor_qmp_requests_pop_any_with_lock(void)
178 QMPRequest *req_obj = NULL;
179 Monitor *mon;
180 MonitorQMP *qmp_mon;
182 qemu_mutex_lock(&monitor_lock);
184 QTAILQ_FOREACH(mon, &mon_list, entry) {
185 if (!monitor_is_qmp(mon)) {
186 continue;
189 qmp_mon = container_of(mon, MonitorQMP, common);
190 qemu_mutex_lock(&qmp_mon->qmp_queue_lock);
191 req_obj = g_queue_pop_head(qmp_mon->qmp_requests);
192 if (req_obj) {
193 /* With the lock of corresponding queue held */
194 break;
196 qemu_mutex_unlock(&qmp_mon->qmp_queue_lock);
199 if (req_obj) {
201 * We found one request on the monitor. Degrade this monitor's
202 * priority to lowest by re-inserting it to end of queue.
204 QTAILQ_REMOVE(&mon_list, mon, entry);
205 QTAILQ_INSERT_TAIL(&mon_list, mon, entry);
208 qemu_mutex_unlock(&monitor_lock);
210 return req_obj;
213 void coroutine_fn monitor_qmp_dispatcher_co(void *data)
215 QMPRequest *req_obj = NULL;
216 QDict *rsp;
217 bool oob_enabled;
218 MonitorQMP *mon;
220 while (true) {
221 assert(qatomic_mb_read(&qmp_dispatcher_co_busy) == true);
224 * Mark the dispatcher as not busy already here so that we
225 * don't miss any new requests coming in the middle of our
226 * processing.
228 qatomic_mb_set(&qmp_dispatcher_co_busy, false);
230 while (!(req_obj = monitor_qmp_requests_pop_any_with_lock())) {
232 * No more requests to process. Wait to be reentered from
233 * handle_qmp_command() when it pushes more requests, or
234 * from monitor_cleanup() when it requests shutdown.
236 if (!qmp_dispatcher_co_shutdown) {
237 qemu_coroutine_yield();
240 * busy must be set to true again by whoever
241 * rescheduled us to avoid double scheduling
243 assert(qatomic_xchg(&qmp_dispatcher_co_busy, false) == true);
247 * qmp_dispatcher_co_shutdown may have changed if we
248 * yielded and were reentered from monitor_cleanup()
250 if (qmp_dispatcher_co_shutdown) {
251 return;
255 trace_monitor_qmp_in_band_dequeue(req_obj,
256 req_obj->mon->qmp_requests->length);
258 if (qatomic_xchg(&qmp_dispatcher_co_busy, true) == true) {
260 * Someone rescheduled us (probably because a new requests
261 * came in), but we didn't actually yield. Do that now,
262 * only to be immediately reentered and removed from the
263 * list of scheduled coroutines.
265 qemu_coroutine_yield();
269 * Move the coroutine from iohandler_ctx to qemu_aio_context for
270 * executing the command handler so that it can make progress if it
271 * involves an AIO_WAIT_WHILE().
273 aio_co_schedule(qemu_get_aio_context(), qmp_dispatcher_co);
274 qemu_coroutine_yield();
277 * @req_obj has a request, we hold req_obj->mon->qmp_queue_lock
280 mon = req_obj->mon;
283 * We need to resume the monitor if handle_qmp_command()
284 * suspended it. Two cases:
285 * 1. OOB enabled: mon->qmp_requests has no more space
286 * Resume right away, so that OOB commands can get executed while
287 * this request is being processed.
288 * 2. OOB disabled: always
289 * Resume only after we're done processing the request,
290 * We need to save qmp_oob_enabled() for later, because
291 * qmp_qmp_capabilities() can change it.
293 oob_enabled = qmp_oob_enabled(mon);
294 if (oob_enabled
295 && mon->qmp_requests->length == QMP_REQ_QUEUE_LEN_MAX - 1) {
296 monitor_resume(&mon->common);
299 qemu_mutex_unlock(&mon->qmp_queue_lock);
301 /* Process request */
302 if (req_obj->req) {
303 if (trace_event_get_state(TRACE_MONITOR_QMP_CMD_IN_BAND)) {
304 QDict *qdict = qobject_to(QDict, req_obj->req);
305 QObject *id = qdict ? qdict_get(qdict, "id") : NULL;
306 GString *id_json;
308 id_json = id ? qobject_to_json(id) : g_string_new(NULL);
309 trace_monitor_qmp_cmd_in_band(id_json->str);
310 g_string_free(id_json, true);
312 monitor_qmp_dispatch(mon, req_obj->req);
313 } else {
314 assert(req_obj->err);
315 trace_monitor_qmp_err_in_band(error_get_pretty(req_obj->err));
316 rsp = qmp_error_response(req_obj->err);
317 req_obj->err = NULL;
318 monitor_qmp_respond(mon, rsp);
319 qobject_unref(rsp);
322 if (!oob_enabled) {
323 monitor_resume(&mon->common);
326 qmp_request_free(req_obj);
329 * Yield and reschedule so the main loop stays responsive.
331 * Move back to iohandler_ctx so that nested event loops for
332 * qemu_aio_context don't start new monitor commands.
334 aio_co_schedule(iohandler_get_aio_context(), qmp_dispatcher_co);
335 qemu_coroutine_yield();
339 static void handle_qmp_command(void *opaque, QObject *req, Error *err)
341 MonitorQMP *mon = opaque;
342 QDict *qdict = qobject_to(QDict, req);
343 QMPRequest *req_obj;
345 assert(!req != !err);
347 if (req && trace_event_get_state_backends(TRACE_HANDLE_QMP_COMMAND)) {
348 GString *req_json = qobject_to_json(req);
349 trace_handle_qmp_command(mon, req_json->str);
350 g_string_free(req_json, true);
353 if (qdict && qmp_is_oob(qdict)) {
354 /* OOB commands are executed immediately */
355 if (trace_event_get_state(TRACE_MONITOR_QMP_CMD_OUT_OF_BAND)) {
356 QObject *id = qdict_get(qdict, "id");
357 GString *id_json;
359 id_json = id ? qobject_to_json(id) : g_string_new(NULL);
360 trace_monitor_qmp_cmd_out_of_band(id_json->str);
361 g_string_free(id_json, true);
363 monitor_qmp_dispatch(mon, req);
364 qobject_unref(req);
365 return;
368 req_obj = g_new0(QMPRequest, 1);
369 req_obj->mon = mon;
370 req_obj->req = req;
371 req_obj->err = err;
373 /* Protect qmp_requests and fetching its length. */
374 qemu_mutex_lock(&mon->qmp_queue_lock);
377 * Suspend the monitor when we can't queue more requests after
378 * this one. Dequeuing in monitor_qmp_dispatcher_co() or
379 * monitor_qmp_cleanup_queue_and_resume() will resume it.
380 * Note that when OOB is disabled, we queue at most one command,
381 * for backward compatibility.
383 if (!qmp_oob_enabled(mon) ||
384 mon->qmp_requests->length == QMP_REQ_QUEUE_LEN_MAX - 1) {
385 monitor_suspend(&mon->common);
389 * Put the request to the end of queue so that requests will be
390 * handled in time order. Ownership for req_obj, req,
391 * etc. will be delivered to the handler side.
393 trace_monitor_qmp_in_band_enqueue(req_obj, mon,
394 mon->qmp_requests->length);
395 assert(mon->qmp_requests->length < QMP_REQ_QUEUE_LEN_MAX);
396 g_queue_push_tail(mon->qmp_requests, req_obj);
397 qemu_mutex_unlock(&mon->qmp_queue_lock);
399 /* Kick the dispatcher routine */
400 if (!qatomic_xchg(&qmp_dispatcher_co_busy, true)) {
401 aio_co_wake(qmp_dispatcher_co);
405 static void monitor_qmp_read(void *opaque, const uint8_t *buf, int size)
407 MonitorQMP *mon = opaque;
409 json_message_parser_feed(&mon->parser, (const char *) buf, size);
412 static QDict *qmp_greeting(MonitorQMP *mon)
414 QList *cap_list = qlist_new();
415 QObject *ver = NULL;
416 QDict *args;
417 QMPCapability cap;
419 args = qdict_new();
420 qmp_marshal_query_version(args, &ver, NULL);
421 qobject_unref(args);
423 for (cap = 0; cap < QMP_CAPABILITY__MAX; cap++) {
424 if (mon->capab_offered[cap]) {
425 qlist_append_str(cap_list, QMPCapability_str(cap));
429 return qdict_from_jsonf_nofail(
430 "{'QMP': {'version': %p, 'capabilities': %p}}",
431 ver, cap_list);
434 static void monitor_qmp_event(void *opaque, QEMUChrEvent event)
436 QDict *data;
437 MonitorQMP *mon = opaque;
439 switch (event) {
440 case CHR_EVENT_OPENED:
441 mon->commands = &qmp_cap_negotiation_commands;
442 monitor_qmp_caps_reset(mon);
443 data = qmp_greeting(mon);
444 qmp_send_response(mon, data);
445 qobject_unref(data);
446 mon_refcount++;
447 break;
448 case CHR_EVENT_CLOSED:
450 * Note: this is only useful when the output of the chardev
451 * backend is still open. For example, when the backend is
452 * stdio, it's possible that stdout is still open when stdin
453 * is closed.
455 monitor_qmp_cleanup_queue_and_resume(mon);
456 json_message_parser_destroy(&mon->parser);
457 json_message_parser_init(&mon->parser, handle_qmp_command,
458 mon, NULL);
459 mon_refcount--;
460 monitor_fdsets_cleanup();
461 break;
462 case CHR_EVENT_BREAK:
463 case CHR_EVENT_MUX_IN:
464 case CHR_EVENT_MUX_OUT:
465 /* Ignore */
466 break;
470 void monitor_data_destroy_qmp(MonitorQMP *mon)
472 json_message_parser_destroy(&mon->parser);
473 qemu_mutex_destroy(&mon->qmp_queue_lock);
474 monitor_qmp_cleanup_req_queue_locked(mon);
475 g_queue_free(mon->qmp_requests);
478 static void monitor_qmp_setup_handlers_bh(void *opaque)
480 MonitorQMP *mon = opaque;
481 GMainContext *context;
483 assert(mon->common.use_io_thread);
484 context = iothread_get_g_main_context(mon_iothread);
485 assert(context);
486 qemu_chr_fe_set_handlers(&mon->common.chr, monitor_can_read,
487 monitor_qmp_read, monitor_qmp_event,
488 NULL, &mon->common, context, true);
489 monitor_list_append(&mon->common);
492 void monitor_init_qmp(Chardev *chr, bool pretty, Error **errp)
494 MonitorQMP *mon = g_new0(MonitorQMP, 1);
496 if (!qemu_chr_fe_init(&mon->common.chr, chr, errp)) {
497 g_free(mon);
498 return;
500 qemu_chr_fe_set_echo(&mon->common.chr, true);
502 /* Note: we run QMP monitor in I/O thread when @chr supports that */
503 monitor_data_init(&mon->common, true, false,
504 qemu_chr_has_feature(chr, QEMU_CHAR_FEATURE_GCONTEXT));
506 mon->pretty = pretty;
508 qemu_mutex_init(&mon->qmp_queue_lock);
509 mon->qmp_requests = g_queue_new();
511 json_message_parser_init(&mon->parser, handle_qmp_command, mon, NULL);
512 if (mon->common.use_io_thread) {
514 * Make sure the old iowatch is gone. It's possible when
515 * e.g. the chardev is in client mode, with wait=on.
517 remove_fd_in_watch(chr);
519 * We can't call qemu_chr_fe_set_handlers() directly here
520 * since chardev might be running in the monitor I/O
521 * thread. Schedule a bottom half.
523 aio_bh_schedule_oneshot(iothread_get_aio_context(mon_iothread),
524 monitor_qmp_setup_handlers_bh, mon);
525 /* The bottom half will add @mon to @mon_list */
526 } else {
527 qemu_chr_fe_set_handlers(&mon->common.chr, monitor_can_read,
528 monitor_qmp_read, monitor_qmp_event,
529 NULL, &mon->common, NULL, true);
530 monitor_list_append(&mon->common);