migration/multifd: Protect accesses to migration_threads
[qemu/armbru.git] / block / qapi.c
blobf34f95e0efaf8a57f2f9b2ae1024a2ce51b10518
1 /*
2 * Block layer qmp and info dump related functions
4 * Copyright (c) 2003-2008 Fabrice Bellard
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
25 #include "qemu/osdep.h"
26 #include "qemu/cutils.h"
27 #include "block/qapi.h"
28 #include "block/block_int.h"
29 #include "block/dirty-bitmap.h"
30 #include "block/throttle-groups.h"
31 #include "block/write-threshold.h"
32 #include "qapi/error.h"
33 #include "qapi/qapi-commands-block-core.h"
34 #include "qapi/qobject-output-visitor.h"
35 #include "qapi/qapi-visit-block-core.h"
36 #include "qapi/qmp/qbool.h"
37 #include "qapi/qmp/qdict.h"
38 #include "qapi/qmp/qlist.h"
39 #include "qapi/qmp/qnum.h"
40 #include "qapi/qmp/qstring.h"
41 #include "qemu/qemu-print.h"
42 #include "sysemu/block-backend.h"
44 BlockDeviceInfo *bdrv_block_device_info(BlockBackend *blk,
45 BlockDriverState *bs,
46 bool flat,
47 Error **errp)
49 ImageInfo **p_image_info;
50 ImageInfo *backing_info;
51 BlockDriverState *bs0, *backing;
52 BlockDeviceInfo *info;
53 ERRP_GUARD();
55 if (!bs->drv) {
56 error_setg(errp, "Block device %s is ejected", bs->node_name);
57 return NULL;
60 bdrv_refresh_filename(bs);
62 info = g_malloc0(sizeof(*info));
63 info->file = g_strdup(bs->filename);
64 info->ro = bdrv_is_read_only(bs);
65 info->drv = g_strdup(bs->drv->format_name);
66 info->encrypted = bs->encrypted;
68 info->cache = g_new(BlockdevCacheInfo, 1);
69 *info->cache = (BlockdevCacheInfo) {
70 .writeback = blk ? blk_enable_write_cache(blk) : true,
71 .direct = !!(bs->open_flags & BDRV_O_NOCACHE),
72 .no_flush = !!(bs->open_flags & BDRV_O_NO_FLUSH),
75 if (bs->node_name[0]) {
76 info->node_name = g_strdup(bs->node_name);
79 backing = bdrv_cow_bs(bs);
80 if (backing) {
81 info->backing_file = g_strdup(backing->filename);
84 if (!QLIST_EMPTY(&bs->dirty_bitmaps)) {
85 info->has_dirty_bitmaps = true;
86 info->dirty_bitmaps = bdrv_query_dirty_bitmaps(bs);
89 info->detect_zeroes = bs->detect_zeroes;
91 if (blk && blk_get_public(blk)->throttle_group_member.throttle_state) {
92 ThrottleConfig cfg;
93 BlockBackendPublic *blkp = blk_get_public(blk);
95 throttle_group_get_config(&blkp->throttle_group_member, &cfg);
97 info->bps = cfg.buckets[THROTTLE_BPS_TOTAL].avg;
98 info->bps_rd = cfg.buckets[THROTTLE_BPS_READ].avg;
99 info->bps_wr = cfg.buckets[THROTTLE_BPS_WRITE].avg;
101 info->iops = cfg.buckets[THROTTLE_OPS_TOTAL].avg;
102 info->iops_rd = cfg.buckets[THROTTLE_OPS_READ].avg;
103 info->iops_wr = cfg.buckets[THROTTLE_OPS_WRITE].avg;
105 info->has_bps_max = cfg.buckets[THROTTLE_BPS_TOTAL].max;
106 info->bps_max = cfg.buckets[THROTTLE_BPS_TOTAL].max;
107 info->has_bps_rd_max = cfg.buckets[THROTTLE_BPS_READ].max;
108 info->bps_rd_max = cfg.buckets[THROTTLE_BPS_READ].max;
109 info->has_bps_wr_max = cfg.buckets[THROTTLE_BPS_WRITE].max;
110 info->bps_wr_max = cfg.buckets[THROTTLE_BPS_WRITE].max;
112 info->has_iops_max = cfg.buckets[THROTTLE_OPS_TOTAL].max;
113 info->iops_max = cfg.buckets[THROTTLE_OPS_TOTAL].max;
114 info->has_iops_rd_max = cfg.buckets[THROTTLE_OPS_READ].max;
115 info->iops_rd_max = cfg.buckets[THROTTLE_OPS_READ].max;
116 info->has_iops_wr_max = cfg.buckets[THROTTLE_OPS_WRITE].max;
117 info->iops_wr_max = cfg.buckets[THROTTLE_OPS_WRITE].max;
119 info->has_bps_max_length = info->has_bps_max;
120 info->bps_max_length =
121 cfg.buckets[THROTTLE_BPS_TOTAL].burst_length;
122 info->has_bps_rd_max_length = info->has_bps_rd_max;
123 info->bps_rd_max_length =
124 cfg.buckets[THROTTLE_BPS_READ].burst_length;
125 info->has_bps_wr_max_length = info->has_bps_wr_max;
126 info->bps_wr_max_length =
127 cfg.buckets[THROTTLE_BPS_WRITE].burst_length;
129 info->has_iops_max_length = info->has_iops_max;
130 info->iops_max_length =
131 cfg.buckets[THROTTLE_OPS_TOTAL].burst_length;
132 info->has_iops_rd_max_length = info->has_iops_rd_max;
133 info->iops_rd_max_length =
134 cfg.buckets[THROTTLE_OPS_READ].burst_length;
135 info->has_iops_wr_max_length = info->has_iops_wr_max;
136 info->iops_wr_max_length =
137 cfg.buckets[THROTTLE_OPS_WRITE].burst_length;
139 info->has_iops_size = cfg.op_size;
140 info->iops_size = cfg.op_size;
142 info->group =
143 g_strdup(throttle_group_get_name(&blkp->throttle_group_member));
146 info->write_threshold = bdrv_write_threshold_get(bs);
148 bs0 = bs;
149 p_image_info = &info->image;
150 info->backing_file_depth = 0;
153 * Skip automatically inserted nodes that the user isn't aware of for
154 * query-block (blk != NULL), but not for query-named-block-nodes
156 bdrv_query_image_info(bs0, p_image_info, flat, blk != NULL, errp);
157 if (*errp) {
158 qapi_free_BlockDeviceInfo(info);
159 return NULL;
162 backing_info = info->image->backing_image;
163 while (backing_info) {
164 info->backing_file_depth++;
165 backing_info = backing_info->backing_image;
168 return info;
172 * Returns 0 on success, with *p_list either set to describe snapshot
173 * information, or NULL because there are no snapshots. Returns -errno on
174 * error, with *p_list untouched.
176 int bdrv_query_snapshot_info_list(BlockDriverState *bs,
177 SnapshotInfoList **p_list,
178 Error **errp)
180 int i, sn_count;
181 QEMUSnapshotInfo *sn_tab = NULL;
182 SnapshotInfoList *head = NULL, **tail = &head;
183 SnapshotInfo *info;
185 sn_count = bdrv_snapshot_list(bs, &sn_tab);
186 if (sn_count < 0) {
187 const char *dev = bdrv_get_device_name(bs);
188 switch (sn_count) {
189 case -ENOMEDIUM:
190 error_setg(errp, "Device '%s' is not inserted", dev);
191 break;
192 case -ENOTSUP:
193 error_setg(errp,
194 "Device '%s' does not support internal snapshots",
195 dev);
196 break;
197 default:
198 error_setg_errno(errp, -sn_count,
199 "Can't list snapshots of device '%s'", dev);
200 break;
202 return sn_count;
205 for (i = 0; i < sn_count; i++) {
206 info = g_new0(SnapshotInfo, 1);
207 info->id = g_strdup(sn_tab[i].id_str);
208 info->name = g_strdup(sn_tab[i].name);
209 info->vm_state_size = sn_tab[i].vm_state_size;
210 info->date_sec = sn_tab[i].date_sec;
211 info->date_nsec = sn_tab[i].date_nsec;
212 info->vm_clock_sec = sn_tab[i].vm_clock_nsec / 1000000000;
213 info->vm_clock_nsec = sn_tab[i].vm_clock_nsec % 1000000000;
214 info->icount = sn_tab[i].icount;
215 info->has_icount = sn_tab[i].icount != -1ULL;
217 QAPI_LIST_APPEND(tail, info);
220 g_free(sn_tab);
221 *p_list = head;
222 return 0;
226 * Helper function for other query info functions. Store information about @bs
227 * in @info, setting @errp on error.
229 static void bdrv_do_query_node_info(BlockDriverState *bs,
230 BlockNodeInfo *info,
231 Error **errp)
233 int64_t size;
234 const char *backing_filename;
235 BlockDriverInfo bdi;
236 int ret;
237 Error *err = NULL;
239 aio_context_acquire(bdrv_get_aio_context(bs));
241 size = bdrv_getlength(bs);
242 if (size < 0) {
243 error_setg_errno(errp, -size, "Can't get image size '%s'",
244 bs->exact_filename);
245 goto out;
248 bdrv_refresh_filename(bs);
250 info->filename = g_strdup(bs->filename);
251 info->format = g_strdup(bdrv_get_format_name(bs));
252 info->virtual_size = size;
253 info->actual_size = bdrv_get_allocated_file_size(bs);
254 info->has_actual_size = info->actual_size >= 0;
255 if (bs->encrypted) {
256 info->encrypted = true;
257 info->has_encrypted = true;
259 if (bdrv_get_info(bs, &bdi) >= 0) {
260 if (bdi.cluster_size != 0) {
261 info->cluster_size = bdi.cluster_size;
262 info->has_cluster_size = true;
264 info->dirty_flag = bdi.is_dirty;
265 info->has_dirty_flag = true;
267 info->format_specific = bdrv_get_specific_info(bs, &err);
268 if (err) {
269 error_propagate(errp, err);
270 goto out;
272 backing_filename = bs->backing_file;
273 if (backing_filename[0] != '\0') {
274 char *backing_filename2;
276 info->backing_filename = g_strdup(backing_filename);
277 backing_filename2 = bdrv_get_full_backing_filename(bs, NULL);
279 /* Always report the full_backing_filename if present, even if it's the
280 * same as backing_filename. That they are same is useful info. */
281 if (backing_filename2) {
282 info->full_backing_filename = g_strdup(backing_filename2);
285 if (bs->backing_format[0]) {
286 info->backing_filename_format = g_strdup(bs->backing_format);
288 g_free(backing_filename2);
291 ret = bdrv_query_snapshot_info_list(bs, &info->snapshots, &err);
292 switch (ret) {
293 case 0:
294 if (info->snapshots) {
295 info->has_snapshots = true;
297 break;
298 /* recoverable error */
299 case -ENOMEDIUM:
300 case -ENOTSUP:
301 error_free(err);
302 break;
303 default:
304 error_propagate(errp, err);
305 goto out;
308 out:
309 aio_context_release(bdrv_get_aio_context(bs));
313 * bdrv_query_block_node_info:
314 * @bs: block node to examine
315 * @p_info: location to store node information
316 * @errp: location to store error information
318 * Store image information about @bs in @p_info.
320 * @p_info will be set only on success. On error, store error in @errp.
322 void bdrv_query_block_node_info(BlockDriverState *bs,
323 BlockNodeInfo **p_info,
324 Error **errp)
326 BlockNodeInfo *info;
327 ERRP_GUARD();
329 info = g_new0(BlockNodeInfo, 1);
330 bdrv_do_query_node_info(bs, info, errp);
331 if (*errp) {
332 qapi_free_BlockNodeInfo(info);
333 return;
336 *p_info = info;
340 * bdrv_query_image_info:
341 * @bs: block node to examine
342 * @p_info: location to store image information
343 * @flat: skip backing node information
344 * @skip_implicit_filters: skip implicit filters in the backing chain
345 * @errp: location to store error information
347 * Store image information in @p_info, potentially recursively covering the
348 * backing chain.
350 * If @flat is true, do not query backing image information, i.e.
351 * (*p_info)->has_backing_image will be set to false and
352 * (*p_info)->backing_image to NULL even when the image does in fact have a
353 * backing image.
355 * If @skip_implicit_filters is true, implicit filter nodes in the backing chain
356 * will be skipped when querying backing image information.
357 * (@skip_implicit_filters is ignored when @flat is true.)
359 * @p_info will be set only on success. On error, store error in @errp.
361 void bdrv_query_image_info(BlockDriverState *bs,
362 ImageInfo **p_info,
363 bool flat,
364 bool skip_implicit_filters,
365 Error **errp)
367 ImageInfo *info;
368 ERRP_GUARD();
370 info = g_new0(ImageInfo, 1);
371 bdrv_do_query_node_info(bs, qapi_ImageInfo_base(info), errp);
372 if (*errp) {
373 goto fail;
376 if (!flat) {
377 BlockDriverState *backing;
380 * Use any filtered child here (for backwards compatibility to when
381 * we always took bs->backing, which might be any filtered child).
383 backing = bdrv_filter_or_cow_bs(bs);
384 if (skip_implicit_filters) {
385 backing = bdrv_skip_implicit_filters(backing);
388 if (backing) {
389 bdrv_query_image_info(backing, &info->backing_image, false,
390 skip_implicit_filters, errp);
391 if (*errp) {
392 goto fail;
397 *p_info = info;
398 return;
400 fail:
401 assert(*errp);
402 qapi_free_ImageInfo(info);
406 * bdrv_query_block_graph_info:
407 * @bs: root node to start from
408 * @p_info: location to store image information
409 * @errp: location to store error information
411 * Store image information about the graph starting from @bs in @p_info.
413 * @p_info will be set only on success. On error, store error in @errp.
415 void bdrv_query_block_graph_info(BlockDriverState *bs,
416 BlockGraphInfo **p_info,
417 Error **errp)
419 BlockGraphInfo *info;
420 BlockChildInfoList **children_list_tail;
421 BdrvChild *c;
422 ERRP_GUARD();
424 info = g_new0(BlockGraphInfo, 1);
425 bdrv_do_query_node_info(bs, qapi_BlockGraphInfo_base(info), errp);
426 if (*errp) {
427 goto fail;
430 children_list_tail = &info->children;
432 QLIST_FOREACH(c, &bs->children, next) {
433 BlockChildInfo *c_info;
435 c_info = g_new0(BlockChildInfo, 1);
436 QAPI_LIST_APPEND(children_list_tail, c_info);
438 c_info->name = g_strdup(c->name);
439 bdrv_query_block_graph_info(c->bs, &c_info->info, errp);
440 if (*errp) {
441 goto fail;
445 *p_info = info;
446 return;
448 fail:
449 assert(*errp != NULL);
450 qapi_free_BlockGraphInfo(info);
453 /* @p_info will be set only on success. */
454 static void bdrv_query_info(BlockBackend *blk, BlockInfo **p_info,
455 Error **errp)
457 BlockInfo *info = g_malloc0(sizeof(*info));
458 BlockDriverState *bs = blk_bs(blk);
459 char *qdev;
461 /* Skip automatically inserted nodes that the user isn't aware of */
462 bs = bdrv_skip_implicit_filters(bs);
464 info->device = g_strdup(blk_name(blk));
465 info->type = g_strdup("unknown");
466 info->locked = blk_dev_is_medium_locked(blk);
467 info->removable = blk_dev_has_removable_media(blk);
469 qdev = blk_get_attached_dev_id(blk);
470 if (qdev && *qdev) {
471 info->qdev = qdev;
472 } else {
473 g_free(qdev);
476 if (blk_dev_has_tray(blk)) {
477 info->has_tray_open = true;
478 info->tray_open = blk_dev_is_tray_open(blk);
481 if (blk_iostatus_is_enabled(blk)) {
482 info->has_io_status = true;
483 info->io_status = blk_iostatus(blk);
486 if (bs && bs->drv) {
487 info->inserted = bdrv_block_device_info(blk, bs, false, errp);
488 if (info->inserted == NULL) {
489 goto err;
493 *p_info = info;
494 return;
496 err:
497 qapi_free_BlockInfo(info);
500 static uint64List *uint64_list(uint64_t *list, int size)
502 int i;
503 uint64List *out_list = NULL;
504 uint64List **tail = &out_list;
506 for (i = 0; i < size; i++) {
507 QAPI_LIST_APPEND(tail, list[i]);
510 return out_list;
513 static BlockLatencyHistogramInfo *
514 bdrv_latency_histogram_stats(BlockLatencyHistogram *hist)
516 BlockLatencyHistogramInfo *info;
518 if (!hist->bins) {
519 return NULL;
522 info = g_new0(BlockLatencyHistogramInfo, 1);
523 info->boundaries = uint64_list(hist->boundaries, hist->nbins - 1);
524 info->bins = uint64_list(hist->bins, hist->nbins);
525 return info;
528 static void bdrv_query_blk_stats(BlockDeviceStats *ds, BlockBackend *blk)
530 BlockAcctStats *stats = blk_get_stats(blk);
531 BlockAcctTimedStats *ts = NULL;
532 BlockLatencyHistogram *hgram;
534 ds->rd_bytes = stats->nr_bytes[BLOCK_ACCT_READ];
535 ds->wr_bytes = stats->nr_bytes[BLOCK_ACCT_WRITE];
536 ds->zone_append_bytes = stats->nr_bytes[BLOCK_ACCT_ZONE_APPEND];
537 ds->unmap_bytes = stats->nr_bytes[BLOCK_ACCT_UNMAP];
538 ds->rd_operations = stats->nr_ops[BLOCK_ACCT_READ];
539 ds->wr_operations = stats->nr_ops[BLOCK_ACCT_WRITE];
540 ds->zone_append_operations = stats->nr_ops[BLOCK_ACCT_ZONE_APPEND];
541 ds->unmap_operations = stats->nr_ops[BLOCK_ACCT_UNMAP];
543 ds->failed_rd_operations = stats->failed_ops[BLOCK_ACCT_READ];
544 ds->failed_wr_operations = stats->failed_ops[BLOCK_ACCT_WRITE];
545 ds->failed_zone_append_operations =
546 stats->failed_ops[BLOCK_ACCT_ZONE_APPEND];
547 ds->failed_flush_operations = stats->failed_ops[BLOCK_ACCT_FLUSH];
548 ds->failed_unmap_operations = stats->failed_ops[BLOCK_ACCT_UNMAP];
550 ds->invalid_rd_operations = stats->invalid_ops[BLOCK_ACCT_READ];
551 ds->invalid_wr_operations = stats->invalid_ops[BLOCK_ACCT_WRITE];
552 ds->invalid_zone_append_operations =
553 stats->invalid_ops[BLOCK_ACCT_ZONE_APPEND];
554 ds->invalid_flush_operations =
555 stats->invalid_ops[BLOCK_ACCT_FLUSH];
556 ds->invalid_unmap_operations = stats->invalid_ops[BLOCK_ACCT_UNMAP];
558 ds->rd_merged = stats->merged[BLOCK_ACCT_READ];
559 ds->wr_merged = stats->merged[BLOCK_ACCT_WRITE];
560 ds->zone_append_merged = stats->merged[BLOCK_ACCT_ZONE_APPEND];
561 ds->unmap_merged = stats->merged[BLOCK_ACCT_UNMAP];
562 ds->flush_operations = stats->nr_ops[BLOCK_ACCT_FLUSH];
563 ds->wr_total_time_ns = stats->total_time_ns[BLOCK_ACCT_WRITE];
564 ds->zone_append_total_time_ns =
565 stats->total_time_ns[BLOCK_ACCT_ZONE_APPEND];
566 ds->rd_total_time_ns = stats->total_time_ns[BLOCK_ACCT_READ];
567 ds->flush_total_time_ns = stats->total_time_ns[BLOCK_ACCT_FLUSH];
568 ds->unmap_total_time_ns = stats->total_time_ns[BLOCK_ACCT_UNMAP];
570 ds->has_idle_time_ns = stats->last_access_time_ns > 0;
571 if (ds->has_idle_time_ns) {
572 ds->idle_time_ns = block_acct_idle_time_ns(stats);
575 ds->account_invalid = stats->account_invalid;
576 ds->account_failed = stats->account_failed;
578 while ((ts = block_acct_interval_next(stats, ts))) {
579 BlockDeviceTimedStats *dev_stats = g_malloc0(sizeof(*dev_stats));
581 TimedAverage *rd = &ts->latency[BLOCK_ACCT_READ];
582 TimedAverage *wr = &ts->latency[BLOCK_ACCT_WRITE];
583 TimedAverage *zap = &ts->latency[BLOCK_ACCT_ZONE_APPEND];
584 TimedAverage *fl = &ts->latency[BLOCK_ACCT_FLUSH];
586 dev_stats->interval_length = ts->interval_length;
588 dev_stats->min_rd_latency_ns = timed_average_min(rd);
589 dev_stats->max_rd_latency_ns = timed_average_max(rd);
590 dev_stats->avg_rd_latency_ns = timed_average_avg(rd);
592 dev_stats->min_wr_latency_ns = timed_average_min(wr);
593 dev_stats->max_wr_latency_ns = timed_average_max(wr);
594 dev_stats->avg_wr_latency_ns = timed_average_avg(wr);
596 dev_stats->min_zone_append_latency_ns = timed_average_min(zap);
597 dev_stats->max_zone_append_latency_ns = timed_average_max(zap);
598 dev_stats->avg_zone_append_latency_ns = timed_average_avg(zap);
600 dev_stats->min_flush_latency_ns = timed_average_min(fl);
601 dev_stats->max_flush_latency_ns = timed_average_max(fl);
602 dev_stats->avg_flush_latency_ns = timed_average_avg(fl);
604 dev_stats->avg_rd_queue_depth =
605 block_acct_queue_depth(ts, BLOCK_ACCT_READ);
606 dev_stats->avg_wr_queue_depth =
607 block_acct_queue_depth(ts, BLOCK_ACCT_WRITE);
608 dev_stats->avg_zone_append_queue_depth =
609 block_acct_queue_depth(ts, BLOCK_ACCT_ZONE_APPEND);
611 QAPI_LIST_PREPEND(ds->timed_stats, dev_stats);
614 hgram = stats->latency_histogram;
615 ds->rd_latency_histogram
616 = bdrv_latency_histogram_stats(&hgram[BLOCK_ACCT_READ]);
617 ds->wr_latency_histogram
618 = bdrv_latency_histogram_stats(&hgram[BLOCK_ACCT_WRITE]);
619 ds->zone_append_latency_histogram
620 = bdrv_latency_histogram_stats(&hgram[BLOCK_ACCT_ZONE_APPEND]);
621 ds->flush_latency_histogram
622 = bdrv_latency_histogram_stats(&hgram[BLOCK_ACCT_FLUSH]);
625 static BlockStats * GRAPH_RDLOCK
626 bdrv_query_bds_stats(BlockDriverState *bs, bool blk_level)
628 BdrvChild *parent_child;
629 BlockDriverState *filter_or_cow_bs;
630 BlockStats *s = NULL;
632 s = g_malloc0(sizeof(*s));
633 s->stats = g_malloc0(sizeof(*s->stats));
635 if (!bs) {
636 return s;
639 /* Skip automatically inserted nodes that the user isn't aware of in
640 * a BlockBackend-level command. Stay at the exact node for a node-level
641 * command. */
642 if (blk_level) {
643 bs = bdrv_skip_implicit_filters(bs);
646 if (bdrv_get_node_name(bs)[0]) {
647 s->node_name = g_strdup(bdrv_get_node_name(bs));
650 s->stats->wr_highest_offset = stat64_get(&bs->wr_highest_offset);
652 s->driver_specific = bdrv_get_specific_stats(bs);
654 parent_child = bdrv_primary_child(bs);
655 if (!parent_child ||
656 !(parent_child->role & (BDRV_CHILD_DATA | BDRV_CHILD_FILTERED)))
658 BdrvChild *c;
661 * Look for a unique data-storing child. We do not need to look for
662 * filtered children, as there would be only one and it would have been
663 * the primary child.
665 parent_child = NULL;
666 QLIST_FOREACH(c, &bs->children, next) {
667 if (c->role & BDRV_CHILD_DATA) {
668 if (parent_child) {
670 * There are multiple data-storing children and we cannot
671 * choose between them.
673 parent_child = NULL;
674 break;
676 parent_child = c;
680 if (parent_child) {
681 s->parent = bdrv_query_bds_stats(parent_child->bs, blk_level);
684 filter_or_cow_bs = bdrv_filter_or_cow_bs(bs);
685 if (blk_level && filter_or_cow_bs) {
687 * Put any filtered or COW child here (for backwards
688 * compatibility to when we put bs0->backing here, which might
689 * be either)
691 s->backing = bdrv_query_bds_stats(filter_or_cow_bs, blk_level);
694 return s;
697 BlockInfoList *qmp_query_block(Error **errp)
699 BlockInfoList *head = NULL, **p_next = &head;
700 BlockBackend *blk;
701 Error *local_err = NULL;
703 for (blk = blk_all_next(NULL); blk; blk = blk_all_next(blk)) {
704 BlockInfoList *info;
706 if (!*blk_name(blk) && !blk_get_attached_dev(blk)) {
707 continue;
710 info = g_malloc0(sizeof(*info));
711 bdrv_query_info(blk, &info->value, &local_err);
712 if (local_err) {
713 error_propagate(errp, local_err);
714 g_free(info);
715 qapi_free_BlockInfoList(head);
716 return NULL;
719 *p_next = info;
720 p_next = &info->next;
723 return head;
726 BlockStatsList *qmp_query_blockstats(bool has_query_nodes,
727 bool query_nodes,
728 Error **errp)
730 BlockStatsList *head = NULL, **tail = &head;
731 BlockBackend *blk;
732 BlockDriverState *bs;
734 GRAPH_RDLOCK_GUARD_MAINLOOP();
736 /* Just to be safe if query_nodes is not always initialized */
737 if (has_query_nodes && query_nodes) {
738 for (bs = bdrv_next_node(NULL); bs; bs = bdrv_next_node(bs)) {
739 AioContext *ctx = bdrv_get_aio_context(bs);
741 aio_context_acquire(ctx);
742 QAPI_LIST_APPEND(tail, bdrv_query_bds_stats(bs, false));
743 aio_context_release(ctx);
745 } else {
746 for (blk = blk_all_next(NULL); blk; blk = blk_all_next(blk)) {
747 AioContext *ctx = blk_get_aio_context(blk);
748 BlockStats *s;
749 char *qdev;
751 if (!*blk_name(blk) && !blk_get_attached_dev(blk)) {
752 continue;
755 aio_context_acquire(ctx);
756 s = bdrv_query_bds_stats(blk_bs(blk), true);
757 s->device = g_strdup(blk_name(blk));
759 qdev = blk_get_attached_dev_id(blk);
760 if (qdev && *qdev) {
761 s->qdev = qdev;
762 } else {
763 g_free(qdev);
766 bdrv_query_blk_stats(s->stats, blk);
767 aio_context_release(ctx);
769 QAPI_LIST_APPEND(tail, s);
773 return head;
776 void bdrv_snapshot_dump(QEMUSnapshotInfo *sn)
778 char clock_buf[128];
779 char icount_buf[128] = {0};
780 int64_t secs;
781 char *sizing = NULL;
783 if (!sn) {
784 qemu_printf("%-10s%-17s%8s%20s%13s%11s",
785 "ID", "TAG", "VM SIZE", "DATE", "VM CLOCK", "ICOUNT");
786 } else {
787 g_autoptr(GDateTime) date = g_date_time_new_from_unix_local(sn->date_sec);
788 g_autofree char *date_buf = g_date_time_format(date, "%Y-%m-%d %H:%M:%S");
790 secs = sn->vm_clock_nsec / 1000000000;
791 snprintf(clock_buf, sizeof(clock_buf),
792 "%02d:%02d:%02d.%03d",
793 (int)(secs / 3600),
794 (int)((secs / 60) % 60),
795 (int)(secs % 60),
796 (int)((sn->vm_clock_nsec / 1000000) % 1000));
797 sizing = size_to_str(sn->vm_state_size);
798 if (sn->icount != -1ULL) {
799 snprintf(icount_buf, sizeof(icount_buf),
800 "%"PRId64, sn->icount);
802 qemu_printf("%-9s %-16s %8s%20s%13s%11s",
803 sn->id_str, sn->name,
804 sizing,
805 date_buf,
806 clock_buf,
807 icount_buf);
809 g_free(sizing);
812 static void dump_qdict(int indentation, QDict *dict);
813 static void dump_qlist(int indentation, QList *list);
815 static void dump_qobject(int comp_indent, QObject *obj)
817 switch (qobject_type(obj)) {
818 case QTYPE_QNUM: {
819 QNum *value = qobject_to(QNum, obj);
820 char *tmp = qnum_to_string(value);
821 qemu_printf("%s", tmp);
822 g_free(tmp);
823 break;
825 case QTYPE_QSTRING: {
826 QString *value = qobject_to(QString, obj);
827 qemu_printf("%s", qstring_get_str(value));
828 break;
830 case QTYPE_QDICT: {
831 QDict *value = qobject_to(QDict, obj);
832 dump_qdict(comp_indent, value);
833 break;
835 case QTYPE_QLIST: {
836 QList *value = qobject_to(QList, obj);
837 dump_qlist(comp_indent, value);
838 break;
840 case QTYPE_QBOOL: {
841 QBool *value = qobject_to(QBool, obj);
842 qemu_printf("%s", qbool_get_bool(value) ? "true" : "false");
843 break;
845 default:
846 abort();
850 static void dump_qlist(int indentation, QList *list)
852 const QListEntry *entry;
853 int i = 0;
855 for (entry = qlist_first(list); entry; entry = qlist_next(entry), i++) {
856 QType type = qobject_type(entry->value);
857 bool composite = (type == QTYPE_QDICT || type == QTYPE_QLIST);
858 qemu_printf("%*s[%i]:%c", indentation * 4, "", i,
859 composite ? '\n' : ' ');
860 dump_qobject(indentation + 1, entry->value);
861 if (!composite) {
862 qemu_printf("\n");
867 static void dump_qdict(int indentation, QDict *dict)
869 const QDictEntry *entry;
871 for (entry = qdict_first(dict); entry; entry = qdict_next(dict, entry)) {
872 QType type = qobject_type(entry->value);
873 bool composite = (type == QTYPE_QDICT || type == QTYPE_QLIST);
874 char *key = g_malloc(strlen(entry->key) + 1);
875 int i;
877 /* replace dashes with spaces in key (variable) names */
878 for (i = 0; entry->key[i]; i++) {
879 key[i] = entry->key[i] == '-' ? ' ' : entry->key[i];
881 key[i] = 0;
882 qemu_printf("%*s%s:%c", indentation * 4, "", key,
883 composite ? '\n' : ' ');
884 dump_qobject(indentation + 1, entry->value);
885 if (!composite) {
886 qemu_printf("\n");
888 g_free(key);
893 * Return whether dumping the given QObject with dump_qobject() would
894 * yield an empty dump, i.e. not print anything.
896 static bool qobject_is_empty_dump(const QObject *obj)
898 switch (qobject_type(obj)) {
899 case QTYPE_QNUM:
900 case QTYPE_QSTRING:
901 case QTYPE_QBOOL:
902 return false;
904 case QTYPE_QDICT:
905 return qdict_size(qobject_to(QDict, obj)) == 0;
907 case QTYPE_QLIST:
908 return qlist_empty(qobject_to(QList, obj));
910 default:
911 abort();
916 * Dumps the given ImageInfoSpecific object in a human-readable form,
917 * prepending an optional prefix if the dump is not empty.
919 void bdrv_image_info_specific_dump(ImageInfoSpecific *info_spec,
920 const char *prefix,
921 int indentation)
923 QObject *obj, *data;
924 Visitor *v = qobject_output_visitor_new(&obj);
926 visit_type_ImageInfoSpecific(v, NULL, &info_spec, &error_abort);
927 visit_complete(v, &obj);
928 data = qdict_get(qobject_to(QDict, obj), "data");
929 if (!qobject_is_empty_dump(data)) {
930 if (prefix) {
931 qemu_printf("%*s%s", indentation * 4, "", prefix);
933 dump_qobject(indentation + 1, data);
935 qobject_unref(obj);
936 visit_free(v);
940 * Print the given @info object in human-readable form. Every field is indented
941 * using the given @indentation (four spaces per indentation level).
943 * When using this to print a whole block graph, @protocol can be set to true to
944 * signify that the given information is associated with a protocol node, i.e.
945 * just data storage for an image, such that the data it presents is not really
946 * a full VM disk. If so, several fields change name: For example, "virtual
947 * size" is printed as "file length".
948 * (Consider a qcow2 image, which is represented by a qcow2 node and a file
949 * node. Printing a "virtual size" for the file node does not make sense,
950 * because without the qcow2 node, it is not really a guest disk, so it does not
951 * have a "virtual size". Therefore, we call it "file length" instead.)
953 * @protocol is ignored when @indentation is 0, because we take that to mean
954 * that the associated node is the root node in the queried block graph, and
955 * thus is always to be interpreted as a standalone guest disk.
957 void bdrv_node_info_dump(BlockNodeInfo *info, int indentation, bool protocol)
959 char *size_buf, *dsize_buf;
960 g_autofree char *ind_s = g_strdup_printf("%*s", indentation * 4, "");
962 if (indentation == 0) {
963 /* Top level, consider this a normal image */
964 protocol = false;
967 if (!info->has_actual_size) {
968 dsize_buf = g_strdup("unavailable");
969 } else {
970 dsize_buf = size_to_str(info->actual_size);
972 size_buf = size_to_str(info->virtual_size);
973 qemu_printf("%s%s: %s\n"
974 "%s%s: %s\n"
975 "%s%s: %s (%" PRId64 " bytes)\n"
976 "%sdisk size: %s\n",
977 ind_s, protocol ? "filename" : "image", info->filename,
978 ind_s, protocol ? "protocol type" : "file format",
979 info->format,
980 ind_s, protocol ? "file length" : "virtual size",
981 size_buf, info->virtual_size,
982 ind_s, dsize_buf);
983 g_free(size_buf);
984 g_free(dsize_buf);
986 if (info->has_encrypted && info->encrypted) {
987 qemu_printf("%sencrypted: yes\n", ind_s);
990 if (info->has_cluster_size) {
991 qemu_printf("%scluster_size: %" PRId64 "\n",
992 ind_s, info->cluster_size);
995 if (info->has_dirty_flag && info->dirty_flag) {
996 qemu_printf("%scleanly shut down: no\n", ind_s);
999 if (info->backing_filename) {
1000 qemu_printf("%sbacking file: %s", ind_s, info->backing_filename);
1001 if (!info->full_backing_filename) {
1002 qemu_printf(" (cannot determine actual path)");
1003 } else if (strcmp(info->backing_filename,
1004 info->full_backing_filename) != 0) {
1005 qemu_printf(" (actual path: %s)", info->full_backing_filename);
1007 qemu_printf("\n");
1008 if (info->backing_filename_format) {
1009 qemu_printf("%sbacking file format: %s\n",
1010 ind_s, info->backing_filename_format);
1014 if (info->has_snapshots) {
1015 SnapshotInfoList *elem;
1017 qemu_printf("%sSnapshot list:\n", ind_s);
1018 qemu_printf("%s", ind_s);
1019 bdrv_snapshot_dump(NULL);
1020 qemu_printf("\n");
1022 /* Ideally bdrv_snapshot_dump() would operate on SnapshotInfoList but
1023 * we convert to the block layer's native QEMUSnapshotInfo for now.
1025 for (elem = info->snapshots; elem; elem = elem->next) {
1026 QEMUSnapshotInfo sn = {
1027 .vm_state_size = elem->value->vm_state_size,
1028 .date_sec = elem->value->date_sec,
1029 .date_nsec = elem->value->date_nsec,
1030 .vm_clock_nsec = elem->value->vm_clock_sec * 1000000000ULL +
1031 elem->value->vm_clock_nsec,
1032 .icount = elem->value->has_icount ?
1033 elem->value->icount : -1ULL,
1036 pstrcpy(sn.id_str, sizeof(sn.id_str), elem->value->id);
1037 pstrcpy(sn.name, sizeof(sn.name), elem->value->name);
1038 qemu_printf("%s", ind_s);
1039 bdrv_snapshot_dump(&sn);
1040 qemu_printf("\n");
1044 if (info->format_specific) {
1045 bdrv_image_info_specific_dump(info->format_specific,
1046 "Format specific information:\n",
1047 indentation);