2 * QEMU block throttling group infrastructure
4 * Copyright (C) Nodalink, EURL. 2014
5 * Copyright (C) Igalia, S.L. 2015
8 * BenoƮt Canet <benoit.canet@nodalink.com>
9 * Alberto Garcia <berto@igalia.com>
11 * This program is free software; you can redistribute it and/or
12 * modify it under the terms of the GNU General Public License as
13 * published by the Free Software Foundation; either version 2 or
14 * (at your option) version 3 of the License.
16 * This program is distributed in the hope that it will be useful,
17 * but WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
19 * GNU General Public License for more details.
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, see <http://www.gnu.org/licenses/>.
25 #include "qemu/osdep.h"
26 #include "sysemu/block-backend.h"
27 #include "block/throttle-groups.h"
28 #include "qemu/throttle-options.h"
29 #include "qemu/main-loop.h"
30 #include "qemu/queue.h"
31 #include "qemu/thread.h"
32 #include "sysemu/qtest.h"
33 #include "qapi/error.h"
34 #include "qapi/qapi-visit-block-core.h"
35 #include "qom/object.h"
36 #include "qom/object_interfaces.h"
38 static void throttle_group_obj_init(Object
*obj
);
39 static void throttle_group_obj_complete(UserCreatable
*obj
, Error
**errp
);
40 static void timer_cb(ThrottleGroupMember
*tgm
, ThrottleDirection direction
);
42 /* The ThrottleGroup structure (with its ThrottleState) is shared
43 * among different ThrottleGroupMembers and it's independent from
44 * AioContext, so in order to use it from different threads it needs
47 * This locking is however handled internally in this file, so it's
48 * transparent to outside users.
50 * The whole ThrottleGroup structure is private and invisible to
51 * outside users, that only use it through its ThrottleState.
53 * In addition to the ThrottleGroup structure, ThrottleGroupMember has
54 * fields that need to be accessed by other members of the group and
55 * therefore also need to be protected by this lock. Once a
56 * ThrottleGroupMember is registered in a group those fields can be accessed
57 * by other threads any time.
59 * Again, all this is handled internally and is mostly transparent to
60 * the outside. The 'throttle_timers' field however has an additional
61 * constraint because it may be temporarily invalid (see for example
62 * blk_set_aio_context()). Therefore in this file a thread will
63 * access some other ThrottleGroupMember's timers only after verifying that
64 * that ThrottleGroupMember has throttled requests in the queue.
66 struct ThrottleGroup
{
69 /* refuse individual property change if initialization is complete */
71 char *name
; /* This is constant during the lifetime of the group */
73 QemuMutex lock
; /* This lock protects the following four fields */
75 QLIST_HEAD(, ThrottleGroupMember
) head
;
76 ThrottleGroupMember
*tokens
[THROTTLE_MAX
];
77 bool any_timer_armed
[THROTTLE_MAX
];
78 QEMUClockType clock_type
;
80 /* This field is protected by the global QEMU mutex */
81 QTAILQ_ENTRY(ThrottleGroup
) list
;
84 /* This is protected by the global QEMU mutex */
85 static QTAILQ_HEAD(, ThrottleGroup
) throttle_groups
=
86 QTAILQ_HEAD_INITIALIZER(throttle_groups
);
89 /* This function reads throttle_groups and must be called under the global
92 static ThrottleGroup
*throttle_group_by_name(const char *name
)
96 /* Look for an existing group with that name */
97 QTAILQ_FOREACH(iter
, &throttle_groups
, list
) {
98 if (!g_strcmp0(name
, iter
->name
)) {
106 /* This function reads throttle_groups and must be called under the global
109 bool throttle_group_exists(const char *name
)
111 return throttle_group_by_name(name
) != NULL
;
114 /* Increments the reference count of a ThrottleGroup given its name.
116 * If no ThrottleGroup is found with the given name a new one is
119 * This function edits throttle_groups and must be called under the global
122 * @name: the name of the ThrottleGroup
123 * @ret: the ThrottleState member of the ThrottleGroup
125 ThrottleState
*throttle_group_incref(const char *name
)
127 ThrottleGroup
*tg
= NULL
;
129 /* Look for an existing group with that name */
130 tg
= throttle_group_by_name(name
);
133 object_ref(OBJECT(tg
));
135 /* Create a new one if not found */
136 /* new ThrottleGroup obj will have a refcnt = 1 */
137 tg
= THROTTLE_GROUP(object_new(TYPE_THROTTLE_GROUP
));
138 tg
->name
= g_strdup(name
);
139 throttle_group_obj_complete(USER_CREATABLE(tg
), &error_abort
);
145 /* Decrease the reference count of a ThrottleGroup.
147 * When the reference count reaches zero the ThrottleGroup is
150 * This function edits throttle_groups and must be called under the global
153 * @ts: The ThrottleGroup to unref, given by its ThrottleState member
155 void throttle_group_unref(ThrottleState
*ts
)
157 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
158 object_unref(OBJECT(tg
));
161 /* Get the name from a ThrottleGroupMember's group. The name (and the pointer)
162 * is guaranteed to remain constant during the lifetime of the group.
164 * @tgm: a ThrottleGroupMember
165 * @ret: the name of the group.
167 const char *throttle_group_get_name(ThrottleGroupMember
*tgm
)
169 ThrottleGroup
*tg
= container_of(tgm
->throttle_state
, ThrottleGroup
, ts
);
173 /* Return the next ThrottleGroupMember in the round-robin sequence, simulating
176 * This assumes that tg->lock is held.
178 * @tgm: the current ThrottleGroupMember
179 * @ret: the next ThrottleGroupMember in the sequence
181 static ThrottleGroupMember
*throttle_group_next_tgm(ThrottleGroupMember
*tgm
)
183 ThrottleState
*ts
= tgm
->throttle_state
;
184 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
185 ThrottleGroupMember
*next
= QLIST_NEXT(tgm
, round_robin
);
188 next
= QLIST_FIRST(&tg
->head
);
195 * Return whether a ThrottleGroupMember has pending requests.
197 * This assumes that tg->lock is held.
199 * @tgm: the ThrottleGroupMember
200 * @direction: the ThrottleDirection
201 * @ret: whether the ThrottleGroupMember has pending requests.
203 static inline bool tgm_has_pending_reqs(ThrottleGroupMember
*tgm
,
204 ThrottleDirection direction
)
206 return tgm
->pending_reqs
[direction
];
209 /* Return the next ThrottleGroupMember in the round-robin sequence with pending
212 * This assumes that tg->lock is held.
214 * @tgm: the current ThrottleGroupMember
215 * @direction: the ThrottleDirection
216 * @ret: the next ThrottleGroupMember with pending requests, or tgm if
219 static ThrottleGroupMember
*next_throttle_token(ThrottleGroupMember
*tgm
,
220 ThrottleDirection direction
)
222 ThrottleState
*ts
= tgm
->throttle_state
;
223 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
224 ThrottleGroupMember
*token
, *start
;
226 /* If this member has its I/O limits disabled then it means that
227 * it's being drained. Skip the round-robin search and return tgm
228 * immediately if it has pending requests. Otherwise we could be
229 * forcing it to wait for other member's throttled requests. */
230 if (tgm_has_pending_reqs(tgm
, direction
) &&
231 qatomic_read(&tgm
->io_limits_disabled
)) {
235 start
= token
= tg
->tokens
[direction
];
237 /* get next bs round in round robin style */
238 token
= throttle_group_next_tgm(token
);
239 while (token
!= start
&& !tgm_has_pending_reqs(token
, direction
)) {
240 token
= throttle_group_next_tgm(token
);
243 /* If no IO are queued for scheduling on the next round robin token
244 * then decide the token is the current tgm because chances are
245 * the current tgm got the current request queued.
247 if (token
== start
&& !tgm_has_pending_reqs(token
, direction
)) {
251 /* Either we return the original TGM, or one with pending requests */
252 assert(token
== tgm
|| tgm_has_pending_reqs(token
, direction
));
257 /* Check if the next I/O request for a ThrottleGroupMember needs to be
258 * throttled or not. If there's no timer set in this group, set one and update
259 * the token accordingly.
261 * This assumes that tg->lock is held.
263 * @tgm: the current ThrottleGroupMember
264 * @direction: the ThrottleDirection
265 * @ret: whether the I/O request needs to be throttled or not
267 static bool throttle_group_schedule_timer(ThrottleGroupMember
*tgm
,
268 ThrottleDirection direction
)
270 ThrottleState
*ts
= tgm
->throttle_state
;
271 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
272 ThrottleTimers
*tt
= &tgm
->throttle_timers
;
275 if (qatomic_read(&tgm
->io_limits_disabled
)) {
279 /* Check if any of the timers in this group is already armed */
280 if (tg
->any_timer_armed
[direction
]) {
284 must_wait
= throttle_schedule_timer(ts
, tt
, direction
);
286 /* If a timer just got armed, set tgm as the current token */
288 tg
->tokens
[direction
] = tgm
;
289 tg
->any_timer_armed
[direction
] = true;
295 /* Start the next pending I/O request for a ThrottleGroupMember. Return whether
296 * any request was actually pending.
298 * @tgm: the current ThrottleGroupMember
299 * @direction: the ThrottleDirection
301 static bool coroutine_fn
throttle_group_co_restart_queue(ThrottleGroupMember
*tgm
,
302 ThrottleDirection direction
)
306 qemu_co_mutex_lock(&tgm
->throttled_reqs_lock
);
307 ret
= qemu_co_queue_next(&tgm
->throttled_reqs
[direction
]);
308 qemu_co_mutex_unlock(&tgm
->throttled_reqs_lock
);
313 /* Look for the next pending I/O request and schedule it.
315 * This assumes that tg->lock is held.
317 * @tgm: the current ThrottleGroupMember
318 * @direction: the ThrottleDirection
320 static void schedule_next_request(ThrottleGroupMember
*tgm
,
321 ThrottleDirection direction
)
323 ThrottleState
*ts
= tgm
->throttle_state
;
324 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
326 ThrottleGroupMember
*token
;
328 /* Check if there's any pending request to schedule next */
329 token
= next_throttle_token(tgm
, direction
);
330 if (!tgm_has_pending_reqs(token
, direction
)) {
334 /* Set a timer for the request if it needs to be throttled */
335 must_wait
= throttle_group_schedule_timer(token
, direction
);
337 /* If it doesn't have to wait, queue it for immediate execution */
339 /* Give preference to requests from the current tgm */
340 if (qemu_in_coroutine() &&
341 throttle_group_co_restart_queue(tgm
, direction
)) {
344 ThrottleTimers
*tt
= &token
->throttle_timers
;
345 int64_t now
= qemu_clock_get_ns(tg
->clock_type
);
346 timer_mod(tt
->timers
[direction
], now
);
347 tg
->any_timer_armed
[direction
] = true;
349 tg
->tokens
[direction
] = token
;
353 /* Check if an I/O request needs to be throttled, wait and set a timer
354 * if necessary, and schedule the next request using a round robin
357 * @tgm: the current ThrottleGroupMember
358 * @bytes: the number of bytes for this I/O
359 * @direction: the ThrottleDirection
361 void coroutine_fn
throttle_group_co_io_limits_intercept(ThrottleGroupMember
*tgm
,
363 ThrottleDirection direction
)
366 ThrottleGroupMember
*token
;
367 ThrottleGroup
*tg
= container_of(tgm
->throttle_state
, ThrottleGroup
, ts
);
370 assert(direction
< THROTTLE_MAX
);
372 qemu_mutex_lock(&tg
->lock
);
374 /* First we check if this I/O has to be throttled. */
375 token
= next_throttle_token(tgm
, direction
);
376 must_wait
= throttle_group_schedule_timer(token
, direction
);
378 /* Wait if there's a timer set or queued requests of this type */
379 if (must_wait
|| tgm
->pending_reqs
[direction
]) {
380 tgm
->pending_reqs
[direction
]++;
381 qemu_mutex_unlock(&tg
->lock
);
382 qemu_co_mutex_lock(&tgm
->throttled_reqs_lock
);
383 qemu_co_queue_wait(&tgm
->throttled_reqs
[direction
],
384 &tgm
->throttled_reqs_lock
);
385 qemu_co_mutex_unlock(&tgm
->throttled_reqs_lock
);
386 qemu_mutex_lock(&tg
->lock
);
387 tgm
->pending_reqs
[direction
]--;
390 /* The I/O will be executed, so do the accounting */
391 throttle_account(tgm
->throttle_state
, direction
, bytes
);
393 /* Schedule the next request */
394 schedule_next_request(tgm
, direction
);
396 qemu_mutex_unlock(&tg
->lock
);
400 ThrottleGroupMember
*tgm
;
401 ThrottleDirection direction
;
404 static void coroutine_fn
throttle_group_restart_queue_entry(void *opaque
)
406 RestartData
*data
= opaque
;
407 ThrottleGroupMember
*tgm
= data
->tgm
;
408 ThrottleState
*ts
= tgm
->throttle_state
;
409 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
410 ThrottleDirection direction
= data
->direction
;
413 empty_queue
= !throttle_group_co_restart_queue(tgm
, direction
);
415 /* If the request queue was empty then we have to take care of
416 * scheduling the next one */
418 qemu_mutex_lock(&tg
->lock
);
419 schedule_next_request(tgm
, direction
);
420 qemu_mutex_unlock(&tg
->lock
);
425 qatomic_dec(&tgm
->restart_pending
);
429 static void throttle_group_restart_queue(ThrottleGroupMember
*tgm
,
430 ThrottleDirection direction
)
433 RestartData
*rd
= g_new0(RestartData
, 1);
436 rd
->direction
= direction
;
438 /* This function is called when a timer is fired or when
439 * throttle_group_restart_tgm() is called. Either way, there can
440 * be no timer pending on this tgm at this point */
441 assert(!timer_pending(tgm
->throttle_timers
.timers
[direction
]));
443 qatomic_inc(&tgm
->restart_pending
);
445 co
= qemu_coroutine_create(throttle_group_restart_queue_entry
, rd
);
446 aio_co_enter(tgm
->aio_context
, co
);
449 void throttle_group_restart_tgm(ThrottleGroupMember
*tgm
)
451 ThrottleDirection dir
;
453 if (tgm
->throttle_state
) {
454 for (dir
= THROTTLE_READ
; dir
< THROTTLE_MAX
; dir
++) {
455 QEMUTimer
*t
= tgm
->throttle_timers
.timers
[dir
];
456 if (timer_pending(t
)) {
457 /* If there's a pending timer on this tgm, fire it now */
461 /* Else run the next request from the queue manually */
462 throttle_group_restart_queue(tgm
, dir
);
468 /* Update the throttle configuration for a particular group. Similar
469 * to throttle_config(), but guarantees atomicity within the
472 * @tgm: a ThrottleGroupMember that is a member of the group
473 * @cfg: the configuration to set
475 void throttle_group_config(ThrottleGroupMember
*tgm
, ThrottleConfig
*cfg
)
477 ThrottleState
*ts
= tgm
->throttle_state
;
478 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
479 qemu_mutex_lock(&tg
->lock
);
480 throttle_config(ts
, tg
->clock_type
, cfg
);
481 qemu_mutex_unlock(&tg
->lock
);
483 throttle_group_restart_tgm(tgm
);
486 /* Get the throttle configuration from a particular group. Similar to
487 * throttle_get_config(), but guarantees atomicity within the
490 * @tgm: a ThrottleGroupMember that is a member of the group
491 * @cfg: the configuration will be written here
493 void throttle_group_get_config(ThrottleGroupMember
*tgm
, ThrottleConfig
*cfg
)
495 ThrottleState
*ts
= tgm
->throttle_state
;
496 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
497 qemu_mutex_lock(&tg
->lock
);
498 throttle_get_config(ts
, cfg
);
499 qemu_mutex_unlock(&tg
->lock
);
502 /* ThrottleTimers callback. This wakes up a request that was waiting
503 * because it had been throttled.
505 * @tgm: the ThrottleGroupMember whose request had been throttled
506 * @direction: the ThrottleDirection
508 static void timer_cb(ThrottleGroupMember
*tgm
, ThrottleDirection direction
)
510 ThrottleState
*ts
= tgm
->throttle_state
;
511 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
513 /* The timer has just been fired, so we can update the flag */
514 qemu_mutex_lock(&tg
->lock
);
515 tg
->any_timer_armed
[direction
] = false;
516 qemu_mutex_unlock(&tg
->lock
);
518 /* Run the request that was waiting for this timer */
519 throttle_group_restart_queue(tgm
, direction
);
522 static void read_timer_cb(void *opaque
)
524 timer_cb(opaque
, THROTTLE_READ
);
527 static void write_timer_cb(void *opaque
)
529 timer_cb(opaque
, THROTTLE_WRITE
);
532 /* Register a ThrottleGroupMember from the throttling group, also initializing
533 * its timers and updating its throttle_state pointer to point to it. If a
534 * throttling group with that name does not exist yet, it will be created.
536 * This function edits throttle_groups and must be called under the global
539 * @tgm: the ThrottleGroupMember to insert
540 * @groupname: the name of the group
541 * @ctx: the AioContext to use
543 void throttle_group_register_tgm(ThrottleGroupMember
*tgm
,
544 const char *groupname
,
547 ThrottleDirection dir
;
548 ThrottleState
*ts
= throttle_group_incref(groupname
);
549 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
551 tgm
->throttle_state
= ts
;
552 tgm
->aio_context
= ctx
;
553 qatomic_set(&tgm
->restart_pending
, 0);
555 QEMU_LOCK_GUARD(&tg
->lock
);
556 /* If the ThrottleGroup is new set this ThrottleGroupMember as the token */
557 for (dir
= THROTTLE_READ
; dir
< THROTTLE_MAX
; dir
++) {
558 if (!tg
->tokens
[dir
]) {
559 tg
->tokens
[dir
] = tgm
;
561 qemu_co_queue_init(&tgm
->throttled_reqs
[dir
]);
564 QLIST_INSERT_HEAD(&tg
->head
, tgm
, round_robin
);
566 throttle_timers_init(&tgm
->throttle_timers
,
572 qemu_co_mutex_init(&tgm
->throttled_reqs_lock
);
575 /* Unregister a ThrottleGroupMember from its group, removing it from the list,
576 * destroying the timers and setting the throttle_state pointer to NULL.
578 * The ThrottleGroupMember must not have pending throttled requests, so the
579 * caller has to drain them first.
581 * The group will be destroyed if it's empty after this operation.
583 * @tgm the ThrottleGroupMember to remove
585 void throttle_group_unregister_tgm(ThrottleGroupMember
*tgm
)
587 ThrottleState
*ts
= tgm
->throttle_state
;
588 ThrottleGroup
*tg
= container_of(ts
, ThrottleGroup
, ts
);
589 ThrottleGroupMember
*token
;
590 ThrottleDirection dir
;
593 /* Discard already unregistered tgm */
597 /* Wait for throttle_group_restart_queue_entry() coroutines to finish */
598 AIO_WAIT_WHILE(tgm
->aio_context
, qatomic_read(&tgm
->restart_pending
) > 0);
600 WITH_QEMU_LOCK_GUARD(&tg
->lock
) {
601 for (dir
= THROTTLE_READ
; dir
< THROTTLE_MAX
; dir
++) {
602 assert(tgm
->pending_reqs
[dir
] == 0);
603 assert(qemu_co_queue_empty(&tgm
->throttled_reqs
[dir
]));
604 assert(!timer_pending(tgm
->throttle_timers
.timers
[dir
]));
605 if (tg
->tokens
[dir
] == tgm
) {
606 token
= throttle_group_next_tgm(tgm
);
607 /* Take care of the case where this is the last tgm in the group */
611 tg
->tokens
[dir
] = token
;
615 /* remove the current tgm from the list */
616 QLIST_REMOVE(tgm
, round_robin
);
617 throttle_timers_destroy(&tgm
->throttle_timers
);
620 throttle_group_unref(&tg
->ts
);
621 tgm
->throttle_state
= NULL
;
624 void throttle_group_attach_aio_context(ThrottleGroupMember
*tgm
,
625 AioContext
*new_context
)
627 ThrottleTimers
*tt
= &tgm
->throttle_timers
;
628 throttle_timers_attach_aio_context(tt
, new_context
);
629 tgm
->aio_context
= new_context
;
632 void throttle_group_detach_aio_context(ThrottleGroupMember
*tgm
)
634 ThrottleGroup
*tg
= container_of(tgm
->throttle_state
, ThrottleGroup
, ts
);
635 ThrottleTimers
*tt
= &tgm
->throttle_timers
;
636 ThrottleDirection dir
;
638 /* Requests must have been drained */
639 for (dir
= THROTTLE_READ
; dir
< THROTTLE_MAX
; dir
++) {
640 assert(tgm
->pending_reqs
[dir
] == 0);
641 assert(qemu_co_queue_empty(&tgm
->throttled_reqs
[dir
]));
644 /* Kick off next ThrottleGroupMember, if necessary */
645 WITH_QEMU_LOCK_GUARD(&tg
->lock
) {
646 for (dir
= THROTTLE_READ
; dir
< THROTTLE_MAX
; dir
++) {
647 if (timer_pending(tt
->timers
[dir
])) {
648 tg
->any_timer_armed
[dir
] = false;
649 schedule_next_request(tgm
, dir
);
654 throttle_timers_detach_aio_context(tt
);
655 tgm
->aio_context
= NULL
;
658 #undef THROTTLE_OPT_PREFIX
659 #define THROTTLE_OPT_PREFIX "x-"
661 /* Helper struct and array for QOM property setter/getter */
673 static ThrottleParamInfo properties
[] = {
675 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL
,
676 THROTTLE_OPS_TOTAL
, AVG
,
679 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL_MAX
,
680 THROTTLE_OPS_TOTAL
, MAX
,
683 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_TOTAL_MAX_LENGTH
,
684 THROTTLE_OPS_TOTAL
, BURST_LENGTH
,
687 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ
,
688 THROTTLE_OPS_READ
, AVG
,
691 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ_MAX
,
692 THROTTLE_OPS_READ
, MAX
,
695 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_READ_MAX_LENGTH
,
696 THROTTLE_OPS_READ
, BURST_LENGTH
,
699 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE
,
700 THROTTLE_OPS_WRITE
, AVG
,
703 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE_MAX
,
704 THROTTLE_OPS_WRITE
, MAX
,
707 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_WRITE_MAX_LENGTH
,
708 THROTTLE_OPS_WRITE
, BURST_LENGTH
,
711 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL
,
712 THROTTLE_BPS_TOTAL
, AVG
,
715 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL_MAX
,
716 THROTTLE_BPS_TOTAL
, MAX
,
719 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_TOTAL_MAX_LENGTH
,
720 THROTTLE_BPS_TOTAL
, BURST_LENGTH
,
723 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ
,
724 THROTTLE_BPS_READ
, AVG
,
727 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ_MAX
,
728 THROTTLE_BPS_READ
, MAX
,
731 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_READ_MAX_LENGTH
,
732 THROTTLE_BPS_READ
, BURST_LENGTH
,
735 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE
,
736 THROTTLE_BPS_WRITE
, AVG
,
739 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE_MAX
,
740 THROTTLE_BPS_WRITE
, MAX
,
743 THROTTLE_OPT_PREFIX QEMU_OPT_BPS_WRITE_MAX_LENGTH
,
744 THROTTLE_BPS_WRITE
, BURST_LENGTH
,
747 THROTTLE_OPT_PREFIX QEMU_OPT_IOPS_SIZE
,
752 /* This function edits throttle_groups and must be called under the global
754 static void throttle_group_obj_init(Object
*obj
)
756 ThrottleGroup
*tg
= THROTTLE_GROUP(obj
);
758 tg
->clock_type
= QEMU_CLOCK_REALTIME
;
759 if (qtest_enabled()) {
760 /* For testing block IO throttling only */
761 tg
->clock_type
= QEMU_CLOCK_VIRTUAL
;
763 tg
->is_initialized
= false;
764 qemu_mutex_init(&tg
->lock
);
765 throttle_init(&tg
->ts
);
766 QLIST_INIT(&tg
->head
);
769 /* This function edits throttle_groups and must be called under the global
771 static void throttle_group_obj_complete(UserCreatable
*obj
, Error
**errp
)
773 ThrottleGroup
*tg
= THROTTLE_GROUP(obj
);
776 /* set group name to object id if it exists */
777 if (!tg
->name
&& tg
->parent_obj
.parent
) {
778 tg
->name
= g_strdup(object_get_canonical_path_component(OBJECT(obj
)));
780 /* We must have a group name at this point */
783 /* error if name is duplicate */
784 if (throttle_group_exists(tg
->name
)) {
785 error_setg(errp
, "A group with this name already exists");
790 throttle_get_config(&tg
->ts
, &cfg
);
791 if (!throttle_is_valid(&cfg
, errp
)) {
794 throttle_config(&tg
->ts
, tg
->clock_type
, &cfg
);
795 QTAILQ_INSERT_TAIL(&throttle_groups
, tg
, list
);
796 tg
->is_initialized
= true;
799 /* This function edits throttle_groups and must be called under the global
801 static void throttle_group_obj_finalize(Object
*obj
)
803 ThrottleGroup
*tg
= THROTTLE_GROUP(obj
);
804 if (tg
->is_initialized
) {
805 QTAILQ_REMOVE(&throttle_groups
, tg
, list
);
807 qemu_mutex_destroy(&tg
->lock
);
811 static void throttle_group_set(Object
*obj
, Visitor
*v
, const char * name
,
812 void *opaque
, Error
**errp
)
815 ThrottleGroup
*tg
= THROTTLE_GROUP(obj
);
817 ThrottleParamInfo
*info
= opaque
;
820 /* If we have finished initialization, don't accept individual property
821 * changes through QOM. Throttle configuration limits must be set in one
822 * transaction, as certain combinations are invalid.
824 if (tg
->is_initialized
) {
825 error_setg(errp
, "Property cannot be set after initialization");
829 if (!visit_type_int64(v
, name
, &value
, errp
)) {
833 error_setg(errp
, "Property values cannot be negative");
838 switch (info
->category
) {
840 cfg
->buckets
[info
->type
].avg
= value
;
843 cfg
->buckets
[info
->type
].max
= value
;
846 if (value
> UINT_MAX
) {
847 error_setg(errp
, "%s value must be in the" "range [0, %u]",
848 info
->name
, UINT_MAX
);
851 cfg
->buckets
[info
->type
].burst_length
= value
;
854 cfg
->op_size
= value
;
859 static void throttle_group_get(Object
*obj
, Visitor
*v
, const char *name
,
860 void *opaque
, Error
**errp
)
862 ThrottleGroup
*tg
= THROTTLE_GROUP(obj
);
864 ThrottleParamInfo
*info
= opaque
;
867 throttle_get_config(&tg
->ts
, &cfg
);
868 switch (info
->category
) {
870 value
= cfg
.buckets
[info
->type
].avg
;
873 value
= cfg
.buckets
[info
->type
].max
;
876 value
= cfg
.buckets
[info
->type
].burst_length
;
883 visit_type_int64(v
, name
, &value
, errp
);
886 static void throttle_group_set_limits(Object
*obj
, Visitor
*v
,
887 const char *name
, void *opaque
,
891 ThrottleGroup
*tg
= THROTTLE_GROUP(obj
);
893 ThrottleLimits
*argp
;
894 Error
*local_err
= NULL
;
896 if (!visit_type_ThrottleLimits(v
, name
, &argp
, errp
)) {
899 qemu_mutex_lock(&tg
->lock
);
900 throttle_get_config(&tg
->ts
, &cfg
);
901 throttle_limits_to_config(argp
, &cfg
, &local_err
);
905 throttle_config(&tg
->ts
, tg
->clock_type
, &cfg
);
908 qemu_mutex_unlock(&tg
->lock
);
909 qapi_free_ThrottleLimits(argp
);
910 error_propagate(errp
, local_err
);
914 static void throttle_group_get_limits(Object
*obj
, Visitor
*v
,
915 const char *name
, void *opaque
,
918 ThrottleGroup
*tg
= THROTTLE_GROUP(obj
);
920 ThrottleLimits arg
= { 0 };
921 ThrottleLimits
*argp
= &arg
;
923 qemu_mutex_lock(&tg
->lock
);
924 throttle_get_config(&tg
->ts
, &cfg
);
925 qemu_mutex_unlock(&tg
->lock
);
927 throttle_config_to_limits(&cfg
, argp
);
929 visit_type_ThrottleLimits(v
, name
, &argp
, errp
);
932 static bool throttle_group_can_be_deleted(UserCreatable
*uc
)
934 return OBJECT(uc
)->ref
== 1;
937 static void throttle_group_obj_class_init(ObjectClass
*klass
, void *class_data
)
940 UserCreatableClass
*ucc
= USER_CREATABLE_CLASS(klass
);
942 ucc
->complete
= throttle_group_obj_complete
;
943 ucc
->can_be_deleted
= throttle_group_can_be_deleted
;
945 /* individual properties */
946 for (i
= 0; i
< sizeof(properties
) / sizeof(ThrottleParamInfo
); i
++) {
947 object_class_property_add(klass
,
952 NULL
, &properties
[i
]);
956 object_class_property_add(klass
,
957 "limits", "ThrottleLimits",
958 throttle_group_get_limits
,
959 throttle_group_set_limits
,
963 static const TypeInfo throttle_group_info
= {
964 .name
= TYPE_THROTTLE_GROUP
,
965 .parent
= TYPE_OBJECT
,
966 .class_init
= throttle_group_obj_class_init
,
967 .instance_size
= sizeof(ThrottleGroup
),
968 .instance_init
= throttle_group_obj_init
,
969 .instance_finalize
= throttle_group_obj_finalize
,
970 .interfaces
= (InterfaceInfo
[]) {
971 { TYPE_USER_CREATABLE
},
976 static void throttle_groups_init(void)
978 type_register_static(&throttle_group_info
);
981 type_init(throttle_groups_init
);