2 * QEMU aio implementation
4 * Copyright IBM, Corp. 2008
7 * Anthony Liguori <aliguori@us.ibm.com>
9 * This work is licensed under the terms of the GNU GPL, version 2. See
10 * the COPYING file in the top-level directory.
17 #include "qemu-common.h"
18 #include "qemu/queue.h"
19 #include "qemu/event_notifier.h"
20 #include "qemu/thread.h"
22 typedef struct BlockDriverAIOCB BlockDriverAIOCB
;
23 typedef void BlockDriverCompletionFunc(void *opaque
, int ret
);
25 typedef struct AIOCBInfo
{
26 void (*cancel
)(BlockDriverAIOCB
*acb
);
30 struct BlockDriverAIOCB
{
31 const AIOCBInfo
*aiocb_info
;
33 BlockDriverCompletionFunc
*cb
;
37 void *qemu_aio_get(const AIOCBInfo
*aiocb_info
, BlockDriverState
*bs
,
38 BlockDriverCompletionFunc
*cb
, void *opaque
);
39 void qemu_aio_release(void *p
);
41 typedef struct AioHandler AioHandler
;
42 typedef void QEMUBHFunc(void *opaque
);
43 typedef void IOHandler(void *opaque
);
45 typedef struct AioContext
{
48 /* The list of registered AIO handlers */
49 QLIST_HEAD(, AioHandler
) aio_handlers
;
51 /* This is a simple lock used to protect the aio_handlers list.
52 * Specifically, it's used to ensure that no callbacks are removed while
53 * we're walking and dispatching callbacks.
57 /* lock to protect between bh's adders and deleter */
59 /* Anchor of the list of Bottom Halves belonging to the context */
60 struct QEMUBH
*first_bh
;
62 /* A simple lock used to protect the first_bh list, and ensure that
63 * no callbacks are removed while we're walking and dispatching callbacks.
67 /* Used for aio_notify. */
68 EventNotifier notifier
;
70 /* GPollFDs for aio_poll() */
73 /* Thread pool for performing work and receiving completion callbacks */
74 struct ThreadPool
*thread_pool
;
77 /* Returns 1 if there are still outstanding AIO requests; 0 otherwise */
78 typedef int (AioFlushEventNotifierHandler
)(EventNotifier
*e
);
81 * aio_context_new: Allocate a new AioContext.
83 * AioContext provide a mini event-loop that can be waited on synchronously.
84 * They also provide bottom halves, a service to execute a piece of code
85 * as soon as possible.
87 AioContext
*aio_context_new(void);
91 * @ctx: The AioContext to operate on.
93 * Add a reference to an AioContext.
95 void aio_context_ref(AioContext
*ctx
);
99 * @ctx: The AioContext to operate on.
101 * Drop a reference to an AioContext.
103 void aio_context_unref(AioContext
*ctx
);
106 * aio_bh_new: Allocate a new bottom half structure.
108 * Bottom halves are lightweight callbacks whose invocation is guaranteed
109 * to be wait-free, thread-safe and signal-safe. The #QEMUBH structure
110 * is opaque and must be allocated prior to its use.
112 QEMUBH
*aio_bh_new(AioContext
*ctx
, QEMUBHFunc
*cb
, void *opaque
);
115 * aio_notify: Force processing of pending events.
117 * Similar to signaling a condition variable, aio_notify forces
118 * aio_wait to exit, so that the next call will re-examine pending events.
119 * The caller of aio_notify will usually call aio_wait again very soon,
120 * or go through another iteration of the GLib main loop. Hence, aio_notify
121 * also has the side effect of recalculating the sets of file descriptors
122 * that the main loop waits for.
124 * Calling aio_notify is rarely necessary, because for example scheduling
125 * a bottom half calls it already.
127 void aio_notify(AioContext
*ctx
);
130 * aio_bh_poll: Poll bottom halves for an AioContext.
132 * These are internal functions used by the QEMU main loop.
133 * And notice that multiple occurrences of aio_bh_poll cannot
134 * be called concurrently
136 int aio_bh_poll(AioContext
*ctx
);
139 * qemu_bh_schedule: Schedule a bottom half.
141 * Scheduling a bottom half interrupts the main loop and causes the
142 * execution of the callback that was passed to qemu_bh_new.
144 * Bottom halves that are scheduled from a bottom half handler are instantly
145 * invoked. This can create an infinite loop if a bottom half handler
148 * @bh: The bottom half to be scheduled.
150 void qemu_bh_schedule(QEMUBH
*bh
);
153 * qemu_bh_cancel: Cancel execution of a bottom half.
155 * Canceling execution of a bottom half undoes the effect of calls to
156 * qemu_bh_schedule without freeing its resources yet. While cancellation
157 * itself is also wait-free and thread-safe, it can of course race with the
158 * loop that executes bottom halves unless you are holding the iothread
159 * mutex. This makes it mostly useless if you are not holding the mutex.
161 * @bh: The bottom half to be canceled.
163 void qemu_bh_cancel(QEMUBH
*bh
);
166 *qemu_bh_delete: Cancel execution of a bottom half and free its resources.
168 * Deleting a bottom half frees the memory that was allocated for it by
169 * qemu_bh_new. It also implies canceling the bottom half if it was
171 * This func is async. The bottom half will do the delete action at the finial
174 * @bh: The bottom half to be deleted.
176 void qemu_bh_delete(QEMUBH
*bh
);
178 /* Return whether there are any pending callbacks from the GSource
179 * attached to the AioContext.
181 * This is used internally in the implementation of the GSource.
183 bool aio_pending(AioContext
*ctx
);
185 /* Progress in completing AIO work to occur. This can issue new pending
186 * aio as a result of executing I/O completion or bh callbacks.
188 * If there is no pending AIO operation or completion (bottom half),
189 * return false. If there are pending AIO operations of bottom halves,
192 * If there are no pending bottom halves, but there are pending AIO
193 * operations, it may not be possible to make any progress without
194 * blocking. If @blocking is true, this function will wait until one
195 * or more AIO events have completed, to ensure something has moved
198 bool aio_poll(AioContext
*ctx
, bool blocking
);
201 /* Returns 1 if there are still outstanding AIO requests; 0 otherwise */
202 typedef int (AioFlushHandler
)(void *opaque
);
204 /* Register a file descriptor and associated callbacks. Behaves very similarly
205 * to qemu_set_fd_handler2. Unlike qemu_set_fd_handler2, these callbacks will
206 * be invoked when using qemu_aio_wait().
208 * Code that invokes AIO completion functions should rely on this function
209 * instead of qemu_set_fd_handler[2].
211 void aio_set_fd_handler(AioContext
*ctx
,
215 AioFlushHandler
*io_flush
,
219 /* Register an event notifier and associated callbacks. Behaves very similarly
220 * to event_notifier_set_handler. Unlike event_notifier_set_handler, these callbacks
221 * will be invoked when using qemu_aio_wait().
223 * Code that invokes AIO completion functions should rely on this function
224 * instead of event_notifier_set_handler.
226 void aio_set_event_notifier(AioContext
*ctx
,
227 EventNotifier
*notifier
,
228 EventNotifierHandler
*io_read
,
229 AioFlushEventNotifierHandler
*io_flush
);
231 /* Return a GSource that lets the main loop poll the file descriptors attached
232 * to this AioContext.
234 GSource
*aio_get_g_source(AioContext
*ctx
);
236 /* Return the ThreadPool bound to this AioContext */
237 struct ThreadPool
*aio_get_thread_pool(AioContext
*ctx
);
239 /* Functions to operate on the main QEMU AioContext. */
241 bool qemu_aio_wait(void);
242 void qemu_aio_set_event_notifier(EventNotifier
*notifier
,
243 EventNotifierHandler
*io_read
,
244 AioFlushEventNotifierHandler
*io_flush
);
247 void qemu_aio_set_fd_handler(int fd
,
250 AioFlushHandler
*io_flush
,