2 * Linux native AIO support.
4 * Copyright (C) 2009 IBM, Corp.
5 * Copyright (C) 2009 Red Hat, Inc.
7 * This work is licensed under the terms of the GNU GPL, version 2 or later.
8 * See the COPYING file in the top-level directory.
10 #include "qemu-common.h"
12 #include "block_int.h"
13 #include "block/raw-posix-aio.h"
15 #include <sys/eventfd.h>
19 * Queue size (per-device).
21 * XXX: eventually we need to communicate this to the guest and/or make it
22 * tunable by the guest. If we get more outstanding requests at a time
23 * than this we will get EAGAIN from io_submit which is communicated to
24 * the guest as an I/O error.
26 #define MAX_EVENTS 128
29 BlockDriverAIOCB common
;
30 struct qemu_laio_state
*ctx
;
36 struct qemu_laio_state
{
42 static inline ssize_t
io_event_ret(struct io_event
*ev
)
44 return (ssize_t
)(((uint64_t)ev
->res2
<< 32) | ev
->res
);
47 static void qemu_laio_completion_cb(void *opaque
)
49 struct qemu_laio_state
*s
= opaque
;
52 struct io_event events
[MAX_EVENTS
];
55 struct timespec ts
= { 0 };
59 ret
= read(s
->efd
, &val
, sizeof(val
));
60 } while (ret
== 1 && errno
== EINTR
);
62 if (ret
== -1 && errno
== EAGAIN
)
69 nevents
= io_getevents(s
->ctx
, val
, MAX_EVENTS
, events
, &ts
);
70 } while (nevents
== -EINTR
);
72 for (i
= 0; i
< nevents
; i
++) {
73 struct iocb
*iocb
= events
[i
].obj
;
74 struct qemu_laiocb
*laiocb
=
75 container_of(iocb
, struct qemu_laiocb
, iocb
);
79 ret
= laiocb
->ret
= io_event_ret(&events
[i
]);
80 if (ret
!= -ECANCELED
) {
81 if (ret
== laiocb
->nbytes
)
86 laiocb
->common
.cb(laiocb
->common
.opaque
, ret
);
89 qemu_aio_release(laiocb
);
94 static int qemu_laio_flush_cb(void *opaque
)
96 struct qemu_laio_state
*s
= opaque
;
98 return (s
->count
> 0) ? 1 : 0;
101 static void laio_cancel(BlockDriverAIOCB
*blockacb
)
103 struct qemu_laiocb
*laiocb
= (struct qemu_laiocb
*)blockacb
;
104 struct io_event event
;
107 if (laiocb
->ret
!= -EINPROGRESS
)
111 * Note that as of Linux 2.6.31 neither the block device code nor any
112 * filesystem implements cancellation of AIO request.
113 * Thus the polling loop below is the normal code path.
115 ret
= io_cancel(laiocb
->ctx
->ctx
, &laiocb
->iocb
, &event
);
117 laiocb
->ret
= -ECANCELED
;
122 * We have to wait for the iocb to finish.
124 * The only way to get the iocb status update is by polling the io context.
125 * We might be able to do this slightly more optimal by removing the
128 while (laiocb
->ret
== -EINPROGRESS
)
129 qemu_laio_completion_cb(laiocb
->ctx
);
132 static AIOPool laio_pool
= {
133 .aiocb_size
= sizeof(struct qemu_laiocb
),
134 .cancel
= laio_cancel
,
137 BlockDriverAIOCB
*laio_submit(BlockDriverState
*bs
, void *aio_ctx
, int fd
,
138 int64_t sector_num
, QEMUIOVector
*qiov
, int nb_sectors
,
139 BlockDriverCompletionFunc
*cb
, void *opaque
, int type
)
141 struct qemu_laio_state
*s
= aio_ctx
;
142 struct qemu_laiocb
*laiocb
;
144 off_t offset
= sector_num
* 512;
146 laiocb
= qemu_aio_get(&laio_pool
, bs
, cb
, opaque
);
149 laiocb
->nbytes
= nb_sectors
* 512;
151 laiocb
->ret
= -EINPROGRESS
;
153 iocbs
= &laiocb
->iocb
;
157 io_prep_pwritev(iocbs
, fd
, qiov
->iov
, qiov
->niov
, offset
);
160 io_prep_preadv(iocbs
, fd
, qiov
->iov
, qiov
->niov
, offset
);
163 fprintf(stderr
, "%s: invalid AIO request type 0x%x.\n",
167 io_set_eventfd(&laiocb
->iocb
, s
->efd
);
170 if (io_submit(s
->ctx
, 1, &iocbs
) < 0)
172 return &laiocb
->common
;
175 qemu_aio_release(laiocb
);
181 void *laio_init(void)
183 struct qemu_laio_state
*s
;
185 s
= qemu_mallocz(sizeof(*s
));
186 s
->efd
= eventfd(0, 0);
189 fcntl(s
->efd
, F_SETFL
, O_NONBLOCK
);
191 if (io_setup(MAX_EVENTS
, &s
->ctx
) != 0)
194 qemu_aio_set_fd_handler(s
->efd
, qemu_laio_completion_cb
,
195 NULL
, qemu_laio_flush_cb
, s
);