4 * Copyright (c) 2015 Red Hat, Inc.
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 #include "qemu/osdep.h"
22 #include "io/channel.h"
23 #include "qapi/error.h"
24 #include "qemu/main-loop.h"
27 bool qio_channel_has_feature(QIOChannel
*ioc
,
28 QIOChannelFeature feature
)
30 return ioc
->features
& (1 << feature
);
34 void qio_channel_set_feature(QIOChannel
*ioc
,
35 QIOChannelFeature feature
)
37 ioc
->features
|= (1 << feature
);
41 void qio_channel_set_name(QIOChannel
*ioc
,
45 ioc
->name
= g_strdup(name
);
49 ssize_t
qio_channel_readv_full(QIOChannel
*ioc
,
50 const struct iovec
*iov
,
56 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
59 !qio_channel_has_feature(ioc
, QIO_CHANNEL_FEATURE_FD_PASS
)) {
60 error_setg_errno(errp
, EINVAL
,
61 "Channel does not support file descriptor passing");
65 return klass
->io_readv(ioc
, iov
, niov
, fds
, nfds
, errp
);
69 ssize_t
qio_channel_writev_full(QIOChannel
*ioc
,
70 const struct iovec
*iov
,
76 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
79 !qio_channel_has_feature(ioc
, QIO_CHANNEL_FEATURE_FD_PASS
)) {
80 error_setg_errno(errp
, EINVAL
,
81 "Channel does not support file descriptor passing");
85 return klass
->io_writev(ioc
, iov
, niov
, fds
, nfds
, errp
);
89 int qio_channel_readv_all_eof(QIOChannel
*ioc
,
90 const struct iovec
*iov
,
95 struct iovec
*local_iov
= g_new(struct iovec
, niov
);
96 struct iovec
*local_iov_head
= local_iov
;
97 unsigned int nlocal_iov
= niov
;
100 nlocal_iov
= iov_copy(local_iov
, nlocal_iov
,
102 0, iov_size(iov
, niov
));
104 while (nlocal_iov
> 0) {
106 len
= qio_channel_readv(ioc
, local_iov
, nlocal_iov
, errp
);
107 if (len
== QIO_CHANNEL_ERR_BLOCK
) {
108 if (qemu_in_coroutine()) {
109 qio_channel_yield(ioc
, G_IO_IN
);
111 qio_channel_wait(ioc
, G_IO_IN
);
114 } else if (len
< 0) {
116 } else if (len
== 0) {
119 "Unexpected end-of-file before all bytes were read");
127 iov_discard_front(&local_iov
, &nlocal_iov
, len
);
133 g_free(local_iov_head
);
137 int qio_channel_readv_all(QIOChannel
*ioc
,
138 const struct iovec
*iov
,
142 int ret
= qio_channel_readv_all_eof(ioc
, iov
, niov
, errp
);
147 "Unexpected end-of-file before all bytes were read");
148 } else if (ret
== 1) {
154 int qio_channel_writev_all(QIOChannel
*ioc
,
155 const struct iovec
*iov
,
160 struct iovec
*local_iov
= g_new(struct iovec
, niov
);
161 struct iovec
*local_iov_head
= local_iov
;
162 unsigned int nlocal_iov
= niov
;
164 nlocal_iov
= iov_copy(local_iov
, nlocal_iov
,
166 0, iov_size(iov
, niov
));
168 while (nlocal_iov
> 0) {
170 len
= qio_channel_writev(ioc
, local_iov
, nlocal_iov
, errp
);
171 if (len
== QIO_CHANNEL_ERR_BLOCK
) {
172 if (qemu_in_coroutine()) {
173 qio_channel_yield(ioc
, G_IO_OUT
);
175 qio_channel_wait(ioc
, G_IO_OUT
);
183 iov_discard_front(&local_iov
, &nlocal_iov
, len
);
188 g_free(local_iov_head
);
192 ssize_t
qio_channel_readv(QIOChannel
*ioc
,
193 const struct iovec
*iov
,
197 return qio_channel_readv_full(ioc
, iov
, niov
, NULL
, NULL
, errp
);
201 ssize_t
qio_channel_writev(QIOChannel
*ioc
,
202 const struct iovec
*iov
,
206 return qio_channel_writev_full(ioc
, iov
, niov
, NULL
, 0, errp
);
210 ssize_t
qio_channel_read(QIOChannel
*ioc
,
215 struct iovec iov
= { .iov_base
= buf
, .iov_len
= buflen
};
216 return qio_channel_readv_full(ioc
, &iov
, 1, NULL
, NULL
, errp
);
220 ssize_t
qio_channel_write(QIOChannel
*ioc
,
225 struct iovec iov
= { .iov_base
= (char *)buf
, .iov_len
= buflen
};
226 return qio_channel_writev_full(ioc
, &iov
, 1, NULL
, 0, errp
);
230 int qio_channel_read_all_eof(QIOChannel
*ioc
,
235 struct iovec iov
= { .iov_base
= buf
, .iov_len
= buflen
};
236 return qio_channel_readv_all_eof(ioc
, &iov
, 1, errp
);
240 int qio_channel_read_all(QIOChannel
*ioc
,
245 struct iovec iov
= { .iov_base
= buf
, .iov_len
= buflen
};
246 return qio_channel_readv_all(ioc
, &iov
, 1, errp
);
250 int qio_channel_write_all(QIOChannel
*ioc
,
255 struct iovec iov
= { .iov_base
= (char *)buf
, .iov_len
= buflen
};
256 return qio_channel_writev_all(ioc
, &iov
, 1, errp
);
260 int qio_channel_set_blocking(QIOChannel
*ioc
,
264 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
265 return klass
->io_set_blocking(ioc
, enabled
, errp
);
269 int qio_channel_close(QIOChannel
*ioc
,
272 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
273 return klass
->io_close(ioc
, errp
);
277 GSource
*qio_channel_create_watch(QIOChannel
*ioc
,
278 GIOCondition condition
)
280 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
281 GSource
*ret
= klass
->io_create_watch(ioc
, condition
);
284 g_source_set_name(ret
, ioc
->name
);
291 void qio_channel_set_aio_fd_handler(QIOChannel
*ioc
,
297 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
299 klass
->io_set_aio_fd_handler(ioc
, ctx
, io_read
, io_write
, opaque
);
302 guint
qio_channel_add_watch_full(QIOChannel
*ioc
,
303 GIOCondition condition
,
306 GDestroyNotify notify
,
307 GMainContext
*context
)
312 source
= qio_channel_create_watch(ioc
, condition
);
314 g_source_set_callback(source
, (GSourceFunc
)func
, user_data
, notify
);
316 id
= g_source_attach(source
, context
);
317 g_source_unref(source
);
322 guint
qio_channel_add_watch(QIOChannel
*ioc
,
323 GIOCondition condition
,
326 GDestroyNotify notify
)
328 return qio_channel_add_watch_full(ioc
, condition
, func
,
329 user_data
, notify
, NULL
);
332 GSource
*qio_channel_add_watch_source(QIOChannel
*ioc
,
333 GIOCondition condition
,
336 GDestroyNotify notify
,
337 GMainContext
*context
)
342 id
= qio_channel_add_watch_full(ioc
, condition
, func
,
343 user_data
, notify
, context
);
344 source
= g_main_context_find_source_by_id(context
, id
);
345 g_source_ref(source
);
350 int qio_channel_shutdown(QIOChannel
*ioc
,
351 QIOChannelShutdown how
,
354 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
356 if (!klass
->io_shutdown
) {
357 error_setg(errp
, "Data path shutdown not supported");
361 return klass
->io_shutdown(ioc
, how
, errp
);
365 void qio_channel_set_delay(QIOChannel
*ioc
,
368 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
370 if (klass
->io_set_delay
) {
371 klass
->io_set_delay(ioc
, enabled
);
376 void qio_channel_set_cork(QIOChannel
*ioc
,
379 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
381 if (klass
->io_set_cork
) {
382 klass
->io_set_cork(ioc
, enabled
);
387 off_t
qio_channel_io_seek(QIOChannel
*ioc
,
392 QIOChannelClass
*klass
= QIO_CHANNEL_GET_CLASS(ioc
);
394 if (!klass
->io_seek
) {
395 error_setg(errp
, "Channel does not support random access");
399 return klass
->io_seek(ioc
, offset
, whence
, errp
);
403 static void qio_channel_set_aio_fd_handlers(QIOChannel
*ioc
);
405 static void qio_channel_restart_read(void *opaque
)
407 QIOChannel
*ioc
= opaque
;
408 Coroutine
*co
= ioc
->read_coroutine
;
410 ioc
->read_coroutine
= NULL
;
411 qio_channel_set_aio_fd_handlers(ioc
);
415 static void qio_channel_restart_write(void *opaque
)
417 QIOChannel
*ioc
= opaque
;
418 Coroutine
*co
= ioc
->write_coroutine
;
420 ioc
->write_coroutine
= NULL
;
421 qio_channel_set_aio_fd_handlers(ioc
);
425 static void qio_channel_set_aio_fd_handlers(QIOChannel
*ioc
)
427 IOHandler
*rd_handler
= NULL
, *wr_handler
= NULL
;
430 if (ioc
->read_coroutine
) {
431 rd_handler
= qio_channel_restart_read
;
433 if (ioc
->write_coroutine
) {
434 wr_handler
= qio_channel_restart_write
;
437 ctx
= ioc
->ctx
? ioc
->ctx
: iohandler_get_aio_context();
438 qio_channel_set_aio_fd_handler(ioc
, ctx
, rd_handler
, wr_handler
, ioc
);
441 void qio_channel_attach_aio_context(QIOChannel
*ioc
,
444 assert(!ioc
->read_coroutine
);
445 assert(!ioc
->write_coroutine
);
449 void qio_channel_detach_aio_context(QIOChannel
*ioc
)
451 ioc
->read_coroutine
= NULL
;
452 ioc
->write_coroutine
= NULL
;
453 qio_channel_set_aio_fd_handlers(ioc
);
457 void coroutine_fn
qio_channel_yield(QIOChannel
*ioc
,
458 GIOCondition condition
)
460 assert(qemu_in_coroutine());
461 if (condition
== G_IO_IN
) {
462 assert(!ioc
->read_coroutine
);
463 ioc
->read_coroutine
= qemu_coroutine_self();
464 } else if (condition
== G_IO_OUT
) {
465 assert(!ioc
->write_coroutine
);
466 ioc
->write_coroutine
= qemu_coroutine_self();
470 qio_channel_set_aio_fd_handlers(ioc
);
471 qemu_coroutine_yield();
475 static gboolean
qio_channel_wait_complete(QIOChannel
*ioc
,
476 GIOCondition condition
,
479 GMainLoop
*loop
= opaque
;
481 g_main_loop_quit(loop
);
486 void qio_channel_wait(QIOChannel
*ioc
,
487 GIOCondition condition
)
489 GMainContext
*ctxt
= g_main_context_new();
490 GMainLoop
*loop
= g_main_loop_new(ctxt
, TRUE
);
493 source
= qio_channel_create_watch(ioc
, condition
);
495 g_source_set_callback(source
,
496 (GSourceFunc
)qio_channel_wait_complete
,
500 g_source_attach(source
, ctxt
);
502 g_main_loop_run(loop
);
504 g_source_unref(source
);
505 g_main_loop_unref(loop
);
506 g_main_context_unref(ctxt
);
510 static void qio_channel_finalize(Object
*obj
)
512 QIOChannel
*ioc
= QIO_CHANNEL(obj
);
518 CloseHandle(ioc
->event
);
523 static const TypeInfo qio_channel_info
= {
524 .parent
= TYPE_OBJECT
,
525 .name
= TYPE_QIO_CHANNEL
,
526 .instance_size
= sizeof(QIOChannel
),
527 .instance_finalize
= qio_channel_finalize
,
529 .class_size
= sizeof(QIOChannelClass
),
533 static void qio_channel_register_types(void)
535 type_register_static(&qio_channel_info
);
539 type_init(qio_channel_register_types
);