2 * QEMU Block driver for NBD
4 * Copyright (C) 2016 Red Hat, Inc.
5 * Copyright (C) 2008 Bull S.A.S.
6 * Author: Laurent Vivier <Laurent.Vivier@bull.net>
9 * Copyright (C) 2007 Anthony Liguori <anthony@codemonkey.ws>
11 * Permission is hereby granted, free of charge, to any person obtaining a copy
12 * of this software and associated documentation files (the "Software"), to deal
13 * in the Software without restriction, including without limitation the rights
14 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
15 * copies of the Software, and to permit persons to whom the Software is
16 * furnished to do so, subject to the following conditions:
18 * The above copyright notice and this permission notice shall be included in
19 * all copies or substantial portions of the Software.
21 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
22 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
23 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
24 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
25 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
26 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
30 #include "qemu/osdep.h"
34 #include "qemu/option.h"
35 #include "qemu/cutils.h"
37 #include "qapi/qapi-visit-sockets.h"
38 #include "qapi/qmp/qstring.h"
40 #include "block/qdict.h"
41 #include "block/nbd.h"
42 #include "block/block_int.h"
44 #define EN_OPTSTR ":exportname="
45 #define MAX_NBD_REQUESTS 16
47 #define HANDLE_TO_INDEX(bs, handle) ((handle) ^ (uint64_t)(intptr_t)(bs))
48 #define INDEX_TO_HANDLE(bs, index) ((index) ^ (uint64_t)(intptr_t)(bs))
52 uint64_t offset
; /* original offset of the request */
53 bool receiving
; /* waiting for connection_co? */
56 typedef struct BDRVNBDState
{
57 QIOChannelSocket
*sioc
; /* The master data channel */
58 QIOChannel
*ioc
; /* The current I/O channel which may differ (eg TLS) */
63 Coroutine
*connection_co
;
66 NBDClientRequest requests
[MAX_NBD_REQUESTS
];
71 /* For nbd_refresh_filename() */
73 char *export
, *tlscredsid
;
76 static void nbd_recv_coroutines_wake_all(BDRVNBDState
*s
)
80 for (i
= 0; i
< MAX_NBD_REQUESTS
; i
++) {
81 NBDClientRequest
*req
= &s
->requests
[i
];
83 if (req
->coroutine
&& req
->receiving
) {
84 aio_co_wake(req
->coroutine
);
89 static void nbd_client_detach_aio_context(BlockDriverState
*bs
)
91 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
93 qio_channel_detach_aio_context(QIO_CHANNEL(s
->ioc
));
96 static void nbd_client_attach_aio_context_bh(void *opaque
)
98 BlockDriverState
*bs
= opaque
;
99 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
102 * The node is still drained, so we know the coroutine has yielded in
103 * nbd_read_eof(), the only place where bs->in_flight can reach 0, or it is
104 * entered for the first time. Both places are safe for entering the
107 qemu_aio_coroutine_enter(bs
->aio_context
, s
->connection_co
);
108 bdrv_dec_in_flight(bs
);
111 static void nbd_client_attach_aio_context(BlockDriverState
*bs
,
112 AioContext
*new_context
)
114 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
116 qio_channel_attach_aio_context(QIO_CHANNEL(s
->ioc
), new_context
);
118 bdrv_inc_in_flight(bs
);
121 * Need to wait here for the BH to run because the BH must run while the
122 * node is still drained.
124 aio_wait_bh_oneshot(new_context
, nbd_client_attach_aio_context_bh
, bs
);
128 static void nbd_teardown_connection(BlockDriverState
*bs
)
130 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
134 /* finish any pending coroutines */
135 qio_channel_shutdown(s
->ioc
,
136 QIO_CHANNEL_SHUTDOWN_BOTH
,
138 BDRV_POLL_WHILE(bs
, s
->connection_co
);
140 nbd_client_detach_aio_context(bs
);
141 object_unref(OBJECT(s
->sioc
));
143 object_unref(OBJECT(s
->ioc
));
147 static coroutine_fn
void nbd_connection_entry(void *opaque
)
149 BDRVNBDState
*s
= opaque
;
152 Error
*local_err
= NULL
;
156 * The NBD client can only really be considered idle when it has
157 * yielded from qio_channel_readv_all_eof(), waiting for data. This is
158 * the point where the additional scheduled coroutine entry happens
159 * after nbd_client_attach_aio_context().
161 * Therefore we keep an additional in_flight reference all the time and
162 * only drop it temporarily here.
164 assert(s
->reply
.handle
== 0);
165 ret
= nbd_receive_reply(s
->bs
, s
->ioc
, &s
->reply
, &local_err
);
168 trace_nbd_read_reply_entry_fail(ret
, error_get_pretty(local_err
));
169 error_free(local_err
);
176 * There's no need for a mutex on the receive side, because the
177 * handler acts as a synchronization point and ensures that only
178 * one coroutine is called until the reply finishes.
180 i
= HANDLE_TO_INDEX(s
, s
->reply
.handle
);
181 if (i
>= MAX_NBD_REQUESTS
||
182 !s
->requests
[i
].coroutine
||
183 !s
->requests
[i
].receiving
||
184 (nbd_reply_is_structured(&s
->reply
) && !s
->info
.structured_reply
))
190 * We're woken up again by the request itself. Note that there
191 * is no race between yielding and reentering connection_co. This
194 * - if the request runs on the same AioContext, it is only
195 * entered after we yield
197 * - if the request runs on a different AioContext, reentering
198 * connection_co happens through a bottom half, which can only
199 * run after we yield.
201 aio_co_wake(s
->requests
[i
].coroutine
);
202 qemu_coroutine_yield();
206 nbd_recv_coroutines_wake_all(s
);
207 bdrv_dec_in_flight(s
->bs
);
209 s
->connection_co
= NULL
;
213 static int nbd_co_send_request(BlockDriverState
*bs
,
217 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
220 qemu_co_mutex_lock(&s
->send_mutex
);
221 while (s
->in_flight
== MAX_NBD_REQUESTS
) {
222 qemu_co_queue_wait(&s
->free_sema
, &s
->send_mutex
);
226 for (i
= 0; i
< MAX_NBD_REQUESTS
; i
++) {
227 if (s
->requests
[i
].coroutine
== NULL
) {
232 g_assert(qemu_in_coroutine());
233 assert(i
< MAX_NBD_REQUESTS
);
235 s
->requests
[i
].coroutine
= qemu_coroutine_self();
236 s
->requests
[i
].offset
= request
->from
;
237 s
->requests
[i
].receiving
= false;
239 request
->handle
= INDEX_TO_HANDLE(s
, i
);
248 qio_channel_set_cork(s
->ioc
, true);
249 rc
= nbd_send_request(s
->ioc
, request
);
250 if (rc
>= 0 && !s
->quit
) {
251 if (qio_channel_writev_all(s
->ioc
, qiov
->iov
, qiov
->niov
,
255 } else if (rc
>= 0) {
258 qio_channel_set_cork(s
->ioc
, false);
260 rc
= nbd_send_request(s
->ioc
, request
);
266 s
->requests
[i
].coroutine
= NULL
;
268 qemu_co_queue_next(&s
->free_sema
);
270 qemu_co_mutex_unlock(&s
->send_mutex
);
274 static inline uint16_t payload_advance16(uint8_t **payload
)
277 return lduw_be_p(*payload
- 2);
280 static inline uint32_t payload_advance32(uint8_t **payload
)
283 return ldl_be_p(*payload
- 4);
286 static inline uint64_t payload_advance64(uint8_t **payload
)
289 return ldq_be_p(*payload
- 8);
292 static int nbd_parse_offset_hole_payload(BDRVNBDState
*s
,
293 NBDStructuredReplyChunk
*chunk
,
294 uint8_t *payload
, uint64_t orig_offset
,
295 QEMUIOVector
*qiov
, Error
**errp
)
300 if (chunk
->length
!= sizeof(offset
) + sizeof(hole_size
)) {
301 error_setg(errp
, "Protocol error: invalid payload for "
302 "NBD_REPLY_TYPE_OFFSET_HOLE");
306 offset
= payload_advance64(&payload
);
307 hole_size
= payload_advance32(&payload
);
309 if (!hole_size
|| offset
< orig_offset
|| hole_size
> qiov
->size
||
310 offset
> orig_offset
+ qiov
->size
- hole_size
) {
311 error_setg(errp
, "Protocol error: server sent chunk exceeding requested"
315 if (s
->info
.min_block
&&
316 !QEMU_IS_ALIGNED(hole_size
, s
->info
.min_block
)) {
317 trace_nbd_structured_read_compliance("hole");
320 qemu_iovec_memset(qiov
, offset
- orig_offset
, 0, hole_size
);
326 * nbd_parse_blockstatus_payload
327 * Based on our request, we expect only one extent in reply, for the
328 * base:allocation context.
330 static int nbd_parse_blockstatus_payload(BDRVNBDState
*s
,
331 NBDStructuredReplyChunk
*chunk
,
332 uint8_t *payload
, uint64_t orig_length
,
333 NBDExtent
*extent
, Error
**errp
)
337 /* The server succeeded, so it must have sent [at least] one extent */
338 if (chunk
->length
< sizeof(context_id
) + sizeof(*extent
)) {
339 error_setg(errp
, "Protocol error: invalid payload for "
340 "NBD_REPLY_TYPE_BLOCK_STATUS");
344 context_id
= payload_advance32(&payload
);
345 if (s
->info
.context_id
!= context_id
) {
346 error_setg(errp
, "Protocol error: unexpected context id %d for "
347 "NBD_REPLY_TYPE_BLOCK_STATUS, when negotiated context "
348 "id is %d", context_id
,
353 extent
->length
= payload_advance32(&payload
);
354 extent
->flags
= payload_advance32(&payload
);
356 if (extent
->length
== 0) {
357 error_setg(errp
, "Protocol error: server sent status chunk with "
363 * A server sending unaligned block status is in violation of the
364 * protocol, but as qemu-nbd 3.1 is such a server (at least for
365 * POSIX files that are not a multiple of 512 bytes, since qemu
366 * rounds files up to 512-byte multiples but lseek(SEEK_HOLE)
367 * still sees an implicit hole beyond the real EOF), it's nicer to
368 * work around the misbehaving server. If the request included
369 * more than the final unaligned block, truncate it back to an
370 * aligned result; if the request was only the final block, round
371 * up to the full block and change the status to fully-allocated
372 * (always a safe status, even if it loses information).
374 if (s
->info
.min_block
&& !QEMU_IS_ALIGNED(extent
->length
,
375 s
->info
.min_block
)) {
376 trace_nbd_parse_blockstatus_compliance("extent length is unaligned");
377 if (extent
->length
> s
->info
.min_block
) {
378 extent
->length
= QEMU_ALIGN_DOWN(extent
->length
,
381 extent
->length
= s
->info
.min_block
;
387 * We used NBD_CMD_FLAG_REQ_ONE, so the server should not have
388 * sent us any more than one extent, nor should it have included
389 * status beyond our request in that extent. However, it's easy
390 * enough to ignore the server's noncompliance without killing the
391 * connection; just ignore trailing extents, and clamp things to
392 * the length of our request.
394 if (chunk
->length
> sizeof(context_id
) + sizeof(*extent
)) {
395 trace_nbd_parse_blockstatus_compliance("more than one extent");
397 if (extent
->length
> orig_length
) {
398 extent
->length
= orig_length
;
399 trace_nbd_parse_blockstatus_compliance("extent length too large");
406 * nbd_parse_error_payload
407 * on success @errp contains message describing nbd error reply
409 static int nbd_parse_error_payload(NBDStructuredReplyChunk
*chunk
,
410 uint8_t *payload
, int *request_ret
,
414 uint16_t message_size
;
416 assert(chunk
->type
& (1 << 15));
418 if (chunk
->length
< sizeof(error
) + sizeof(message_size
)) {
420 "Protocol error: invalid payload for structured error");
424 error
= nbd_errno_to_system_errno(payload_advance32(&payload
));
426 error_setg(errp
, "Protocol error: server sent structured error chunk "
431 *request_ret
= -error
;
432 message_size
= payload_advance16(&payload
);
434 if (message_size
> chunk
->length
- sizeof(error
) - sizeof(message_size
)) {
435 error_setg(errp
, "Protocol error: server sent structured error chunk "
436 "with incorrect message size");
440 /* TODO: Add a trace point to mention the server complaint */
442 /* TODO handle ERROR_OFFSET */
447 static int nbd_co_receive_offset_data_payload(BDRVNBDState
*s
,
448 uint64_t orig_offset
,
449 QEMUIOVector
*qiov
, Error
**errp
)
451 QEMUIOVector sub_qiov
;
455 NBDStructuredReplyChunk
*chunk
= &s
->reply
.structured
;
457 assert(nbd_reply_is_structured(&s
->reply
));
459 /* The NBD spec requires at least one byte of payload */
460 if (chunk
->length
<= sizeof(offset
)) {
461 error_setg(errp
, "Protocol error: invalid payload for "
462 "NBD_REPLY_TYPE_OFFSET_DATA");
466 if (nbd_read64(s
->ioc
, &offset
, "OFFSET_DATA offset", errp
) < 0) {
470 data_size
= chunk
->length
- sizeof(offset
);
472 if (offset
< orig_offset
|| data_size
> qiov
->size
||
473 offset
> orig_offset
+ qiov
->size
- data_size
) {
474 error_setg(errp
, "Protocol error: server sent chunk exceeding requested"
478 if (s
->info
.min_block
&& !QEMU_IS_ALIGNED(data_size
, s
->info
.min_block
)) {
479 trace_nbd_structured_read_compliance("data");
482 qemu_iovec_init(&sub_qiov
, qiov
->niov
);
483 qemu_iovec_concat(&sub_qiov
, qiov
, offset
- orig_offset
, data_size
);
484 ret
= qio_channel_readv_all(s
->ioc
, sub_qiov
.iov
, sub_qiov
.niov
, errp
);
485 qemu_iovec_destroy(&sub_qiov
);
487 return ret
< 0 ? -EIO
: 0;
490 #define NBD_MAX_MALLOC_PAYLOAD 1000
491 static coroutine_fn
int nbd_co_receive_structured_payload(
492 BDRVNBDState
*s
, void **payload
, Error
**errp
)
497 assert(nbd_reply_is_structured(&s
->reply
));
499 len
= s
->reply
.structured
.length
;
505 if (payload
== NULL
) {
506 error_setg(errp
, "Unexpected structured payload");
510 if (len
> NBD_MAX_MALLOC_PAYLOAD
) {
511 error_setg(errp
, "Payload too large");
515 *payload
= g_new(char, len
);
516 ret
= nbd_read(s
->ioc
, *payload
, len
, "structured payload", errp
);
527 * nbd_co_do_receive_one_chunk
529 * set request_ret to received reply error
530 * if qiov is not NULL: read payload to @qiov
531 * for structured reply chunk:
532 * if error chunk: read payload, set @request_ret, do not set @payload
533 * else if offset_data chunk: read payload data to @qiov, do not set @payload
534 * else: read payload to @payload
536 * If function fails, @errp contains corresponding error message, and the
537 * connection with the server is suspect. If it returns 0, then the
538 * transaction succeeded (although @request_ret may be a negative errno
539 * corresponding to the server's error reply), and errp is unchanged.
541 static coroutine_fn
int nbd_co_do_receive_one_chunk(
542 BDRVNBDState
*s
, uint64_t handle
, bool only_structured
,
543 int *request_ret
, QEMUIOVector
*qiov
, void **payload
, Error
**errp
)
546 int i
= HANDLE_TO_INDEX(s
, handle
);
547 void *local_payload
= NULL
;
548 NBDStructuredReplyChunk
*chunk
;
555 /* Wait until we're woken up by nbd_connection_entry. */
556 s
->requests
[i
].receiving
= true;
557 qemu_coroutine_yield();
558 s
->requests
[i
].receiving
= false;
560 error_setg(errp
, "Connection closed");
565 assert(s
->reply
.handle
== handle
);
567 if (nbd_reply_is_simple(&s
->reply
)) {
568 if (only_structured
) {
569 error_setg(errp
, "Protocol error: simple reply when structured "
570 "reply chunk was expected");
574 *request_ret
= -nbd_errno_to_system_errno(s
->reply
.simple
.error
);
575 if (*request_ret
< 0 || !qiov
) {
579 return qio_channel_readv_all(s
->ioc
, qiov
->iov
, qiov
->niov
,
580 errp
) < 0 ? -EIO
: 0;
583 /* handle structured reply chunk */
584 assert(s
->info
.structured_reply
);
585 chunk
= &s
->reply
.structured
;
587 if (chunk
->type
== NBD_REPLY_TYPE_NONE
) {
588 if (!(chunk
->flags
& NBD_REPLY_FLAG_DONE
)) {
589 error_setg(errp
, "Protocol error: NBD_REPLY_TYPE_NONE chunk without"
590 " NBD_REPLY_FLAG_DONE flag set");
594 error_setg(errp
, "Protocol error: NBD_REPLY_TYPE_NONE chunk with"
601 if (chunk
->type
== NBD_REPLY_TYPE_OFFSET_DATA
) {
603 error_setg(errp
, "Unexpected NBD_REPLY_TYPE_OFFSET_DATA chunk");
607 return nbd_co_receive_offset_data_payload(s
, s
->requests
[i
].offset
,
611 if (nbd_reply_type_is_error(chunk
->type
)) {
612 payload
= &local_payload
;
615 ret
= nbd_co_receive_structured_payload(s
, payload
, errp
);
620 if (nbd_reply_type_is_error(chunk
->type
)) {
621 ret
= nbd_parse_error_payload(chunk
, local_payload
, request_ret
, errp
);
622 g_free(local_payload
);
630 * nbd_co_receive_one_chunk
631 * Read reply, wake up connection_co and set s->quit if needed.
632 * Return value is a fatal error code or normal nbd reply error code
634 static coroutine_fn
int nbd_co_receive_one_chunk(
635 BDRVNBDState
*s
, uint64_t handle
, bool only_structured
,
636 int *request_ret
, QEMUIOVector
*qiov
, NBDReply
*reply
, void **payload
,
639 int ret
= nbd_co_do_receive_one_chunk(s
, handle
, only_structured
,
640 request_ret
, qiov
, payload
, errp
);
645 /* For assert at loop start in nbd_connection_entry */
652 if (s
->connection_co
) {
653 aio_co_wake(s
->connection_co
);
659 typedef struct NBDReplyChunkIter
{
663 bool done
, only_structured
;
666 static void nbd_iter_channel_error(NBDReplyChunkIter
*iter
,
667 int ret
, Error
**local_err
)
673 error_propagate(&iter
->err
, *local_err
);
675 error_free(*local_err
);
681 static void nbd_iter_request_error(NBDReplyChunkIter
*iter
, int ret
)
685 if (!iter
->request_ret
) {
686 iter
->request_ret
= ret
;
691 * NBD_FOREACH_REPLY_CHUNK
692 * The pointer stored in @payload requires g_free() to free it.
694 #define NBD_FOREACH_REPLY_CHUNK(s, iter, handle, structured, \
695 qiov, reply, payload) \
696 for (iter = (NBDReplyChunkIter) { .only_structured = structured }; \
697 nbd_reply_chunk_iter_receive(s, &iter, handle, qiov, reply, payload);)
700 * nbd_reply_chunk_iter_receive
701 * The pointer stored in @payload requires g_free() to free it.
703 static bool nbd_reply_chunk_iter_receive(BDRVNBDState
*s
,
704 NBDReplyChunkIter
*iter
,
706 QEMUIOVector
*qiov
, NBDReply
*reply
,
709 int ret
, request_ret
;
710 NBDReply local_reply
;
711 NBDStructuredReplyChunk
*chunk
;
712 Error
*local_err
= NULL
;
714 error_setg(&local_err
, "Connection closed");
715 nbd_iter_channel_error(iter
, -EIO
, &local_err
);
720 /* Previous iteration was last. */
725 reply
= &local_reply
;
728 ret
= nbd_co_receive_one_chunk(s
, handle
, iter
->only_structured
,
729 &request_ret
, qiov
, reply
, payload
,
732 nbd_iter_channel_error(iter
, ret
, &local_err
);
733 } else if (request_ret
< 0) {
734 nbd_iter_request_error(iter
, request_ret
);
737 /* Do not execute the body of NBD_FOREACH_REPLY_CHUNK for simple reply. */
738 if (nbd_reply_is_simple(reply
) || s
->quit
) {
742 chunk
= &reply
->structured
;
743 iter
->only_structured
= true;
745 if (chunk
->type
== NBD_REPLY_TYPE_NONE
) {
746 /* NBD_REPLY_FLAG_DONE is already checked in nbd_co_receive_one_chunk */
747 assert(chunk
->flags
& NBD_REPLY_FLAG_DONE
);
751 if (chunk
->flags
& NBD_REPLY_FLAG_DONE
) {
752 /* This iteration is last. */
756 /* Execute the loop body */
760 s
->requests
[HANDLE_TO_INDEX(s
, handle
)].coroutine
= NULL
;
762 qemu_co_mutex_lock(&s
->send_mutex
);
764 qemu_co_queue_next(&s
->free_sema
);
765 qemu_co_mutex_unlock(&s
->send_mutex
);
770 static int nbd_co_receive_return_code(BDRVNBDState
*s
, uint64_t handle
,
771 int *request_ret
, Error
**errp
)
773 NBDReplyChunkIter iter
;
775 NBD_FOREACH_REPLY_CHUNK(s
, iter
, handle
, false, NULL
, NULL
, NULL
) {
776 /* nbd_reply_chunk_iter_receive does all the work */
779 error_propagate(errp
, iter
.err
);
780 *request_ret
= iter
.request_ret
;
784 static int nbd_co_receive_cmdread_reply(BDRVNBDState
*s
, uint64_t handle
,
785 uint64_t offset
, QEMUIOVector
*qiov
,
786 int *request_ret
, Error
**errp
)
788 NBDReplyChunkIter iter
;
790 void *payload
= NULL
;
791 Error
*local_err
= NULL
;
793 NBD_FOREACH_REPLY_CHUNK(s
, iter
, handle
, s
->info
.structured_reply
,
794 qiov
, &reply
, &payload
)
797 NBDStructuredReplyChunk
*chunk
= &reply
.structured
;
799 assert(nbd_reply_is_structured(&reply
));
801 switch (chunk
->type
) {
802 case NBD_REPLY_TYPE_OFFSET_DATA
:
804 * special cased in nbd_co_receive_one_chunk, data is already
808 case NBD_REPLY_TYPE_OFFSET_HOLE
:
809 ret
= nbd_parse_offset_hole_payload(s
, &reply
.structured
, payload
,
810 offset
, qiov
, &local_err
);
813 nbd_iter_channel_error(&iter
, ret
, &local_err
);
817 if (!nbd_reply_type_is_error(chunk
->type
)) {
818 /* not allowed reply type */
820 error_setg(&local_err
,
821 "Unexpected reply type: %d (%s) for CMD_READ",
822 chunk
->type
, nbd_reply_type_lookup(chunk
->type
));
823 nbd_iter_channel_error(&iter
, -EINVAL
, &local_err
);
831 error_propagate(errp
, iter
.err
);
832 *request_ret
= iter
.request_ret
;
836 static int nbd_co_receive_blockstatus_reply(BDRVNBDState
*s
,
837 uint64_t handle
, uint64_t length
,
839 int *request_ret
, Error
**errp
)
841 NBDReplyChunkIter iter
;
843 void *payload
= NULL
;
844 Error
*local_err
= NULL
;
845 bool received
= false;
847 assert(!extent
->length
);
848 NBD_FOREACH_REPLY_CHUNK(s
, iter
, handle
, false, NULL
, &reply
, &payload
) {
850 NBDStructuredReplyChunk
*chunk
= &reply
.structured
;
852 assert(nbd_reply_is_structured(&reply
));
854 switch (chunk
->type
) {
855 case NBD_REPLY_TYPE_BLOCK_STATUS
:
858 error_setg(&local_err
, "Several BLOCK_STATUS chunks in reply");
859 nbd_iter_channel_error(&iter
, -EINVAL
, &local_err
);
863 ret
= nbd_parse_blockstatus_payload(s
, &reply
.structured
,
864 payload
, length
, extent
,
868 nbd_iter_channel_error(&iter
, ret
, &local_err
);
872 if (!nbd_reply_type_is_error(chunk
->type
)) {
874 error_setg(&local_err
,
875 "Unexpected reply type: %d (%s) "
876 "for CMD_BLOCK_STATUS",
877 chunk
->type
, nbd_reply_type_lookup(chunk
->type
));
878 nbd_iter_channel_error(&iter
, -EINVAL
, &local_err
);
886 if (!extent
->length
&& !iter
.request_ret
) {
887 error_setg(&local_err
, "Server did not reply with any status extents");
888 nbd_iter_channel_error(&iter
, -EIO
, &local_err
);
891 error_propagate(errp
, iter
.err
);
892 *request_ret
= iter
.request_ret
;
896 static int nbd_co_request(BlockDriverState
*bs
, NBDRequest
*request
,
897 QEMUIOVector
*write_qiov
)
899 int ret
, request_ret
;
900 Error
*local_err
= NULL
;
901 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
903 assert(request
->type
!= NBD_CMD_READ
);
905 assert(request
->type
== NBD_CMD_WRITE
);
906 assert(request
->len
== iov_size(write_qiov
->iov
, write_qiov
->niov
));
908 assert(request
->type
!= NBD_CMD_WRITE
);
910 ret
= nbd_co_send_request(bs
, request
, write_qiov
);
915 ret
= nbd_co_receive_return_code(s
, request
->handle
,
916 &request_ret
, &local_err
);
918 trace_nbd_co_request_fail(request
->from
, request
->len
, request
->handle
,
919 request
->flags
, request
->type
,
920 nbd_cmd_lookup(request
->type
),
921 ret
, error_get_pretty(local_err
));
922 error_free(local_err
);
924 return ret
? ret
: request_ret
;
927 static int nbd_client_co_preadv(BlockDriverState
*bs
, uint64_t offset
,
928 uint64_t bytes
, QEMUIOVector
*qiov
, int flags
)
930 int ret
, request_ret
;
931 Error
*local_err
= NULL
;
932 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
933 NBDRequest request
= {
934 .type
= NBD_CMD_READ
,
939 assert(bytes
<= NBD_MAX_BUFFER_SIZE
);
946 * Work around the fact that the block layer doesn't do
947 * byte-accurate sizing yet - if the read exceeds the server's
948 * advertised size because the block layer rounded size up, then
949 * truncate the request to the server and tail-pad with zero.
951 if (offset
>= s
->info
.size
) {
952 assert(bytes
< BDRV_SECTOR_SIZE
);
953 qemu_iovec_memset(qiov
, 0, 0, bytes
);
956 if (offset
+ bytes
> s
->info
.size
) {
957 uint64_t slop
= offset
+ bytes
- s
->info
.size
;
959 assert(slop
< BDRV_SECTOR_SIZE
);
960 qemu_iovec_memset(qiov
, bytes
- slop
, 0, slop
);
964 ret
= nbd_co_send_request(bs
, &request
, NULL
);
969 ret
= nbd_co_receive_cmdread_reply(s
, request
.handle
, offset
, qiov
,
970 &request_ret
, &local_err
);
972 trace_nbd_co_request_fail(request
.from
, request
.len
, request
.handle
,
973 request
.flags
, request
.type
,
974 nbd_cmd_lookup(request
.type
),
975 ret
, error_get_pretty(local_err
));
976 error_free(local_err
);
978 return ret
? ret
: request_ret
;
981 static int nbd_client_co_pwritev(BlockDriverState
*bs
, uint64_t offset
,
982 uint64_t bytes
, QEMUIOVector
*qiov
, int flags
)
984 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
985 NBDRequest request
= {
986 .type
= NBD_CMD_WRITE
,
991 assert(!(s
->info
.flags
& NBD_FLAG_READ_ONLY
));
992 if (flags
& BDRV_REQ_FUA
) {
993 assert(s
->info
.flags
& NBD_FLAG_SEND_FUA
);
994 request
.flags
|= NBD_CMD_FLAG_FUA
;
997 assert(bytes
<= NBD_MAX_BUFFER_SIZE
);
1002 return nbd_co_request(bs
, &request
, qiov
);
1005 static int nbd_client_co_pwrite_zeroes(BlockDriverState
*bs
, int64_t offset
,
1006 int bytes
, BdrvRequestFlags flags
)
1008 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
1009 NBDRequest request
= {
1010 .type
= NBD_CMD_WRITE_ZEROES
,
1015 assert(!(s
->info
.flags
& NBD_FLAG_READ_ONLY
));
1016 if (!(s
->info
.flags
& NBD_FLAG_SEND_WRITE_ZEROES
)) {
1020 if (flags
& BDRV_REQ_FUA
) {
1021 assert(s
->info
.flags
& NBD_FLAG_SEND_FUA
);
1022 request
.flags
|= NBD_CMD_FLAG_FUA
;
1024 if (!(flags
& BDRV_REQ_MAY_UNMAP
)) {
1025 request
.flags
|= NBD_CMD_FLAG_NO_HOLE
;
1031 return nbd_co_request(bs
, &request
, NULL
);
1034 static int nbd_client_co_flush(BlockDriverState
*bs
)
1036 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
1037 NBDRequest request
= { .type
= NBD_CMD_FLUSH
};
1039 if (!(s
->info
.flags
& NBD_FLAG_SEND_FLUSH
)) {
1046 return nbd_co_request(bs
, &request
, NULL
);
1049 static int nbd_client_co_pdiscard(BlockDriverState
*bs
, int64_t offset
,
1052 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
1053 NBDRequest request
= {
1054 .type
= NBD_CMD_TRIM
,
1059 assert(!(s
->info
.flags
& NBD_FLAG_READ_ONLY
));
1060 if (!(s
->info
.flags
& NBD_FLAG_SEND_TRIM
) || !bytes
) {
1064 return nbd_co_request(bs
, &request
, NULL
);
1067 static int coroutine_fn
nbd_client_co_block_status(
1068 BlockDriverState
*bs
, bool want_zero
, int64_t offset
, int64_t bytes
,
1069 int64_t *pnum
, int64_t *map
, BlockDriverState
**file
)
1071 int ret
, request_ret
;
1072 NBDExtent extent
= { 0 };
1073 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
1074 Error
*local_err
= NULL
;
1076 NBDRequest request
= {
1077 .type
= NBD_CMD_BLOCK_STATUS
,
1079 .len
= MIN(MIN_NON_ZERO(QEMU_ALIGN_DOWN(INT_MAX
,
1080 bs
->bl
.request_alignment
),
1082 MIN(bytes
, s
->info
.size
- offset
)),
1083 .flags
= NBD_CMD_FLAG_REQ_ONE
,
1086 if (!s
->info
.base_allocation
) {
1090 return BDRV_BLOCK_DATA
| BDRV_BLOCK_OFFSET_VALID
;
1094 * Work around the fact that the block layer doesn't do
1095 * byte-accurate sizing yet - if the status request exceeds the
1096 * server's advertised size because the block layer rounded size
1097 * up, we truncated the request to the server (above), or are
1098 * called on just the hole.
1100 if (offset
>= s
->info
.size
) {
1102 assert(bytes
< BDRV_SECTOR_SIZE
);
1103 /* Intentionally don't report offset_valid for the hole */
1104 return BDRV_BLOCK_ZERO
;
1107 if (s
->info
.min_block
) {
1108 assert(QEMU_IS_ALIGNED(request
.len
, s
->info
.min_block
));
1110 ret
= nbd_co_send_request(bs
, &request
, NULL
);
1115 ret
= nbd_co_receive_blockstatus_reply(s
, request
.handle
, bytes
,
1116 &extent
, &request_ret
, &local_err
);
1118 trace_nbd_co_request_fail(request
.from
, request
.len
, request
.handle
,
1119 request
.flags
, request
.type
,
1120 nbd_cmd_lookup(request
.type
),
1121 ret
, error_get_pretty(local_err
));
1122 error_free(local_err
);
1124 if (ret
< 0 || request_ret
< 0) {
1125 return ret
? ret
: request_ret
;
1128 assert(extent
.length
);
1129 *pnum
= extent
.length
;
1132 return (extent
.flags
& NBD_STATE_HOLE
? 0 : BDRV_BLOCK_DATA
) |
1133 (extent
.flags
& NBD_STATE_ZERO
? BDRV_BLOCK_ZERO
: 0) |
1134 BDRV_BLOCK_OFFSET_VALID
;
1137 static void nbd_client_close(BlockDriverState
*bs
)
1139 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
1140 NBDRequest request
= { .type
= NBD_CMD_DISC
};
1144 nbd_send_request(s
->ioc
, &request
);
1146 nbd_teardown_connection(bs
);
1149 static QIOChannelSocket
*nbd_establish_connection(SocketAddress
*saddr
,
1152 QIOChannelSocket
*sioc
;
1153 Error
*local_err
= NULL
;
1155 sioc
= qio_channel_socket_new();
1156 qio_channel_set_name(QIO_CHANNEL(sioc
), "nbd-client");
1158 qio_channel_socket_connect_sync(sioc
, saddr
, &local_err
);
1160 object_unref(OBJECT(sioc
));
1161 error_propagate(errp
, local_err
);
1165 qio_channel_set_delay(QIO_CHANNEL(sioc
), false);
1170 static int nbd_client_connect(BlockDriverState
*bs
,
1171 SocketAddress
*saddr
,
1173 QCryptoTLSCreds
*tlscreds
,
1174 const char *hostname
,
1175 const char *x_dirty_bitmap
,
1178 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
1182 * establish TCP connection, return error if it fails
1183 * TODO: Configurable retry-until-timeout behaviour.
1185 QIOChannelSocket
*sioc
= nbd_establish_connection(saddr
, errp
);
1188 return -ECONNREFUSED
;
1192 trace_nbd_client_connect(export
);
1193 qio_channel_set_blocking(QIO_CHANNEL(sioc
), true, NULL
);
1195 s
->info
.request_sizes
= true;
1196 s
->info
.structured_reply
= true;
1197 s
->info
.base_allocation
= true;
1198 s
->info
.x_dirty_bitmap
= g_strdup(x_dirty_bitmap
);
1199 s
->info
.name
= g_strdup(export
?: "");
1200 ret
= nbd_receive_negotiate(QIO_CHANNEL(sioc
), tlscreds
, hostname
,
1201 &s
->ioc
, &s
->info
, errp
);
1202 g_free(s
->info
.x_dirty_bitmap
);
1203 g_free(s
->info
.name
);
1205 object_unref(OBJECT(sioc
));
1208 if (x_dirty_bitmap
&& !s
->info
.base_allocation
) {
1209 error_setg(errp
, "requested x-dirty-bitmap %s not found",
1214 if (s
->info
.flags
& NBD_FLAG_READ_ONLY
) {
1215 ret
= bdrv_apply_auto_read_only(bs
, "NBD export is read-only", errp
);
1220 if (s
->info
.flags
& NBD_FLAG_SEND_FUA
) {
1221 bs
->supported_write_flags
= BDRV_REQ_FUA
;
1222 bs
->supported_zero_flags
|= BDRV_REQ_FUA
;
1224 if (s
->info
.flags
& NBD_FLAG_SEND_WRITE_ZEROES
) {
1225 bs
->supported_zero_flags
|= BDRV_REQ_MAY_UNMAP
;
1231 s
->ioc
= QIO_CHANNEL(sioc
);
1232 object_ref(OBJECT(s
->ioc
));
1236 * Now that we're connected, set the socket to be non-blocking and
1237 * kick the reply mechanism.
1239 qio_channel_set_blocking(QIO_CHANNEL(sioc
), false, NULL
);
1240 s
->connection_co
= qemu_coroutine_create(nbd_connection_entry
, s
);
1241 bdrv_inc_in_flight(bs
);
1242 nbd_client_attach_aio_context(bs
, bdrv_get_aio_context(bs
));
1244 trace_nbd_client_connect_success(export
);
1250 * We have connected, but must fail for other reasons. The
1251 * connection is still blocking; send NBD_CMD_DISC as a courtesy
1255 NBDRequest request
= { .type
= NBD_CMD_DISC
};
1257 nbd_send_request(s
->ioc
?: QIO_CHANNEL(sioc
), &request
);
1259 object_unref(OBJECT(sioc
));
1265 static int nbd_client_init(BlockDriverState
*bs
,
1266 SocketAddress
*saddr
,
1268 QCryptoTLSCreds
*tlscreds
,
1269 const char *hostname
,
1270 const char *x_dirty_bitmap
,
1273 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
1276 qemu_co_mutex_init(&s
->send_mutex
);
1277 qemu_co_queue_init(&s
->free_sema
);
1279 return nbd_client_connect(bs
, saddr
, export
, tlscreds
, hostname
,
1280 x_dirty_bitmap
, errp
);
1283 static int nbd_parse_uri(const char *filename
, QDict
*options
)
1287 QueryParams
*qp
= NULL
;
1291 uri
= uri_parse(filename
);
1297 if (!g_strcmp0(uri
->scheme
, "nbd")) {
1299 } else if (!g_strcmp0(uri
->scheme
, "nbd+tcp")) {
1301 } else if (!g_strcmp0(uri
->scheme
, "nbd+unix")) {
1308 p
= uri
->path
? uri
->path
: "/";
1309 p
+= strspn(p
, "/");
1311 qdict_put_str(options
, "export", p
);
1314 qp
= query_params_parse(uri
->query
);
1315 if (qp
->n
> 1 || (is_unix
&& !qp
->n
) || (!is_unix
&& qp
->n
)) {
1321 /* nbd+unix:///export?socket=path */
1322 if (uri
->server
|| uri
->port
|| strcmp(qp
->p
[0].name
, "socket")) {
1326 qdict_put_str(options
, "server.type", "unix");
1327 qdict_put_str(options
, "server.path", qp
->p
[0].value
);
1332 /* nbd[+tcp]://host[:port]/export */
1338 /* strip braces from literal IPv6 address */
1339 if (uri
->server
[0] == '[') {
1340 host
= qstring_from_substr(uri
->server
, 1,
1341 strlen(uri
->server
) - 1);
1343 host
= qstring_from_str(uri
->server
);
1346 qdict_put_str(options
, "server.type", "inet");
1347 qdict_put(options
, "server.host", host
);
1349 port_str
= g_strdup_printf("%d", uri
->port
?: NBD_DEFAULT_PORT
);
1350 qdict_put_str(options
, "server.port", port_str
);
1356 query_params_free(qp
);
1362 static bool nbd_has_filename_options_conflict(QDict
*options
, Error
**errp
)
1364 const QDictEntry
*e
;
1366 for (e
= qdict_first(options
); e
; e
= qdict_next(options
, e
)) {
1367 if (!strcmp(e
->key
, "host") ||
1368 !strcmp(e
->key
, "port") ||
1369 !strcmp(e
->key
, "path") ||
1370 !strcmp(e
->key
, "export") ||
1371 strstart(e
->key
, "server.", NULL
))
1373 error_setg(errp
, "Option '%s' cannot be used with a file name",
1382 static void nbd_parse_filename(const char *filename
, QDict
*options
,
1387 const char *host_spec
;
1388 const char *unixpath
;
1390 if (nbd_has_filename_options_conflict(options
, errp
)) {
1394 if (strstr(filename
, "://")) {
1395 int ret
= nbd_parse_uri(filename
, options
);
1397 error_setg(errp
, "No valid URL specified");
1402 file
= g_strdup(filename
);
1404 export_name
= strstr(file
, EN_OPTSTR
);
1406 if (export_name
[strlen(EN_OPTSTR
)] == 0) {
1409 export_name
[0] = 0; /* truncate 'file' */
1410 export_name
+= strlen(EN_OPTSTR
);
1412 qdict_put_str(options
, "export", export_name
);
1415 /* extract the host_spec - fail if it's not nbd:... */
1416 if (!strstart(file
, "nbd:", &host_spec
)) {
1417 error_setg(errp
, "File name string for NBD must start with 'nbd:'");
1425 /* are we a UNIX or TCP socket? */
1426 if (strstart(host_spec
, "unix:", &unixpath
)) {
1427 qdict_put_str(options
, "server.type", "unix");
1428 qdict_put_str(options
, "server.path", unixpath
);
1430 InetSocketAddress
*addr
= g_new(InetSocketAddress
, 1);
1432 if (inet_parse(addr
, host_spec
, errp
)) {
1436 qdict_put_str(options
, "server.type", "inet");
1437 qdict_put_str(options
, "server.host", addr
->host
);
1438 qdict_put_str(options
, "server.port", addr
->port
);
1440 qapi_free_InetSocketAddress(addr
);
1447 static bool nbd_process_legacy_socket_options(QDict
*output_options
,
1448 QemuOpts
*legacy_opts
,
1451 const char *path
= qemu_opt_get(legacy_opts
, "path");
1452 const char *host
= qemu_opt_get(legacy_opts
, "host");
1453 const char *port
= qemu_opt_get(legacy_opts
, "port");
1454 const QDictEntry
*e
;
1456 if (!path
&& !host
&& !port
) {
1460 for (e
= qdict_first(output_options
); e
; e
= qdict_next(output_options
, e
))
1462 if (strstart(e
->key
, "server.", NULL
)) {
1463 error_setg(errp
, "Cannot use 'server' and path/host/port at the "
1470 error_setg(errp
, "path and host may not be used at the same time");
1474 error_setg(errp
, "port may not be used without host");
1478 qdict_put_str(output_options
, "server.type", "unix");
1479 qdict_put_str(output_options
, "server.path", path
);
1481 qdict_put_str(output_options
, "server.type", "inet");
1482 qdict_put_str(output_options
, "server.host", host
);
1483 qdict_put_str(output_options
, "server.port",
1484 port
?: stringify(NBD_DEFAULT_PORT
));
1490 static SocketAddress
*nbd_config(BDRVNBDState
*s
, QDict
*options
,
1493 SocketAddress
*saddr
= NULL
;
1496 Error
*local_err
= NULL
;
1498 qdict_extract_subqdict(options
, &addr
, "server.");
1499 if (!qdict_size(addr
)) {
1500 error_setg(errp
, "NBD server address missing");
1504 iv
= qobject_input_visitor_new_flat_confused(addr
, errp
);
1509 visit_type_SocketAddress(iv
, NULL
, &saddr
, &local_err
);
1511 error_propagate(errp
, local_err
);
1516 qobject_unref(addr
);
1521 static QCryptoTLSCreds
*nbd_get_tls_creds(const char *id
, Error
**errp
)
1524 QCryptoTLSCreds
*creds
;
1526 obj
= object_resolve_path_component(
1527 object_get_objects_root(), id
);
1529 error_setg(errp
, "No TLS credentials with id '%s'",
1533 creds
= (QCryptoTLSCreds
*)
1534 object_dynamic_cast(obj
, TYPE_QCRYPTO_TLS_CREDS
);
1536 error_setg(errp
, "Object with id '%s' is not TLS credentials",
1541 if (creds
->endpoint
!= QCRYPTO_TLS_CREDS_ENDPOINT_CLIENT
) {
1543 "Expecting TLS credentials with a client endpoint");
1551 static QemuOptsList nbd_runtime_opts
= {
1553 .head
= QTAILQ_HEAD_INITIALIZER(nbd_runtime_opts
.head
),
1557 .type
= QEMU_OPT_STRING
,
1558 .help
= "TCP host to connect to",
1562 .type
= QEMU_OPT_STRING
,
1563 .help
= "TCP port to connect to",
1567 .type
= QEMU_OPT_STRING
,
1568 .help
= "Unix socket path to connect to",
1572 .type
= QEMU_OPT_STRING
,
1573 .help
= "Name of the NBD export to open",
1576 .name
= "tls-creds",
1577 .type
= QEMU_OPT_STRING
,
1578 .help
= "ID of the TLS credentials to use",
1581 .name
= "x-dirty-bitmap",
1582 .type
= QEMU_OPT_STRING
,
1583 .help
= "experimental: expose named dirty bitmap in place of "
1586 { /* end of list */ }
1590 static int nbd_open(BlockDriverState
*bs
, QDict
*options
, int flags
,
1593 BDRVNBDState
*s
= bs
->opaque
;
1594 QemuOpts
*opts
= NULL
;
1595 Error
*local_err
= NULL
;
1596 QCryptoTLSCreds
*tlscreds
= NULL
;
1597 const char *hostname
= NULL
;
1600 opts
= qemu_opts_create(&nbd_runtime_opts
, NULL
, 0, &error_abort
);
1601 qemu_opts_absorb_qdict(opts
, options
, &local_err
);
1603 error_propagate(errp
, local_err
);
1607 /* Translate @host, @port, and @path to a SocketAddress */
1608 if (!nbd_process_legacy_socket_options(options
, opts
, errp
)) {
1612 /* Pop the config into our state object. Exit if invalid. */
1613 s
->saddr
= nbd_config(s
, options
, errp
);
1618 s
->export
= g_strdup(qemu_opt_get(opts
, "export"));
1620 s
->tlscredsid
= g_strdup(qemu_opt_get(opts
, "tls-creds"));
1621 if (s
->tlscredsid
) {
1622 tlscreds
= nbd_get_tls_creds(s
->tlscredsid
, errp
);
1627 /* TODO SOCKET_ADDRESS_KIND_FD where fd has AF_INET or AF_INET6 */
1628 if (s
->saddr
->type
!= SOCKET_ADDRESS_TYPE_INET
) {
1629 error_setg(errp
, "TLS only supported over IP sockets");
1632 hostname
= s
->saddr
->u
.inet
.host
;
1636 ret
= nbd_client_init(bs
, s
->saddr
, s
->export
, tlscreds
, hostname
,
1637 qemu_opt_get(opts
, "x-dirty-bitmap"), errp
);
1641 object_unref(OBJECT(tlscreds
));
1644 qapi_free_SocketAddress(s
->saddr
);
1646 g_free(s
->tlscredsid
);
1648 qemu_opts_del(opts
);
1652 static int nbd_co_flush(BlockDriverState
*bs
)
1654 return nbd_client_co_flush(bs
);
1657 static void nbd_refresh_limits(BlockDriverState
*bs
, Error
**errp
)
1659 BDRVNBDState
*s
= (BDRVNBDState
*)bs
->opaque
;
1660 uint32_t min
= s
->info
.min_block
;
1661 uint32_t max
= MIN_NON_ZERO(NBD_MAX_BUFFER_SIZE
, s
->info
.max_block
);
1664 * If the server did not advertise an alignment:
1665 * - a size that is not sector-aligned implies that an alignment
1666 * of 1 can be used to access those tail bytes
1667 * - advertisement of block status requires an alignment of 1, so
1668 * that we don't violate block layer constraints that block
1669 * status is always aligned (as we can't control whether the
1670 * server will report sub-sector extents, such as a hole at EOF
1671 * on an unaligned POSIX file)
1672 * - otherwise, assume the server is so old that we are safer avoiding
1673 * sub-sector requests
1676 min
= (!QEMU_IS_ALIGNED(s
->info
.size
, BDRV_SECTOR_SIZE
) ||
1677 s
->info
.base_allocation
) ? 1 : BDRV_SECTOR_SIZE
;
1680 bs
->bl
.request_alignment
= min
;
1681 bs
->bl
.max_pdiscard
= max
;
1682 bs
->bl
.max_pwrite_zeroes
= max
;
1683 bs
->bl
.max_transfer
= max
;
1685 if (s
->info
.opt_block
&&
1686 s
->info
.opt_block
> bs
->bl
.opt_transfer
) {
1687 bs
->bl
.opt_transfer
= s
->info
.opt_block
;
1691 static void nbd_close(BlockDriverState
*bs
)
1693 BDRVNBDState
*s
= bs
->opaque
;
1695 nbd_client_close(bs
);
1697 qapi_free_SocketAddress(s
->saddr
);
1699 g_free(s
->tlscredsid
);
1702 static int64_t nbd_getlength(BlockDriverState
*bs
)
1704 BDRVNBDState
*s
= bs
->opaque
;
1706 return s
->info
.size
;
1709 static void nbd_refresh_filename(BlockDriverState
*bs
)
1711 BDRVNBDState
*s
= bs
->opaque
;
1712 const char *host
= NULL
, *port
= NULL
, *path
= NULL
;
1714 if (s
->saddr
->type
== SOCKET_ADDRESS_TYPE_INET
) {
1715 const InetSocketAddress
*inet
= &s
->saddr
->u
.inet
;
1716 if (!inet
->has_ipv4
&& !inet
->has_ipv6
&& !inet
->has_to
) {
1720 } else if (s
->saddr
->type
== SOCKET_ADDRESS_TYPE_UNIX
) {
1721 path
= s
->saddr
->u
.q_unix
.path
;
1722 } /* else can't represent as pseudo-filename */
1724 if (path
&& s
->export
) {
1725 snprintf(bs
->exact_filename
, sizeof(bs
->exact_filename
),
1726 "nbd+unix:///%s?socket=%s", s
->export
, path
);
1727 } else if (path
&& !s
->export
) {
1728 snprintf(bs
->exact_filename
, sizeof(bs
->exact_filename
),
1729 "nbd+unix://?socket=%s", path
);
1730 } else if (host
&& s
->export
) {
1731 snprintf(bs
->exact_filename
, sizeof(bs
->exact_filename
),
1732 "nbd://%s:%s/%s", host
, port
, s
->export
);
1733 } else if (host
&& !s
->export
) {
1734 snprintf(bs
->exact_filename
, sizeof(bs
->exact_filename
),
1735 "nbd://%s:%s", host
, port
);
1739 static char *nbd_dirname(BlockDriverState
*bs
, Error
**errp
)
1741 /* The generic bdrv_dirname() implementation is able to work out some
1742 * directory name for NBD nodes, but that would be wrong. So far there is no
1743 * specification for how "export paths" would work, so NBD does not have
1744 * directory names. */
1745 error_setg(errp
, "Cannot generate a base directory for NBD nodes");
1749 static const char *const nbd_strong_runtime_opts
[] = {
1760 static BlockDriver bdrv_nbd
= {
1761 .format_name
= "nbd",
1762 .protocol_name
= "nbd",
1763 .instance_size
= sizeof(BDRVNBDState
),
1764 .bdrv_parse_filename
= nbd_parse_filename
,
1765 .bdrv_file_open
= nbd_open
,
1766 .bdrv_co_preadv
= nbd_client_co_preadv
,
1767 .bdrv_co_pwritev
= nbd_client_co_pwritev
,
1768 .bdrv_co_pwrite_zeroes
= nbd_client_co_pwrite_zeroes
,
1769 .bdrv_close
= nbd_close
,
1770 .bdrv_co_flush_to_os
= nbd_co_flush
,
1771 .bdrv_co_pdiscard
= nbd_client_co_pdiscard
,
1772 .bdrv_refresh_limits
= nbd_refresh_limits
,
1773 .bdrv_getlength
= nbd_getlength
,
1774 .bdrv_detach_aio_context
= nbd_client_detach_aio_context
,
1775 .bdrv_attach_aio_context
= nbd_client_attach_aio_context
,
1776 .bdrv_refresh_filename
= nbd_refresh_filename
,
1777 .bdrv_co_block_status
= nbd_client_co_block_status
,
1778 .bdrv_dirname
= nbd_dirname
,
1779 .strong_runtime_opts
= nbd_strong_runtime_opts
,
1782 static BlockDriver bdrv_nbd_tcp
= {
1783 .format_name
= "nbd",
1784 .protocol_name
= "nbd+tcp",
1785 .instance_size
= sizeof(BDRVNBDState
),
1786 .bdrv_parse_filename
= nbd_parse_filename
,
1787 .bdrv_file_open
= nbd_open
,
1788 .bdrv_co_preadv
= nbd_client_co_preadv
,
1789 .bdrv_co_pwritev
= nbd_client_co_pwritev
,
1790 .bdrv_co_pwrite_zeroes
= nbd_client_co_pwrite_zeroes
,
1791 .bdrv_close
= nbd_close
,
1792 .bdrv_co_flush_to_os
= nbd_co_flush
,
1793 .bdrv_co_pdiscard
= nbd_client_co_pdiscard
,
1794 .bdrv_refresh_limits
= nbd_refresh_limits
,
1795 .bdrv_getlength
= nbd_getlength
,
1796 .bdrv_detach_aio_context
= nbd_client_detach_aio_context
,
1797 .bdrv_attach_aio_context
= nbd_client_attach_aio_context
,
1798 .bdrv_refresh_filename
= nbd_refresh_filename
,
1799 .bdrv_co_block_status
= nbd_client_co_block_status
,
1800 .bdrv_dirname
= nbd_dirname
,
1801 .strong_runtime_opts
= nbd_strong_runtime_opts
,
1804 static BlockDriver bdrv_nbd_unix
= {
1805 .format_name
= "nbd",
1806 .protocol_name
= "nbd+unix",
1807 .instance_size
= sizeof(BDRVNBDState
),
1808 .bdrv_parse_filename
= nbd_parse_filename
,
1809 .bdrv_file_open
= nbd_open
,
1810 .bdrv_co_preadv
= nbd_client_co_preadv
,
1811 .bdrv_co_pwritev
= nbd_client_co_pwritev
,
1812 .bdrv_co_pwrite_zeroes
= nbd_client_co_pwrite_zeroes
,
1813 .bdrv_close
= nbd_close
,
1814 .bdrv_co_flush_to_os
= nbd_co_flush
,
1815 .bdrv_co_pdiscard
= nbd_client_co_pdiscard
,
1816 .bdrv_refresh_limits
= nbd_refresh_limits
,
1817 .bdrv_getlength
= nbd_getlength
,
1818 .bdrv_detach_aio_context
= nbd_client_detach_aio_context
,
1819 .bdrv_attach_aio_context
= nbd_client_attach_aio_context
,
1820 .bdrv_refresh_filename
= nbd_refresh_filename
,
1821 .bdrv_co_block_status
= nbd_client_co_block_status
,
1822 .bdrv_dirname
= nbd_dirname
,
1823 .strong_runtime_opts
= nbd_strong_runtime_opts
,
1826 static void bdrv_nbd_init(void)
1828 bdrv_register(&bdrv_nbd
);
1829 bdrv_register(&bdrv_nbd_tcp
);
1830 bdrv_register(&bdrv_nbd_unix
);
1833 block_init(bdrv_nbd_init
);