1 #include "qemu-common.h"
3 #include "qemu/sockets.h"
4 #include "block/coroutine.h"
5 #include "migration/migration.h"
6 #include "migration/qemu-file.h"
9 #define IO_BUF_SIZE 32768
10 #define MAX_IOV_SIZE MIN(IOV_MAX, 64)
13 const QEMUFileOps
*ops
;
19 int64_t pos
; /* start of buffer when writing, end of buffer
22 int buf_size
; /* 0 when writing */
23 uint8_t buf
[IO_BUF_SIZE
];
25 struct iovec iov
[MAX_IOV_SIZE
];
31 typedef struct QEMUFileStdio
{
36 typedef struct QEMUFileSocket
{
41 static ssize_t
socket_writev_buffer(void *opaque
, struct iovec
*iov
, int iovcnt
,
44 QEMUFileSocket
*s
= opaque
;
46 ssize_t size
= iov_size(iov
, iovcnt
);
48 len
= iov_send(s
->fd
, iov
, iovcnt
, 0, size
);
50 len
= -socket_error();
55 static int socket_get_fd(void *opaque
)
57 QEMUFileSocket
*s
= opaque
;
62 static int socket_get_buffer(void *opaque
, uint8_t *buf
, int64_t pos
, int size
)
64 QEMUFileSocket
*s
= opaque
;
68 len
= qemu_recv(s
->fd
, buf
, size
, 0);
72 if (socket_error() == EAGAIN
) {
73 yield_until_fd_readable(s
->fd
);
74 } else if (socket_error() != EINTR
) {
80 len
= -socket_error();
85 static int socket_close(void *opaque
)
87 QEMUFileSocket
*s
= opaque
;
93 static int stdio_get_fd(void *opaque
)
95 QEMUFileStdio
*s
= opaque
;
97 return fileno(s
->stdio_file
);
100 static int stdio_put_buffer(void *opaque
, const uint8_t *buf
, int64_t pos
,
103 QEMUFileStdio
*s
= opaque
;
106 res
= fwrite(buf
, 1, size
, s
->stdio_file
);
114 static int stdio_get_buffer(void *opaque
, uint8_t *buf
, int64_t pos
, int size
)
116 QEMUFileStdio
*s
= opaque
;
117 FILE *fp
= s
->stdio_file
;
122 bytes
= fread(buf
, 1, size
, fp
);
123 if (bytes
!= 0 || !ferror(fp
)) {
126 if (errno
== EAGAIN
) {
127 yield_until_fd_readable(fileno(fp
));
128 } else if (errno
!= EINTR
) {
135 static int stdio_pclose(void *opaque
)
137 QEMUFileStdio
*s
= opaque
;
139 ret
= pclose(s
->stdio_file
);
142 } else if (!WIFEXITED(ret
) || WEXITSTATUS(ret
) != 0) {
143 /* close succeeded, but non-zero exit code: */
144 ret
= -EIO
; /* fake errno value */
150 static int stdio_fclose(void *opaque
)
152 QEMUFileStdio
*s
= opaque
;
155 if (s
->file
->ops
->put_buffer
|| s
->file
->ops
->writev_buffer
) {
156 int fd
= fileno(s
->stdio_file
);
159 ret
= fstat(fd
, &st
);
160 if (ret
== 0 && S_ISREG(st
.st_mode
)) {
162 * If the file handle is a regular file make sure the
163 * data is flushed to disk before signaling success.
172 if (fclose(s
->stdio_file
) == EOF
) {
179 static const QEMUFileOps stdio_pipe_read_ops
= {
180 .get_fd
= stdio_get_fd
,
181 .get_buffer
= stdio_get_buffer
,
182 .close
= stdio_pclose
185 static const QEMUFileOps stdio_pipe_write_ops
= {
186 .get_fd
= stdio_get_fd
,
187 .put_buffer
= stdio_put_buffer
,
188 .close
= stdio_pclose
191 QEMUFile
*qemu_popen_cmd(const char *command
, const char *mode
)
196 if (mode
== NULL
|| (mode
[0] != 'r' && mode
[0] != 'w') || mode
[1] != 0) {
197 fprintf(stderr
, "qemu_popen: Argument validity check failed\n");
201 stdio_file
= popen(command
, mode
);
202 if (stdio_file
== NULL
) {
206 s
= g_malloc0(sizeof(QEMUFileStdio
));
208 s
->stdio_file
= stdio_file
;
210 if (mode
[0] == 'r') {
211 s
->file
= qemu_fopen_ops(s
, &stdio_pipe_read_ops
);
213 s
->file
= qemu_fopen_ops(s
, &stdio_pipe_write_ops
);
218 static const QEMUFileOps stdio_file_read_ops
= {
219 .get_fd
= stdio_get_fd
,
220 .get_buffer
= stdio_get_buffer
,
221 .close
= stdio_fclose
224 static const QEMUFileOps stdio_file_write_ops
= {
225 .get_fd
= stdio_get_fd
,
226 .put_buffer
= stdio_put_buffer
,
227 .close
= stdio_fclose
230 static ssize_t
unix_writev_buffer(void *opaque
, struct iovec
*iov
, int iovcnt
,
233 QEMUFileSocket
*s
= opaque
;
235 ssize_t size
= iov_size(iov
, iovcnt
);
241 /* Find the next start position; skip all full-sized vector elements */
242 while (offset
>= iov
[0].iov_len
) {
243 offset
-= iov
[0].iov_len
;
247 /* skip `offset' bytes from the (now) first element, undo it on exit */
249 iov
[0].iov_base
+= offset
;
250 iov
[0].iov_len
-= offset
;
253 len
= writev(s
->fd
, iov
, iovcnt
);
254 } while (len
== -1 && errno
== EINTR
);
259 /* Undo the changes above */
260 iov
[0].iov_base
-= offset
;
261 iov
[0].iov_len
+= offset
;
263 /* Prepare for the next iteration */
272 static int unix_get_buffer(void *opaque
, uint8_t *buf
, int64_t pos
, int size
)
274 QEMUFileSocket
*s
= opaque
;
278 len
= read(s
->fd
, buf
, size
);
282 if (errno
== EAGAIN
) {
283 yield_until_fd_readable(s
->fd
);
284 } else if (errno
!= EINTR
) {
295 static int unix_close(void *opaque
)
297 QEMUFileSocket
*s
= opaque
;
303 static const QEMUFileOps unix_read_ops
= {
304 .get_fd
= socket_get_fd
,
305 .get_buffer
= unix_get_buffer
,
309 static const QEMUFileOps unix_write_ops
= {
310 .get_fd
= socket_get_fd
,
311 .writev_buffer
= unix_writev_buffer
,
315 QEMUFile
*qemu_fdopen(int fd
, const char *mode
)
320 (mode
[0] != 'r' && mode
[0] != 'w') ||
321 mode
[1] != 'b' || mode
[2] != 0) {
322 fprintf(stderr
, "qemu_fdopen: Argument validity check failed\n");
326 s
= g_malloc0(sizeof(QEMUFileSocket
));
329 if (mode
[0] == 'r') {
330 s
->file
= qemu_fopen_ops(s
, &unix_read_ops
);
332 s
->file
= qemu_fopen_ops(s
, &unix_write_ops
);
337 static const QEMUFileOps socket_read_ops
= {
338 .get_fd
= socket_get_fd
,
339 .get_buffer
= socket_get_buffer
,
340 .close
= socket_close
343 static const QEMUFileOps socket_write_ops
= {
344 .get_fd
= socket_get_fd
,
345 .writev_buffer
= socket_writev_buffer
,
346 .close
= socket_close
349 bool qemu_file_mode_is_not_valid(const char *mode
)
352 (mode
[0] != 'r' && mode
[0] != 'w') ||
353 mode
[1] != 'b' || mode
[2] != 0) {
354 fprintf(stderr
, "qemu_fopen: Argument validity check failed\n");
361 QEMUFile
*qemu_fopen_socket(int fd
, const char *mode
)
365 if (qemu_file_mode_is_not_valid(mode
)) {
369 s
= g_malloc0(sizeof(QEMUFileSocket
));
371 if (mode
[0] == 'w') {
372 qemu_set_block(s
->fd
);
373 s
->file
= qemu_fopen_ops(s
, &socket_write_ops
);
375 s
->file
= qemu_fopen_ops(s
, &socket_read_ops
);
380 QEMUFile
*qemu_fopen(const char *filename
, const char *mode
)
384 if (qemu_file_mode_is_not_valid(mode
)) {
388 s
= g_malloc0(sizeof(QEMUFileStdio
));
390 s
->stdio_file
= fopen(filename
, mode
);
391 if (!s
->stdio_file
) {
395 if (mode
[0] == 'w') {
396 s
->file
= qemu_fopen_ops(s
, &stdio_file_write_ops
);
398 s
->file
= qemu_fopen_ops(s
, &stdio_file_read_ops
);
406 QEMUFile
*qemu_fopen_ops(void *opaque
, const QEMUFileOps
*ops
)
410 f
= g_malloc0(sizeof(QEMUFile
));
418 * Get last error for stream f
420 * Return negative error value if there has been an error on previous
421 * operations, return 0 if no error happened.
424 int qemu_file_get_error(QEMUFile
*f
)
426 return f
->last_error
;
429 void qemu_file_set_error(QEMUFile
*f
, int ret
)
431 if (f
->last_error
== 0) {
436 static inline bool qemu_file_is_writable(QEMUFile
*f
)
438 return f
->ops
->writev_buffer
|| f
->ops
->put_buffer
;
442 * Flushes QEMUFile buffer
444 * If there is writev_buffer QEMUFileOps it uses it otherwise uses
447 void qemu_fflush(QEMUFile
*f
)
451 if (!qemu_file_is_writable(f
)) {
455 if (f
->ops
->writev_buffer
) {
457 ret
= f
->ops
->writev_buffer(f
->opaque
, f
->iov
, f
->iovcnt
, f
->pos
);
460 if (f
->buf_index
> 0) {
461 ret
= f
->ops
->put_buffer(f
->opaque
, f
->buf
, f
->pos
, f
->buf_index
);
470 qemu_file_set_error(f
, ret
);
474 void ram_control_before_iterate(QEMUFile
*f
, uint64_t flags
)
478 if (f
->ops
->before_ram_iterate
) {
479 ret
= f
->ops
->before_ram_iterate(f
, f
->opaque
, flags
);
481 qemu_file_set_error(f
, ret
);
486 void ram_control_after_iterate(QEMUFile
*f
, uint64_t flags
)
490 if (f
->ops
->after_ram_iterate
) {
491 ret
= f
->ops
->after_ram_iterate(f
, f
->opaque
, flags
);
493 qemu_file_set_error(f
, ret
);
498 void ram_control_load_hook(QEMUFile
*f
, uint64_t flags
)
502 if (f
->ops
->hook_ram_load
) {
503 ret
= f
->ops
->hook_ram_load(f
, f
->opaque
, flags
);
505 qemu_file_set_error(f
, ret
);
508 qemu_file_set_error(f
, ret
);
512 size_t ram_control_save_page(QEMUFile
*f
, ram_addr_t block_offset
,
513 ram_addr_t offset
, size_t size
, int *bytes_sent
)
515 if (f
->ops
->save_page
) {
516 int ret
= f
->ops
->save_page(f
, f
->opaque
, block_offset
,
517 offset
, size
, bytes_sent
);
519 if (ret
!= RAM_SAVE_CONTROL_DELAYED
) {
520 if (bytes_sent
&& *bytes_sent
> 0) {
521 qemu_update_position(f
, *bytes_sent
);
522 } else if (ret
< 0) {
523 qemu_file_set_error(f
, ret
);
530 return RAM_SAVE_CONTROL_NOT_SUPP
;
534 * Attempt to fill the buffer from the underlying file
535 * Returns the number of bytes read, or negative value for an error.
537 * Note that it can return a partially full buffer even in a not error/not EOF
538 * case if the underlying file descriptor gives a short read, and that can
539 * happen even on a blocking fd.
541 static ssize_t
qemu_fill_buffer(QEMUFile
*f
)
546 assert(!qemu_file_is_writable(f
));
548 pending
= f
->buf_size
- f
->buf_index
;
550 memmove(f
->buf
, f
->buf
+ f
->buf_index
, pending
);
553 f
->buf_size
= pending
;
555 len
= f
->ops
->get_buffer(f
->opaque
, f
->buf
+ pending
, f
->pos
,
556 IO_BUF_SIZE
- pending
);
560 } else if (len
== 0) {
561 qemu_file_set_error(f
, -EIO
);
562 } else if (len
!= -EAGAIN
) {
563 qemu_file_set_error(f
, len
);
569 int qemu_get_fd(QEMUFile
*f
)
571 if (f
->ops
->get_fd
) {
572 return f
->ops
->get_fd(f
->opaque
);
577 void qemu_update_position(QEMUFile
*f
, size_t size
)
584 * Returns negative error value if any error happened on previous operations or
585 * while closing the file. Returns 0 or positive number on success.
587 * The meaning of return value on success depends on the specific backend
590 int qemu_fclose(QEMUFile
*f
)
594 ret
= qemu_file_get_error(f
);
597 int ret2
= f
->ops
->close(f
->opaque
);
602 /* If any error was spotted before closing, we should report it
603 * instead of the close() return value.
609 trace_qemu_file_fclose();
613 static void add_to_iovec(QEMUFile
*f
, const uint8_t *buf
, int size
)
615 /* check for adjacent buffer and coalesce them */
616 if (f
->iovcnt
> 0 && buf
== f
->iov
[f
->iovcnt
- 1].iov_base
+
617 f
->iov
[f
->iovcnt
- 1].iov_len
) {
618 f
->iov
[f
->iovcnt
- 1].iov_len
+= size
;
620 f
->iov
[f
->iovcnt
].iov_base
= (uint8_t *)buf
;
621 f
->iov
[f
->iovcnt
++].iov_len
= size
;
624 if (f
->iovcnt
>= MAX_IOV_SIZE
) {
629 void qemu_put_buffer_async(QEMUFile
*f
, const uint8_t *buf
, int size
)
631 if (!f
->ops
->writev_buffer
) {
632 qemu_put_buffer(f
, buf
, size
);
640 f
->bytes_xfer
+= size
;
641 add_to_iovec(f
, buf
, size
);
644 void qemu_put_buffer(QEMUFile
*f
, const uint8_t *buf
, int size
)
653 l
= IO_BUF_SIZE
- f
->buf_index
;
657 memcpy(f
->buf
+ f
->buf_index
, buf
, l
);
659 if (f
->ops
->writev_buffer
) {
660 add_to_iovec(f
, f
->buf
+ f
->buf_index
, l
);
663 if (f
->buf_index
== IO_BUF_SIZE
) {
666 if (qemu_file_get_error(f
)) {
674 void qemu_put_byte(QEMUFile
*f
, int v
)
680 f
->buf
[f
->buf_index
] = v
;
682 if (f
->ops
->writev_buffer
) {
683 add_to_iovec(f
, f
->buf
+ f
->buf_index
, 1);
686 if (f
->buf_index
== IO_BUF_SIZE
) {
691 void qemu_file_skip(QEMUFile
*f
, int size
)
693 if (f
->buf_index
+ size
<= f
->buf_size
) {
694 f
->buf_index
+= size
;
699 * Read 'size' bytes from file (at 'offset') into buf without moving the
702 * It will return size bytes unless there was an error, in which case it will
703 * return as many as it managed to read (assuming blocking fd's which
704 * all current QEMUFile are)
706 int qemu_peek_buffer(QEMUFile
*f
, uint8_t *buf
, int size
, size_t offset
)
711 assert(!qemu_file_is_writable(f
));
712 assert(offset
< IO_BUF_SIZE
);
713 assert(size
<= IO_BUF_SIZE
- offset
);
715 /* The 1st byte to read from */
716 index
= f
->buf_index
+ offset
;
717 /* The number of available bytes starting at index */
718 pending
= f
->buf_size
- index
;
721 * qemu_fill_buffer might return just a few bytes, even when there isn't
722 * an error, so loop collecting them until we get enough.
724 while (pending
< size
) {
725 int received
= qemu_fill_buffer(f
);
731 index
= f
->buf_index
+ offset
;
732 pending
= f
->buf_size
- index
;
738 if (size
> pending
) {
742 memcpy(buf
, f
->buf
+ index
, size
);
747 * Read 'size' bytes of data from the file into buf.
748 * 'size' can be larger than the internal buffer.
750 * It will return size bytes unless there was an error, in which case it will
751 * return as many as it managed to read (assuming blocking fd's which
752 * all current QEMUFile are)
754 int qemu_get_buffer(QEMUFile
*f
, uint8_t *buf
, int size
)
759 while (pending
> 0) {
762 res
= qemu_peek_buffer(f
, buf
, MIN(pending
, IO_BUF_SIZE
), 0);
766 qemu_file_skip(f
, res
);
775 * Peeks a single byte from the buffer; this isn't guaranteed to work if
776 * offset leaves a gap after the previous read/peeked data.
778 int qemu_peek_byte(QEMUFile
*f
, int offset
)
780 int index
= f
->buf_index
+ offset
;
782 assert(!qemu_file_is_writable(f
));
783 assert(offset
< IO_BUF_SIZE
);
785 if (index
>= f
->buf_size
) {
787 index
= f
->buf_index
+ offset
;
788 if (index
>= f
->buf_size
) {
792 return f
->buf
[index
];
795 int qemu_get_byte(QEMUFile
*f
)
799 result
= qemu_peek_byte(f
, 0);
800 qemu_file_skip(f
, 1);
804 int64_t qemu_ftell(QEMUFile
*f
)
810 int qemu_file_rate_limit(QEMUFile
*f
)
812 if (qemu_file_get_error(f
)) {
815 if (f
->xfer_limit
> 0 && f
->bytes_xfer
> f
->xfer_limit
) {
821 int64_t qemu_file_get_rate_limit(QEMUFile
*f
)
823 return f
->xfer_limit
;
826 void qemu_file_set_rate_limit(QEMUFile
*f
, int64_t limit
)
828 f
->xfer_limit
= limit
;
831 void qemu_file_reset_rate_limit(QEMUFile
*f
)
836 void qemu_put_be16(QEMUFile
*f
, unsigned int v
)
838 qemu_put_byte(f
, v
>> 8);
842 void qemu_put_be32(QEMUFile
*f
, unsigned int v
)
844 qemu_put_byte(f
, v
>> 24);
845 qemu_put_byte(f
, v
>> 16);
846 qemu_put_byte(f
, v
>> 8);
850 void qemu_put_be64(QEMUFile
*f
, uint64_t v
)
852 qemu_put_be32(f
, v
>> 32);
856 unsigned int qemu_get_be16(QEMUFile
*f
)
859 v
= qemu_get_byte(f
) << 8;
860 v
|= qemu_get_byte(f
);
864 unsigned int qemu_get_be32(QEMUFile
*f
)
867 v
= qemu_get_byte(f
) << 24;
868 v
|= qemu_get_byte(f
) << 16;
869 v
|= qemu_get_byte(f
) << 8;
870 v
|= qemu_get_byte(f
);
874 uint64_t qemu_get_be64(QEMUFile
*f
)
877 v
= (uint64_t)qemu_get_be32(f
) << 32;
878 v
|= qemu_get_be32(f
);
882 #define QSB_CHUNK_SIZE (1 << 10)
883 #define QSB_MAX_CHUNK_SIZE (16 * QSB_CHUNK_SIZE)
886 * Create a QEMUSizedBuffer
887 * This type of buffer uses scatter-gather lists internally and
888 * can grow to any size. Any data array in the scatter-gather list
889 * can hold different amount of bytes.
891 * @buffer: Optional buffer to copy into the QSB
892 * @len: size of initial buffer; if @buffer is given, buffer must
893 * hold at least len bytes
895 * Returns a pointer to a QEMUSizedBuffer or NULL on allocation failure
897 QEMUSizedBuffer
*qsb_create(const uint8_t *buffer
, size_t len
)
899 QEMUSizedBuffer
*qsb
;
900 size_t alloc_len
, num_chunks
, i
, to_copy
;
901 size_t chunk_size
= (len
> QSB_MAX_CHUNK_SIZE
)
905 num_chunks
= DIV_ROUND_UP(len
? len
: QSB_CHUNK_SIZE
, chunk_size
);
906 alloc_len
= num_chunks
* chunk_size
;
908 qsb
= g_try_new0(QEMUSizedBuffer
, 1);
913 qsb
->iov
= g_try_new0(struct iovec
, num_chunks
);
919 qsb
->n_iov
= num_chunks
;
921 for (i
= 0; i
< num_chunks
; i
++) {
922 qsb
->iov
[i
].iov_base
= g_try_malloc0(chunk_size
);
923 if (!qsb
->iov
[i
].iov_base
) {
924 /* qsb_free is safe since g_free can cope with NULL */
929 qsb
->iov
[i
].iov_len
= chunk_size
;
931 to_copy
= (len
- qsb
->used
) > chunk_size
932 ? chunk_size
: (len
- qsb
->used
);
933 memcpy(qsb
->iov
[i
].iov_base
, &buffer
[qsb
->used
], to_copy
);
934 qsb
->used
+= to_copy
;
938 qsb
->size
= alloc_len
;
944 * Free the QEMUSizedBuffer
946 * @qsb: The QEMUSizedBuffer to free
948 void qsb_free(QEMUSizedBuffer
*qsb
)
956 for (i
= 0; i
< qsb
->n_iov
; i
++) {
957 g_free(qsb
->iov
[i
].iov_base
);
964 * Get the number of used bytes in the QEMUSizedBuffer
966 * @qsb: A QEMUSizedBuffer
968 * Returns the number of bytes currently used in this buffer
970 size_t qsb_get_length(const QEMUSizedBuffer
*qsb
)
976 * Set the length of the buffer; the primary usage of this
977 * function is to truncate the number of used bytes in the buffer.
978 * The size will not be extended beyond the current number of
979 * allocated bytes in the QEMUSizedBuffer.
981 * @qsb: A QEMUSizedBuffer
982 * @new_len: The new length of bytes in the buffer
984 * Returns the number of bytes the buffer was truncated or extended
987 size_t qsb_set_length(QEMUSizedBuffer
*qsb
, size_t new_len
)
989 if (new_len
<= qsb
->size
) {
992 qsb
->used
= qsb
->size
;
998 * Get the iovec that holds the data for a given position @pos.
1000 * @qsb: A QEMUSizedBuffer
1001 * @pos: The index of a byte in the buffer
1002 * @d_off: Pointer to an offset that this function will indicate
1003 * at what position within the returned iovec the byte
1006 * Returns the index of the iovec that holds the byte at the given
1007 * index @pos in the byte stream; a negative number if the iovec
1008 * for the given position @pos does not exist.
1010 static ssize_t
qsb_get_iovec(const QEMUSizedBuffer
*qsb
,
1011 off_t pos
, off_t
*d_off
)
1016 if (pos
> qsb
->used
) {
1020 for (i
= 0; i
< qsb
->n_iov
; i
++) {
1021 if (curr
+ qsb
->iov
[i
].iov_len
> pos
) {
1022 *d_off
= pos
- curr
;
1025 curr
+= qsb
->iov
[i
].iov_len
;
1031 * Convert the QEMUSizedBuffer into a flat buffer.
1033 * Note: If at all possible, try to avoid this function since it
1034 * may unnecessarily copy memory around.
1036 * @qsb: pointer to QEMUSizedBuffer
1037 * @start: offset to start at
1038 * @count: number of bytes to copy
1039 * @buf: a pointer to a buffer to write into (at least @count bytes)
1041 * Returns the number of bytes copied into the output buffer
1043 ssize_t
qsb_get_buffer(const QEMUSizedBuffer
*qsb
, off_t start
,
1044 size_t count
, uint8_t *buffer
)
1046 const struct iovec
*iov
;
1047 size_t to_copy
, all_copy
;
1053 if (start
> qsb
->used
) {
1057 all_copy
= qsb
->used
- start
;
1058 if (all_copy
> count
) {
1064 index
= qsb_get_iovec(qsb
, start
, &s_off
);
1069 while (all_copy
> 0) {
1070 iov
= &qsb
->iov
[index
];
1074 to_copy
= iov
->iov_len
- s_off
;
1075 if (to_copy
> all_copy
) {
1078 memcpy(&buffer
[d_off
], &s
[s_off
], to_copy
);
1081 all_copy
-= to_copy
;
1091 * Grow the QEMUSizedBuffer to the given size and allocate
1094 * @qsb: A QEMUSizedBuffer
1095 * @new_size: The new size of the buffer
1098 * a negative error code in case of memory allocation failure
1100 * the new size of the buffer. The returned size may be greater or equal
1103 static ssize_t
qsb_grow(QEMUSizedBuffer
*qsb
, size_t new_size
)
1105 size_t needed_chunks
, i
;
1107 if (qsb
->size
< new_size
) {
1108 struct iovec
*new_iov
;
1109 size_t size_diff
= new_size
- qsb
->size
;
1110 size_t chunk_size
= (size_diff
> QSB_MAX_CHUNK_SIZE
)
1111 ? QSB_MAX_CHUNK_SIZE
: QSB_CHUNK_SIZE
;
1113 needed_chunks
= DIV_ROUND_UP(size_diff
, chunk_size
);
1115 new_iov
= g_try_new(struct iovec
, qsb
->n_iov
+ needed_chunks
);
1116 if (new_iov
== NULL
) {
1120 /* Allocate new chunks as needed into new_iov */
1121 for (i
= qsb
->n_iov
; i
< qsb
->n_iov
+ needed_chunks
; i
++) {
1122 new_iov
[i
].iov_base
= g_try_malloc0(chunk_size
);
1123 new_iov
[i
].iov_len
= chunk_size
;
1124 if (!new_iov
[i
].iov_base
) {
1127 /* Free previously allocated new chunks */
1128 for (j
= qsb
->n_iov
; j
< i
; j
++) {
1129 g_free(new_iov
[j
].iov_base
);
1138 * Now we can't get any allocation errors, copy over to new iov
1141 for (i
= 0; i
< qsb
->n_iov
; i
++) {
1142 new_iov
[i
] = qsb
->iov
[i
];
1145 qsb
->n_iov
+= needed_chunks
;
1148 qsb
->size
+= (needed_chunks
* chunk_size
);
1155 * Write into the QEMUSizedBuffer at a given position and a given
1156 * number of bytes. This function will automatically grow the
1159 * @qsb: A QEMUSizedBuffer
1160 * @source: A byte array to copy data from
1161 * @pos: The position within the @qsb to write data to
1162 * @size: The number of bytes to copy into the @qsb
1164 * Returns @size or a negative error code in case of memory allocation failure,
1165 * or with an invalid 'pos'
1167 ssize_t
qsb_write_at(QEMUSizedBuffer
*qsb
, const uint8_t *source
,
1168 off_t pos
, size_t count
)
1170 ssize_t rc
= qsb_grow(qsb
, pos
+ count
);
1172 size_t all_copy
= count
;
1173 const struct iovec
*iov
;
1176 off_t d_off
, s_off
= 0;
1182 if (pos
+ count
> qsb
->used
) {
1183 qsb
->used
= pos
+ count
;
1186 index
= qsb_get_iovec(qsb
, pos
, &d_off
);
1191 while (all_copy
> 0) {
1192 iov
= &qsb
->iov
[index
];
1194 dest
= iov
->iov_base
;
1196 to_copy
= iov
->iov_len
- d_off
;
1197 if (to_copy
> all_copy
) {
1201 memcpy(&dest
[d_off
], &source
[s_off
], to_copy
);
1204 all_copy
-= to_copy
;
1214 * Create a deep copy of the given QEMUSizedBuffer.
1216 * @qsb: A QEMUSizedBuffer
1218 * Returns a clone of @qsb or NULL on allocation failure
1220 QEMUSizedBuffer
*qsb_clone(const QEMUSizedBuffer
*qsb
)
1222 QEMUSizedBuffer
*out
= qsb_create(NULL
, qsb_get_length(qsb
));
1231 for (i
= 0; i
< qsb
->n_iov
; i
++) {
1232 res
= qsb_write_at(out
, qsb
->iov
[i
].iov_base
,
1233 pos
, qsb
->iov
[i
].iov_len
);
1244 typedef struct QEMUBuffer
{
1245 QEMUSizedBuffer
*qsb
;
1249 static int buf_get_buffer(void *opaque
, uint8_t *buf
, int64_t pos
, int size
)
1251 QEMUBuffer
*s
= opaque
;
1252 ssize_t len
= qsb_get_length(s
->qsb
) - pos
;
1261 return qsb_get_buffer(s
->qsb
, pos
, len
, buf
);
1264 static int buf_put_buffer(void *opaque
, const uint8_t *buf
,
1265 int64_t pos
, int size
)
1267 QEMUBuffer
*s
= opaque
;
1269 return qsb_write_at(s
->qsb
, buf
, pos
, size
);
1272 static int buf_close(void *opaque
)
1274 QEMUBuffer
*s
= opaque
;
1283 const QEMUSizedBuffer
*qemu_buf_get(QEMUFile
*f
)
1294 static const QEMUFileOps buf_read_ops
= {
1295 .get_buffer
= buf_get_buffer
,
1299 static const QEMUFileOps buf_write_ops
= {
1300 .put_buffer
= buf_put_buffer
,
1304 QEMUFile
*qemu_bufopen(const char *mode
, QEMUSizedBuffer
*input
)
1308 if (mode
== NULL
|| (mode
[0] != 'r' && mode
[0] != 'w') ||
1310 error_report("qemu_bufopen: Argument validity check failed");
1314 s
= g_malloc0(sizeof(QEMUBuffer
));
1315 if (mode
[0] == 'r') {
1319 if (s
->qsb
== NULL
) {
1320 s
->qsb
= qsb_create(NULL
, 0);
1324 error_report("qemu_bufopen: qsb_create failed");
1329 if (mode
[0] == 'r') {
1330 s
->file
= qemu_fopen_ops(s
, &buf_read_ops
);
1332 s
->file
= qemu_fopen_ops(s
, &buf_write_ops
);