2 * Block driver for the VMDK format
4 * Copyright (c) 2004 Fabrice Bellard
5 * Copyright (c) 2005 Filip Navara
7 * Permission is hereby granted, free of charge, to any person obtaining a copy
8 * of this software and associated documentation files (the "Software"), to deal
9 * in the Software without restriction, including without limitation the rights
10 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
11 * copies of the Software, and to permit persons to whom the Software is
12 * furnished to do so, subject to the following conditions:
14 * The above copyright notice and this permission notice shall be included in
15 * all copies or substantial portions of the Software.
17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
18 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
19 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
20 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
21 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
22 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
27 #include "block_int.h"
29 #define VMDK3_MAGIC (('C' << 24) | ('O' << 16) | ('W' << 8) | 'D')
30 #define VMDK4_MAGIC (('K' << 24) | ('D' << 16) | ('M' << 8) | 'V')
35 uint32_t disk_sectors
;
37 uint32_t l1dir_offset
;
39 uint32_t file_sectors
;
42 uint32_t sectors_per_track
;
52 int32_t num_gtes_per_gte
;
58 } __attribute__((packed
)) VMDK4Header
;
60 #define L2_CACHE_SIZE 16
62 typedef struct BDRVVmdkState
{
64 int64_t l1_table_offset
;
65 int64_t l1_backup_table_offset
;
67 uint32_t *l1_backup_table
;
69 uint32_t l1_entry_sectors
;
73 uint32_t l2_cache_offsets
[L2_CACHE_SIZE
];
74 uint32_t l2_cache_counts
[L2_CACHE_SIZE
];
76 unsigned int cluster_sectors
;
80 static int vmdk_probe(const uint8_t *buf
, int buf_size
, const char *filename
)
86 magic
= be32_to_cpu(*(uint32_t *)buf
);
87 if (magic
== VMDK3_MAGIC
||
96 #define SECTOR_SIZE 512
97 #define DESC_SIZE 20*SECTOR_SIZE // 20 sectors of 512 bytes each
98 #define HEADER_SIZE 512 // first sector of 512 bytes
100 static uint32_t vmdk_read_cid(BlockDriverState
*bs
, int parent
)
102 BDRVVmdkState
*s
= bs
->opaque
;
103 char desc
[DESC_SIZE
];
105 char *p_name
, *cid_str
;
108 /* the descriptor offset = 0x200 */
109 if (bdrv_pread(s
->hd
, 0x200, desc
, DESC_SIZE
) != DESC_SIZE
)
113 cid_str
= "parentCID";
114 cid_str_size
= sizeof("parentCID");
117 cid_str_size
= sizeof("CID");
120 if ((p_name
= strstr(desc
,cid_str
)) != 0) {
121 p_name
+= cid_str_size
;
122 sscanf(p_name
,"%x",&cid
);
128 static int vmdk_write_cid(BlockDriverState
*bs
, uint32_t cid
)
130 BDRVVmdkState
*s
= bs
->opaque
;
131 char desc
[DESC_SIZE
], tmp_desc
[DESC_SIZE
];
132 char *p_name
, *tmp_str
;
134 /* the descriptor offset = 0x200 */
135 if (bdrv_pread(s
->hd
, 0x200, desc
, DESC_SIZE
) != DESC_SIZE
)
138 tmp_str
= strstr(desc
,"parentCID");
139 strcpy(tmp_desc
, tmp_str
);
140 if ((p_name
= strstr(desc
,"CID")) != 0) {
141 p_name
+= sizeof("CID");
142 sprintf(p_name
,"%x\n",cid
);
143 strcat(desc
,tmp_desc
);
146 if (bdrv_pwrite(s
->hd
, 0x200, desc
, DESC_SIZE
) != DESC_SIZE
)
151 static int vmdk_is_cid_valid(BlockDriverState
*bs
)
154 BDRVVmdkState
*s
= bs
->opaque
;
155 BlockDriverState
*p_bs
= s
->hd
->backing_hd
;
159 cur_pcid
= vmdk_read_cid(p_bs
,0);
160 if (s
->parent_cid
!= cur_pcid
)
169 static int vmdk_snapshot_create(const char *filename
, const char *backing_file
)
173 char *p_name
, *gd_buf
, *rgd_buf
;
174 const char *real_filename
, *temp_str
;
176 uint32_t gde_entries
, gd_size
;
177 int64_t gd_offset
, rgd_offset
, capacity
, gt_size
;
178 char p_desc
[DESC_SIZE
], s_desc
[DESC_SIZE
], hdr
[HEADER_SIZE
];
179 char *desc_template
=
180 "# Disk DescriptorFile\n"
184 "createType=\"monolithicSparse\"\n"
185 "parentFileNameHint=\"%s\"\n"
187 "# Extent description\n"
188 "RW %lu SPARSE \"%s\"\n"
190 "# The Disk Data Base \n"
194 snp_fd
= open(filename
, O_RDWR
| O_CREAT
| O_TRUNC
| O_BINARY
| O_LARGEFILE
, 0644);
197 p_fd
= open(backing_file
, O_RDONLY
| O_BINARY
| O_LARGEFILE
);
203 /* read the header */
204 if (lseek(p_fd
, 0x0, SEEK_SET
) == -1)
206 if (read(p_fd
, hdr
, HEADER_SIZE
) != HEADER_SIZE
)
209 /* write the header */
210 if (lseek(snp_fd
, 0x0, SEEK_SET
) == -1)
212 if (write(snp_fd
, hdr
, HEADER_SIZE
) == -1)
215 memset(&header
, 0, sizeof(header
));
216 memcpy(&header
,&hdr
[4], sizeof(header
)); // skip the VMDK4_MAGIC
218 ftruncate(snp_fd
, header
.grain_offset
<< 9);
219 /* the descriptor offset = 0x200 */
220 if (lseek(p_fd
, 0x200, SEEK_SET
) == -1)
222 if (read(p_fd
, p_desc
, DESC_SIZE
) != DESC_SIZE
)
225 if ((p_name
= strstr(p_desc
,"CID")) != 0) {
226 p_name
+= sizeof("CID");
227 sscanf(p_name
,"%x",&p_cid
);
230 real_filename
= filename
;
231 if ((temp_str
= strrchr(real_filename
, '\\')) != NULL
)
232 real_filename
= temp_str
+ 1;
233 if ((temp_str
= strrchr(real_filename
, '/')) != NULL
)
234 real_filename
= temp_str
+ 1;
235 if ((temp_str
= strrchr(real_filename
, ':')) != NULL
)
236 real_filename
= temp_str
+ 1;
238 sprintf(s_desc
, desc_template
, p_cid
, p_cid
, backing_file
239 , (uint32_t)header
.capacity
, real_filename
);
241 /* write the descriptor */
242 if (lseek(snp_fd
, 0x200, SEEK_SET
) == -1)
244 if (write(snp_fd
, s_desc
, strlen(s_desc
)) == -1)
247 gd_offset
= header
.gd_offset
* SECTOR_SIZE
; // offset of GD table
248 rgd_offset
= header
.rgd_offset
* SECTOR_SIZE
; // offset of RGD table
249 capacity
= header
.capacity
* SECTOR_SIZE
; // Extent size
251 * Each GDE span 32M disk, means:
252 * 512 GTE per GT, each GTE points to grain
254 gt_size
= (int64_t)header
.num_gtes_per_gte
* header
.granularity
* SECTOR_SIZE
;
257 gde_entries
= (uint32_t)(capacity
/ gt_size
); // number of gde/rgde
258 gd_size
= gde_entries
* sizeof(uint32_t);
261 rgd_buf
= qemu_malloc(gd_size
);
264 if (lseek(p_fd
, rgd_offset
, SEEK_SET
) == -1)
266 if (read(p_fd
, rgd_buf
, gd_size
) != gd_size
)
268 if (lseek(snp_fd
, rgd_offset
, SEEK_SET
) == -1)
270 if (write(snp_fd
, rgd_buf
, gd_size
) == -1)
275 gd_buf
= qemu_malloc(gd_size
);
278 if (lseek(p_fd
, gd_offset
, SEEK_SET
) == -1)
280 if (read(p_fd
, gd_buf
, gd_size
) != gd_size
)
282 if (lseek(snp_fd
, gd_offset
, SEEK_SET
) == -1)
284 if (write(snp_fd
, gd_buf
, gd_size
) == -1)
302 static void vmdk_parent_close(BlockDriverState
*bs
)
305 bdrv_close(bs
->backing_hd
);
309 static int vmdk_parent_open(BlockDriverState
*bs
, const char * filename
)
311 BDRVVmdkState
*s
= bs
->opaque
;
313 char desc
[DESC_SIZE
];
314 char parent_img_name
[1024];
316 /* the descriptor offset = 0x200 */
317 if (bdrv_pread(s
->hd
, 0x200, desc
, DESC_SIZE
) != DESC_SIZE
)
320 if ((p_name
= strstr(desc
,"parentFileNameHint")) != 0) {
322 struct stat file_buf
;
324 p_name
+= sizeof("parentFileNameHint") + 1;
325 if ((end_name
= strchr(p_name
,'\"')) == 0)
328 strncpy(s
->hd
->backing_file
, p_name
, end_name
- p_name
);
329 if (stat(s
->hd
->backing_file
, &file_buf
) != 0) {
330 path_combine(parent_img_name
, sizeof(parent_img_name
),
331 filename
, s
->hd
->backing_file
);
333 strcpy(parent_img_name
, s
->hd
->backing_file
);
336 s
->hd
->backing_hd
= bdrv_new("");
337 if (!s
->hd
->backing_hd
) {
342 if (bdrv_open(s
->hd
->backing_hd
, parent_img_name
, 0) < 0)
349 static int vmdk_open(BlockDriverState
*bs
, const char *filename
, int flags
)
351 BDRVVmdkState
*s
= bs
->opaque
;
355 ret
= bdrv_file_open(&s
->hd
, filename
, flags
);
358 if (bdrv_pread(s
->hd
, 0, &magic
, sizeof(magic
)) != sizeof(magic
))
361 magic
= be32_to_cpu(magic
);
362 if (magic
== VMDK3_MAGIC
) {
365 if (bdrv_pread(s
->hd
, sizeof(magic
), &header
, sizeof(header
)) != sizeof(header
))
367 s
->cluster_sectors
= le32_to_cpu(header
.granularity
);
370 bs
->total_sectors
= le32_to_cpu(header
.disk_sectors
);
371 s
->l1_table_offset
= le32_to_cpu(header
.l1dir_offset
) << 9;
372 s
->l1_backup_table_offset
= 0;
373 s
->l1_entry_sectors
= s
->l2_size
* s
->cluster_sectors
;
374 } else if (magic
== VMDK4_MAGIC
) {
377 if (bdrv_pread(s
->hd
, sizeof(magic
), &header
, sizeof(header
)) != sizeof(header
))
379 bs
->total_sectors
= le64_to_cpu(header
.capacity
);
380 s
->cluster_sectors
= le64_to_cpu(header
.granularity
);
381 s
->l2_size
= le32_to_cpu(header
.num_gtes_per_gte
);
382 s
->l1_entry_sectors
= s
->l2_size
* s
->cluster_sectors
;
383 if (s
->l1_entry_sectors
<= 0)
385 s
->l1_size
= (bs
->total_sectors
+ s
->l1_entry_sectors
- 1)
386 / s
->l1_entry_sectors
;
387 s
->l1_table_offset
= le64_to_cpu(header
.rgd_offset
) << 9;
388 s
->l1_backup_table_offset
= le64_to_cpu(header
.gd_offset
) << 9;
390 // try to open parent images, if exist
391 if (vmdk_parent_open(bs
, filename
) != 0)
393 // write the CID once after the image creation
394 s
->parent_cid
= vmdk_read_cid(bs
,1);
399 /* read the L1 table */
400 l1_size
= s
->l1_size
* sizeof(uint32_t);
401 s
->l1_table
= qemu_malloc(l1_size
);
404 if (bdrv_pread(s
->hd
, s
->l1_table_offset
, s
->l1_table
, l1_size
) != l1_size
)
406 for(i
= 0; i
< s
->l1_size
; i
++) {
407 le32_to_cpus(&s
->l1_table
[i
]);
410 if (s
->l1_backup_table_offset
) {
411 s
->l1_backup_table
= qemu_malloc(l1_size
);
412 if (!s
->l1_backup_table
)
414 if (bdrv_pread(s
->hd
, s
->l1_backup_table_offset
, s
->l1_backup_table
, l1_size
) != l1_size
)
416 for(i
= 0; i
< s
->l1_size
; i
++) {
417 le32_to_cpus(&s
->l1_backup_table
[i
]);
421 s
->l2_cache
= qemu_malloc(s
->l2_size
* L2_CACHE_SIZE
* sizeof(uint32_t));
426 qemu_free(s
->l1_backup_table
);
427 qemu_free(s
->l1_table
);
428 qemu_free(s
->l2_cache
);
433 static uint64_t get_cluster_offset(BlockDriverState
*bs
, uint64_t offset
, int allocate
);
435 static int get_whole_cluster(BlockDriverState
*bs
, uint64_t cluster_offset
,
436 uint64_t offset
, int allocate
)
438 uint64_t parent_cluster_offset
;
439 BDRVVmdkState
*s
= bs
->opaque
;
440 uint8_t whole_grain
[s
->cluster_sectors
*512]; // 128 sectors * 512 bytes each = grain size 64KB
442 // we will be here if it's first write on non-exist grain(cluster).
443 // try to read from parent image, if exist
444 if (s
->hd
->backing_hd
) {
445 BDRVVmdkState
*ps
= s
->hd
->backing_hd
->opaque
;
447 if (!vmdk_is_cid_valid(bs
))
449 parent_cluster_offset
= get_cluster_offset(s
->hd
->backing_hd
, offset
, allocate
);
450 if (bdrv_pread(ps
->hd
, parent_cluster_offset
, whole_grain
, ps
->cluster_sectors
*512) !=
451 ps
->cluster_sectors
*512)
454 if (bdrv_pwrite(s
->hd
, cluster_offset
<< 9, whole_grain
, sizeof(whole_grain
)) !=
461 static uint64_t get_cluster_offset(BlockDriverState
*bs
,
462 uint64_t offset
, int allocate
)
464 BDRVVmdkState
*s
= bs
->opaque
;
465 unsigned int l1_index
, l2_offset
, l2_index
;
467 uint32_t min_count
, *l2_table
, tmp
;
468 uint64_t cluster_offset
;
470 l1_index
= (offset
>> 9) / s
->l1_entry_sectors
;
471 if (l1_index
>= s
->l1_size
)
473 l2_offset
= s
->l1_table
[l1_index
];
476 for(i
= 0; i
< L2_CACHE_SIZE
; i
++) {
477 if (l2_offset
== s
->l2_cache_offsets
[i
]) {
478 /* increment the hit count */
479 if (++s
->l2_cache_counts
[i
] == 0xffffffff) {
480 for(j
= 0; j
< L2_CACHE_SIZE
; j
++) {
481 s
->l2_cache_counts
[j
] >>= 1;
484 l2_table
= s
->l2_cache
+ (i
* s
->l2_size
);
488 /* not found: load a new entry in the least used one */
490 min_count
= 0xffffffff;
491 for(i
= 0; i
< L2_CACHE_SIZE
; i
++) {
492 if (s
->l2_cache_counts
[i
] < min_count
) {
493 min_count
= s
->l2_cache_counts
[i
];
497 l2_table
= s
->l2_cache
+ (min_index
* s
->l2_size
);
498 if (bdrv_pread(s
->hd
, (int64_t)l2_offset
* 512, l2_table
, s
->l2_size
* sizeof(uint32_t)) !=
499 s
->l2_size
* sizeof(uint32_t))
502 s
->l2_cache_offsets
[min_index
] = l2_offset
;
503 s
->l2_cache_counts
[min_index
] = 1;
505 l2_index
= ((offset
>> 9) / s
->cluster_sectors
) % s
->l2_size
;
506 cluster_offset
= le32_to_cpu(l2_table
[l2_index
]);
507 if (!cluster_offset
) {
508 struct stat file_buf
;
512 stat(s
->hd
->filename
, &file_buf
);
513 cluster_offset
= file_buf
.st_size
;
514 bdrv_truncate(s
->hd
, cluster_offset
+ (s
->cluster_sectors
<< 9));
516 cluster_offset
>>= 9;
517 /* update L2 table */
518 tmp
= cpu_to_le32(cluster_offset
);
519 l2_table
[l2_index
] = tmp
;
520 if (bdrv_pwrite(s
->hd
, ((int64_t)l2_offset
* 512) + (l2_index
* sizeof(tmp
)),
521 &tmp
, sizeof(tmp
)) != sizeof(tmp
))
523 /* update backup L2 table */
524 if (s
->l1_backup_table_offset
!= 0) {
525 l2_offset
= s
->l1_backup_table
[l1_index
];
526 if (bdrv_pwrite(s
->hd
, ((int64_t)l2_offset
* 512) + (l2_index
* sizeof(tmp
)),
527 &tmp
, sizeof(tmp
)) != sizeof(tmp
))
531 if (get_whole_cluster(bs
, cluster_offset
, offset
, allocate
) == -1)
534 cluster_offset
<<= 9;
535 return cluster_offset
;
538 static int vmdk_is_allocated(BlockDriverState
*bs
, int64_t sector_num
,
539 int nb_sectors
, int *pnum
)
541 BDRVVmdkState
*s
= bs
->opaque
;
542 int index_in_cluster
, n
;
543 uint64_t cluster_offset
;
545 cluster_offset
= get_cluster_offset(bs
, sector_num
<< 9, 0);
546 index_in_cluster
= sector_num
% s
->cluster_sectors
;
547 n
= s
->cluster_sectors
- index_in_cluster
;
551 return (cluster_offset
!= 0);
554 static int vmdk_read(BlockDriverState
*bs
, int64_t sector_num
,
555 uint8_t *buf
, int nb_sectors
)
557 BDRVVmdkState
*s
= bs
->opaque
;
558 int index_in_cluster
, n
, ret
;
559 uint64_t cluster_offset
;
561 while (nb_sectors
> 0) {
562 cluster_offset
= get_cluster_offset(bs
, sector_num
<< 9, 0);
563 index_in_cluster
= sector_num
% s
->cluster_sectors
;
564 n
= s
->cluster_sectors
- index_in_cluster
;
567 if (!cluster_offset
) {
568 // try to read from parent image, if exist
569 if (s
->hd
->backing_hd
) {
570 if (!vmdk_is_cid_valid(bs
))
572 ret
= bdrv_read(s
->hd
->backing_hd
, sector_num
, buf
, n
);
576 memset(buf
, 0, 512 * n
);
579 if(bdrv_pread(s
->hd
, cluster_offset
+ index_in_cluster
* 512, buf
, n
* 512) != n
* 512)
589 static int vmdk_write(BlockDriverState
*bs
, int64_t sector_num
,
590 const uint8_t *buf
, int nb_sectors
)
592 BDRVVmdkState
*s
= bs
->opaque
;
593 int index_in_cluster
, n
;
594 uint64_t cluster_offset
;
595 static int cid_update
= 0;
597 while (nb_sectors
> 0) {
598 index_in_cluster
= sector_num
& (s
->cluster_sectors
- 1);
599 n
= s
->cluster_sectors
- index_in_cluster
;
602 cluster_offset
= get_cluster_offset(bs
, sector_num
<< 9, 1);
605 if (bdrv_pwrite(s
->hd
, cluster_offset
+ index_in_cluster
* 512, buf
, n
* 512) != n
* 512)
611 // update CID on the first write every time the virtual disk is opened
613 vmdk_write_cid(bs
, time(NULL
));
620 static int vmdk_create(const char *filename
, int64_t total_size
,
621 const char *backing_file
, int flags
)
625 uint32_t tmp
, magic
, grains
, gd_size
, gt_size
, gt_count
;
626 char *desc_template
=
627 "# Disk DescriptorFile\n"
630 "parentCID=ffffffff\n"
631 "createType=\"monolithicSparse\"\n"
633 "# Extent description\n"
634 "RW %lu SPARSE \"%s\"\n"
636 "# The Disk Data Base \n"
639 "ddb.virtualHWVersion = \"4\"\n"
640 "ddb.geometry.cylinders = \"%lu\"\n"
641 "ddb.geometry.heads = \"16\"\n"
642 "ddb.geometry.sectors = \"63\"\n"
643 "ddb.adapterType = \"ide\"\n";
645 const char *real_filename
, *temp_str
;
647 /* XXX: add support for backing file */
649 return vmdk_snapshot_create(filename
, backing_file
);
652 fd
= open(filename
, O_WRONLY
| O_CREAT
| O_TRUNC
| O_BINARY
| O_LARGEFILE
,
656 magic
= cpu_to_be32(VMDK4_MAGIC
);
657 memset(&header
, 0, sizeof(header
));
658 header
.version
= cpu_to_le32(1);
659 header
.flags
= cpu_to_le32(3); /* ?? */
660 header
.capacity
= cpu_to_le64(total_size
);
661 header
.granularity
= cpu_to_le64(128);
662 header
.num_gtes_per_gte
= cpu_to_le32(512);
664 grains
= (total_size
+ header
.granularity
- 1) / header
.granularity
;
665 gt_size
= ((header
.num_gtes_per_gte
* sizeof(uint32_t)) + 511) >> 9;
666 gt_count
= (grains
+ header
.num_gtes_per_gte
- 1) / header
.num_gtes_per_gte
;
667 gd_size
= (gt_count
* sizeof(uint32_t) + 511) >> 9;
669 header
.desc_offset
= 1;
670 header
.desc_size
= 20;
671 header
.rgd_offset
= header
.desc_offset
+ header
.desc_size
;
672 header
.gd_offset
= header
.rgd_offset
+ gd_size
+ (gt_size
* gt_count
);
673 header
.grain_offset
=
674 ((header
.gd_offset
+ gd_size
+ (gt_size
* gt_count
) +
675 header
.granularity
- 1) / header
.granularity
) *
678 header
.desc_offset
= cpu_to_le64(header
.desc_offset
);
679 header
.desc_size
= cpu_to_le64(header
.desc_size
);
680 header
.rgd_offset
= cpu_to_le64(header
.rgd_offset
);
681 header
.gd_offset
= cpu_to_le64(header
.gd_offset
);
682 header
.grain_offset
= cpu_to_le64(header
.grain_offset
);
684 header
.check_bytes
[0] = 0xa;
685 header
.check_bytes
[1] = 0x20;
686 header
.check_bytes
[2] = 0xd;
687 header
.check_bytes
[3] = 0xa;
689 /* write all the data */
690 write(fd
, &magic
, sizeof(magic
));
691 write(fd
, &header
, sizeof(header
));
693 ftruncate(fd
, header
.grain_offset
<< 9);
695 /* write grain directory */
696 lseek(fd
, le64_to_cpu(header
.rgd_offset
) << 9, SEEK_SET
);
697 for (i
= 0, tmp
= header
.rgd_offset
+ gd_size
;
698 i
< gt_count
; i
++, tmp
+= gt_size
)
699 write(fd
, &tmp
, sizeof(tmp
));
701 /* write backup grain directory */
702 lseek(fd
, le64_to_cpu(header
.gd_offset
) << 9, SEEK_SET
);
703 for (i
= 0, tmp
= header
.gd_offset
+ gd_size
;
704 i
< gt_count
; i
++, tmp
+= gt_size
)
705 write(fd
, &tmp
, sizeof(tmp
));
707 /* compose the descriptor */
708 real_filename
= filename
;
709 if ((temp_str
= strrchr(real_filename
, '\\')) != NULL
)
710 real_filename
= temp_str
+ 1;
711 if ((temp_str
= strrchr(real_filename
, '/')) != NULL
)
712 real_filename
= temp_str
+ 1;
713 if ((temp_str
= strrchr(real_filename
, ':')) != NULL
)
714 real_filename
= temp_str
+ 1;
715 sprintf(desc
, desc_template
, time(NULL
), (unsigned long)total_size
,
716 real_filename
, total_size
/ (63 * 16));
718 /* write the descriptor */
719 lseek(fd
, le64_to_cpu(header
.desc_offset
) << 9, SEEK_SET
);
720 write(fd
, desc
, strlen(desc
));
726 static void vmdk_close(BlockDriverState
*bs
)
728 BDRVVmdkState
*s
= bs
->opaque
;
730 qemu_free(s
->l1_table
);
731 qemu_free(s
->l2_cache
);
733 // try to close parent image, if exist
734 vmdk_parent_close(s
->hd
);
737 static void vmdk_flush(BlockDriverState
*bs
)
739 BDRVVmdkState
*s
= bs
->opaque
;
743 BlockDriver bdrv_vmdk
= {
745 sizeof(BDRVVmdkState
),