16740 libtopo: 'psn' may be used uninitialized
[illumos-gate.git] / usr / src / uts / common / fs / zfs / dmu_send.c
blobb7135df3fa95592f1e9171cfa1f72a533d0b01a1
1 /*
2 * CDDL HEADER START
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
19 * CDDL HEADER END
22 * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved.
23 * Copyright 2011 Nexenta Systems, Inc. All rights reserved.
24 * Copyright (c) 2011, 2015 by Delphix. All rights reserved.
25 * Copyright (c) 2014, Joyent, Inc. All rights reserved.
26 * Copyright 2014 HybridCluster. All rights reserved.
27 * Copyright 2016 RackTop Systems.
28 * Copyright (c) 2014 Integros [integros.com]
31 #include <sys/dmu.h>
32 #include <sys/dmu_impl.h>
33 #include <sys/dmu_tx.h>
34 #include <sys/dbuf.h>
35 #include <sys/dnode.h>
36 #include <sys/zfs_context.h>
37 #include <sys/dmu_objset.h>
38 #include <sys/dmu_traverse.h>
39 #include <sys/dsl_dataset.h>
40 #include <sys/dsl_dir.h>
41 #include <sys/dsl_prop.h>
42 #include <sys/dsl_pool.h>
43 #include <sys/dsl_synctask.h>
44 #include <sys/zfs_ioctl.h>
45 #include <sys/zap.h>
46 #include <sys/zio_checksum.h>
47 #include <sys/zfs_znode.h>
48 #include <zfs_fletcher.h>
49 #include <sys/avl.h>
50 #include <sys/ddt.h>
51 #include <sys/zfs_onexit.h>
52 #include <sys/dmu_send.h>
53 #include <sys/dsl_destroy.h>
54 #include <sys/blkptr.h>
55 #include <sys/dsl_bookmark.h>
56 #include <sys/zfeature.h>
57 #include <sys/bqueue.h>
59 /* Set this tunable to TRUE to replace corrupt data with 0x2f5baddb10c */
60 int zfs_send_corrupt_data = B_FALSE;
61 int zfs_send_queue_length = SPA_MAXBLOCKSIZE;
62 /* Set this tunable to FALSE to disable setting of DRR_FLAG_FREERECORDS */
63 int zfs_send_set_freerecords_bit = B_TRUE;
64 /* Set this tunable to FALSE is disable sending unmodified spill blocks. */
65 int zfs_send_unmodified_spill_blocks = B_TRUE;
68 * Use this to override the recordsize calculation for fast zfs send estimates.
70 uint64_t zfs_override_estimate_recordsize = 0;
72 #define BP_SPAN(datablkszsec, indblkshift, level) \
73 (((uint64_t)datablkszsec) << (SPA_MINBLOCKSHIFT + \
74 (level) * (indblkshift - SPA_BLKPTRSHIFT)))
76 struct send_thread_arg {
77 bqueue_t q;
78 dsl_dataset_t *ds; /* Dataset to traverse */
79 uint64_t fromtxg; /* Traverse from this txg */
80 int flags; /* flags to pass to traverse_dataset */
81 int error_code;
82 boolean_t cancel;
83 zbookmark_phys_t resume;
86 struct send_block_record {
87 boolean_t eos_marker; /* Marks the end of the stream */
88 blkptr_t bp;
89 zbookmark_phys_t zb;
90 uint8_t indblkshift;
91 uint16_t datablkszsec;
92 bqueue_node_t ln;
95 static int do_dump(dmu_sendarg_t *dsa, struct send_block_record *data);
97 static int
98 dump_bytes(dmu_sendarg_t *dsp, void *buf, int len)
100 dsl_dataset_t *ds = dmu_objset_ds(dsp->dsa_os);
101 ssize_t resid; /* have to get resid to get detailed errno */
104 * The code does not rely on len being a multiple of 8. We keep
105 * this assertion because of the corresponding assertion in
106 * receive_read(). Keeping this assertion ensures that we do not
107 * inadvertently break backwards compatibility (causing the assertion
108 * in receive_read() to trigger on old software). Newer feature flags
109 * (such as raw send) may break this assertion since they were
110 * introduced after the requirement was made obsolete.
113 ASSERT(len % 8 == 0 ||
114 (dsp->dsa_featureflags & DMU_BACKUP_FEATURE_RAW) != 0);
116 dsp->dsa_err = vn_rdwr(UIO_WRITE, dsp->dsa_vp,
117 (caddr_t)buf, len,
118 0, UIO_SYSSPACE, FAPPEND, RLIM64_INFINITY, CRED(), &resid);
120 mutex_enter(&ds->ds_sendstream_lock);
121 *dsp->dsa_off += len;
122 mutex_exit(&ds->ds_sendstream_lock);
124 return (dsp->dsa_err);
128 * For all record types except BEGIN, fill in the checksum (overlaid in
129 * drr_u.drr_checksum.drr_checksum). The checksum verifies everything
130 * up to the start of the checksum itself.
132 static int
133 dump_record(dmu_sendarg_t *dsp, void *payload, int payload_len)
135 ASSERT3U(offsetof(dmu_replay_record_t, drr_u.drr_checksum.drr_checksum),
136 ==, sizeof (dmu_replay_record_t) - sizeof (zio_cksum_t));
137 (void) fletcher_4_incremental_native(dsp->dsa_drr,
138 offsetof(dmu_replay_record_t, drr_u.drr_checksum.drr_checksum),
139 &dsp->dsa_zc);
140 if (dsp->dsa_drr->drr_type == DRR_BEGIN) {
141 dsp->dsa_sent_begin = B_TRUE;
142 } else {
143 ASSERT(ZIO_CHECKSUM_IS_ZERO(&dsp->dsa_drr->drr_u.
144 drr_checksum.drr_checksum));
145 dsp->dsa_drr->drr_u.drr_checksum.drr_checksum = dsp->dsa_zc;
147 if (dsp->dsa_drr->drr_type == DRR_END) {
148 dsp->dsa_sent_end = B_TRUE;
150 (void) fletcher_4_incremental_native(&dsp->dsa_drr->
151 drr_u.drr_checksum.drr_checksum,
152 sizeof (zio_cksum_t), &dsp->dsa_zc);
153 if (dump_bytes(dsp, dsp->dsa_drr, sizeof (dmu_replay_record_t)) != 0)
154 return (SET_ERROR(EINTR));
155 if (payload_len != 0) {
156 (void) fletcher_4_incremental_native(payload, payload_len,
157 &dsp->dsa_zc);
158 if (dump_bytes(dsp, payload, payload_len) != 0)
159 return (SET_ERROR(EINTR));
161 return (0);
165 * Fill in the drr_free struct, or perform aggregation if the previous record is
166 * also a free record, and the two are adjacent.
168 * Note that we send free records even for a full send, because we want to be
169 * able to receive a full send as a clone, which requires a list of all the free
170 * and freeobject records that were generated on the source.
172 static int
173 dump_free(dmu_sendarg_t *dsp, uint64_t object, uint64_t offset,
174 uint64_t length)
176 struct drr_free *drrf = &(dsp->dsa_drr->drr_u.drr_free);
179 * When we receive a free record, dbuf_free_range() assumes
180 * that the receiving system doesn't have any dbufs in the range
181 * being freed. This is always true because there is a one-record
182 * constraint: we only send one WRITE record for any given
183 * object,offset. We know that the one-record constraint is
184 * true because we always send data in increasing order by
185 * object,offset.
187 * If the increasing-order constraint ever changes, we should find
188 * another way to assert that the one-record constraint is still
189 * satisfied.
191 ASSERT(object > dsp->dsa_last_data_object ||
192 (object == dsp->dsa_last_data_object &&
193 offset > dsp->dsa_last_data_offset));
196 * If there is a pending op, but it's not PENDING_FREE, push it out,
197 * since free block aggregation can only be done for blocks of the
198 * same type (i.e., DRR_FREE records can only be aggregated with
199 * other DRR_FREE records. DRR_FREEOBJECTS records can only be
200 * aggregated with other DRR_FREEOBJECTS records.
202 if (dsp->dsa_pending_op != PENDING_NONE &&
203 dsp->dsa_pending_op != PENDING_FREE) {
204 if (dump_record(dsp, NULL, 0) != 0)
205 return (SET_ERROR(EINTR));
206 dsp->dsa_pending_op = PENDING_NONE;
209 if (dsp->dsa_pending_op == PENDING_FREE) {
211 * There should never be a PENDING_FREE if length is
212 * DMU_OBJECT_END (because dump_dnode is the only place where
213 * this function is called with a DMU_OBJECT_END, and only after
214 * flushing any pending record).
216 ASSERT(length != DMU_OBJECT_END);
218 * Check to see whether this free block can be aggregated
219 * with pending one.
221 if (drrf->drr_object == object && drrf->drr_offset +
222 drrf->drr_length == offset) {
223 if (offset + length < offset)
224 drrf->drr_length = DMU_OBJECT_END;
225 else
226 drrf->drr_length += length;
227 return (0);
228 } else {
229 /* not a continuation. Push out pending record */
230 if (dump_record(dsp, NULL, 0) != 0)
231 return (SET_ERROR(EINTR));
232 dsp->dsa_pending_op = PENDING_NONE;
235 /* create a FREE record and make it pending */
236 bzero(dsp->dsa_drr, sizeof (dmu_replay_record_t));
237 dsp->dsa_drr->drr_type = DRR_FREE;
238 drrf->drr_object = object;
239 drrf->drr_offset = offset;
240 if (offset + length < offset)
241 drrf->drr_length = DMU_OBJECT_END;
242 else
243 drrf->drr_length = length;
244 drrf->drr_toguid = dsp->dsa_toguid;
245 if (length == DMU_OBJECT_END) {
246 if (dump_record(dsp, NULL, 0) != 0)
247 return (SET_ERROR(EINTR));
248 } else {
249 dsp->dsa_pending_op = PENDING_FREE;
252 return (0);
255 static int
256 dump_write(dmu_sendarg_t *dsp, dmu_object_type_t type, uint64_t object,
257 uint64_t offset, int lsize, int psize, const blkptr_t *bp, void *data)
259 uint64_t payload_size;
260 boolean_t raw = (dsp->dsa_featureflags & DMU_BACKUP_FEATURE_RAW);
261 struct drr_write *drrw = &(dsp->dsa_drr->drr_u.drr_write);
264 * We send data in increasing object, offset order.
265 * See comment in dump_free() for details.
267 ASSERT(object > dsp->dsa_last_data_object ||
268 (object == dsp->dsa_last_data_object &&
269 offset > dsp->dsa_last_data_offset));
270 dsp->dsa_last_data_object = object;
271 dsp->dsa_last_data_offset = offset + lsize - 1;
274 * If there is any kind of pending aggregation (currently either
275 * a grouping of free objects or free blocks), push it out to
276 * the stream, since aggregation can't be done across operations
277 * of different types.
279 if (dsp->dsa_pending_op != PENDING_NONE) {
280 if (dump_record(dsp, NULL, 0) != 0)
281 return (SET_ERROR(EINTR));
282 dsp->dsa_pending_op = PENDING_NONE;
284 /* write a WRITE record */
285 bzero(dsp->dsa_drr, sizeof (dmu_replay_record_t));
286 dsp->dsa_drr->drr_type = DRR_WRITE;
287 drrw->drr_object = object;
288 drrw->drr_type = type;
289 drrw->drr_offset = offset;
290 drrw->drr_toguid = dsp->dsa_toguid;
291 drrw->drr_logical_size = lsize;
293 /* only set the compression fields if the buf is compressed or raw */
294 if (raw || lsize != psize) {
295 ASSERT(!BP_IS_EMBEDDED(bp));
296 ASSERT3S(psize, >, 0);
298 if (raw) {
299 ASSERT(BP_IS_PROTECTED(bp));
302 * This is a raw protected block so we need to pass
303 * along everything the receiving side will need to
304 * interpret this block, including the byteswap, salt,
305 * IV, and MAC.
307 if (BP_SHOULD_BYTESWAP(bp))
308 drrw->drr_flags |= DRR_RAW_BYTESWAP;
309 zio_crypt_decode_params_bp(bp, drrw->drr_salt,
310 drrw->drr_iv);
311 zio_crypt_decode_mac_bp(bp, drrw->drr_mac);
312 } else {
313 /* this is a compressed block */
314 ASSERT(dsp->dsa_featureflags &
315 DMU_BACKUP_FEATURE_COMPRESSED);
316 ASSERT(!BP_SHOULD_BYTESWAP(bp));
317 ASSERT(!DMU_OT_IS_METADATA(BP_GET_TYPE(bp)));
318 ASSERT3U(BP_GET_COMPRESS(bp), !=, ZIO_COMPRESS_OFF);
319 ASSERT3S(lsize, >=, psize);
322 /* set fields common to compressed and raw sends */
323 drrw->drr_compressiontype = BP_GET_COMPRESS(bp);
324 drrw->drr_compressed_size = psize;
325 payload_size = drrw->drr_compressed_size;
326 } else {
327 payload_size = drrw->drr_logical_size;
330 if (bp == NULL || BP_IS_EMBEDDED(bp) || (BP_IS_PROTECTED(bp) && !raw)) {
332 * There's no pre-computed checksum for partial-block writes,
333 * embedded BP's, or encrypted BP's that are being sent as
334 * plaintext, so (like fletcher4-checkummed blocks) userland
335 * will have to compute a dedup-capable checksum itself.
337 drrw->drr_checksumtype = ZIO_CHECKSUM_OFF;
338 } else {
339 drrw->drr_checksumtype = BP_GET_CHECKSUM(bp);
340 if (zio_checksum_table[drrw->drr_checksumtype].ci_flags &
341 ZCHECKSUM_FLAG_DEDUP)
342 drrw->drr_flags |= DRR_CHECKSUM_DEDUP;
343 DDK_SET_LSIZE(&drrw->drr_key, BP_GET_LSIZE(bp));
344 DDK_SET_PSIZE(&drrw->drr_key, BP_GET_PSIZE(bp));
345 DDK_SET_COMPRESS(&drrw->drr_key, BP_GET_COMPRESS(bp));
346 DDK_SET_CRYPT(&drrw->drr_key, BP_IS_PROTECTED(bp));
347 drrw->drr_key.ddk_cksum = bp->blk_cksum;
350 if (dump_record(dsp, data, payload_size) != 0)
351 return (SET_ERROR(EINTR));
352 return (0);
355 static int
356 dump_write_embedded(dmu_sendarg_t *dsp, uint64_t object, uint64_t offset,
357 int blksz, const blkptr_t *bp)
359 char buf[BPE_PAYLOAD_SIZE];
360 struct drr_write_embedded *drrw =
361 &(dsp->dsa_drr->drr_u.drr_write_embedded);
363 if (dsp->dsa_pending_op != PENDING_NONE) {
364 if (dump_record(dsp, NULL, 0) != 0)
365 return (EINTR);
366 dsp->dsa_pending_op = PENDING_NONE;
369 ASSERT(BP_IS_EMBEDDED(bp));
371 bzero(dsp->dsa_drr, sizeof (dmu_replay_record_t));
372 dsp->dsa_drr->drr_type = DRR_WRITE_EMBEDDED;
373 drrw->drr_object = object;
374 drrw->drr_offset = offset;
375 drrw->drr_length = blksz;
376 drrw->drr_toguid = dsp->dsa_toguid;
377 drrw->drr_compression = BP_GET_COMPRESS(bp);
378 drrw->drr_etype = BPE_GET_ETYPE(bp);
379 drrw->drr_lsize = BPE_GET_LSIZE(bp);
380 drrw->drr_psize = BPE_GET_PSIZE(bp);
382 decode_embedded_bp_compressed(bp, buf);
384 if (dump_record(dsp, buf, P2ROUNDUP(drrw->drr_psize, 8)) != 0)
385 return (EINTR);
386 return (0);
389 static int
390 dump_spill(dmu_sendarg_t *dsp, const blkptr_t *bp, uint64_t object, void *data)
392 struct drr_spill *drrs = &(dsp->dsa_drr->drr_u.drr_spill);
393 uint64_t blksz = BP_GET_LSIZE(bp);
394 uint64_t payload_size = blksz;
396 if (dsp->dsa_pending_op != PENDING_NONE) {
397 if (dump_record(dsp, NULL, 0) != 0)
398 return (SET_ERROR(EINTR));
399 dsp->dsa_pending_op = PENDING_NONE;
402 /* write a SPILL record */
403 bzero(dsp->dsa_drr, sizeof (dmu_replay_record_t));
404 dsp->dsa_drr->drr_type = DRR_SPILL;
405 drrs->drr_object = object;
406 drrs->drr_length = blksz;
407 drrs->drr_toguid = dsp->dsa_toguid;
409 /* See comment in dump_dnode() for full details */
410 if (zfs_send_unmodified_spill_blocks &&
411 (bp->blk_birth <= dsp->dsa_fromtxg)) {
412 drrs->drr_flags |= DRR_SPILL_UNMODIFIED;
415 /* handle raw send fields */
416 if (dsp->dsa_featureflags & DMU_BACKUP_FEATURE_RAW) {
417 ASSERT(BP_IS_PROTECTED(bp));
419 if (BP_SHOULD_BYTESWAP(bp))
420 drrs->drr_flags |= DRR_RAW_BYTESWAP;
421 drrs->drr_compressiontype = BP_GET_COMPRESS(bp);
422 drrs->drr_compressed_size = BP_GET_PSIZE(bp);
423 zio_crypt_decode_params_bp(bp, drrs->drr_salt, drrs->drr_iv);
424 zio_crypt_decode_mac_bp(bp, drrs->drr_mac);
425 payload_size = drrs->drr_compressed_size;
428 if (dump_record(dsp, data, payload_size) != 0)
429 return (SET_ERROR(EINTR));
430 return (0);
433 static int
434 dump_freeobjects(dmu_sendarg_t *dsp, uint64_t firstobj, uint64_t numobjs)
436 struct drr_freeobjects *drrfo = &(dsp->dsa_drr->drr_u.drr_freeobjects);
439 * If there is a pending op, but it's not PENDING_FREEOBJECTS,
440 * push it out, since free block aggregation can only be done for
441 * blocks of the same type (i.e., DRR_FREE records can only be
442 * aggregated with other DRR_FREE records. DRR_FREEOBJECTS records
443 * can only be aggregated with other DRR_FREEOBJECTS records.
445 if (dsp->dsa_pending_op != PENDING_NONE &&
446 dsp->dsa_pending_op != PENDING_FREEOBJECTS) {
447 if (dump_record(dsp, NULL, 0) != 0)
448 return (SET_ERROR(EINTR));
449 dsp->dsa_pending_op = PENDING_NONE;
451 if (dsp->dsa_pending_op == PENDING_FREEOBJECTS) {
453 * See whether this free object array can be aggregated
454 * with pending one
456 if (drrfo->drr_firstobj + drrfo->drr_numobjs == firstobj) {
457 drrfo->drr_numobjs += numobjs;
458 return (0);
459 } else {
460 /* can't be aggregated. Push out pending record */
461 if (dump_record(dsp, NULL, 0) != 0)
462 return (SET_ERROR(EINTR));
463 dsp->dsa_pending_op = PENDING_NONE;
467 /* write a FREEOBJECTS record */
468 bzero(dsp->dsa_drr, sizeof (dmu_replay_record_t));
469 dsp->dsa_drr->drr_type = DRR_FREEOBJECTS;
470 drrfo->drr_firstobj = firstobj;
471 drrfo->drr_numobjs = numobjs;
472 drrfo->drr_toguid = dsp->dsa_toguid;
474 dsp->dsa_pending_op = PENDING_FREEOBJECTS;
476 return (0);
479 static int
480 dump_dnode(dmu_sendarg_t *dsp, const blkptr_t *bp, uint64_t object,
481 dnode_phys_t *dnp)
483 struct drr_object *drro = &(dsp->dsa_drr->drr_u.drr_object);
484 int bonuslen;
486 if (object < dsp->dsa_resume_object) {
488 * Note: when resuming, we will visit all the dnodes in
489 * the block of dnodes that we are resuming from. In
490 * this case it's unnecessary to send the dnodes prior to
491 * the one we are resuming from. We should be at most one
492 * block's worth of dnodes behind the resume point.
494 ASSERT3U(dsp->dsa_resume_object - object, <,
495 1 << (DNODE_BLOCK_SHIFT - DNODE_SHIFT));
496 return (0);
499 if (dnp == NULL || dnp->dn_type == DMU_OT_NONE)
500 return (dump_freeobjects(dsp, object, 1));
502 if (dsp->dsa_pending_op != PENDING_NONE) {
503 if (dump_record(dsp, NULL, 0) != 0)
504 return (SET_ERROR(EINTR));
505 dsp->dsa_pending_op = PENDING_NONE;
508 /* write an OBJECT record */
509 bzero(dsp->dsa_drr, sizeof (dmu_replay_record_t));
510 dsp->dsa_drr->drr_type = DRR_OBJECT;
511 drro->drr_object = object;
512 drro->drr_type = dnp->dn_type;
513 drro->drr_bonustype = dnp->dn_bonustype;
514 drro->drr_blksz = dnp->dn_datablkszsec << SPA_MINBLOCKSHIFT;
515 drro->drr_bonuslen = dnp->dn_bonuslen;
516 drro->drr_dn_slots = dnp->dn_extra_slots + 1;
517 drro->drr_checksumtype = dnp->dn_checksum;
518 drro->drr_compress = dnp->dn_compress;
519 drro->drr_toguid = dsp->dsa_toguid;
521 if (!(dsp->dsa_featureflags & DMU_BACKUP_FEATURE_LARGE_BLOCKS) &&
522 drro->drr_blksz > SPA_OLD_MAXBLOCKSIZE)
523 drro->drr_blksz = SPA_OLD_MAXBLOCKSIZE;
525 bonuslen = P2ROUNDUP(dnp->dn_bonuslen, 8);
527 if ((dsp->dsa_featureflags & DMU_BACKUP_FEATURE_RAW)) {
528 ASSERT(BP_IS_ENCRYPTED(bp));
530 if (BP_SHOULD_BYTESWAP(bp))
531 drro->drr_flags |= DRR_RAW_BYTESWAP;
533 /* needed for reconstructing dnp on recv side */
534 drro->drr_maxblkid = dnp->dn_maxblkid;
535 drro->drr_indblkshift = dnp->dn_indblkshift;
536 drro->drr_nlevels = dnp->dn_nlevels;
537 drro->drr_nblkptr = dnp->dn_nblkptr;
540 * Since we encrypt the entire bonus area, the (raw) part
541 * beyond the bonuslen is actually nonzero, so we need
542 * to send it.
544 if (bonuslen != 0) {
545 drro->drr_raw_bonuslen = DN_MAX_BONUS_LEN(dnp);
546 bonuslen = drro->drr_raw_bonuslen;
551 * DRR_OBJECT_SPILL is set for every dnode which references a
552 * spill block. This allows the receiving pool to definitively
553 * determine when a spill block should be kept or freed.
555 if (dnp->dn_flags & DNODE_FLAG_SPILL_BLKPTR)
556 drro->drr_flags |= DRR_OBJECT_SPILL;
558 if (dump_record(dsp, DN_BONUS(dnp), bonuslen) != 0)
559 return (SET_ERROR(EINTR));
561 /* Free anything past the end of the file. */
562 if (dump_free(dsp, object, (dnp->dn_maxblkid + 1) *
563 (dnp->dn_datablkszsec << SPA_MINBLOCKSHIFT), DMU_OBJECT_END) != 0)
564 return (SET_ERROR(EINTR));
567 * Send DRR_SPILL records for unmodified spill blocks. This is useful
568 * because changing certain attributes of the object (e.g. blocksize)
569 * can cause old versions of ZFS to incorrectly remove a spill block.
570 * Including these records in the stream forces an up to date version
571 * to always be written ensuring they're never lost. Current versions
572 * of the code which understand the DRR_FLAG_SPILL_BLOCK feature can
573 * ignore these unmodified spill blocks.
575 if (zfs_send_unmodified_spill_blocks &&
576 (dnp->dn_flags & DNODE_FLAG_SPILL_BLKPTR) &&
577 (DN_SPILL_BLKPTR(dnp)->blk_birth <= dsp->dsa_fromtxg)) {
578 struct send_block_record record;
580 bzero(&record, sizeof (struct send_block_record));
581 record.eos_marker = B_FALSE;
582 record.bp = *DN_SPILL_BLKPTR(dnp);
583 SET_BOOKMARK(&(record.zb), dmu_objset_id(dsp->dsa_os),
584 object, 0, DMU_SPILL_BLKID);
586 if (do_dump(dsp, &record) != 0)
587 return (SET_ERROR(EINTR));
590 if (dsp->dsa_err != 0)
591 return (SET_ERROR(EINTR));
592 return (0);
595 static int
596 dump_object_range(dmu_sendarg_t *dsp, const blkptr_t *bp, uint64_t firstobj,
597 uint64_t numslots)
599 struct drr_object_range *drror =
600 &(dsp->dsa_drr->drr_u.drr_object_range);
602 /* we only use this record type for raw sends */
603 ASSERT(BP_IS_PROTECTED(bp));
604 ASSERT(dsp->dsa_featureflags & DMU_BACKUP_FEATURE_RAW);
605 ASSERT3U(BP_GET_COMPRESS(bp), ==, ZIO_COMPRESS_OFF);
606 ASSERT3U(BP_GET_TYPE(bp), ==, DMU_OT_DNODE);
607 ASSERT0(BP_GET_LEVEL(bp));
609 if (dsp->dsa_pending_op != PENDING_NONE) {
610 if (dump_record(dsp, NULL, 0) != 0)
611 return (SET_ERROR(EINTR));
612 dsp->dsa_pending_op = PENDING_NONE;
615 bzero(dsp->dsa_drr, sizeof (dmu_replay_record_t));
616 dsp->dsa_drr->drr_type = DRR_OBJECT_RANGE;
617 drror->drr_firstobj = firstobj;
618 drror->drr_numslots = numslots;
619 drror->drr_toguid = dsp->dsa_toguid;
620 if (BP_SHOULD_BYTESWAP(bp))
621 drror->drr_flags |= DRR_RAW_BYTESWAP;
622 zio_crypt_decode_params_bp(bp, drror->drr_salt, drror->drr_iv);
623 zio_crypt_decode_mac_bp(bp, drror->drr_mac);
625 if (dump_record(dsp, NULL, 0) != 0)
626 return (SET_ERROR(EINTR));
627 return (0);
630 static boolean_t
631 backup_do_embed(dmu_sendarg_t *dsp, const blkptr_t *bp)
633 if (!BP_IS_EMBEDDED(bp))
634 return (B_FALSE);
637 * Compression function must be legacy, or explicitly enabled.
639 if ((BP_GET_COMPRESS(bp) >= ZIO_COMPRESS_LEGACY_FUNCTIONS &&
640 !(dsp->dsa_featureflags & DMU_BACKUP_FEATURE_LZ4)))
641 return (B_FALSE);
644 * Embed type must be explicitly enabled.
646 switch (BPE_GET_ETYPE(bp)) {
647 case BP_EMBEDDED_TYPE_DATA:
648 if (dsp->dsa_featureflags & DMU_BACKUP_FEATURE_EMBED_DATA)
649 return (B_TRUE);
650 break;
651 default:
652 return (B_FALSE);
654 return (B_FALSE);
658 * This is the callback function to traverse_dataset that acts as the worker
659 * thread for dmu_send_impl.
661 /*ARGSUSED*/
662 static int
663 send_cb(spa_t *spa, zilog_t *zilog, const blkptr_t *bp,
664 const zbookmark_phys_t *zb, const struct dnode_phys *dnp, void *arg)
666 struct send_thread_arg *sta = arg;
667 struct send_block_record *record;
668 uint64_t record_size;
669 int err = 0;
671 ASSERT(zb->zb_object == DMU_META_DNODE_OBJECT ||
672 zb->zb_object >= sta->resume.zb_object);
673 ASSERT3P(sta->ds, !=, NULL);
675 if (sta->cancel)
676 return (SET_ERROR(EINTR));
678 if (bp == NULL) {
679 ASSERT3U(zb->zb_level, ==, ZB_DNODE_LEVEL);
680 return (0);
681 } else if (zb->zb_level < 0) {
682 return (0);
685 record = kmem_zalloc(sizeof (struct send_block_record), KM_SLEEP);
686 record->eos_marker = B_FALSE;
687 record->bp = *bp;
688 record->zb = *zb;
689 record->indblkshift = dnp->dn_indblkshift;
690 record->datablkszsec = dnp->dn_datablkszsec;
691 record_size = dnp->dn_datablkszsec << SPA_MINBLOCKSHIFT;
692 bqueue_enqueue(&sta->q, record, record_size);
694 return (err);
698 * This function kicks off the traverse_dataset. It also handles setting the
699 * error code of the thread in case something goes wrong, and pushes the End of
700 * Stream record when the traverse_dataset call has finished. If there is no
701 * dataset to traverse, the thread immediately pushes End of Stream marker.
703 static void
704 send_traverse_thread(void *arg)
706 struct send_thread_arg *st_arg = arg;
707 int err;
708 struct send_block_record *data;
710 if (st_arg->ds != NULL) {
711 err = traverse_dataset_resume(st_arg->ds,
712 st_arg->fromtxg, &st_arg->resume,
713 st_arg->flags, send_cb, st_arg);
715 if (err != EINTR)
716 st_arg->error_code = err;
718 data = kmem_zalloc(sizeof (*data), KM_SLEEP);
719 data->eos_marker = B_TRUE;
720 bqueue_enqueue(&st_arg->q, data, 1);
721 thread_exit();
725 * This function actually handles figuring out what kind of record needs to be
726 * dumped, reading the data (which has hopefully been prefetched), and calling
727 * the appropriate helper function.
729 static int
730 do_dump(dmu_sendarg_t *dsa, struct send_block_record *data)
732 dsl_dataset_t *ds = dmu_objset_ds(dsa->dsa_os);
733 const blkptr_t *bp = &data->bp;
734 const zbookmark_phys_t *zb = &data->zb;
735 uint8_t indblkshift = data->indblkshift;
736 uint16_t dblkszsec = data->datablkszsec;
737 spa_t *spa = ds->ds_dir->dd_pool->dp_spa;
738 dmu_object_type_t type = bp ? BP_GET_TYPE(bp) : DMU_OT_NONE;
739 int err = 0;
741 ASSERT3U(zb->zb_level, >=, 0);
743 ASSERT(zb->zb_object == DMU_META_DNODE_OBJECT ||
744 zb->zb_object >= dsa->dsa_resume_object);
747 * All bps of an encrypted os should have the encryption bit set.
748 * If this is not true it indicates tampering and we report an error.
750 if (dsa->dsa_os->os_encrypted &&
751 !BP_IS_HOLE(bp) && !BP_USES_CRYPT(bp)) {
752 spa_log_error(spa, zb);
753 zfs_panic_recover("unencrypted block in encrypted "
754 "object set %llu", ds->ds_object);
755 return (SET_ERROR(EIO));
758 if (zb->zb_object != DMU_META_DNODE_OBJECT &&
759 DMU_OBJECT_IS_SPECIAL(zb->zb_object)) {
760 return (0);
761 } else if (BP_IS_HOLE(bp) &&
762 zb->zb_object == DMU_META_DNODE_OBJECT) {
763 uint64_t span = BP_SPAN(dblkszsec, indblkshift, zb->zb_level);
764 uint64_t dnobj = (zb->zb_blkid * span) >> DNODE_SHIFT;
765 err = dump_freeobjects(dsa, dnobj, span >> DNODE_SHIFT);
766 } else if (BP_IS_HOLE(bp)) {
767 uint64_t span = BP_SPAN(dblkszsec, indblkshift, zb->zb_level);
768 uint64_t offset = zb->zb_blkid * span;
769 /* Don't dump free records for offsets > DMU_OBJECT_END */
770 if (zb->zb_blkid == 0 || span <= DMU_OBJECT_END / zb->zb_blkid)
771 err = dump_free(dsa, zb->zb_object, offset, span);
772 } else if (zb->zb_level > 0 || type == DMU_OT_OBJSET) {
773 return (0);
774 } else if (type == DMU_OT_DNODE) {
775 int epb = BP_GET_LSIZE(bp) >> DNODE_SHIFT;
776 arc_flags_t aflags = ARC_FLAG_WAIT;
777 arc_buf_t *abuf;
778 enum zio_flag zioflags = ZIO_FLAG_CANFAIL;
780 if (dsa->dsa_featureflags & DMU_BACKUP_FEATURE_RAW) {
781 ASSERT(BP_IS_ENCRYPTED(bp));
782 ASSERT3U(BP_GET_COMPRESS(bp), ==, ZIO_COMPRESS_OFF);
783 zioflags |= ZIO_FLAG_RAW;
786 ASSERT0(zb->zb_level);
788 if (arc_read(NULL, spa, bp, arc_getbuf_func, &abuf,
789 ZIO_PRIORITY_ASYNC_READ, zioflags, &aflags, zb) != 0)
790 return (SET_ERROR(EIO));
792 dnode_phys_t *blk = abuf->b_data;
793 uint64_t dnobj = zb->zb_blkid * epb;
796 * Raw sends require sending encryption parameters for the
797 * block of dnodes. Regular sends do not need to send this
798 * info.
800 if (dsa->dsa_featureflags & DMU_BACKUP_FEATURE_RAW) {
801 ASSERT(arc_is_encrypted(abuf));
802 err = dump_object_range(dsa, bp, dnobj, epb);
805 if (err == 0) {
806 for (int i = 0; i < epb;
807 i += blk[i].dn_extra_slots + 1) {
808 err = dump_dnode(dsa, bp, dnobj + i, blk + i);
809 if (err != 0)
810 break;
813 arc_buf_destroy(abuf, &abuf);
814 } else if (type == DMU_OT_SA) {
815 arc_flags_t aflags = ARC_FLAG_WAIT;
816 arc_buf_t *abuf;
817 enum zio_flag zioflags = ZIO_FLAG_CANFAIL;
819 if (dsa->dsa_featureflags & DMU_BACKUP_FEATURE_RAW) {
820 ASSERT(BP_IS_PROTECTED(bp));
821 zioflags |= ZIO_FLAG_RAW;
824 if (arc_read(NULL, spa, bp, arc_getbuf_func, &abuf,
825 ZIO_PRIORITY_ASYNC_READ, zioflags, &aflags, zb) != 0)
826 return (SET_ERROR(EIO));
828 err = dump_spill(dsa, bp, zb->zb_object, abuf->b_data);
829 arc_buf_destroy(abuf, &abuf);
830 } else if (backup_do_embed(dsa, bp)) {
831 /* it's an embedded level-0 block of a regular object */
832 int blksz = dblkszsec << SPA_MINBLOCKSHIFT;
833 ASSERT0(zb->zb_level);
834 err = dump_write_embedded(dsa, zb->zb_object,
835 zb->zb_blkid * blksz, blksz, bp);
836 } else {
837 /* it's a level-0 block of a regular object */
838 arc_flags_t aflags = ARC_FLAG_WAIT;
839 arc_buf_t *abuf;
840 int blksz = dblkszsec << SPA_MINBLOCKSHIFT;
841 uint64_t offset;
844 * If we have large blocks stored on disk but the send flags
845 * don't allow us to send large blocks, we split the data from
846 * the arc buf into chunks.
848 boolean_t split_large_blocks = blksz > SPA_OLD_MAXBLOCKSIZE &&
849 !(dsa->dsa_featureflags & DMU_BACKUP_FEATURE_LARGE_BLOCKS);
852 * Raw sends require that we always get raw data as it exists
853 * on disk, so we assert that we are not splitting blocks here.
855 boolean_t request_raw =
856 (dsa->dsa_featureflags & DMU_BACKUP_FEATURE_RAW) != 0;
859 * We should only request compressed data from the ARC if all
860 * the following are true:
861 * - stream compression was requested
862 * - we aren't splitting large blocks into smaller chunks
863 * - the data won't need to be byteswapped before sending
864 * - this isn't an embedded block
865 * - this isn't metadata (if receiving on a different endian
866 * system it can be byteswapped more easily)
868 boolean_t request_compressed =
869 (dsa->dsa_featureflags & DMU_BACKUP_FEATURE_COMPRESSED) &&
870 !split_large_blocks && !BP_SHOULD_BYTESWAP(bp) &&
871 !BP_IS_EMBEDDED(bp) && !DMU_OT_IS_METADATA(BP_GET_TYPE(bp));
873 IMPLY(request_raw, !split_large_blocks);
874 IMPLY(request_raw, BP_IS_PROTECTED(bp));
875 ASSERT0(zb->zb_level);
876 ASSERT(zb->zb_object > dsa->dsa_resume_object ||
877 (zb->zb_object == dsa->dsa_resume_object &&
878 zb->zb_blkid * blksz >= dsa->dsa_resume_offset));
880 ASSERT0(zb->zb_level);
881 ASSERT(zb->zb_object > dsa->dsa_resume_object ||
882 (zb->zb_object == dsa->dsa_resume_object &&
883 zb->zb_blkid * blksz >= dsa->dsa_resume_offset));
885 ASSERT3U(blksz, ==, BP_GET_LSIZE(bp));
887 enum zio_flag zioflags = ZIO_FLAG_CANFAIL;
888 if (request_raw)
889 zioflags |= ZIO_FLAG_RAW;
890 else if (request_compressed)
891 zioflags |= ZIO_FLAG_RAW_COMPRESS;
893 if (arc_read(NULL, spa, bp, arc_getbuf_func, &abuf,
894 ZIO_PRIORITY_ASYNC_READ, zioflags, &aflags, zb) != 0) {
895 if (zfs_send_corrupt_data) {
896 /* Send a block filled with 0x"zfs badd bloc" */
897 abuf = arc_alloc_buf(spa, &abuf, ARC_BUFC_DATA,
898 blksz);
899 uint64_t *ptr;
900 for (ptr = abuf->b_data;
901 (char *)ptr < (char *)abuf->b_data + blksz;
902 ptr++)
903 *ptr = 0x2f5baddb10cULL;
904 } else {
905 return (SET_ERROR(EIO));
909 offset = zb->zb_blkid * blksz;
911 if (split_large_blocks) {
912 ASSERT0(arc_is_encrypted(abuf));
913 ASSERT3U(arc_get_compression(abuf), ==,
914 ZIO_COMPRESS_OFF);
915 char *buf = abuf->b_data;
916 while (blksz > 0 && err == 0) {
917 int n = MIN(blksz, SPA_OLD_MAXBLOCKSIZE);
918 err = dump_write(dsa, type, zb->zb_object,
919 offset, n, n, NULL, buf);
920 offset += n;
921 buf += n;
922 blksz -= n;
924 } else {
925 err = dump_write(dsa, type, zb->zb_object, offset,
926 blksz, arc_buf_size(abuf), bp, abuf->b_data);
928 arc_buf_destroy(abuf, &abuf);
931 ASSERT(err == 0 || err == EINTR);
932 return (err);
936 * Pop the new data off the queue, and free the old data.
938 static struct send_block_record *
939 get_next_record(bqueue_t *bq, struct send_block_record *data)
941 struct send_block_record *tmp = bqueue_dequeue(bq);
942 kmem_free(data, sizeof (*data));
943 return (tmp);
947 * Actually do the bulk of the work in a zfs send.
949 * Note: Releases dp using the specified tag.
951 static int
952 dmu_send_impl(void *tag, dsl_pool_t *dp, dsl_dataset_t *to_ds,
953 zfs_bookmark_phys_t *ancestor_zb, boolean_t is_clone,
954 boolean_t embedok, boolean_t large_block_ok, boolean_t compressok,
955 boolean_t rawok, int outfd, uint64_t resumeobj, uint64_t resumeoff,
956 vnode_t *vp, offset_t *off)
958 objset_t *os;
959 dmu_replay_record_t *drr;
960 dmu_sendarg_t *dsp;
961 int err;
962 uint64_t fromtxg = 0;
963 uint64_t featureflags = 0;
964 struct send_thread_arg to_arg = { 0 };
966 err = dmu_objset_from_ds(to_ds, &os);
967 if (err != 0) {
968 dsl_pool_rele(dp, tag);
969 return (err);
973 * If this is a non-raw send of an encrypted ds, we can ensure that
974 * the objset_phys_t is authenticated. This is safe because this is
975 * either a snapshot or we have owned the dataset, ensuring that
976 * it can't be modified.
978 if (!rawok && os->os_encrypted &&
979 arc_is_unauthenticated(os->os_phys_buf)) {
980 zbookmark_phys_t zb;
982 SET_BOOKMARK(&zb, to_ds->ds_object, ZB_ROOT_OBJECT,
983 ZB_ROOT_LEVEL, ZB_ROOT_BLKID);
984 err = arc_untransform(os->os_phys_buf, os->os_spa,
985 &zb, B_FALSE);
986 if (err != 0) {
987 dsl_pool_rele(dp, tag);
988 return (err);
991 ASSERT0(arc_is_unauthenticated(os->os_phys_buf));
994 drr = kmem_zalloc(sizeof (dmu_replay_record_t), KM_SLEEP);
995 drr->drr_type = DRR_BEGIN;
996 drr->drr_u.drr_begin.drr_magic = DMU_BACKUP_MAGIC;
997 DMU_SET_STREAM_HDRTYPE(drr->drr_u.drr_begin.drr_versioninfo,
998 DMU_SUBSTREAM);
1000 #ifdef _KERNEL
1001 if (dmu_objset_type(os) == DMU_OST_ZFS) {
1002 uint64_t version;
1003 if (zfs_get_zplprop(os, ZFS_PROP_VERSION, &version) != 0) {
1004 kmem_free(drr, sizeof (dmu_replay_record_t));
1005 dsl_pool_rele(dp, tag);
1006 return (SET_ERROR(EINVAL));
1008 if (version >= ZPL_VERSION_SA) {
1009 featureflags |= DMU_BACKUP_FEATURE_SA_SPILL;
1012 #endif
1014 /* raw sends imply large_block_ok */
1015 if ((large_block_ok || rawok) &&
1016 to_ds->ds_feature_inuse[SPA_FEATURE_LARGE_BLOCKS])
1017 featureflags |= DMU_BACKUP_FEATURE_LARGE_BLOCKS;
1018 if (to_ds->ds_feature_inuse[SPA_FEATURE_LARGE_DNODE])
1019 featureflags |= DMU_BACKUP_FEATURE_LARGE_DNODE;
1021 /* encrypted datasets will not have embedded blocks */
1022 if ((embedok || rawok) && !os->os_encrypted &&
1023 spa_feature_is_active(dp->dp_spa, SPA_FEATURE_EMBEDDED_DATA)) {
1024 featureflags |= DMU_BACKUP_FEATURE_EMBED_DATA;
1027 /* raw send implies compressok */
1028 if (compressok || rawok)
1029 featureflags |= DMU_BACKUP_FEATURE_COMPRESSED;
1030 if (rawok && os->os_encrypted)
1031 featureflags |= DMU_BACKUP_FEATURE_RAW;
1033 if ((featureflags &
1034 (DMU_BACKUP_FEATURE_EMBED_DATA | DMU_BACKUP_FEATURE_COMPRESSED |
1035 DMU_BACKUP_FEATURE_RAW)) != 0 &&
1036 spa_feature_is_active(dp->dp_spa, SPA_FEATURE_LZ4_COMPRESS)) {
1037 featureflags |= DMU_BACKUP_FEATURE_LZ4;
1040 if (resumeobj != 0 || resumeoff != 0) {
1041 featureflags |= DMU_BACKUP_FEATURE_RESUMING;
1044 DMU_SET_FEATUREFLAGS(drr->drr_u.drr_begin.drr_versioninfo,
1045 featureflags);
1047 drr->drr_u.drr_begin.drr_creation_time =
1048 dsl_dataset_phys(to_ds)->ds_creation_time;
1049 drr->drr_u.drr_begin.drr_type = dmu_objset_type(os);
1050 if (is_clone)
1051 drr->drr_u.drr_begin.drr_flags |= DRR_FLAG_CLONE;
1052 drr->drr_u.drr_begin.drr_toguid = dsl_dataset_phys(to_ds)->ds_guid;
1053 if (dsl_dataset_phys(to_ds)->ds_flags & DS_FLAG_CI_DATASET)
1054 drr->drr_u.drr_begin.drr_flags |= DRR_FLAG_CI_DATA;
1055 if (zfs_send_set_freerecords_bit)
1056 drr->drr_u.drr_begin.drr_flags |= DRR_FLAG_FREERECORDS;
1058 drr->drr_u.drr_begin.drr_flags |= DRR_FLAG_SPILL_BLOCK;
1060 if (ancestor_zb != NULL) {
1061 drr->drr_u.drr_begin.drr_fromguid =
1062 ancestor_zb->zbm_guid;
1063 fromtxg = ancestor_zb->zbm_creation_txg;
1065 dsl_dataset_name(to_ds, drr->drr_u.drr_begin.drr_toname);
1066 if (!to_ds->ds_is_snapshot) {
1067 (void) strlcat(drr->drr_u.drr_begin.drr_toname, "@--head--",
1068 sizeof (drr->drr_u.drr_begin.drr_toname));
1071 dsp = kmem_zalloc(sizeof (dmu_sendarg_t), KM_SLEEP);
1073 dsp->dsa_drr = drr;
1074 dsp->dsa_vp = vp;
1075 dsp->dsa_outfd = outfd;
1076 dsp->dsa_proc = curproc;
1077 dsp->dsa_os = os;
1078 dsp->dsa_off = off;
1079 dsp->dsa_toguid = dsl_dataset_phys(to_ds)->ds_guid;
1080 dsp->dsa_fromtxg = fromtxg;
1081 dsp->dsa_pending_op = PENDING_NONE;
1082 dsp->dsa_featureflags = featureflags;
1083 dsp->dsa_resume_object = resumeobj;
1084 dsp->dsa_resume_offset = resumeoff;
1086 mutex_enter(&to_ds->ds_sendstream_lock);
1087 list_insert_head(&to_ds->ds_sendstreams, dsp);
1088 mutex_exit(&to_ds->ds_sendstream_lock);
1090 dsl_dataset_long_hold(to_ds, FTAG);
1091 dsl_pool_rele(dp, tag);
1093 void *payload = NULL;
1094 size_t payload_len = 0;
1095 /* handle features that require a DRR_BEGIN payload */
1096 if (featureflags &
1097 (DMU_BACKUP_FEATURE_RESUMING | DMU_BACKUP_FEATURE_RAW)) {
1098 nvlist_t *keynvl = NULL;
1099 nvlist_t *nvl = fnvlist_alloc();
1101 if (featureflags & DMU_BACKUP_FEATURE_RESUMING) {
1102 dmu_object_info_t to_doi;
1103 err = dmu_object_info(os, resumeobj, &to_doi);
1104 if (err != 0) {
1105 fnvlist_free(nvl);
1106 goto out;
1109 SET_BOOKMARK(&to_arg.resume, to_ds->ds_object,
1110 resumeobj, 0,
1111 resumeoff / to_doi.doi_data_block_size);
1113 fnvlist_add_uint64(nvl, "resume_object", resumeobj);
1114 fnvlist_add_uint64(nvl, "resume_offset", resumeoff);
1117 if (featureflags & DMU_BACKUP_FEATURE_RAW) {
1118 uint64_t ivset_guid = (ancestor_zb != NULL) ?
1119 ancestor_zb->zbm_ivset_guid : 0;
1121 ASSERT(os->os_encrypted);
1123 err = dsl_crypto_populate_key_nvlist(to_ds,
1124 ivset_guid, &keynvl);
1125 if (err != 0) {
1126 fnvlist_free(nvl);
1127 goto out;
1130 fnvlist_add_nvlist(nvl, "crypt_keydata", keynvl);
1133 payload = fnvlist_pack(nvl, &payload_len);
1134 drr->drr_payloadlen = payload_len;
1135 fnvlist_free(keynvl);
1136 fnvlist_free(nvl);
1139 err = dump_record(dsp, payload, payload_len);
1140 fnvlist_pack_free(payload, payload_len);
1141 if (err != 0) {
1142 err = dsp->dsa_err;
1143 goto out;
1146 err = bqueue_init(&to_arg.q,
1147 MAX(zfs_send_queue_length, 2 * zfs_max_recordsize),
1148 offsetof(struct send_block_record, ln));
1149 to_arg.error_code = 0;
1150 to_arg.cancel = B_FALSE;
1151 to_arg.ds = to_ds;
1152 to_arg.fromtxg = fromtxg;
1153 to_arg.flags = TRAVERSE_PRE | TRAVERSE_PREFETCH;
1154 if (rawok)
1155 to_arg.flags |= TRAVERSE_NO_DECRYPT;
1156 (void) thread_create(NULL, 0, send_traverse_thread, &to_arg, 0, curproc,
1157 TS_RUN, minclsyspri);
1159 struct send_block_record *to_data;
1160 to_data = bqueue_dequeue(&to_arg.q);
1162 while (!to_data->eos_marker && err == 0) {
1163 err = do_dump(dsp, to_data);
1164 to_data = get_next_record(&to_arg.q, to_data);
1165 if (issig(JUSTLOOKING) && issig(FORREAL))
1166 err = EINTR;
1169 if (err != 0) {
1170 to_arg.cancel = B_TRUE;
1171 while (!to_data->eos_marker) {
1172 to_data = get_next_record(&to_arg.q, to_data);
1175 kmem_free(to_data, sizeof (*to_data));
1177 bqueue_destroy(&to_arg.q);
1179 if (err == 0 && to_arg.error_code != 0)
1180 err = to_arg.error_code;
1182 if (err != 0)
1183 goto out;
1185 if (dsp->dsa_pending_op != PENDING_NONE)
1186 if (dump_record(dsp, NULL, 0) != 0)
1187 err = SET_ERROR(EINTR);
1189 if (err != 0) {
1190 if (err == EINTR && dsp->dsa_err != 0)
1191 err = dsp->dsa_err;
1192 goto out;
1195 bzero(drr, sizeof (dmu_replay_record_t));
1196 drr->drr_type = DRR_END;
1197 drr->drr_u.drr_end.drr_checksum = dsp->dsa_zc;
1198 drr->drr_u.drr_end.drr_toguid = dsp->dsa_toguid;
1200 if (dump_record(dsp, NULL, 0) != 0)
1201 err = dsp->dsa_err;
1202 out:
1203 mutex_enter(&to_ds->ds_sendstream_lock);
1204 list_remove(&to_ds->ds_sendstreams, dsp);
1205 mutex_exit(&to_ds->ds_sendstream_lock);
1207 VERIFY(err != 0 || (dsp->dsa_sent_begin && dsp->dsa_sent_end));
1209 kmem_free(drr, sizeof (dmu_replay_record_t));
1210 kmem_free(dsp, sizeof (dmu_sendarg_t));
1212 dsl_dataset_long_rele(to_ds, FTAG);
1214 return (err);
1218 dmu_send_obj(const char *pool, uint64_t tosnap, uint64_t fromsnap,
1219 boolean_t embedok, boolean_t large_block_ok, boolean_t compressok,
1220 boolean_t rawok, int outfd, vnode_t *vp, offset_t *off)
1222 dsl_pool_t *dp;
1223 dsl_dataset_t *ds;
1224 dsl_dataset_t *fromds = NULL;
1225 ds_hold_flags_t dsflags;
1226 int err;
1228 dsflags = (rawok) ? DS_HOLD_FLAG_NONE : DS_HOLD_FLAG_DECRYPT;
1229 err = dsl_pool_hold(pool, FTAG, &dp);
1230 if (err != 0)
1231 return (err);
1233 err = dsl_dataset_hold_obj_flags(dp, tosnap, dsflags, FTAG, &ds);
1234 if (err != 0) {
1235 dsl_pool_rele(dp, FTAG);
1236 return (err);
1239 if (fromsnap != 0) {
1240 zfs_bookmark_phys_t zb = { 0 };
1241 boolean_t is_clone;
1243 err = dsl_dataset_hold_obj(dp, fromsnap, FTAG, &fromds);
1244 if (err != 0) {
1245 dsl_dataset_rele_flags(ds, dsflags, FTAG);
1246 dsl_pool_rele(dp, FTAG);
1247 return (err);
1249 if (!dsl_dataset_is_before(ds, fromds, 0)) {
1250 err = SET_ERROR(EXDEV);
1251 dsl_dataset_rele(fromds, FTAG);
1252 dsl_dataset_rele_flags(ds, dsflags, FTAG);
1253 dsl_pool_rele(dp, FTAG);
1254 return (err);
1257 zb.zbm_creation_time =
1258 dsl_dataset_phys(fromds)->ds_creation_time;
1259 zb.zbm_creation_txg = dsl_dataset_phys(fromds)->ds_creation_txg;
1260 zb.zbm_guid = dsl_dataset_phys(fromds)->ds_guid;
1262 if (dsl_dataset_is_zapified(fromds)) {
1263 (void) zap_lookup(dp->dp_meta_objset,
1264 fromds->ds_object, DS_FIELD_IVSET_GUID, 8, 1,
1265 &zb.zbm_ivset_guid);
1268 is_clone = (fromds->ds_dir != ds->ds_dir);
1269 dsl_dataset_rele(fromds, FTAG);
1270 err = dmu_send_impl(FTAG, dp, ds, &zb, is_clone,
1271 embedok, large_block_ok, compressok, rawok, outfd,
1272 0, 0, vp, off);
1273 } else {
1274 err = dmu_send_impl(FTAG, dp, ds, NULL, B_FALSE,
1275 embedok, large_block_ok, compressok, rawok, outfd,
1276 0, 0, vp, off);
1278 dsl_dataset_rele_flags(ds, dsflags, FTAG);
1279 return (err);
1283 dmu_send(const char *tosnap, const char *fromsnap, boolean_t embedok,
1284 boolean_t large_block_ok, boolean_t compressok, boolean_t rawok,
1285 int outfd, uint64_t resumeobj, uint64_t resumeoff, vnode_t *vp,
1286 offset_t *off)
1288 dsl_pool_t *dp;
1289 dsl_dataset_t *ds;
1290 int err;
1291 ds_hold_flags_t dsflags;
1292 boolean_t owned = B_FALSE;
1294 dsflags = (rawok) ? DS_HOLD_FLAG_NONE : DS_HOLD_FLAG_DECRYPT;
1295 if (fromsnap != NULL && strpbrk(fromsnap, "@#") == NULL)
1296 return (SET_ERROR(EINVAL));
1298 err = dsl_pool_hold(tosnap, FTAG, &dp);
1299 if (err != 0)
1300 return (err);
1301 if (strchr(tosnap, '@') == NULL && spa_writeable(dp->dp_spa)) {
1303 * We are sending a filesystem or volume. Ensure
1304 * that it doesn't change by owning the dataset.
1306 err = dsl_dataset_own(dp, tosnap, dsflags, FTAG, &ds);
1307 owned = B_TRUE;
1308 } else {
1309 err = dsl_dataset_hold_flags(dp, tosnap, dsflags, FTAG, &ds);
1311 if (err != 0) {
1312 dsl_pool_rele(dp, FTAG);
1313 return (err);
1316 if (fromsnap != NULL) {
1317 zfs_bookmark_phys_t zb = { 0 };
1318 boolean_t is_clone = B_FALSE;
1319 int fsnamelen = strchr(tosnap, '@') - tosnap;
1322 * If the fromsnap is in a different filesystem, then
1323 * mark the send stream as a clone.
1325 if (strncmp(tosnap, fromsnap, fsnamelen) != 0 ||
1326 (fromsnap[fsnamelen] != '@' &&
1327 fromsnap[fsnamelen] != '#')) {
1328 is_clone = B_TRUE;
1331 if (strchr(fromsnap, '@')) {
1332 dsl_dataset_t *fromds;
1333 err = dsl_dataset_hold(dp, fromsnap, FTAG, &fromds);
1334 if (err == 0) {
1335 if (!dsl_dataset_is_before(ds, fromds, 0))
1336 err = SET_ERROR(EXDEV);
1337 zb.zbm_creation_time =
1338 dsl_dataset_phys(fromds)->ds_creation_time;
1339 zb.zbm_creation_txg =
1340 dsl_dataset_phys(fromds)->ds_creation_txg;
1341 zb.zbm_guid = dsl_dataset_phys(fromds)->ds_guid;
1342 is_clone = (ds->ds_dir != fromds->ds_dir);
1344 if (dsl_dataset_is_zapified(fromds)) {
1345 (void) zap_lookup(dp->dp_meta_objset,
1346 fromds->ds_object,
1347 DS_FIELD_IVSET_GUID, 8, 1,
1348 &zb.zbm_ivset_guid);
1350 dsl_dataset_rele(fromds, FTAG);
1352 } else {
1353 err = dsl_bookmark_lookup(dp, fromsnap, ds, &zb);
1355 if (err != 0) {
1356 if (owned)
1357 dsl_dataset_disown(ds, dsflags, FTAG);
1358 else
1359 dsl_dataset_rele_flags(ds, dsflags, FTAG);
1361 dsl_pool_rele(dp, FTAG);
1362 return (err);
1364 err = dmu_send_impl(FTAG, dp, ds, &zb, is_clone,
1365 embedok, large_block_ok, compressok, rawok,
1366 outfd, resumeobj, resumeoff, vp, off);
1367 } else {
1368 err = dmu_send_impl(FTAG, dp, ds, NULL, B_FALSE,
1369 embedok, large_block_ok, compressok, rawok,
1370 outfd, resumeobj, resumeoff, vp, off);
1372 if (owned)
1373 dsl_dataset_disown(ds, dsflags, FTAG);
1374 else
1375 dsl_dataset_rele_flags(ds, dsflags, FTAG);
1377 return (err);
1380 static int
1381 dmu_adjust_send_estimate_for_indirects(dsl_dataset_t *ds, uint64_t uncompressed,
1382 uint64_t compressed, boolean_t stream_compressed, uint64_t *sizep)
1384 int err = 0;
1385 uint64_t size;
1387 * Assume that space (both on-disk and in-stream) is dominated by
1388 * data. We will adjust for indirect blocks and the copies property,
1389 * but ignore per-object space used (eg, dnodes and DRR_OBJECT records).
1391 uint64_t recordsize;
1392 uint64_t record_count;
1393 objset_t *os;
1394 VERIFY0(dmu_objset_from_ds(ds, &os));
1396 /* Assume all (uncompressed) blocks are recordsize. */
1397 if (zfs_override_estimate_recordsize != 0) {
1398 recordsize = zfs_override_estimate_recordsize;
1399 } else if (os->os_phys->os_type == DMU_OST_ZVOL) {
1400 err = dsl_prop_get_int_ds(ds,
1401 zfs_prop_to_name(ZFS_PROP_VOLBLOCKSIZE), &recordsize);
1402 } else {
1403 err = dsl_prop_get_int_ds(ds,
1404 zfs_prop_to_name(ZFS_PROP_RECORDSIZE), &recordsize);
1406 if (err != 0)
1407 return (err);
1408 record_count = uncompressed / recordsize;
1411 * If we're estimating a send size for a compressed stream, use the
1412 * compressed data size to estimate the stream size. Otherwise, use the
1413 * uncompressed data size.
1415 size = stream_compressed ? compressed : uncompressed;
1418 * Subtract out approximate space used by indirect blocks.
1419 * Assume most space is used by data blocks (non-indirect, non-dnode).
1420 * Assume no ditto blocks or internal fragmentation.
1422 * Therefore, space used by indirect blocks is sizeof(blkptr_t) per
1423 * block.
1425 size -= record_count * sizeof (blkptr_t);
1427 /* Add in the space for the record associated with each block. */
1428 size += record_count * sizeof (dmu_replay_record_t);
1430 *sizep = size;
1432 return (0);
1436 dmu_send_estimate(dsl_dataset_t *ds, dsl_dataset_t *fromds,
1437 boolean_t stream_compressed, uint64_t *sizep)
1439 dsl_pool_t *dp = ds->ds_dir->dd_pool;
1440 int err;
1441 uint64_t uncomp, comp;
1443 ASSERT(dsl_pool_config_held(dp));
1445 /* tosnap must be a snapshot */
1446 if (!ds->ds_is_snapshot)
1447 return (SET_ERROR(EINVAL));
1449 /* fromsnap, if provided, must be a snapshot */
1450 if (fromds != NULL && !fromds->ds_is_snapshot)
1451 return (SET_ERROR(EINVAL));
1454 * fromsnap must be an earlier snapshot from the same fs as tosnap,
1455 * or the origin's fs.
1457 if (fromds != NULL && !dsl_dataset_is_before(ds, fromds, 0))
1458 return (SET_ERROR(EXDEV));
1460 /* Get compressed and uncompressed size estimates of changed data. */
1461 if (fromds == NULL) {
1462 uncomp = dsl_dataset_phys(ds)->ds_uncompressed_bytes;
1463 comp = dsl_dataset_phys(ds)->ds_compressed_bytes;
1464 } else {
1465 uint64_t used;
1466 err = dsl_dataset_space_written(fromds, ds,
1467 &used, &comp, &uncomp);
1468 if (err != 0)
1469 return (err);
1472 err = dmu_adjust_send_estimate_for_indirects(ds, uncomp, comp,
1473 stream_compressed, sizep);
1475 * Add the size of the BEGIN and END records to the estimate.
1477 *sizep += 2 * sizeof (dmu_replay_record_t);
1478 return (err);
1481 struct calculate_send_arg {
1482 uint64_t uncompressed;
1483 uint64_t compressed;
1487 * Simple callback used to traverse the blocks of a snapshot and sum their
1488 * uncompressed and compressed sizes.
1490 /* ARGSUSED */
1491 static int
1492 dmu_calculate_send_traversal(spa_t *spa, zilog_t *zilog, const blkptr_t *bp,
1493 const zbookmark_phys_t *zb, const dnode_phys_t *dnp, void *arg)
1495 struct calculate_send_arg *space = arg;
1496 if (bp != NULL && !BP_IS_HOLE(bp)) {
1497 space->uncompressed += BP_GET_UCSIZE(bp);
1498 space->compressed += BP_GET_PSIZE(bp);
1500 return (0);
1504 * Given a desination snapshot and a TXG, calculate the approximate size of a
1505 * send stream sent from that TXG. from_txg may be zero, indicating that the
1506 * whole snapshot will be sent.
1509 dmu_send_estimate_from_txg(dsl_dataset_t *ds, uint64_t from_txg,
1510 boolean_t stream_compressed, uint64_t *sizep)
1512 dsl_pool_t *dp = ds->ds_dir->dd_pool;
1513 int err;
1514 struct calculate_send_arg size = { 0 };
1516 ASSERT(dsl_pool_config_held(dp));
1518 /* tosnap must be a snapshot */
1519 if (!ds->ds_is_snapshot)
1520 return (SET_ERROR(EINVAL));
1522 /* verify that from_txg is before the provided snapshot was taken */
1523 if (from_txg >= dsl_dataset_phys(ds)->ds_creation_txg) {
1524 return (SET_ERROR(EXDEV));
1528 * traverse the blocks of the snapshot with birth times after
1529 * from_txg, summing their uncompressed size
1531 err = traverse_dataset(ds, from_txg,
1532 TRAVERSE_POST | TRAVERSE_NO_DECRYPT,
1533 dmu_calculate_send_traversal, &size);
1534 if (err)
1535 return (err);
1537 err = dmu_adjust_send_estimate_for_indirects(ds, size.uncompressed,
1538 size.compressed, stream_compressed, sizep);
1539 return (err);