rtld - do not allow both dynamic DTV index and static TLS offset
[dragonfly.git] / sys / kern / subr_diskiocom.c
blobd873dfd9577649fe209d5671d668aed6da94a0ae
1 /*
2 * Copyright (c) 2012 The DragonFly Project. All rights reserved.
4 * This code is derived from software contributed to The DragonFly Project
5 * by Matthew Dillon <dillon@backplane.com>
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in
15 * the documentation and/or other materials provided with the
16 * distribution.
17 * 3. Neither the name of The DragonFly Project nor the names of its
18 * contributors may be used to endorse or promote products derived
19 * from this software without specific, prior written permission.
21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
24 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
25 * COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
26 * INCIDENTAL, SPECIAL, EXEMPLARY OR CONSEQUENTIAL DAMAGES (INCLUDING,
27 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
28 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
29 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
30 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
31 * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
32 * SUCH DAMAGE.
34 #include <sys/param.h>
35 #include <sys/systm.h>
36 #include <sys/kernel.h>
37 #include <sys/proc.h>
38 #include <sys/sysctl.h>
39 #include <sys/buf.h>
40 #include <sys/conf.h>
41 #include <sys/disklabel.h>
42 #include <sys/disklabel32.h>
43 #include <sys/disklabel64.h>
44 #include <sys/diskslice.h>
45 #include <sys/diskmbr.h>
46 #include <sys/disk.h>
47 #include <sys/malloc.h>
48 #include <sys/device.h>
49 #include <sys/devfs.h>
50 #include <sys/thread.h>
51 #include <sys/queue.h>
52 #include <sys/lock.h>
53 #include <sys/stat.h>
54 #include <sys/uuid.h>
55 #include <sys/dmsg.h>
57 #include <sys/buf2.h>
58 #include <sys/msgport2.h>
60 struct dios_open {
61 int openrd;
62 int openwr;
65 struct dios_io {
66 int count;
67 int eof;
68 kdmsg_data_t data;
71 static MALLOC_DEFINE(M_DMSG_DISK, "dmsg_disk", "disk dmsg");
73 static int blk_active;
74 SYSCTL_INT(_debug, OID_AUTO, blk_active, CTLFLAG_RW, &blk_active, 0,
75 "Number of active iocom IOs");
77 static int disk_iocom_reconnect(struct disk *dp, struct file *fp);
78 static int disk_rcvdmsg(kdmsg_msg_t *msg);
80 static void disk_blk_open(struct disk *dp, kdmsg_msg_t *msg);
81 static void disk_blk_read(struct disk *dp, kdmsg_msg_t *msg);
82 static void disk_blk_write(struct disk *dp, kdmsg_msg_t *msg);
83 static void disk_blk_flush(struct disk *dp, kdmsg_msg_t *msg);
84 static void disk_blk_freeblks(struct disk *dp, kdmsg_msg_t *msg);
85 static void diskiodone(struct bio *bio);
87 void
88 disk_iocom_init(struct disk *dp)
90 kdmsg_iocom_init(&dp->d_iocom, dp,
91 KDMSG_IOCOMF_AUTOCONN |
92 KDMSG_IOCOMF_AUTORXSPAN |
93 KDMSG_IOCOMF_AUTOTXSPAN,
94 M_DMSG_DISK, disk_rcvdmsg);
97 void
98 disk_iocom_update(struct disk *dp)
102 void
103 disk_iocom_uninit(struct disk *dp)
105 kdmsg_iocom_uninit(&dp->d_iocom);
109 disk_iocom_ioctl(struct disk *dp, u_long cmd, void *data)
111 struct file *fp;
112 struct disk_ioc_recluster *recl;
113 int error;
115 switch(cmd) {
116 case DIOCRECLUSTER:
117 recl = data;
118 fp = holdfp(curthread, recl->fd, -1);
119 if (fp) {
120 error = disk_iocom_reconnect(dp, fp);
121 } else {
122 error = EINVAL;
124 break;
125 default:
126 error = EOPNOTSUPP;
127 break;
129 return error;
132 static
134 disk_iocom_reconnect(struct disk *dp, struct file *fp)
136 char devname[64];
138 ksnprintf(devname, sizeof(devname), "%s%d",
139 dev_dname(dp->d_rawdev), dkunit(dp->d_rawdev));
141 kdmsg_iocom_reconnect(&dp->d_iocom, fp, devname);
143 dp->d_iocom.auto_lnk_conn.proto_version = DMSG_SPAN_PROTO_1;
144 dp->d_iocom.auto_lnk_conn.peer_type = DMSG_PEER_BLOCK;
145 dp->d_iocom.auto_lnk_conn.peer_mask = 1LLU << DMSG_PEER_BLOCK;
146 dp->d_iocom.auto_lnk_conn.peer_mask = (uint64_t)-1;
147 #if 0
148 if (dp->d_info.d_serialno) {
149 ksnprintf(dp->d_iocom.auto_lnk_conn.peer_label,
150 sizeof(dp->d_iocom.auto_lnk_conn.peer_label),
151 "%s/%s", hostname, dp->d_info.d_serialno);
152 } else {
153 ksnprintf(dp->d_iocom.auto_lnk_conn.peer_label,
154 sizeof(dp->d_iocom.auto_lnk_conn.peer_label),
155 "%s/%s", hostname, devname);
157 #endif
158 ksnprintf(dp->d_iocom.auto_lnk_conn.peer_label,
159 sizeof(dp->d_iocom.auto_lnk_conn.peer_label),
160 "%s/%s", hostname, devname);
162 dp->d_iocom.auto_lnk_span.proto_version = DMSG_SPAN_PROTO_1;
163 dp->d_iocom.auto_lnk_span.peer_type = DMSG_PEER_BLOCK;
164 dp->d_iocom.auto_lnk_span.media.block.bytes =
165 dp->d_info.d_media_size;
166 dp->d_iocom.auto_lnk_span.media.block.blksize =
167 dp->d_info.d_media_blksize;
168 ksnprintf(dp->d_iocom.auto_lnk_span.peer_label,
169 sizeof(dp->d_iocom.auto_lnk_span.peer_label),
170 "%s", dp->d_iocom.auto_lnk_conn.peer_label);
171 if (dp->d_info.d_serialno) {
172 ksnprintf(dp->d_iocom.auto_lnk_span.pfs_label,
173 sizeof(dp->d_iocom.auto_lnk_span.pfs_label),
174 "%s", dp->d_info.d_serialno);
175 } else {
177 * If no serial number is available generate a dummy serial
178 * number from the host and device name and pray. This will
179 * allow e.g. /dev/vn* to look meaningful on a remote machine.
181 ksnprintf(dp->d_iocom.auto_lnk_span.pfs_label,
182 sizeof(dp->d_iocom.auto_lnk_span.pfs_label),
183 "%s.%s", hostname, devname);
186 kdmsg_iocom_autoinitiate(&dp->d_iocom, NULL);
188 return (0);
191 static int
192 disk_rcvdmsg(kdmsg_msg_t *msg)
194 struct disk *dp = msg->state->iocom->handle;
197 * Handle debug messages (these might not be in transactions)
199 switch(msg->any.head.cmd & DMSGF_CMDSWMASK) {
200 case DMSG_DBG_SHELL:
202 * Execute shell command (not supported atm)
204 kdmsg_msg_reply(msg, DMSG_ERR_NOSUPP);
205 return(0);
206 case DMSG_DBG_SHELL | DMSGF_REPLY:
207 if (msg->aux_data) {
208 msg->aux_data[msg->aux_size - 1] = 0;
209 kprintf("diskiocom: DEBUGMSG: %s\n", msg->aux_data);
211 return(0);
215 * All remaining messages must be in a transaction.
217 * NOTE! We currently don't care if the transaction is just
218 * the span transaction (for disk probes) or if it is the
219 * BLK_OPEN transaction.
221 * NOTE! We are switching on the first message's command. The
222 * actual message command within the transaction may be
223 * different (if streaming within a transaction).
225 if (msg->state == &msg->state->iocom->state0) {
226 kdmsg_msg_reply(msg, DMSG_ERR_NOSUPP);
227 return(0);
230 switch(msg->state->rxcmd & DMSGF_CMDSWMASK) {
231 case DMSG_BLK_OPEN:
232 disk_blk_open(dp, msg);
233 break;
234 case DMSG_BLK_READ:
236 * not reached normally but leave in for completeness
238 disk_blk_read(dp, msg);
239 break;
240 case DMSG_BLK_WRITE:
241 disk_blk_write(dp, msg);
242 break;
243 case DMSG_BLK_FLUSH:
244 disk_blk_flush(dp, msg);
245 break;
246 case DMSG_BLK_FREEBLKS:
247 disk_blk_freeblks(dp, msg);
248 break;
249 default:
250 if ((msg->any.head.cmd & DMSGF_REPLY) == 0) {
251 if (msg->any.head.cmd & DMSGF_DELETE)
252 kdmsg_msg_reply(msg, DMSG_ERR_NOSUPP);
253 else
254 kdmsg_msg_result(msg, DMSG_ERR_NOSUPP);
256 break;
258 return (0);
261 static
262 void
263 disk_blk_open(struct disk *dp, kdmsg_msg_t *msg)
265 struct dios_open *openst;
266 int error = DMSG_ERR_NOSUPP;
267 int fflags;
269 openst = msg->state->any.any;
270 if ((msg->any.head.cmd & DMSGF_CMDSWMASK) == DMSG_BLK_OPEN) {
271 if (openst == NULL) {
272 openst = kmalloc(sizeof(*openst), M_DEVBUF,
273 M_WAITOK | M_ZERO);
274 msg->state->any.any = openst;
276 fflags = 0;
277 if (msg->any.blk_open.modes & DMSG_BLKOPEN_RD)
278 fflags = FREAD;
279 if (msg->any.blk_open.modes & DMSG_BLKOPEN_WR)
280 fflags |= FWRITE;
281 error = dev_dopen(dp->d_rawdev, fflags, S_IFCHR,
282 proc0.p_ucred, NULL, NULL);
283 if (error) {
284 error = DMSG_ERR_IO;
285 } else {
286 if (msg->any.blk_open.modes & DMSG_BLKOPEN_RD)
287 ++openst->openrd;
288 if (msg->any.blk_open.modes & DMSG_BLKOPEN_WR)
289 ++openst->openwr;
292 #if 0
293 if ((msg->any.head.cmd & DMSGF_CMDSWMASK) == DMSG_BLK_CLOSE &&
294 openst) {
295 fflags = 0;
296 if ((msg->any.blk_open.modes & DMSG_BLKOPEN_RD) &&
297 openst->openrd) {
298 fflags = FREAD;
300 if ((msg->any.blk_open.modes & DMSG_BLKOPEN_WR) &&
301 openst->openwr) {
302 fflags |= FWRITE;
304 error = dev_dclose(dp->d_rawdev, fflags, S_IFCHR, NULL);
305 if (error) {
306 error = DMSG_ERR_IO;
307 } else {
308 if (msg->any.blk_open.modes & DMSG_BLKOPEN_RD)
309 --openst->openrd;
310 if (msg->any.blk_open.modes & DMSG_BLKOPEN_WR)
311 --openst->openwr;
314 #endif
315 if (msg->any.head.cmd & DMSGF_DELETE) {
316 if (openst) {
317 while (openst->openrd && openst->openwr) {
318 --openst->openrd;
319 --openst->openwr;
320 dev_dclose(dp->d_rawdev, FREAD|FWRITE, S_IFCHR, NULL);
322 while (openst->openrd) {
323 --openst->openrd;
324 dev_dclose(dp->d_rawdev, FREAD, S_IFCHR, NULL);
326 while (openst->openwr) {
327 --openst->openwr;
328 dev_dclose(dp->d_rawdev, FWRITE, S_IFCHR, NULL);
330 kfree(openst, M_DEVBUF);
331 msg->state->any.any = NULL;
333 kdmsg_msg_reply(msg, error);
334 } else {
335 kdmsg_msg_result(msg, error);
339 static
340 void
341 disk_blk_read(struct disk *dp, kdmsg_msg_t *msg)
343 struct dios_io *iost;
344 struct buf *bp;
345 struct bio *bio;
346 int error = DMSG_ERR_NOSUPP;
347 int reterr = 1;
350 * Only DMSG_BLK_READ commands imply read ops.
352 iost = msg->state->any.any;
353 if ((msg->any.head.cmd & DMSGF_CMDSWMASK) == DMSG_BLK_READ) {
354 if (msg->any.blk_read.bytes < DEV_BSIZE ||
355 msg->any.blk_read.bytes > MAXPHYS) {
356 error = DMSG_ERR_PARAM;
357 goto done;
359 if (iost == NULL) {
360 iost = kmalloc(sizeof(*iost), M_DEVBUF,
361 M_WAITOK | M_ZERO);
362 msg->state->any.any = iost;
364 reterr = 0;
365 bp = getpbuf_mem(NULL);
366 KKASSERT(msg->any.blk_read.bytes <= bp->b_bufsize);
367 bio = &bp->b_bio1;
368 bp->b_cmd = BUF_CMD_READ;
369 bp->b_bcount = msg->any.blk_read.bytes;
370 bp->b_resid = bp->b_bcount;
371 bio->bio_offset = msg->any.blk_read.offset;
372 bio->bio_caller_info1.ptr = msg->state;
373 bio->bio_done = diskiodone;
375 /* kdmsg_state_hold(msg->state); */
376 atomic_add_int(&blk_active, 1);
377 atomic_add_int(&iost->count, 1);
378 if (msg->any.head.cmd & DMSGF_DELETE)
379 iost->eof = 1;
380 BUF_KERNPROC(bp);
381 dev_dstrategy(dp->d_rawdev, bio);
383 done:
384 if (reterr) {
385 if (msg->any.head.cmd & DMSGF_DELETE) {
386 if (iost && iost->count == 0) {
387 kfree(iost, M_DEVBUF);
388 msg->state->any.any = NULL;
390 kdmsg_msg_reply(msg, error);
391 } else {
392 kdmsg_msg_result(msg, error);
397 static
398 void
399 disk_blk_write(struct disk *dp, kdmsg_msg_t *msg)
401 struct dios_io *iost;
402 struct buf *bp;
403 struct bio *bio;
404 int error = DMSG_ERR_NOSUPP;
405 int reterr = 1;
408 * Only DMSG_BLK_WRITE commands imply read ops.
410 iost = msg->state->any.any;
411 if ((msg->any.head.cmd & DMSGF_CMDSWMASK) == DMSG_BLK_WRITE) {
412 if (msg->any.blk_write.bytes < DEV_BSIZE ||
413 msg->any.blk_write.bytes > MAXPHYS) {
414 error = DMSG_ERR_PARAM;
415 goto done;
417 if (iost == NULL) {
418 iost = kmalloc(sizeof(*iost), M_DEVBUF,
419 M_WAITOK | M_ZERO);
420 msg->state->any.any = iost;
424 * Issue WRITE. Short data implies zeros. Try to optimize
425 * the buffer cache buffer for the case where we can just
426 * use the message's data pointer.
428 reterr = 0;
429 if (msg->aux_size >= msg->any.blk_write.bytes)
430 bp = getpbuf(NULL);
431 else
432 bp = getpbuf_mem(NULL);
433 KKASSERT(msg->any.blk_write.bytes <= bp->b_bufsize);
434 bio = &bp->b_bio1;
435 bp->b_cmd = BUF_CMD_WRITE;
436 bp->b_bcount = msg->any.blk_write.bytes;
437 bp->b_resid = bp->b_bcount;
438 if (msg->aux_size >= msg->any.blk_write.bytes) {
439 bp->b_data = msg->aux_data;
440 kdmsg_detach_aux_data(msg, &iost->data);
441 } else {
442 bcopy(msg->aux_data, bp->b_data, msg->aux_size);
443 bzero(bp->b_data + msg->aux_size,
444 msg->any.blk_write.bytes - msg->aux_size);
445 bzero(&iost->data, sizeof(iost->data));
447 bio->bio_offset = msg->any.blk_write.offset;
448 bio->bio_caller_info1.ptr = msg->state;
449 bio->bio_done = diskiodone;
451 /* kdmsg_state_hold(msg->state); */
452 atomic_add_int(&blk_active, 1);
453 atomic_add_int(&iost->count, 1);
454 if (msg->any.head.cmd & DMSGF_DELETE)
455 iost->eof = 1;
456 BUF_KERNPROC(bp);
457 dev_dstrategy(dp->d_rawdev, bio);
459 done:
460 if (reterr) {
461 if (msg->any.head.cmd & DMSGF_DELETE) {
462 if (iost && iost->count == 0) {
463 kfree(iost, M_DEVBUF);
464 msg->state->any.any = NULL;
466 kdmsg_msg_reply(msg, error);
467 } else {
468 kdmsg_msg_result(msg, error);
473 static
474 void
475 disk_blk_flush(struct disk *dp, kdmsg_msg_t *msg)
477 struct dios_io *iost;
478 struct buf *bp;
479 struct bio *bio;
480 int error = DMSG_ERR_NOSUPP;
481 int reterr = 1;
484 * Only DMSG_BLK_FLUSH commands imply read ops.
486 iost = msg->state->any.any;
487 if ((msg->any.head.cmd & DMSGF_CMDSWMASK) == DMSG_BLK_FLUSH) {
488 if (iost == NULL) {
489 iost = kmalloc(sizeof(*iost), M_DEVBUF,
490 M_WAITOK | M_ZERO);
491 msg->state->any.any = iost;
493 reterr = 0;
494 bp = getpbuf(NULL);
495 bio = &bp->b_bio1;
496 bp->b_cmd = BUF_CMD_FLUSH;
497 bp->b_bcount = msg->any.blk_flush.bytes;
498 bp->b_resid = 0;
499 bio->bio_offset = msg->any.blk_flush.offset;
500 bio->bio_caller_info1.ptr = msg->state;
501 bio->bio_done = diskiodone;
503 /* kdmsg_state_hold(msg->state); */
504 atomic_add_int(&blk_active, 1);
505 atomic_add_int(&iost->count, 1);
506 if (msg->any.head.cmd & DMSGF_DELETE)
507 iost->eof = 1;
508 BUF_KERNPROC(bp);
509 dev_dstrategy(dp->d_rawdev, bio);
511 if (reterr) {
512 if (msg->any.head.cmd & DMSGF_DELETE) {
513 if (iost && iost->count == 0) {
514 kfree(iost, M_DEVBUF);
515 msg->state->any.any = NULL;
517 kdmsg_msg_reply(msg, error);
518 } else {
519 kdmsg_msg_result(msg, error);
524 static
525 void
526 disk_blk_freeblks(struct disk *dp, kdmsg_msg_t *msg)
528 struct dios_io *iost;
529 struct buf *bp;
530 struct bio *bio;
531 int error = DMSG_ERR_NOSUPP;
532 int reterr = 1;
535 * Only DMSG_BLK_FREEBLKS commands imply read ops.
537 iost = msg->state->any.any;
538 if ((msg->any.head.cmd & DMSGF_CMDSWMASK) == DMSG_BLK_FREEBLKS) {
539 if (iost == NULL) {
540 iost = kmalloc(sizeof(*iost), M_DEVBUF,
541 M_WAITOK | M_ZERO);
542 msg->state->any.any = iost;
544 reterr = 0;
545 bp = getpbuf(NULL);
546 bio = &bp->b_bio1;
547 bp->b_cmd = BUF_CMD_FREEBLKS;
548 bp->b_bcount = msg->any.blk_freeblks.bytes;
549 bp->b_resid = 0;
550 bio->bio_offset = msg->any.blk_freeblks.offset;
551 bio->bio_caller_info1.ptr = msg->state;
552 bio->bio_done = diskiodone;
554 /* kdmsg_state_hold(msg->state); */
555 atomic_add_int(&blk_active, 1);
556 atomic_add_int(&iost->count, 1);
557 if (msg->any.head.cmd & DMSGF_DELETE)
558 iost->eof = 1;
559 BUF_KERNPROC(bp);
560 dev_dstrategy(dp->d_rawdev, bio);
562 if (reterr) {
563 if (msg->any.head.cmd & DMSGF_DELETE) {
564 if (iost && iost->count == 0) {
565 kfree(iost, M_DEVBUF);
566 msg->state->any.any = NULL;
568 kdmsg_msg_reply(msg, error);
569 } else {
570 kdmsg_msg_result(msg, error);
575 static
576 void
577 diskiodone(struct bio *bio)
579 struct buf *bp = bio->bio_buf;
580 kdmsg_state_t *state = bio->bio_caller_info1.ptr;
581 kdmsg_msg_t *rmsg;
582 struct dios_io *iost = state->any.any;
583 int error;
584 int resid = 0;
585 int bytes;
586 uint32_t cmd;
587 void *data;
589 cmd = DMSG_LNK_ERROR;
590 data = NULL;
591 bytes = 0;
593 switch(bp->b_cmd) {
594 case BUF_CMD_READ:
595 cmd = DMSG_LNK_ERROR;
596 data = bp->b_data;
597 bytes = bp->b_bcount;
598 /* fall through */
599 case BUF_CMD_WRITE:
600 if (bp->b_flags & B_ERROR) {
601 error = bp->b_error;
602 } else {
603 error = 0;
604 resid = bp->b_resid;
606 kdmsg_free_aux_data(&iost->data);
607 break;
608 case BUF_CMD_FLUSH:
609 case BUF_CMD_FREEBLKS:
610 if (bp->b_flags & B_ERROR)
611 error = bp->b_error;
612 else
613 error = 0;
614 break;
615 default:
616 panic("diskiodone: Unknown bio cmd = %d\n",
617 bio->bio_buf->b_cmd);
618 error = 0; /* avoid compiler warning */
619 break; /* NOT REACHED */
623 * Convert error to DMSG_ERR_* code.
625 if (error)
626 error = DMSG_ERR_IO;
629 * Convert LNK_ERROR or BLK_ERROR if non-zero resid. READS will
630 * have already converted cmd to BLK_ERROR and set up data to return.
632 if (resid && cmd == DMSG_LNK_ERROR)
633 cmd = DMSG_BLK_ERROR;
634 /* XXX txcmd is delayed so this won't work for streaming */
635 if ((state->txcmd & DMSGF_CREATE) == 0) /* assume serialized */
636 cmd |= DMSGF_CREATE;
637 if (iost->eof) {
638 if (atomic_fetchadd_int(&iost->count, -1) == 1)
639 cmd |= DMSGF_DELETE;
640 } else {
641 atomic_add_int(&iost->count, -1);
643 atomic_add_int(&blk_active, -1);
644 cmd |= DMSGF_REPLY;
647 * Allocate a basic or extended reply. Be careful not to populate
648 * extended header fields unless we allocated an extended reply.
650 rmsg = kdmsg_msg_alloc(state, cmd, NULL, 0);
651 if (data) {
652 rmsg->aux_data = kmalloc(bytes, state->iocom->mmsg, M_INTWAIT);
653 rmsg->aux_size = bytes;
654 rmsg->flags |= KDMSG_FLAG_AUXALLOC;
655 bcopy(data, rmsg->aux_data, bytes);
657 rmsg->any.blk_error.head.error = error;
658 if ((cmd & DMSGF_BASECMDMASK) == DMSG_BLK_ERROR)
659 rmsg->any.blk_error.resid = resid;
660 bio->bio_caller_info1.ptr = NULL;
661 /* kdmsg_state_drop(state); */
662 kdmsg_msg_write(rmsg);
663 if (bp->b_flags & B_PAGING) {
664 relpbuf(bp, NULL);
665 } else {
666 bp->b_flags |= B_INVAL | B_AGE;
667 relpbuf(bp, NULL);