2 * Copyright (c) 1998 Michael Smith <msmith@freebsd.org>
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26 * $FreeBSD: src/sys/boot/common/bcache.c,v 1.12 2003/08/25 23:30:41 obrien Exp $
27 * $DragonFly: src/sys/boot/common/bcache.c,v 1.3 2003/11/10 06:08:31 dillon Exp $
31 * Simple LRU block cache
36 #include <bitstring.h>
38 #include "bootstrap.h"
40 /* #define BCACHE_DEBUG */
43 #define BCACHE_TIMEOUT 10
44 # define DEBUG(fmt, args...) printf("%s: " fmt "\n" , __func__ , ## args)
46 #define BCACHE_TIMEOUT 2
47 # define DEBUG(fmt, args...)
58 static struct bcachectl
*bcache_ctl
;
59 static caddr_t bcache_data
;
60 static bitstr_t
*bcache_miss
;
61 static u_int bcache_nblks
;
62 static u_int bcache_blksize
;
63 static u_int bcache_hits
, bcache_misses
, bcache_ops
, bcache_bypasses
;
64 static u_int bcache_flushes
;
65 static u_int bcache_bcount
;
67 static void bcache_invalidate(daddr_t blkno
);
68 static void bcache_insert(caddr_t buf
, daddr_t blkno
);
69 static int bcache_lookup(caddr_t buf
, daddr_t blkno
);
72 * Initialise the cache for (nblks) of (bsize).
75 bcache_init(u_int nblks
, size_t bsize
)
77 /* discard any old contents */
78 if (bcache_data
!= NULL
) {
84 /* Allocate control structures */
86 bcache_blksize
= bsize
;
87 bcache_data
= malloc(bcache_nblks
* bcache_blksize
);
88 bcache_ctl
= (struct bcachectl
*)malloc(bcache_nblks
* sizeof(struct bcachectl
));
89 bcache_miss
= bit_alloc((bcache_nblks
+ 1) / 2);
90 if ((bcache_data
== NULL
) || (bcache_ctl
== NULL
) || (bcache_miss
== NULL
)) {
114 /* Flush the cache */
115 for (i
= 0; i
< bcache_nblks
; i
++) {
116 bcache_ctl
[i
].bc_count
= -1;
117 bcache_ctl
[i
].bc_blkno
= -1;
122 * Handle a write request; write directly to the disk, and populate the
123 * cache with the new values.
126 write_strategy(void *devdata
, int unit
, int rw
, daddr_t blk
, size_t size
,
127 char *buf
, size_t *rsize
)
129 struct bcache_devdata
*dd
= (struct bcache_devdata
*)devdata
;
133 nblk
= size
/ bcache_blksize
;
135 /* Invalidate the blocks being written */
136 for (i
= 0; i
< nblk
; i
++) {
137 bcache_invalidate(blk
+ i
);
140 /* Write the blocks */
141 err
= dd
->dv_strategy(dd
->dv_devdata
, rw
, blk
, size
, buf
, rsize
);
143 /* Populate the block cache with the new data */
145 for (i
= 0; i
< nblk
; i
++) {
146 bcache_insert(buf
+ (i
* bcache_blksize
),blk
+ i
);
154 * Handle a read request; fill in parts of the request that can
155 * be satisfied by the cache, use the supplied strategy routine to do
156 * device I/O and then use the I/O results to populate the cache.
159 read_strategy(void *devdata
, int unit
, int rw
, daddr_t blk
, size_t size
,
160 char *buf
, size_t *rsize
)
162 struct bcache_devdata
*dd
= (struct bcache_devdata
*)devdata
;
164 daddr_t p_blk
, i
, j
, nblk
;
167 nblk
= size
/ bcache_blksize
;
170 /* Satisfy any cache hits up front */
171 for (i
= 0; i
< nblk
; i
++) {
172 if (bcache_lookup(buf
+ (bcache_blksize
* i
), blk
+ i
)) {
173 bit_set(bcache_miss
, i
); /* cache miss */
176 bit_clear(bcache_miss
, i
); /* cache hit */
181 /* Go back and fill in any misses XXX optimise */
185 for (i
= 0; i
< nblk
; i
++) {
186 if (bit_test(bcache_miss
, i
)) {
187 /* miss, add to pending transfer */
190 p_buf
= buf
+ (bcache_blksize
* i
);
195 } else if (p_blk
!= -1) {
196 /* hit, complete pending transfer */
197 result
= dd
->dv_strategy(dd
->dv_devdata
, rw
, p_blk
, p_size
* bcache_blksize
, p_buf
, NULL
);
200 for (j
= 0; j
< p_size
; j
++)
201 bcache_insert(p_buf
+ (j
* bcache_blksize
), p_blk
+ j
);
206 /* pending transfer left */
207 result
= dd
->dv_strategy(dd
->dv_devdata
, rw
, p_blk
, p_size
* bcache_blksize
, p_buf
, NULL
);
210 for (j
= 0; j
< p_size
; j
++)
211 bcache_insert(p_buf
+ (j
* bcache_blksize
), p_blk
+ j
);
215 if ((result
== 0) && (rsize
!= NULL
))
221 * Requests larger than 1/2 the cache size will be bypassed and go
222 * directly to the disk. XXX tune this.
225 bcache_strategy(void *devdata
, int unit
, int rw
, daddr_t blk
, size_t size
,
226 char *buf
, size_t *rsize
)
228 static int bcache_unit
= -1;
229 struct bcache_devdata
*dd
= (struct bcache_devdata
*)devdata
;
233 if(bcache_unit
!= unit
) {
238 /* bypass large requests, or when the cache is inactive */
239 if ((bcache_data
== NULL
) || ((size
* 2 / bcache_blksize
) > bcache_nblks
)) {
240 DEBUG("bypass %d from %d", size
/ bcache_blksize
, blk
);
242 return(dd
->dv_strategy(dd
->dv_devdata
, rw
, blk
, size
, buf
, rsize
));
247 return read_strategy(devdata
, unit
, rw
, blk
, size
, buf
, rsize
);
249 return write_strategy(devdata
, unit
, rw
, blk
, size
, buf
, rsize
);
256 * Insert a block into the cache. Retire the oldest block to do so, if required.
258 * XXX the LRU algorithm will fail after 2^31 blocks have been transferred.
261 bcache_insert(caddr_t buf
, daddr_t blkno
)
268 cand
= 0; /* assume the first block */
269 ocount
= bcache_ctl
[0].bc_count
;
271 /* find the oldest block */
272 for (i
= 1; i
< bcache_nblks
; i
++) {
273 if (bcache_ctl
[i
].bc_blkno
== blkno
) {
274 /* reuse old entry */
278 if (bcache_ctl
[i
].bc_count
< ocount
) {
279 ocount
= bcache_ctl
[i
].bc_count
;
284 DEBUG("insert blk %d -> %d @ %d # %d", blkno
, cand
, now
, bcache_bcount
);
285 bcopy(buf
, bcache_data
+ (bcache_blksize
* cand
), bcache_blksize
);
286 bcache_ctl
[cand
].bc_blkno
= blkno
;
287 bcache_ctl
[cand
].bc_stamp
= now
;
288 bcache_ctl
[cand
].bc_count
= bcache_bcount
++;
292 * Look for a block in the cache. Blocks more than BCACHE_TIMEOUT seconds old
293 * may be stale (removable media) and thus are discarded. Copy the block out
294 * if successful and return zero, or return nonzero on failure.
297 bcache_lookup(caddr_t buf
, daddr_t blkno
)
304 for (i
= 0; i
< bcache_nblks
; i
++)
306 if ((bcache_ctl
[i
].bc_blkno
== blkno
) && ((bcache_ctl
[i
].bc_stamp
+ BCACHE_TIMEOUT
) >= now
)) {
307 bcopy(bcache_data
+ (bcache_blksize
* i
), buf
, bcache_blksize
);
308 DEBUG("hit blk %d <- %d (now %d then %d)", blkno
, i
, now
, bcache_ctl
[i
].bc_stamp
);
315 * Invalidate a block from the cache.
318 bcache_invalidate(daddr_t blkno
)
322 for (i
= 0; i
< bcache_nblks
; i
++) {
323 if (bcache_ctl
[i
].bc_blkno
== blkno
) {
324 bcache_ctl
[i
].bc_count
= -1;
325 bcache_ctl
[i
].bc_blkno
= -1;
326 DEBUG("invalidate blk %d", blkno
);
332 COMMAND_SET(bcachestat
, "bcachestat", "get disk block cache stats", command_bcache
);
335 command_bcache(int argc
, char *argv
[])
339 for (i
= 0; i
< bcache_nblks
; i
++) {
340 printf("%08x %04x %04x|", bcache_ctl
[i
].bc_blkno
, (unsigned int)bcache_ctl
[i
].bc_stamp
& 0xffff, bcache_ctl
[i
].bc_count
& 0xffff);
341 if (((i
+ 1) % 4) == 0)
344 printf("\n%d ops %d bypasses %d hits %d misses %d flushes\n", bcache_ops
, bcache_bypasses
, bcache_hits
, bcache_misses
, bcache_flushes
);