2 linear.c : Multiple Devices driver for Linux
3 Copyright (C) 1994-96 Marc ZYNGIER
4 <zyngier@ufr-info-p7.ibp.fr> or
7 Linear mode management functions.
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 2, or (at your option)
14 You should have received a copy of the GNU General Public License
15 (for example /usr/src/linux/COPYING); if not, write to the Free
16 Software Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
19 #include <linux/module.h>
21 #include <linux/raid/md.h>
22 #include <linux/slab.h>
23 #include <linux/raid/linear.h>
25 #define MAJOR_NR MD_MAJOR
27 #define MD_PERSONALITY
30 * find which device holds a particular offset
32 static inline dev_info_t
*which_dev(mddev_t
*mddev
, sector_t sector
)
35 linear_conf_t
*conf
= mddev_to_conf(mddev
);
36 sector_t block
= sector
>> 1;
39 * sector_div(a,b) returns the remainer and sets a to a/b
41 block
>>= conf
->preshift
;
42 (void)sector_div(block
, conf
->hash_spacing
);
43 hash
= conf
->hash_table
[block
];
45 while ((sector
>>1) >= (hash
->size
+ hash
->offset
))
51 * linear_mergeable_bvec -- tell bio layer if two requests can be merged
53 * @bio: the buffer head that's been built up so far
54 * @biovec: the request that could be merged to it.
56 * Return amount of bytes we can take at this offset
58 static int linear_mergeable_bvec(struct request_queue
*q
, struct bio
*bio
, struct bio_vec
*biovec
)
60 mddev_t
*mddev
= q
->queuedata
;
62 unsigned long maxsectors
, bio_sectors
= bio
->bi_size
>> 9;
63 sector_t sector
= bio
->bi_sector
+ get_start_sect(bio
->bi_bdev
);
65 dev0
= which_dev(mddev
, sector
);
66 maxsectors
= (dev0
->size
<< 1) - (sector
- (dev0
->offset
<<1));
68 if (maxsectors
< bio_sectors
)
71 maxsectors
-= bio_sectors
;
73 if (maxsectors
<= (PAGE_SIZE
>> 9 ) && bio_sectors
== 0)
74 return biovec
->bv_len
;
75 /* The bytes available at this offset could be really big,
76 * so we cap at 2^31 to avoid overflow */
77 if (maxsectors
> (1 << (31-9)))
79 return maxsectors
<< 9;
82 static void linear_unplug(struct request_queue
*q
)
84 mddev_t
*mddev
= q
->queuedata
;
85 linear_conf_t
*conf
= mddev_to_conf(mddev
);
88 for (i
=0; i
< mddev
->raid_disks
; i
++) {
89 struct request_queue
*r_queue
= bdev_get_queue(conf
->disks
[i
].rdev
->bdev
);
94 static int linear_congested(void *data
, int bits
)
96 mddev_t
*mddev
= data
;
97 linear_conf_t
*conf
= mddev_to_conf(mddev
);
100 for (i
= 0; i
< mddev
->raid_disks
&& !ret
; i
++) {
101 struct request_queue
*q
= bdev_get_queue(conf
->disks
[i
].rdev
->bdev
);
102 ret
|= bdi_congested(&q
->backing_dev_info
, bits
);
107 static linear_conf_t
*linear_conf(mddev_t
*mddev
, int raid_disks
)
113 sector_t min_spacing
;
114 sector_t curr_offset
;
115 struct list_head
*tmp
;
117 conf
= kzalloc (sizeof (*conf
) + raid_disks
*sizeof(dev_info_t
),
123 conf
->array_size
= 0;
125 rdev_for_each(rdev
, tmp
, mddev
) {
126 int j
= rdev
->raid_disk
;
127 dev_info_t
*disk
= conf
->disks
+ j
;
129 if (j
< 0 || j
> raid_disks
|| disk
->rdev
) {
130 printk("linear: disk numbering problem. Aborting!\n");
136 blk_queue_stack_limits(mddev
->queue
,
137 rdev
->bdev
->bd_disk
->queue
);
138 /* as we don't honour merge_bvec_fn, we must never risk
139 * violating it, so limit ->max_sector to one PAGE, as
140 * a one page request is never in violation.
142 if (rdev
->bdev
->bd_disk
->queue
->merge_bvec_fn
&&
143 mddev
->queue
->max_sectors
> (PAGE_SIZE
>>9))
144 blk_queue_max_sectors(mddev
->queue
, PAGE_SIZE
>>9);
146 disk
->size
= rdev
->size
;
147 conf
->array_size
+= rdev
->size
;
151 if (cnt
!= raid_disks
) {
152 printk("linear: not enough drives present. Aborting!\n");
156 min_spacing
= conf
->array_size
;
157 sector_div(min_spacing
, PAGE_SIZE
/sizeof(struct dev_info
*));
159 /* min_spacing is the minimum spacing that will fit the hash
160 * table in one PAGE. This may be much smaller than needed.
161 * We find the smallest non-terminal set of consecutive devices
162 * that is larger than min_spacing as use the size of that as
165 conf
->hash_spacing
= conf
->array_size
;
166 for (i
=0; i
< cnt
-1 ; i
++) {
169 for (j
= i
; j
< cnt
- 1 && sz
< min_spacing
; j
++)
170 sz
+= conf
->disks
[j
].size
;
171 if (sz
>= min_spacing
&& sz
< conf
->hash_spacing
)
172 conf
->hash_spacing
= sz
;
175 /* hash_spacing may be too large for sector_div to work with,
176 * so we might need to pre-shift
179 if (sizeof(sector_t
) > sizeof(u32
)) {
180 sector_t space
= conf
->hash_spacing
;
181 while (space
> (sector_t
)(~(u32
)0)) {
187 * This code was restructured to work around a gcc-2.95.3 internal
188 * compiler error. Alter it with care.
195 sz
= conf
->array_size
>> conf
->preshift
;
196 sz
+= 1; /* force round-up */
197 base
= conf
->hash_spacing
>> conf
->preshift
;
198 round
= sector_div(sz
, base
);
199 nb_zone
= sz
+ (round
? 1 : 0);
201 BUG_ON(nb_zone
> PAGE_SIZE
/ sizeof(struct dev_info
*));
203 conf
->hash_table
= kmalloc (sizeof (struct dev_info
*) * nb_zone
,
205 if (!conf
->hash_table
)
209 * Here we generate the linear hash table
210 * First calculate the device offsets.
212 conf
->disks
[0].offset
= 0;
213 for (i
= 1; i
< raid_disks
; i
++)
214 conf
->disks
[i
].offset
=
215 conf
->disks
[i
-1].offset
+
216 conf
->disks
[i
-1].size
;
218 table
= conf
->hash_table
;
221 for (curr_offset
= 0;
222 curr_offset
< conf
->array_size
;
223 curr_offset
+= conf
->hash_spacing
) {
225 while (i
< raid_disks
-1 &&
226 curr_offset
>= conf
->disks
[i
+1].offset
)
229 *table
++ = conf
->disks
+ i
;
232 if (conf
->preshift
) {
233 conf
->hash_spacing
>>= conf
->preshift
;
234 /* round hash_spacing up so that when we divide by it,
235 * we err on the side of "too-low", which is safest.
237 conf
->hash_spacing
++;
240 BUG_ON(table
- conf
->hash_table
> nb_zone
);
249 static int linear_run (mddev_t
*mddev
)
253 mddev
->queue
->queue_lock
= &mddev
->queue
->__queue_lock
;
254 conf
= linear_conf(mddev
, mddev
->raid_disks
);
258 mddev
->private = conf
;
259 mddev
->array_size
= conf
->array_size
;
261 blk_queue_merge_bvec(mddev
->queue
, linear_mergeable_bvec
);
262 mddev
->queue
->unplug_fn
= linear_unplug
;
263 mddev
->queue
->backing_dev_info
.congested_fn
= linear_congested
;
264 mddev
->queue
->backing_dev_info
.congested_data
= mddev
;
268 static int linear_add(mddev_t
*mddev
, mdk_rdev_t
*rdev
)
270 /* Adding a drive to a linear array allows the array to grow.
271 * It is permitted if the new drive has a matching superblock
272 * already on it, with raid_disk equal to raid_disks.
273 * It is achieved by creating a new linear_private_data structure
274 * and swapping it in in-place of the current one.
275 * The current one is never freed until the array is stopped.
278 linear_conf_t
*newconf
;
280 if (rdev
->saved_raid_disk
!= mddev
->raid_disks
)
283 rdev
->raid_disk
= rdev
->saved_raid_disk
;
285 newconf
= linear_conf(mddev
,mddev
->raid_disks
+1);
290 newconf
->prev
= mddev_to_conf(mddev
);
291 mddev
->private = newconf
;
293 mddev
->array_size
= newconf
->array_size
;
294 set_capacity(mddev
->gendisk
, mddev
->array_size
<< 1);
298 static int linear_stop (mddev_t
*mddev
)
300 linear_conf_t
*conf
= mddev_to_conf(mddev
);
302 blk_sync_queue(mddev
->queue
); /* the unplug fn references 'conf'*/
304 linear_conf_t
*t
= conf
->prev
;
305 kfree(conf
->hash_table
);
313 static int linear_make_request (struct request_queue
*q
, struct bio
*bio
)
315 const int rw
= bio_data_dir(bio
);
316 mddev_t
*mddev
= q
->queuedata
;
320 if (unlikely(bio_barrier(bio
))) {
321 bio_endio(bio
, -EOPNOTSUPP
);
325 disk_stat_inc(mddev
->gendisk
, ios
[rw
]);
326 disk_stat_add(mddev
->gendisk
, sectors
[rw
], bio_sectors(bio
));
328 tmp_dev
= which_dev(mddev
, bio
->bi_sector
);
329 block
= bio
->bi_sector
>> 1;
331 if (unlikely(block
>= (tmp_dev
->size
+ tmp_dev
->offset
)
332 || block
< tmp_dev
->offset
)) {
333 char b
[BDEVNAME_SIZE
];
335 printk("linear_make_request: Block %llu out of bounds on "
336 "dev %s size %llu offset %llu\n",
337 (unsigned long long)block
,
338 bdevname(tmp_dev
->rdev
->bdev
, b
),
339 (unsigned long long)tmp_dev
->size
,
340 (unsigned long long)tmp_dev
->offset
);
344 if (unlikely(bio
->bi_sector
+ (bio
->bi_size
>> 9) >
345 (tmp_dev
->offset
+ tmp_dev
->size
)<<1)) {
346 /* This bio crosses a device boundary, so we have to
350 bp
= bio_split(bio
, bio_split_pool
,
351 ((tmp_dev
->offset
+ tmp_dev
->size
)<<1) - bio
->bi_sector
);
352 if (linear_make_request(q
, &bp
->bio1
))
353 generic_make_request(&bp
->bio1
);
354 if (linear_make_request(q
, &bp
->bio2
))
355 generic_make_request(&bp
->bio2
);
356 bio_pair_release(bp
);
360 bio
->bi_bdev
= tmp_dev
->rdev
->bdev
;
361 bio
->bi_sector
= bio
->bi_sector
- (tmp_dev
->offset
<< 1) + tmp_dev
->rdev
->data_offset
;
366 static void linear_status (struct seq_file
*seq
, mddev_t
*mddev
)
372 linear_conf_t
*conf
= mddev_to_conf(mddev
);
375 seq_printf(seq
, " ");
376 for (j
= 0; j
< mddev
->raid_disks
; j
++)
378 char b
[BDEVNAME_SIZE
];
379 s
+= conf
->smallest_size
;
380 seq_printf(seq
, "[%s",
381 bdevname(conf
->hash_table
[j
][0].rdev
->bdev
,b
));
383 while (s
> conf
->hash_table
[j
][0].offset
+
384 conf
->hash_table
[j
][0].size
)
385 seq_printf(seq
, "/%s] ",
386 bdevname(conf
->hash_table
[j
][1].rdev
->bdev
,b
));
388 seq_printf(seq
, "] ");
390 seq_printf(seq
, "\n");
392 seq_printf(seq
, " %dk rounding", mddev
->chunk_size
/1024);
396 static struct mdk_personality linear_personality
=
399 .level
= LEVEL_LINEAR
,
400 .owner
= THIS_MODULE
,
401 .make_request
= linear_make_request
,
404 .status
= linear_status
,
405 .hot_add_disk
= linear_add
,
408 static int __init
linear_init (void)
410 return register_md_personality (&linear_personality
);
413 static void linear_exit (void)
415 unregister_md_personality (&linear_personality
);
419 module_init(linear_init
);
420 module_exit(linear_exit
);
421 MODULE_LICENSE("GPL");
422 MODULE_ALIAS("md-personality-1"); /* LINEAR - deprecated*/
423 MODULE_ALIAS("md-linear");
424 MODULE_ALIAS("md-level--1");