ARM: fix ioremap/iounmap for !CONFIG_MMU
[linux-2.6/libata-dev.git] / drivers / block / rbd.c
bloba6278e7e61a00bfde01bcb6726524ffb3a06514e
1 /*
2 rbd.c -- Export ceph rados objects as a Linux block device
5 based on drivers/block/osdblk.c:
7 Copyright 2009 Red Hat, Inc.
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation.
13 This program is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 GNU General Public License for more details.
18 You should have received a copy of the GNU General Public License
19 along with this program; see the file COPYING. If not, write to
20 the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
24 For usage instructions, please refer to:
26 Documentation/ABI/testing/sysfs-bus-rbd
30 #include <linux/ceph/libceph.h>
31 #include <linux/ceph/osd_client.h>
32 #include <linux/ceph/mon_client.h>
33 #include <linux/ceph/decode.h>
34 #include <linux/parser.h>
36 #include <linux/kernel.h>
37 #include <linux/device.h>
38 #include <linux/module.h>
39 #include <linux/fs.h>
40 #include <linux/blkdev.h>
42 #include "rbd_types.h"
44 #define DRV_NAME "rbd"
45 #define DRV_NAME_LONG "rbd (rados block device)"
47 #define RBD_MINORS_PER_MAJOR 256 /* max minors per blkdev */
49 #define RBD_MAX_MD_NAME_LEN (96 + sizeof(RBD_SUFFIX))
50 #define RBD_MAX_POOL_NAME_LEN 64
51 #define RBD_MAX_SNAP_NAME_LEN 32
52 #define RBD_MAX_OPT_LEN 1024
54 #define RBD_SNAP_HEAD_NAME "-"
56 #define DEV_NAME_LEN 32
58 #define RBD_NOTIFY_TIMEOUT_DEFAULT 10
61 * block device image metadata (in-memory version)
63 struct rbd_image_header {
64 u64 image_size;
65 char block_name[32];
66 __u8 obj_order;
67 __u8 crypt_type;
68 __u8 comp_type;
69 struct rw_semaphore snap_rwsem;
70 struct ceph_snap_context *snapc;
71 size_t snap_names_len;
72 u64 snap_seq;
73 u32 total_snaps;
75 char *snap_names;
76 u64 *snap_sizes;
78 u64 obj_version;
81 struct rbd_options {
82 int notify_timeout;
86 * an instance of the client. multiple devices may share a client.
88 struct rbd_client {
89 struct ceph_client *client;
90 struct rbd_options *rbd_opts;
91 struct kref kref;
92 struct list_head node;
95 struct rbd_req_coll;
98 * a single io request
100 struct rbd_request {
101 struct request *rq; /* blk layer request */
102 struct bio *bio; /* cloned bio */
103 struct page **pages; /* list of used pages */
104 u64 len;
105 int coll_index;
106 struct rbd_req_coll *coll;
109 struct rbd_req_status {
110 int done;
111 int rc;
112 u64 bytes;
116 * a collection of requests
118 struct rbd_req_coll {
119 int total;
120 int num_done;
121 struct kref kref;
122 struct rbd_req_status status[0];
125 struct rbd_snap {
126 struct device dev;
127 const char *name;
128 size_t size;
129 struct list_head node;
130 u64 id;
134 * a single device
136 struct rbd_device {
137 int id; /* blkdev unique id */
139 int major; /* blkdev assigned major */
140 struct gendisk *disk; /* blkdev's gendisk and rq */
141 struct request_queue *q;
143 struct ceph_client *client;
144 struct rbd_client *rbd_client;
146 char name[DEV_NAME_LEN]; /* blkdev name, e.g. rbd3 */
148 spinlock_t lock; /* queue lock */
150 struct rbd_image_header header;
151 char obj[RBD_MAX_OBJ_NAME_LEN]; /* rbd image name */
152 int obj_len;
153 char obj_md_name[RBD_MAX_MD_NAME_LEN]; /* hdr nm. */
154 char pool_name[RBD_MAX_POOL_NAME_LEN];
155 int poolid;
157 struct ceph_osd_event *watch_event;
158 struct ceph_osd_request *watch_request;
160 char snap_name[RBD_MAX_SNAP_NAME_LEN];
161 u32 cur_snap; /* index+1 of current snapshot within snap context
162 0 - for the head */
163 int read_only;
165 struct list_head node;
167 /* list of snapshots */
168 struct list_head snaps;
170 /* sysfs related */
171 struct device dev;
174 static struct bus_type rbd_bus_type = {
175 .name = "rbd",
178 static spinlock_t node_lock; /* protects client get/put */
180 static DEFINE_MUTEX(ctl_mutex); /* Serialize open/close/setup/teardown */
181 static LIST_HEAD(rbd_dev_list); /* devices */
182 static LIST_HEAD(rbd_client_list); /* clients */
184 static int __rbd_init_snaps_header(struct rbd_device *rbd_dev);
185 static void rbd_dev_release(struct device *dev);
186 static ssize_t rbd_snap_add(struct device *dev,
187 struct device_attribute *attr,
188 const char *buf,
189 size_t count);
190 static void __rbd_remove_snap_dev(struct rbd_device *rbd_dev,
191 struct rbd_snap *snap);
194 static struct rbd_device *dev_to_rbd(struct device *dev)
196 return container_of(dev, struct rbd_device, dev);
199 static struct device *rbd_get_dev(struct rbd_device *rbd_dev)
201 return get_device(&rbd_dev->dev);
204 static void rbd_put_dev(struct rbd_device *rbd_dev)
206 put_device(&rbd_dev->dev);
209 static int __rbd_update_snaps(struct rbd_device *rbd_dev);
211 static int rbd_open(struct block_device *bdev, fmode_t mode)
213 struct gendisk *disk = bdev->bd_disk;
214 struct rbd_device *rbd_dev = disk->private_data;
216 rbd_get_dev(rbd_dev);
218 set_device_ro(bdev, rbd_dev->read_only);
220 if ((mode & FMODE_WRITE) && rbd_dev->read_only)
221 return -EROFS;
223 return 0;
226 static int rbd_release(struct gendisk *disk, fmode_t mode)
228 struct rbd_device *rbd_dev = disk->private_data;
230 rbd_put_dev(rbd_dev);
232 return 0;
235 static const struct block_device_operations rbd_bd_ops = {
236 .owner = THIS_MODULE,
237 .open = rbd_open,
238 .release = rbd_release,
242 * Initialize an rbd client instance.
243 * We own *opt.
245 static struct rbd_client *rbd_client_create(struct ceph_options *opt,
246 struct rbd_options *rbd_opts)
248 struct rbd_client *rbdc;
249 int ret = -ENOMEM;
251 dout("rbd_client_create\n");
252 rbdc = kmalloc(sizeof(struct rbd_client), GFP_KERNEL);
253 if (!rbdc)
254 goto out_opt;
256 kref_init(&rbdc->kref);
257 INIT_LIST_HEAD(&rbdc->node);
259 rbdc->client = ceph_create_client(opt, rbdc, 0, 0);
260 if (IS_ERR(rbdc->client))
261 goto out_rbdc;
262 opt = NULL; /* Now rbdc->client is responsible for opt */
264 ret = ceph_open_session(rbdc->client);
265 if (ret < 0)
266 goto out_err;
268 rbdc->rbd_opts = rbd_opts;
270 spin_lock(&node_lock);
271 list_add_tail(&rbdc->node, &rbd_client_list);
272 spin_unlock(&node_lock);
274 dout("rbd_client_create created %p\n", rbdc);
275 return rbdc;
277 out_err:
278 ceph_destroy_client(rbdc->client);
279 out_rbdc:
280 kfree(rbdc);
281 out_opt:
282 if (opt)
283 ceph_destroy_options(opt);
284 return ERR_PTR(ret);
288 * Find a ceph client with specific addr and configuration.
290 static struct rbd_client *__rbd_client_find(struct ceph_options *opt)
292 struct rbd_client *client_node;
294 if (opt->flags & CEPH_OPT_NOSHARE)
295 return NULL;
297 list_for_each_entry(client_node, &rbd_client_list, node)
298 if (ceph_compare_options(opt, client_node->client) == 0)
299 return client_node;
300 return NULL;
304 * mount options
306 enum {
307 Opt_notify_timeout,
308 Opt_last_int,
309 /* int args above */
310 Opt_last_string,
311 /* string args above */
314 static match_table_t rbdopt_tokens = {
315 {Opt_notify_timeout, "notify_timeout=%d"},
316 /* int args above */
317 /* string args above */
318 {-1, NULL}
321 static int parse_rbd_opts_token(char *c, void *private)
323 struct rbd_options *rbdopt = private;
324 substring_t argstr[MAX_OPT_ARGS];
325 int token, intval, ret;
327 token = match_token((char *)c, rbdopt_tokens, argstr);
328 if (token < 0)
329 return -EINVAL;
331 if (token < Opt_last_int) {
332 ret = match_int(&argstr[0], &intval);
333 if (ret < 0) {
334 pr_err("bad mount option arg (not int) "
335 "at '%s'\n", c);
336 return ret;
338 dout("got int token %d val %d\n", token, intval);
339 } else if (token > Opt_last_int && token < Opt_last_string) {
340 dout("got string token %d val %s\n", token,
341 argstr[0].from);
342 } else {
343 dout("got token %d\n", token);
346 switch (token) {
347 case Opt_notify_timeout:
348 rbdopt->notify_timeout = intval;
349 break;
350 default:
351 BUG_ON(token);
353 return 0;
357 * Get a ceph client with specific addr and configuration, if one does
358 * not exist create it.
360 static int rbd_get_client(struct rbd_device *rbd_dev, const char *mon_addr,
361 char *options)
363 struct rbd_client *rbdc;
364 struct ceph_options *opt;
365 int ret;
366 struct rbd_options *rbd_opts;
368 rbd_opts = kzalloc(sizeof(*rbd_opts), GFP_KERNEL);
369 if (!rbd_opts)
370 return -ENOMEM;
372 rbd_opts->notify_timeout = RBD_NOTIFY_TIMEOUT_DEFAULT;
374 ret = ceph_parse_options(&opt, options, mon_addr,
375 mon_addr + strlen(mon_addr), parse_rbd_opts_token, rbd_opts);
376 if (ret < 0)
377 goto done_err;
379 spin_lock(&node_lock);
380 rbdc = __rbd_client_find(opt);
381 if (rbdc) {
382 ceph_destroy_options(opt);
383 kfree(rbd_opts);
385 /* using an existing client */
386 kref_get(&rbdc->kref);
387 rbd_dev->rbd_client = rbdc;
388 rbd_dev->client = rbdc->client;
389 spin_unlock(&node_lock);
390 return 0;
392 spin_unlock(&node_lock);
394 rbdc = rbd_client_create(opt, rbd_opts);
395 if (IS_ERR(rbdc)) {
396 ret = PTR_ERR(rbdc);
397 goto done_err;
400 rbd_dev->rbd_client = rbdc;
401 rbd_dev->client = rbdc->client;
402 return 0;
403 done_err:
404 kfree(rbd_opts);
405 return ret;
409 * Destroy ceph client
411 * Caller must hold node_lock.
413 static void rbd_client_release(struct kref *kref)
415 struct rbd_client *rbdc = container_of(kref, struct rbd_client, kref);
417 dout("rbd_release_client %p\n", rbdc);
418 list_del(&rbdc->node);
420 ceph_destroy_client(rbdc->client);
421 kfree(rbdc->rbd_opts);
422 kfree(rbdc);
426 * Drop reference to ceph client node. If it's not referenced anymore, release
427 * it.
429 static void rbd_put_client(struct rbd_device *rbd_dev)
431 spin_lock(&node_lock);
432 kref_put(&rbd_dev->rbd_client->kref, rbd_client_release);
433 spin_unlock(&node_lock);
434 rbd_dev->rbd_client = NULL;
435 rbd_dev->client = NULL;
439 * Destroy requests collection
441 static void rbd_coll_release(struct kref *kref)
443 struct rbd_req_coll *coll =
444 container_of(kref, struct rbd_req_coll, kref);
446 dout("rbd_coll_release %p\n", coll);
447 kfree(coll);
451 * Create a new header structure, translate header format from the on-disk
452 * header.
454 static int rbd_header_from_disk(struct rbd_image_header *header,
455 struct rbd_image_header_ondisk *ondisk,
456 int allocated_snaps,
457 gfp_t gfp_flags)
459 int i;
460 u32 snap_count = le32_to_cpu(ondisk->snap_count);
461 int ret = -ENOMEM;
463 if (memcmp(ondisk, RBD_HEADER_TEXT, sizeof(RBD_HEADER_TEXT))) {
464 return -ENXIO;
467 init_rwsem(&header->snap_rwsem);
468 header->snap_names_len = le64_to_cpu(ondisk->snap_names_len);
469 header->snapc = kmalloc(sizeof(struct ceph_snap_context) +
470 snap_count *
471 sizeof(struct rbd_image_snap_ondisk),
472 gfp_flags);
473 if (!header->snapc)
474 return -ENOMEM;
475 if (snap_count) {
476 header->snap_names = kmalloc(header->snap_names_len,
477 GFP_KERNEL);
478 if (!header->snap_names)
479 goto err_snapc;
480 header->snap_sizes = kmalloc(snap_count * sizeof(u64),
481 GFP_KERNEL);
482 if (!header->snap_sizes)
483 goto err_names;
484 } else {
485 header->snap_names = NULL;
486 header->snap_sizes = NULL;
488 memcpy(header->block_name, ondisk->block_name,
489 sizeof(ondisk->block_name));
491 header->image_size = le64_to_cpu(ondisk->image_size);
492 header->obj_order = ondisk->options.order;
493 header->crypt_type = ondisk->options.crypt_type;
494 header->comp_type = ondisk->options.comp_type;
496 atomic_set(&header->snapc->nref, 1);
497 header->snap_seq = le64_to_cpu(ondisk->snap_seq);
498 header->snapc->num_snaps = snap_count;
499 header->total_snaps = snap_count;
501 if (snap_count &&
502 allocated_snaps == snap_count) {
503 for (i = 0; i < snap_count; i++) {
504 header->snapc->snaps[i] =
505 le64_to_cpu(ondisk->snaps[i].id);
506 header->snap_sizes[i] =
507 le64_to_cpu(ondisk->snaps[i].image_size);
510 /* copy snapshot names */
511 memcpy(header->snap_names, &ondisk->snaps[i],
512 header->snap_names_len);
515 return 0;
517 err_names:
518 kfree(header->snap_names);
519 err_snapc:
520 kfree(header->snapc);
521 return ret;
524 static int snap_index(struct rbd_image_header *header, int snap_num)
526 return header->total_snaps - snap_num;
529 static u64 cur_snap_id(struct rbd_device *rbd_dev)
531 struct rbd_image_header *header = &rbd_dev->header;
533 if (!rbd_dev->cur_snap)
534 return 0;
536 return header->snapc->snaps[snap_index(header, rbd_dev->cur_snap)];
539 static int snap_by_name(struct rbd_image_header *header, const char *snap_name,
540 u64 *seq, u64 *size)
542 int i;
543 char *p = header->snap_names;
545 for (i = 0; i < header->total_snaps; i++, p += strlen(p) + 1) {
546 if (strcmp(snap_name, p) == 0)
547 break;
549 if (i == header->total_snaps)
550 return -ENOENT;
551 if (seq)
552 *seq = header->snapc->snaps[i];
554 if (size)
555 *size = header->snap_sizes[i];
557 return i;
560 static int rbd_header_set_snap(struct rbd_device *dev,
561 const char *snap_name,
562 u64 *size)
564 struct rbd_image_header *header = &dev->header;
565 struct ceph_snap_context *snapc = header->snapc;
566 int ret = -ENOENT;
568 down_write(&header->snap_rwsem);
570 if (!snap_name ||
571 !*snap_name ||
572 strcmp(snap_name, "-") == 0 ||
573 strcmp(snap_name, RBD_SNAP_HEAD_NAME) == 0) {
574 if (header->total_snaps)
575 snapc->seq = header->snap_seq;
576 else
577 snapc->seq = 0;
578 dev->cur_snap = 0;
579 dev->read_only = 0;
580 if (size)
581 *size = header->image_size;
582 } else {
583 ret = snap_by_name(header, snap_name, &snapc->seq, size);
584 if (ret < 0)
585 goto done;
587 dev->cur_snap = header->total_snaps - ret;
588 dev->read_only = 1;
591 ret = 0;
592 done:
593 up_write(&header->snap_rwsem);
594 return ret;
597 static void rbd_header_free(struct rbd_image_header *header)
599 kfree(header->snapc);
600 kfree(header->snap_names);
601 kfree(header->snap_sizes);
605 * get the actual striped segment name, offset and length
607 static u64 rbd_get_segment(struct rbd_image_header *header,
608 const char *block_name,
609 u64 ofs, u64 len,
610 char *seg_name, u64 *segofs)
612 u64 seg = ofs >> header->obj_order;
614 if (seg_name)
615 snprintf(seg_name, RBD_MAX_SEG_NAME_LEN,
616 "%s.%012llx", block_name, seg);
618 ofs = ofs & ((1 << header->obj_order) - 1);
619 len = min_t(u64, len, (1 << header->obj_order) - ofs);
621 if (segofs)
622 *segofs = ofs;
624 return len;
627 static int rbd_get_num_segments(struct rbd_image_header *header,
628 u64 ofs, u64 len)
630 u64 start_seg = ofs >> header->obj_order;
631 u64 end_seg = (ofs + len - 1) >> header->obj_order;
632 return end_seg - start_seg + 1;
636 * returns the size of an object in the image
638 static u64 rbd_obj_bytes(struct rbd_image_header *header)
640 return 1 << header->obj_order;
644 * bio helpers
647 static void bio_chain_put(struct bio *chain)
649 struct bio *tmp;
651 while (chain) {
652 tmp = chain;
653 chain = chain->bi_next;
654 bio_put(tmp);
659 * zeros a bio chain, starting at specific offset
661 static void zero_bio_chain(struct bio *chain, int start_ofs)
663 struct bio_vec *bv;
664 unsigned long flags;
665 void *buf;
666 int i;
667 int pos = 0;
669 while (chain) {
670 bio_for_each_segment(bv, chain, i) {
671 if (pos + bv->bv_len > start_ofs) {
672 int remainder = max(start_ofs - pos, 0);
673 buf = bvec_kmap_irq(bv, &flags);
674 memset(buf + remainder, 0,
675 bv->bv_len - remainder);
676 bvec_kunmap_irq(buf, &flags);
678 pos += bv->bv_len;
681 chain = chain->bi_next;
686 * bio_chain_clone - clone a chain of bios up to a certain length.
687 * might return a bio_pair that will need to be released.
689 static struct bio *bio_chain_clone(struct bio **old, struct bio **next,
690 struct bio_pair **bp,
691 int len, gfp_t gfpmask)
693 struct bio *tmp, *old_chain = *old, *new_chain = NULL, *tail = NULL;
694 int total = 0;
696 if (*bp) {
697 bio_pair_release(*bp);
698 *bp = NULL;
701 while (old_chain && (total < len)) {
702 tmp = bio_kmalloc(gfpmask, old_chain->bi_max_vecs);
703 if (!tmp)
704 goto err_out;
706 if (total + old_chain->bi_size > len) {
707 struct bio_pair *bp;
710 * this split can only happen with a single paged bio,
711 * split_bio will BUG_ON if this is not the case
713 dout("bio_chain_clone split! total=%d remaining=%d"
714 "bi_size=%d\n",
715 (int)total, (int)len-total,
716 (int)old_chain->bi_size);
718 /* split the bio. We'll release it either in the next
719 call, or it will have to be released outside */
720 bp = bio_split(old_chain, (len - total) / 512ULL);
721 if (!bp)
722 goto err_out;
724 __bio_clone(tmp, &bp->bio1);
726 *next = &bp->bio2;
727 } else {
728 __bio_clone(tmp, old_chain);
729 *next = old_chain->bi_next;
732 tmp->bi_bdev = NULL;
733 gfpmask &= ~__GFP_WAIT;
734 tmp->bi_next = NULL;
736 if (!new_chain) {
737 new_chain = tail = tmp;
738 } else {
739 tail->bi_next = tmp;
740 tail = tmp;
742 old_chain = old_chain->bi_next;
744 total += tmp->bi_size;
747 BUG_ON(total < len);
749 if (tail)
750 tail->bi_next = NULL;
752 *old = old_chain;
754 return new_chain;
756 err_out:
757 dout("bio_chain_clone with err\n");
758 bio_chain_put(new_chain);
759 return NULL;
763 * helpers for osd request op vectors.
765 static int rbd_create_rw_ops(struct ceph_osd_req_op **ops,
766 int num_ops,
767 int opcode,
768 u32 payload_len)
770 *ops = kzalloc(sizeof(struct ceph_osd_req_op) * (num_ops + 1),
771 GFP_NOIO);
772 if (!*ops)
773 return -ENOMEM;
774 (*ops)[0].op = opcode;
776 * op extent offset and length will be set later on
777 * in calc_raw_layout()
779 (*ops)[0].payload_len = payload_len;
780 return 0;
783 static void rbd_destroy_ops(struct ceph_osd_req_op *ops)
785 kfree(ops);
788 static void rbd_coll_end_req_index(struct request *rq,
789 struct rbd_req_coll *coll,
790 int index,
791 int ret, u64 len)
793 struct request_queue *q;
794 int min, max, i;
796 dout("rbd_coll_end_req_index %p index %d ret %d len %lld\n",
797 coll, index, ret, len);
799 if (!rq)
800 return;
802 if (!coll) {
803 blk_end_request(rq, ret, len);
804 return;
807 q = rq->q;
809 spin_lock_irq(q->queue_lock);
810 coll->status[index].done = 1;
811 coll->status[index].rc = ret;
812 coll->status[index].bytes = len;
813 max = min = coll->num_done;
814 while (max < coll->total && coll->status[max].done)
815 max++;
817 for (i = min; i<max; i++) {
818 __blk_end_request(rq, coll->status[i].rc,
819 coll->status[i].bytes);
820 coll->num_done++;
821 kref_put(&coll->kref, rbd_coll_release);
823 spin_unlock_irq(q->queue_lock);
826 static void rbd_coll_end_req(struct rbd_request *req,
827 int ret, u64 len)
829 rbd_coll_end_req_index(req->rq, req->coll, req->coll_index, ret, len);
833 * Send ceph osd request
835 static int rbd_do_request(struct request *rq,
836 struct rbd_device *dev,
837 struct ceph_snap_context *snapc,
838 u64 snapid,
839 const char *obj, u64 ofs, u64 len,
840 struct bio *bio,
841 struct page **pages,
842 int num_pages,
843 int flags,
844 struct ceph_osd_req_op *ops,
845 int num_reply,
846 struct rbd_req_coll *coll,
847 int coll_index,
848 void (*rbd_cb)(struct ceph_osd_request *req,
849 struct ceph_msg *msg),
850 struct ceph_osd_request **linger_req,
851 u64 *ver)
853 struct ceph_osd_request *req;
854 struct ceph_file_layout *layout;
855 int ret;
856 u64 bno;
857 struct timespec mtime = CURRENT_TIME;
858 struct rbd_request *req_data;
859 struct ceph_osd_request_head *reqhead;
860 struct rbd_image_header *header = &dev->header;
862 req_data = kzalloc(sizeof(*req_data), GFP_NOIO);
863 if (!req_data) {
864 if (coll)
865 rbd_coll_end_req_index(rq, coll, coll_index,
866 -ENOMEM, len);
867 return -ENOMEM;
870 if (coll) {
871 req_data->coll = coll;
872 req_data->coll_index = coll_index;
875 dout("rbd_do_request obj=%s ofs=%lld len=%lld\n", obj, len, ofs);
877 down_read(&header->snap_rwsem);
879 req = ceph_osdc_alloc_request(&dev->client->osdc, flags,
880 snapc,
881 ops,
882 false,
883 GFP_NOIO, pages, bio);
884 if (!req) {
885 up_read(&header->snap_rwsem);
886 ret = -ENOMEM;
887 goto done_pages;
890 req->r_callback = rbd_cb;
892 req_data->rq = rq;
893 req_data->bio = bio;
894 req_data->pages = pages;
895 req_data->len = len;
897 req->r_priv = req_data;
899 reqhead = req->r_request->front.iov_base;
900 reqhead->snapid = cpu_to_le64(CEPH_NOSNAP);
902 strncpy(req->r_oid, obj, sizeof(req->r_oid));
903 req->r_oid_len = strlen(req->r_oid);
905 layout = &req->r_file_layout;
906 memset(layout, 0, sizeof(*layout));
907 layout->fl_stripe_unit = cpu_to_le32(1 << RBD_MAX_OBJ_ORDER);
908 layout->fl_stripe_count = cpu_to_le32(1);
909 layout->fl_object_size = cpu_to_le32(1 << RBD_MAX_OBJ_ORDER);
910 layout->fl_pg_preferred = cpu_to_le32(-1);
911 layout->fl_pg_pool = cpu_to_le32(dev->poolid);
912 ceph_calc_raw_layout(&dev->client->osdc, layout, snapid,
913 ofs, &len, &bno, req, ops);
915 ceph_osdc_build_request(req, ofs, &len,
916 ops,
917 snapc,
918 &mtime,
919 req->r_oid, req->r_oid_len);
920 up_read(&header->snap_rwsem);
922 if (linger_req) {
923 ceph_osdc_set_request_linger(&dev->client->osdc, req);
924 *linger_req = req;
927 ret = ceph_osdc_start_request(&dev->client->osdc, req, false);
928 if (ret < 0)
929 goto done_err;
931 if (!rbd_cb) {
932 ret = ceph_osdc_wait_request(&dev->client->osdc, req);
933 if (ver)
934 *ver = le64_to_cpu(req->r_reassert_version.version);
935 dout("reassert_ver=%lld\n",
936 le64_to_cpu(req->r_reassert_version.version));
937 ceph_osdc_put_request(req);
939 return ret;
941 done_err:
942 bio_chain_put(req_data->bio);
943 ceph_osdc_put_request(req);
944 done_pages:
945 rbd_coll_end_req(req_data, ret, len);
946 kfree(req_data);
947 return ret;
951 * Ceph osd op callback
953 static void rbd_req_cb(struct ceph_osd_request *req, struct ceph_msg *msg)
955 struct rbd_request *req_data = req->r_priv;
956 struct ceph_osd_reply_head *replyhead;
957 struct ceph_osd_op *op;
958 __s32 rc;
959 u64 bytes;
960 int read_op;
962 /* parse reply */
963 replyhead = msg->front.iov_base;
964 WARN_ON(le32_to_cpu(replyhead->num_ops) == 0);
965 op = (void *)(replyhead + 1);
966 rc = le32_to_cpu(replyhead->result);
967 bytes = le64_to_cpu(op->extent.length);
968 read_op = (le32_to_cpu(op->op) == CEPH_OSD_OP_READ);
970 dout("rbd_req_cb bytes=%lld readop=%d rc=%d\n", bytes, read_op, rc);
972 if (rc == -ENOENT && read_op) {
973 zero_bio_chain(req_data->bio, 0);
974 rc = 0;
975 } else if (rc == 0 && read_op && bytes < req_data->len) {
976 zero_bio_chain(req_data->bio, bytes);
977 bytes = req_data->len;
980 rbd_coll_end_req(req_data, rc, bytes);
982 if (req_data->bio)
983 bio_chain_put(req_data->bio);
985 ceph_osdc_put_request(req);
986 kfree(req_data);
989 static void rbd_simple_req_cb(struct ceph_osd_request *req, struct ceph_msg *msg)
991 ceph_osdc_put_request(req);
995 * Do a synchronous ceph osd operation
997 static int rbd_req_sync_op(struct rbd_device *dev,
998 struct ceph_snap_context *snapc,
999 u64 snapid,
1000 int opcode,
1001 int flags,
1002 struct ceph_osd_req_op *orig_ops,
1003 int num_reply,
1004 const char *obj,
1005 u64 ofs, u64 len,
1006 char *buf,
1007 struct ceph_osd_request **linger_req,
1008 u64 *ver)
1010 int ret;
1011 struct page **pages;
1012 int num_pages;
1013 struct ceph_osd_req_op *ops = orig_ops;
1014 u32 payload_len;
1016 num_pages = calc_pages_for(ofs , len);
1017 pages = ceph_alloc_page_vector(num_pages, GFP_KERNEL);
1018 if (IS_ERR(pages))
1019 return PTR_ERR(pages);
1021 if (!orig_ops) {
1022 payload_len = (flags & CEPH_OSD_FLAG_WRITE ? len : 0);
1023 ret = rbd_create_rw_ops(&ops, 1, opcode, payload_len);
1024 if (ret < 0)
1025 goto done;
1027 if ((flags & CEPH_OSD_FLAG_WRITE) && buf) {
1028 ret = ceph_copy_to_page_vector(pages, buf, ofs, len);
1029 if (ret < 0)
1030 goto done_ops;
1034 ret = rbd_do_request(NULL, dev, snapc, snapid,
1035 obj, ofs, len, NULL,
1036 pages, num_pages,
1037 flags,
1038 ops,
1040 NULL, 0,
1041 NULL,
1042 linger_req, ver);
1043 if (ret < 0)
1044 goto done_ops;
1046 if ((flags & CEPH_OSD_FLAG_READ) && buf)
1047 ret = ceph_copy_from_page_vector(pages, buf, ofs, ret);
1049 done_ops:
1050 if (!orig_ops)
1051 rbd_destroy_ops(ops);
1052 done:
1053 ceph_release_page_vector(pages, num_pages);
1054 return ret;
1058 * Do an asynchronous ceph osd operation
1060 static int rbd_do_op(struct request *rq,
1061 struct rbd_device *rbd_dev ,
1062 struct ceph_snap_context *snapc,
1063 u64 snapid,
1064 int opcode, int flags, int num_reply,
1065 u64 ofs, u64 len,
1066 struct bio *bio,
1067 struct rbd_req_coll *coll,
1068 int coll_index)
1070 char *seg_name;
1071 u64 seg_ofs;
1072 u64 seg_len;
1073 int ret;
1074 struct ceph_osd_req_op *ops;
1075 u32 payload_len;
1077 seg_name = kmalloc(RBD_MAX_SEG_NAME_LEN + 1, GFP_NOIO);
1078 if (!seg_name)
1079 return -ENOMEM;
1081 seg_len = rbd_get_segment(&rbd_dev->header,
1082 rbd_dev->header.block_name,
1083 ofs, len,
1084 seg_name, &seg_ofs);
1086 payload_len = (flags & CEPH_OSD_FLAG_WRITE ? seg_len : 0);
1088 ret = rbd_create_rw_ops(&ops, 1, opcode, payload_len);
1089 if (ret < 0)
1090 goto done;
1092 /* we've taken care of segment sizes earlier when we
1093 cloned the bios. We should never have a segment
1094 truncated at this point */
1095 BUG_ON(seg_len < len);
1097 ret = rbd_do_request(rq, rbd_dev, snapc, snapid,
1098 seg_name, seg_ofs, seg_len,
1099 bio,
1100 NULL, 0,
1101 flags,
1102 ops,
1103 num_reply,
1104 coll, coll_index,
1105 rbd_req_cb, 0, NULL);
1107 rbd_destroy_ops(ops);
1108 done:
1109 kfree(seg_name);
1110 return ret;
1114 * Request async osd write
1116 static int rbd_req_write(struct request *rq,
1117 struct rbd_device *rbd_dev,
1118 struct ceph_snap_context *snapc,
1119 u64 ofs, u64 len,
1120 struct bio *bio,
1121 struct rbd_req_coll *coll,
1122 int coll_index)
1124 return rbd_do_op(rq, rbd_dev, snapc, CEPH_NOSNAP,
1125 CEPH_OSD_OP_WRITE,
1126 CEPH_OSD_FLAG_WRITE | CEPH_OSD_FLAG_ONDISK,
1128 ofs, len, bio, coll, coll_index);
1132 * Request async osd read
1134 static int rbd_req_read(struct request *rq,
1135 struct rbd_device *rbd_dev,
1136 u64 snapid,
1137 u64 ofs, u64 len,
1138 struct bio *bio,
1139 struct rbd_req_coll *coll,
1140 int coll_index)
1142 return rbd_do_op(rq, rbd_dev, NULL,
1143 (snapid ? snapid : CEPH_NOSNAP),
1144 CEPH_OSD_OP_READ,
1145 CEPH_OSD_FLAG_READ,
1147 ofs, len, bio, coll, coll_index);
1151 * Request sync osd read
1153 static int rbd_req_sync_read(struct rbd_device *dev,
1154 struct ceph_snap_context *snapc,
1155 u64 snapid,
1156 const char *obj,
1157 u64 ofs, u64 len,
1158 char *buf,
1159 u64 *ver)
1161 return rbd_req_sync_op(dev, NULL,
1162 (snapid ? snapid : CEPH_NOSNAP),
1163 CEPH_OSD_OP_READ,
1164 CEPH_OSD_FLAG_READ,
1165 NULL,
1166 1, obj, ofs, len, buf, NULL, ver);
1170 * Request sync osd watch
1172 static int rbd_req_sync_notify_ack(struct rbd_device *dev,
1173 u64 ver,
1174 u64 notify_id,
1175 const char *obj)
1177 struct ceph_osd_req_op *ops;
1178 struct page **pages = NULL;
1179 int ret;
1181 ret = rbd_create_rw_ops(&ops, 1, CEPH_OSD_OP_NOTIFY_ACK, 0);
1182 if (ret < 0)
1183 return ret;
1185 ops[0].watch.ver = cpu_to_le64(dev->header.obj_version);
1186 ops[0].watch.cookie = notify_id;
1187 ops[0].watch.flag = 0;
1189 ret = rbd_do_request(NULL, dev, NULL, CEPH_NOSNAP,
1190 obj, 0, 0, NULL,
1191 pages, 0,
1192 CEPH_OSD_FLAG_READ,
1193 ops,
1195 NULL, 0,
1196 rbd_simple_req_cb, 0, NULL);
1198 rbd_destroy_ops(ops);
1199 return ret;
1202 static void rbd_watch_cb(u64 ver, u64 notify_id, u8 opcode, void *data)
1204 struct rbd_device *dev = (struct rbd_device *)data;
1205 int rc;
1207 if (!dev)
1208 return;
1210 dout("rbd_watch_cb %s notify_id=%lld opcode=%d\n", dev->obj_md_name,
1211 notify_id, (int)opcode);
1212 mutex_lock_nested(&ctl_mutex, SINGLE_DEPTH_NESTING);
1213 rc = __rbd_update_snaps(dev);
1214 mutex_unlock(&ctl_mutex);
1215 if (rc)
1216 pr_warning(DRV_NAME "%d got notification but failed to update"
1217 " snaps: %d\n", dev->major, rc);
1219 rbd_req_sync_notify_ack(dev, ver, notify_id, dev->obj_md_name);
1223 * Request sync osd watch
1225 static int rbd_req_sync_watch(struct rbd_device *dev,
1226 const char *obj,
1227 u64 ver)
1229 struct ceph_osd_req_op *ops;
1230 struct ceph_osd_client *osdc = &dev->client->osdc;
1232 int ret = rbd_create_rw_ops(&ops, 1, CEPH_OSD_OP_WATCH, 0);
1233 if (ret < 0)
1234 return ret;
1236 ret = ceph_osdc_create_event(osdc, rbd_watch_cb, 0,
1237 (void *)dev, &dev->watch_event);
1238 if (ret < 0)
1239 goto fail;
1241 ops[0].watch.ver = cpu_to_le64(ver);
1242 ops[0].watch.cookie = cpu_to_le64(dev->watch_event->cookie);
1243 ops[0].watch.flag = 1;
1245 ret = rbd_req_sync_op(dev, NULL,
1246 CEPH_NOSNAP,
1248 CEPH_OSD_FLAG_WRITE | CEPH_OSD_FLAG_ONDISK,
1249 ops,
1250 1, obj, 0, 0, NULL,
1251 &dev->watch_request, NULL);
1253 if (ret < 0)
1254 goto fail_event;
1256 rbd_destroy_ops(ops);
1257 return 0;
1259 fail_event:
1260 ceph_osdc_cancel_event(dev->watch_event);
1261 dev->watch_event = NULL;
1262 fail:
1263 rbd_destroy_ops(ops);
1264 return ret;
1268 * Request sync osd unwatch
1270 static int rbd_req_sync_unwatch(struct rbd_device *dev,
1271 const char *obj)
1273 struct ceph_osd_req_op *ops;
1275 int ret = rbd_create_rw_ops(&ops, 1, CEPH_OSD_OP_WATCH, 0);
1276 if (ret < 0)
1277 return ret;
1279 ops[0].watch.ver = 0;
1280 ops[0].watch.cookie = cpu_to_le64(dev->watch_event->cookie);
1281 ops[0].watch.flag = 0;
1283 ret = rbd_req_sync_op(dev, NULL,
1284 CEPH_NOSNAP,
1286 CEPH_OSD_FLAG_WRITE | CEPH_OSD_FLAG_ONDISK,
1287 ops,
1288 1, obj, 0, 0, NULL, NULL, NULL);
1290 rbd_destroy_ops(ops);
1291 ceph_osdc_cancel_event(dev->watch_event);
1292 dev->watch_event = NULL;
1293 return ret;
1296 struct rbd_notify_info {
1297 struct rbd_device *dev;
1300 static void rbd_notify_cb(u64 ver, u64 notify_id, u8 opcode, void *data)
1302 struct rbd_device *dev = (struct rbd_device *)data;
1303 if (!dev)
1304 return;
1306 dout("rbd_notify_cb %s notify_id=%lld opcode=%d\n", dev->obj_md_name,
1307 notify_id, (int)opcode);
1311 * Request sync osd notify
1313 static int rbd_req_sync_notify(struct rbd_device *dev,
1314 const char *obj)
1316 struct ceph_osd_req_op *ops;
1317 struct ceph_osd_client *osdc = &dev->client->osdc;
1318 struct ceph_osd_event *event;
1319 struct rbd_notify_info info;
1320 int payload_len = sizeof(u32) + sizeof(u32);
1321 int ret;
1323 ret = rbd_create_rw_ops(&ops, 1, CEPH_OSD_OP_NOTIFY, payload_len);
1324 if (ret < 0)
1325 return ret;
1327 info.dev = dev;
1329 ret = ceph_osdc_create_event(osdc, rbd_notify_cb, 1,
1330 (void *)&info, &event);
1331 if (ret < 0)
1332 goto fail;
1334 ops[0].watch.ver = 1;
1335 ops[0].watch.flag = 1;
1336 ops[0].watch.cookie = event->cookie;
1337 ops[0].watch.prot_ver = RADOS_NOTIFY_VER;
1338 ops[0].watch.timeout = 12;
1340 ret = rbd_req_sync_op(dev, NULL,
1341 CEPH_NOSNAP,
1343 CEPH_OSD_FLAG_WRITE | CEPH_OSD_FLAG_ONDISK,
1344 ops,
1345 1, obj, 0, 0, NULL, NULL, NULL);
1346 if (ret < 0)
1347 goto fail_event;
1349 ret = ceph_osdc_wait_event(event, CEPH_OSD_TIMEOUT_DEFAULT);
1350 dout("ceph_osdc_wait_event returned %d\n", ret);
1351 rbd_destroy_ops(ops);
1352 return 0;
1354 fail_event:
1355 ceph_osdc_cancel_event(event);
1356 fail:
1357 rbd_destroy_ops(ops);
1358 return ret;
1362 * Request sync osd read
1364 static int rbd_req_sync_exec(struct rbd_device *dev,
1365 const char *obj,
1366 const char *cls,
1367 const char *method,
1368 const char *data,
1369 int len,
1370 u64 *ver)
1372 struct ceph_osd_req_op *ops;
1373 int cls_len = strlen(cls);
1374 int method_len = strlen(method);
1375 int ret = rbd_create_rw_ops(&ops, 1, CEPH_OSD_OP_CALL,
1376 cls_len + method_len + len);
1377 if (ret < 0)
1378 return ret;
1380 ops[0].cls.class_name = cls;
1381 ops[0].cls.class_len = (__u8)cls_len;
1382 ops[0].cls.method_name = method;
1383 ops[0].cls.method_len = (__u8)method_len;
1384 ops[0].cls.argc = 0;
1385 ops[0].cls.indata = data;
1386 ops[0].cls.indata_len = len;
1388 ret = rbd_req_sync_op(dev, NULL,
1389 CEPH_NOSNAP,
1391 CEPH_OSD_FLAG_WRITE | CEPH_OSD_FLAG_ONDISK,
1392 ops,
1393 1, obj, 0, 0, NULL, NULL, ver);
1395 rbd_destroy_ops(ops);
1397 dout("cls_exec returned %d\n", ret);
1398 return ret;
1401 static struct rbd_req_coll *rbd_alloc_coll(int num_reqs)
1403 struct rbd_req_coll *coll =
1404 kzalloc(sizeof(struct rbd_req_coll) +
1405 sizeof(struct rbd_req_status) * num_reqs,
1406 GFP_ATOMIC);
1408 if (!coll)
1409 return NULL;
1410 coll->total = num_reqs;
1411 kref_init(&coll->kref);
1412 return coll;
1416 * block device queue callback
1418 static void rbd_rq_fn(struct request_queue *q)
1420 struct rbd_device *rbd_dev = q->queuedata;
1421 struct request *rq;
1422 struct bio_pair *bp = NULL;
1424 rq = blk_fetch_request(q);
1426 while (1) {
1427 struct bio *bio;
1428 struct bio *rq_bio, *next_bio = NULL;
1429 bool do_write;
1430 int size, op_size = 0;
1431 u64 ofs;
1432 int num_segs, cur_seg = 0;
1433 struct rbd_req_coll *coll;
1435 /* peek at request from block layer */
1436 if (!rq)
1437 break;
1439 dout("fetched request\n");
1441 /* filter out block requests we don't understand */
1442 if ((rq->cmd_type != REQ_TYPE_FS)) {
1443 __blk_end_request_all(rq, 0);
1444 goto next;
1447 /* deduce our operation (read, write) */
1448 do_write = (rq_data_dir(rq) == WRITE);
1450 size = blk_rq_bytes(rq);
1451 ofs = blk_rq_pos(rq) * 512ULL;
1452 rq_bio = rq->bio;
1453 if (do_write && rbd_dev->read_only) {
1454 __blk_end_request_all(rq, -EROFS);
1455 goto next;
1458 spin_unlock_irq(q->queue_lock);
1460 dout("%s 0x%x bytes at 0x%llx\n",
1461 do_write ? "write" : "read",
1462 size, blk_rq_pos(rq) * 512ULL);
1464 num_segs = rbd_get_num_segments(&rbd_dev->header, ofs, size);
1465 coll = rbd_alloc_coll(num_segs);
1466 if (!coll) {
1467 spin_lock_irq(q->queue_lock);
1468 __blk_end_request_all(rq, -ENOMEM);
1469 goto next;
1472 do {
1473 /* a bio clone to be passed down to OSD req */
1474 dout("rq->bio->bi_vcnt=%d\n", rq->bio->bi_vcnt);
1475 op_size = rbd_get_segment(&rbd_dev->header,
1476 rbd_dev->header.block_name,
1477 ofs, size,
1478 NULL, NULL);
1479 kref_get(&coll->kref);
1480 bio = bio_chain_clone(&rq_bio, &next_bio, &bp,
1481 op_size, GFP_ATOMIC);
1482 if (!bio) {
1483 rbd_coll_end_req_index(rq, coll, cur_seg,
1484 -ENOMEM, op_size);
1485 goto next_seg;
1489 /* init OSD command: write or read */
1490 if (do_write)
1491 rbd_req_write(rq, rbd_dev,
1492 rbd_dev->header.snapc,
1493 ofs,
1494 op_size, bio,
1495 coll, cur_seg);
1496 else
1497 rbd_req_read(rq, rbd_dev,
1498 cur_snap_id(rbd_dev),
1499 ofs,
1500 op_size, bio,
1501 coll, cur_seg);
1503 next_seg:
1504 size -= op_size;
1505 ofs += op_size;
1507 cur_seg++;
1508 rq_bio = next_bio;
1509 } while (size > 0);
1510 kref_put(&coll->kref, rbd_coll_release);
1512 if (bp)
1513 bio_pair_release(bp);
1514 spin_lock_irq(q->queue_lock);
1515 next:
1516 rq = blk_fetch_request(q);
1521 * a queue callback. Makes sure that we don't create a bio that spans across
1522 * multiple osd objects. One exception would be with a single page bios,
1523 * which we handle later at bio_chain_clone
1525 static int rbd_merge_bvec(struct request_queue *q, struct bvec_merge_data *bmd,
1526 struct bio_vec *bvec)
1528 struct rbd_device *rbd_dev = q->queuedata;
1529 unsigned int chunk_sectors = 1 << (rbd_dev->header.obj_order - 9);
1530 sector_t sector = bmd->bi_sector + get_start_sect(bmd->bi_bdev);
1531 unsigned int bio_sectors = bmd->bi_size >> 9;
1532 int max;
1534 max = (chunk_sectors - ((sector & (chunk_sectors - 1))
1535 + bio_sectors)) << 9;
1536 if (max < 0)
1537 max = 0; /* bio_add cannot handle a negative return */
1538 if (max <= bvec->bv_len && bio_sectors == 0)
1539 return bvec->bv_len;
1540 return max;
1543 static void rbd_free_disk(struct rbd_device *rbd_dev)
1545 struct gendisk *disk = rbd_dev->disk;
1547 if (!disk)
1548 return;
1550 rbd_header_free(&rbd_dev->header);
1552 if (disk->flags & GENHD_FL_UP)
1553 del_gendisk(disk);
1554 if (disk->queue)
1555 blk_cleanup_queue(disk->queue);
1556 put_disk(disk);
1560 * reload the ondisk the header
1562 static int rbd_read_header(struct rbd_device *rbd_dev,
1563 struct rbd_image_header *header)
1565 ssize_t rc;
1566 struct rbd_image_header_ondisk *dh;
1567 int snap_count = 0;
1568 u64 snap_names_len = 0;
1569 u64 ver;
1571 while (1) {
1572 int len = sizeof(*dh) +
1573 snap_count * sizeof(struct rbd_image_snap_ondisk) +
1574 snap_names_len;
1576 rc = -ENOMEM;
1577 dh = kmalloc(len, GFP_KERNEL);
1578 if (!dh)
1579 return -ENOMEM;
1581 rc = rbd_req_sync_read(rbd_dev,
1582 NULL, CEPH_NOSNAP,
1583 rbd_dev->obj_md_name,
1584 0, len,
1585 (char *)dh, &ver);
1586 if (rc < 0)
1587 goto out_dh;
1589 rc = rbd_header_from_disk(header, dh, snap_count, GFP_KERNEL);
1590 if (rc < 0) {
1591 if (rc == -ENXIO) {
1592 pr_warning("unrecognized header format"
1593 " for image %s", rbd_dev->obj);
1595 goto out_dh;
1598 if (snap_count != header->total_snaps) {
1599 snap_count = header->total_snaps;
1600 snap_names_len = header->snap_names_len;
1601 rbd_header_free(header);
1602 kfree(dh);
1603 continue;
1605 break;
1607 header->obj_version = ver;
1609 out_dh:
1610 kfree(dh);
1611 return rc;
1615 * create a snapshot
1617 static int rbd_header_add_snap(struct rbd_device *dev,
1618 const char *snap_name,
1619 gfp_t gfp_flags)
1621 int name_len = strlen(snap_name);
1622 u64 new_snapid;
1623 int ret;
1624 void *data, *p, *e;
1625 u64 ver;
1627 /* we should create a snapshot only if we're pointing at the head */
1628 if (dev->cur_snap)
1629 return -EINVAL;
1631 ret = ceph_monc_create_snapid(&dev->client->monc, dev->poolid,
1632 &new_snapid);
1633 dout("created snapid=%lld\n", new_snapid);
1634 if (ret < 0)
1635 return ret;
1637 data = kmalloc(name_len + 16, gfp_flags);
1638 if (!data)
1639 return -ENOMEM;
1641 p = data;
1642 e = data + name_len + 16;
1644 ceph_encode_string_safe(&p, e, snap_name, name_len, bad);
1645 ceph_encode_64_safe(&p, e, new_snapid, bad);
1647 ret = rbd_req_sync_exec(dev, dev->obj_md_name, "rbd", "snap_add",
1648 data, p - data, &ver);
1650 kfree(data);
1652 if (ret < 0)
1653 return ret;
1655 dev->header.snapc->seq = new_snapid;
1657 return 0;
1658 bad:
1659 return -ERANGE;
1662 static void __rbd_remove_all_snaps(struct rbd_device *rbd_dev)
1664 struct rbd_snap *snap;
1666 while (!list_empty(&rbd_dev->snaps)) {
1667 snap = list_first_entry(&rbd_dev->snaps, struct rbd_snap, node);
1668 __rbd_remove_snap_dev(rbd_dev, snap);
1673 * only read the first part of the ondisk header, without the snaps info
1675 static int __rbd_update_snaps(struct rbd_device *rbd_dev)
1677 int ret;
1678 struct rbd_image_header h;
1679 u64 snap_seq;
1680 int follow_seq = 0;
1682 ret = rbd_read_header(rbd_dev, &h);
1683 if (ret < 0)
1684 return ret;
1686 /* resized? */
1687 set_capacity(rbd_dev->disk, h.image_size / 512ULL);
1689 down_write(&rbd_dev->header.snap_rwsem);
1691 snap_seq = rbd_dev->header.snapc->seq;
1692 if (rbd_dev->header.total_snaps &&
1693 rbd_dev->header.snapc->snaps[0] == snap_seq)
1694 /* pointing at the head, will need to follow that
1695 if head moves */
1696 follow_seq = 1;
1698 kfree(rbd_dev->header.snapc);
1699 kfree(rbd_dev->header.snap_names);
1700 kfree(rbd_dev->header.snap_sizes);
1702 rbd_dev->header.total_snaps = h.total_snaps;
1703 rbd_dev->header.snapc = h.snapc;
1704 rbd_dev->header.snap_names = h.snap_names;
1705 rbd_dev->header.snap_names_len = h.snap_names_len;
1706 rbd_dev->header.snap_sizes = h.snap_sizes;
1707 if (follow_seq)
1708 rbd_dev->header.snapc->seq = rbd_dev->header.snapc->snaps[0];
1709 else
1710 rbd_dev->header.snapc->seq = snap_seq;
1712 ret = __rbd_init_snaps_header(rbd_dev);
1714 up_write(&rbd_dev->header.snap_rwsem);
1716 return ret;
1719 static int rbd_init_disk(struct rbd_device *rbd_dev)
1721 struct gendisk *disk;
1722 struct request_queue *q;
1723 int rc;
1724 u64 total_size = 0;
1726 /* contact OSD, request size info about the object being mapped */
1727 rc = rbd_read_header(rbd_dev, &rbd_dev->header);
1728 if (rc)
1729 return rc;
1731 /* no need to lock here, as rbd_dev is not registered yet */
1732 rc = __rbd_init_snaps_header(rbd_dev);
1733 if (rc)
1734 return rc;
1736 rc = rbd_header_set_snap(rbd_dev, rbd_dev->snap_name, &total_size);
1737 if (rc)
1738 return rc;
1740 /* create gendisk info */
1741 rc = -ENOMEM;
1742 disk = alloc_disk(RBD_MINORS_PER_MAJOR);
1743 if (!disk)
1744 goto out;
1746 snprintf(disk->disk_name, sizeof(disk->disk_name), DRV_NAME "%d",
1747 rbd_dev->id);
1748 disk->major = rbd_dev->major;
1749 disk->first_minor = 0;
1750 disk->fops = &rbd_bd_ops;
1751 disk->private_data = rbd_dev;
1753 /* init rq */
1754 rc = -ENOMEM;
1755 q = blk_init_queue(rbd_rq_fn, &rbd_dev->lock);
1756 if (!q)
1757 goto out_disk;
1759 /* set io sizes to object size */
1760 blk_queue_max_hw_sectors(q, rbd_obj_bytes(&rbd_dev->header) / 512ULL);
1761 blk_queue_max_segment_size(q, rbd_obj_bytes(&rbd_dev->header));
1762 blk_queue_io_min(q, rbd_obj_bytes(&rbd_dev->header));
1763 blk_queue_io_opt(q, rbd_obj_bytes(&rbd_dev->header));
1765 blk_queue_merge_bvec(q, rbd_merge_bvec);
1766 disk->queue = q;
1768 q->queuedata = rbd_dev;
1770 rbd_dev->disk = disk;
1771 rbd_dev->q = q;
1773 /* finally, announce the disk to the world */
1774 set_capacity(disk, total_size / 512ULL);
1775 add_disk(disk);
1777 pr_info("%s: added with size 0x%llx\n",
1778 disk->disk_name, (unsigned long long)total_size);
1779 return 0;
1781 out_disk:
1782 put_disk(disk);
1783 out:
1784 return rc;
1788 sysfs
1791 static ssize_t rbd_size_show(struct device *dev,
1792 struct device_attribute *attr, char *buf)
1794 struct rbd_device *rbd_dev = dev_to_rbd(dev);
1796 return sprintf(buf, "%llu\n", (unsigned long long)rbd_dev->header.image_size);
1799 static ssize_t rbd_major_show(struct device *dev,
1800 struct device_attribute *attr, char *buf)
1802 struct rbd_device *rbd_dev = dev_to_rbd(dev);
1804 return sprintf(buf, "%d\n", rbd_dev->major);
1807 static ssize_t rbd_client_id_show(struct device *dev,
1808 struct device_attribute *attr, char *buf)
1810 struct rbd_device *rbd_dev = dev_to_rbd(dev);
1812 return sprintf(buf, "client%lld\n", ceph_client_id(rbd_dev->client));
1815 static ssize_t rbd_pool_show(struct device *dev,
1816 struct device_attribute *attr, char *buf)
1818 struct rbd_device *rbd_dev = dev_to_rbd(dev);
1820 return sprintf(buf, "%s\n", rbd_dev->pool_name);
1823 static ssize_t rbd_name_show(struct device *dev,
1824 struct device_attribute *attr, char *buf)
1826 struct rbd_device *rbd_dev = dev_to_rbd(dev);
1828 return sprintf(buf, "%s\n", rbd_dev->obj);
1831 static ssize_t rbd_snap_show(struct device *dev,
1832 struct device_attribute *attr,
1833 char *buf)
1835 struct rbd_device *rbd_dev = dev_to_rbd(dev);
1837 return sprintf(buf, "%s\n", rbd_dev->snap_name);
1840 static ssize_t rbd_image_refresh(struct device *dev,
1841 struct device_attribute *attr,
1842 const char *buf,
1843 size_t size)
1845 struct rbd_device *rbd_dev = dev_to_rbd(dev);
1846 int rc;
1847 int ret = size;
1849 mutex_lock_nested(&ctl_mutex, SINGLE_DEPTH_NESTING);
1851 rc = __rbd_update_snaps(rbd_dev);
1852 if (rc < 0)
1853 ret = rc;
1855 mutex_unlock(&ctl_mutex);
1856 return ret;
1859 static DEVICE_ATTR(size, S_IRUGO, rbd_size_show, NULL);
1860 static DEVICE_ATTR(major, S_IRUGO, rbd_major_show, NULL);
1861 static DEVICE_ATTR(client_id, S_IRUGO, rbd_client_id_show, NULL);
1862 static DEVICE_ATTR(pool, S_IRUGO, rbd_pool_show, NULL);
1863 static DEVICE_ATTR(name, S_IRUGO, rbd_name_show, NULL);
1864 static DEVICE_ATTR(refresh, S_IWUSR, NULL, rbd_image_refresh);
1865 static DEVICE_ATTR(current_snap, S_IRUGO, rbd_snap_show, NULL);
1866 static DEVICE_ATTR(create_snap, S_IWUSR, NULL, rbd_snap_add);
1868 static struct attribute *rbd_attrs[] = {
1869 &dev_attr_size.attr,
1870 &dev_attr_major.attr,
1871 &dev_attr_client_id.attr,
1872 &dev_attr_pool.attr,
1873 &dev_attr_name.attr,
1874 &dev_attr_current_snap.attr,
1875 &dev_attr_refresh.attr,
1876 &dev_attr_create_snap.attr,
1877 NULL
1880 static struct attribute_group rbd_attr_group = {
1881 .attrs = rbd_attrs,
1884 static const struct attribute_group *rbd_attr_groups[] = {
1885 &rbd_attr_group,
1886 NULL
1889 static void rbd_sysfs_dev_release(struct device *dev)
1893 static struct device_type rbd_device_type = {
1894 .name = "rbd",
1895 .groups = rbd_attr_groups,
1896 .release = rbd_sysfs_dev_release,
1901 sysfs - snapshots
1904 static ssize_t rbd_snap_size_show(struct device *dev,
1905 struct device_attribute *attr,
1906 char *buf)
1908 struct rbd_snap *snap = container_of(dev, struct rbd_snap, dev);
1910 return sprintf(buf, "%lld\n", (long long)snap->size);
1913 static ssize_t rbd_snap_id_show(struct device *dev,
1914 struct device_attribute *attr,
1915 char *buf)
1917 struct rbd_snap *snap = container_of(dev, struct rbd_snap, dev);
1919 return sprintf(buf, "%lld\n", (long long)snap->id);
1922 static DEVICE_ATTR(snap_size, S_IRUGO, rbd_snap_size_show, NULL);
1923 static DEVICE_ATTR(snap_id, S_IRUGO, rbd_snap_id_show, NULL);
1925 static struct attribute *rbd_snap_attrs[] = {
1926 &dev_attr_snap_size.attr,
1927 &dev_attr_snap_id.attr,
1928 NULL,
1931 static struct attribute_group rbd_snap_attr_group = {
1932 .attrs = rbd_snap_attrs,
1935 static void rbd_snap_dev_release(struct device *dev)
1937 struct rbd_snap *snap = container_of(dev, struct rbd_snap, dev);
1938 kfree(snap->name);
1939 kfree(snap);
1942 static const struct attribute_group *rbd_snap_attr_groups[] = {
1943 &rbd_snap_attr_group,
1944 NULL
1947 static struct device_type rbd_snap_device_type = {
1948 .groups = rbd_snap_attr_groups,
1949 .release = rbd_snap_dev_release,
1952 static void __rbd_remove_snap_dev(struct rbd_device *rbd_dev,
1953 struct rbd_snap *snap)
1955 list_del(&snap->node);
1956 device_unregister(&snap->dev);
1959 static int rbd_register_snap_dev(struct rbd_device *rbd_dev,
1960 struct rbd_snap *snap,
1961 struct device *parent)
1963 struct device *dev = &snap->dev;
1964 int ret;
1966 dev->type = &rbd_snap_device_type;
1967 dev->parent = parent;
1968 dev->release = rbd_snap_dev_release;
1969 dev_set_name(dev, "snap_%s", snap->name);
1970 ret = device_register(dev);
1972 return ret;
1975 static int __rbd_add_snap_dev(struct rbd_device *rbd_dev,
1976 int i, const char *name,
1977 struct rbd_snap **snapp)
1979 int ret;
1980 struct rbd_snap *snap = kzalloc(sizeof(*snap), GFP_KERNEL);
1981 if (!snap)
1982 return -ENOMEM;
1983 snap->name = kstrdup(name, GFP_KERNEL);
1984 snap->size = rbd_dev->header.snap_sizes[i];
1985 snap->id = rbd_dev->header.snapc->snaps[i];
1986 if (device_is_registered(&rbd_dev->dev)) {
1987 ret = rbd_register_snap_dev(rbd_dev, snap,
1988 &rbd_dev->dev);
1989 if (ret < 0)
1990 goto err;
1992 *snapp = snap;
1993 return 0;
1994 err:
1995 kfree(snap->name);
1996 kfree(snap);
1997 return ret;
2001 * search for the previous snap in a null delimited string list
2003 const char *rbd_prev_snap_name(const char *name, const char *start)
2005 if (name < start + 2)
2006 return NULL;
2008 name -= 2;
2009 while (*name) {
2010 if (name == start)
2011 return start;
2012 name--;
2014 return name + 1;
2018 * compare the old list of snapshots that we have to what's in the header
2019 * and update it accordingly. Note that the header holds the snapshots
2020 * in a reverse order (from newest to oldest) and we need to go from
2021 * older to new so that we don't get a duplicate snap name when
2022 * doing the process (e.g., removed snapshot and recreated a new
2023 * one with the same name.
2025 static int __rbd_init_snaps_header(struct rbd_device *rbd_dev)
2027 const char *name, *first_name;
2028 int i = rbd_dev->header.total_snaps;
2029 struct rbd_snap *snap, *old_snap = NULL;
2030 int ret;
2031 struct list_head *p, *n;
2033 first_name = rbd_dev->header.snap_names;
2034 name = first_name + rbd_dev->header.snap_names_len;
2036 list_for_each_prev_safe(p, n, &rbd_dev->snaps) {
2037 u64 cur_id;
2039 old_snap = list_entry(p, struct rbd_snap, node);
2041 if (i)
2042 cur_id = rbd_dev->header.snapc->snaps[i - 1];
2044 if (!i || old_snap->id < cur_id) {
2045 /* old_snap->id was skipped, thus was removed */
2046 __rbd_remove_snap_dev(rbd_dev, old_snap);
2047 continue;
2049 if (old_snap->id == cur_id) {
2050 /* we have this snapshot already */
2051 i--;
2052 name = rbd_prev_snap_name(name, first_name);
2053 continue;
2055 for (; i > 0;
2056 i--, name = rbd_prev_snap_name(name, first_name)) {
2057 if (!name) {
2058 WARN_ON(1);
2059 return -EINVAL;
2061 cur_id = rbd_dev->header.snapc->snaps[i];
2062 /* snapshot removal? handle it above */
2063 if (cur_id >= old_snap->id)
2064 break;
2065 /* a new snapshot */
2066 ret = __rbd_add_snap_dev(rbd_dev, i - 1, name, &snap);
2067 if (ret < 0)
2068 return ret;
2070 /* note that we add it backward so using n and not p */
2071 list_add(&snap->node, n);
2072 p = &snap->node;
2075 /* we're done going over the old snap list, just add what's left */
2076 for (; i > 0; i--) {
2077 name = rbd_prev_snap_name(name, first_name);
2078 if (!name) {
2079 WARN_ON(1);
2080 return -EINVAL;
2082 ret = __rbd_add_snap_dev(rbd_dev, i - 1, name, &snap);
2083 if (ret < 0)
2084 return ret;
2085 list_add(&snap->node, &rbd_dev->snaps);
2088 return 0;
2092 static void rbd_root_dev_release(struct device *dev)
2096 static struct device rbd_root_dev = {
2097 .init_name = "rbd",
2098 .release = rbd_root_dev_release,
2101 static int rbd_bus_add_dev(struct rbd_device *rbd_dev)
2103 int ret = -ENOMEM;
2104 struct device *dev;
2105 struct rbd_snap *snap;
2107 mutex_lock_nested(&ctl_mutex, SINGLE_DEPTH_NESTING);
2108 dev = &rbd_dev->dev;
2110 dev->bus = &rbd_bus_type;
2111 dev->type = &rbd_device_type;
2112 dev->parent = &rbd_root_dev;
2113 dev->release = rbd_dev_release;
2114 dev_set_name(dev, "%d", rbd_dev->id);
2115 ret = device_register(dev);
2116 if (ret < 0)
2117 goto done_free;
2119 list_for_each_entry(snap, &rbd_dev->snaps, node) {
2120 ret = rbd_register_snap_dev(rbd_dev, snap,
2121 &rbd_dev->dev);
2122 if (ret < 0)
2123 break;
2126 mutex_unlock(&ctl_mutex);
2127 return 0;
2128 done_free:
2129 mutex_unlock(&ctl_mutex);
2130 return ret;
2133 static void rbd_bus_del_dev(struct rbd_device *rbd_dev)
2135 device_unregister(&rbd_dev->dev);
2138 static int rbd_init_watch_dev(struct rbd_device *rbd_dev)
2140 int ret, rc;
2142 do {
2143 ret = rbd_req_sync_watch(rbd_dev, rbd_dev->obj_md_name,
2144 rbd_dev->header.obj_version);
2145 if (ret == -ERANGE) {
2146 mutex_lock_nested(&ctl_mutex, SINGLE_DEPTH_NESTING);
2147 rc = __rbd_update_snaps(rbd_dev);
2148 mutex_unlock(&ctl_mutex);
2149 if (rc < 0)
2150 return rc;
2152 } while (ret == -ERANGE);
2154 return ret;
2157 static ssize_t rbd_add(struct bus_type *bus,
2158 const char *buf,
2159 size_t count)
2161 struct ceph_osd_client *osdc;
2162 struct rbd_device *rbd_dev;
2163 ssize_t rc = -ENOMEM;
2164 int irc, new_id = 0;
2165 struct list_head *tmp;
2166 char *mon_dev_name;
2167 char *options;
2169 if (!try_module_get(THIS_MODULE))
2170 return -ENODEV;
2172 mon_dev_name = kmalloc(RBD_MAX_OPT_LEN, GFP_KERNEL);
2173 if (!mon_dev_name)
2174 goto err_out_mod;
2176 options = kmalloc(RBD_MAX_OPT_LEN, GFP_KERNEL);
2177 if (!options)
2178 goto err_mon_dev;
2180 /* new rbd_device object */
2181 rbd_dev = kzalloc(sizeof(*rbd_dev), GFP_KERNEL);
2182 if (!rbd_dev)
2183 goto err_out_opt;
2185 /* static rbd_device initialization */
2186 spin_lock_init(&rbd_dev->lock);
2187 INIT_LIST_HEAD(&rbd_dev->node);
2188 INIT_LIST_HEAD(&rbd_dev->snaps);
2190 init_rwsem(&rbd_dev->header.snap_rwsem);
2192 /* generate unique id: find highest unique id, add one */
2193 mutex_lock_nested(&ctl_mutex, SINGLE_DEPTH_NESTING);
2195 list_for_each(tmp, &rbd_dev_list) {
2196 struct rbd_device *rbd_dev;
2198 rbd_dev = list_entry(tmp, struct rbd_device, node);
2199 if (rbd_dev->id >= new_id)
2200 new_id = rbd_dev->id + 1;
2203 rbd_dev->id = new_id;
2205 /* add to global list */
2206 list_add_tail(&rbd_dev->node, &rbd_dev_list);
2208 /* parse add command */
2209 if (sscanf(buf, "%" __stringify(RBD_MAX_OPT_LEN) "s "
2210 "%" __stringify(RBD_MAX_OPT_LEN) "s "
2211 "%" __stringify(RBD_MAX_POOL_NAME_LEN) "s "
2212 "%" __stringify(RBD_MAX_OBJ_NAME_LEN) "s"
2213 "%" __stringify(RBD_MAX_SNAP_NAME_LEN) "s",
2214 mon_dev_name, options, rbd_dev->pool_name,
2215 rbd_dev->obj, rbd_dev->snap_name) < 4) {
2216 rc = -EINVAL;
2217 goto err_out_slot;
2220 if (rbd_dev->snap_name[0] == 0)
2221 rbd_dev->snap_name[0] = '-';
2223 rbd_dev->obj_len = strlen(rbd_dev->obj);
2224 snprintf(rbd_dev->obj_md_name, sizeof(rbd_dev->obj_md_name), "%s%s",
2225 rbd_dev->obj, RBD_SUFFIX);
2227 /* initialize rest of new object */
2228 snprintf(rbd_dev->name, DEV_NAME_LEN, DRV_NAME "%d", rbd_dev->id);
2229 rc = rbd_get_client(rbd_dev, mon_dev_name, options);
2230 if (rc < 0)
2231 goto err_out_slot;
2233 mutex_unlock(&ctl_mutex);
2235 /* pick the pool */
2236 osdc = &rbd_dev->client->osdc;
2237 rc = ceph_pg_poolid_by_name(osdc->osdmap, rbd_dev->pool_name);
2238 if (rc < 0)
2239 goto err_out_client;
2240 rbd_dev->poolid = rc;
2242 /* register our block device */
2243 irc = register_blkdev(0, rbd_dev->name);
2244 if (irc < 0) {
2245 rc = irc;
2246 goto err_out_client;
2248 rbd_dev->major = irc;
2250 rc = rbd_bus_add_dev(rbd_dev);
2251 if (rc)
2252 goto err_out_blkdev;
2254 /* set up and announce blkdev mapping */
2255 rc = rbd_init_disk(rbd_dev);
2256 if (rc)
2257 goto err_out_bus;
2259 rc = rbd_init_watch_dev(rbd_dev);
2260 if (rc)
2261 goto err_out_bus;
2263 return count;
2265 err_out_bus:
2266 mutex_lock_nested(&ctl_mutex, SINGLE_DEPTH_NESTING);
2267 list_del_init(&rbd_dev->node);
2268 mutex_unlock(&ctl_mutex);
2270 /* this will also clean up rest of rbd_dev stuff */
2272 rbd_bus_del_dev(rbd_dev);
2273 kfree(options);
2274 kfree(mon_dev_name);
2275 return rc;
2277 err_out_blkdev:
2278 unregister_blkdev(rbd_dev->major, rbd_dev->name);
2279 err_out_client:
2280 rbd_put_client(rbd_dev);
2281 mutex_lock_nested(&ctl_mutex, SINGLE_DEPTH_NESTING);
2282 err_out_slot:
2283 list_del_init(&rbd_dev->node);
2284 mutex_unlock(&ctl_mutex);
2286 kfree(rbd_dev);
2287 err_out_opt:
2288 kfree(options);
2289 err_mon_dev:
2290 kfree(mon_dev_name);
2291 err_out_mod:
2292 dout("Error adding device %s\n", buf);
2293 module_put(THIS_MODULE);
2294 return rc;
2297 static struct rbd_device *__rbd_get_dev(unsigned long id)
2299 struct list_head *tmp;
2300 struct rbd_device *rbd_dev;
2302 list_for_each(tmp, &rbd_dev_list) {
2303 rbd_dev = list_entry(tmp, struct rbd_device, node);
2304 if (rbd_dev->id == id)
2305 return rbd_dev;
2307 return NULL;
2310 static void rbd_dev_release(struct device *dev)
2312 struct rbd_device *rbd_dev =
2313 container_of(dev, struct rbd_device, dev);
2315 if (rbd_dev->watch_request)
2316 ceph_osdc_unregister_linger_request(&rbd_dev->client->osdc,
2317 rbd_dev->watch_request);
2318 if (rbd_dev->watch_event)
2319 rbd_req_sync_unwatch(rbd_dev, rbd_dev->obj_md_name);
2321 rbd_put_client(rbd_dev);
2323 /* clean up and free blkdev */
2324 rbd_free_disk(rbd_dev);
2325 unregister_blkdev(rbd_dev->major, rbd_dev->name);
2326 kfree(rbd_dev);
2328 /* release module ref */
2329 module_put(THIS_MODULE);
2332 static ssize_t rbd_remove(struct bus_type *bus,
2333 const char *buf,
2334 size_t count)
2336 struct rbd_device *rbd_dev = NULL;
2337 int target_id, rc;
2338 unsigned long ul;
2339 int ret = count;
2341 rc = strict_strtoul(buf, 10, &ul);
2342 if (rc)
2343 return rc;
2345 /* convert to int; abort if we lost anything in the conversion */
2346 target_id = (int) ul;
2347 if (target_id != ul)
2348 return -EINVAL;
2350 mutex_lock_nested(&ctl_mutex, SINGLE_DEPTH_NESTING);
2352 rbd_dev = __rbd_get_dev(target_id);
2353 if (!rbd_dev) {
2354 ret = -ENOENT;
2355 goto done;
2358 list_del_init(&rbd_dev->node);
2360 __rbd_remove_all_snaps(rbd_dev);
2361 rbd_bus_del_dev(rbd_dev);
2363 done:
2364 mutex_unlock(&ctl_mutex);
2365 return ret;
2368 static ssize_t rbd_snap_add(struct device *dev,
2369 struct device_attribute *attr,
2370 const char *buf,
2371 size_t count)
2373 struct rbd_device *rbd_dev = dev_to_rbd(dev);
2374 int ret;
2375 char *name = kmalloc(count + 1, GFP_KERNEL);
2376 if (!name)
2377 return -ENOMEM;
2379 snprintf(name, count, "%s", buf);
2381 mutex_lock_nested(&ctl_mutex, SINGLE_DEPTH_NESTING);
2383 ret = rbd_header_add_snap(rbd_dev,
2384 name, GFP_KERNEL);
2385 if (ret < 0)
2386 goto err_unlock;
2388 ret = __rbd_update_snaps(rbd_dev);
2389 if (ret < 0)
2390 goto err_unlock;
2392 /* shouldn't hold ctl_mutex when notifying.. notify might
2393 trigger a watch callback that would need to get that mutex */
2394 mutex_unlock(&ctl_mutex);
2396 /* make a best effort, don't error if failed */
2397 rbd_req_sync_notify(rbd_dev, rbd_dev->obj_md_name);
2399 ret = count;
2400 kfree(name);
2401 return ret;
2403 err_unlock:
2404 mutex_unlock(&ctl_mutex);
2405 kfree(name);
2406 return ret;
2409 static struct bus_attribute rbd_bus_attrs[] = {
2410 __ATTR(add, S_IWUSR, NULL, rbd_add),
2411 __ATTR(remove, S_IWUSR, NULL, rbd_remove),
2412 __ATTR_NULL
2416 * create control files in sysfs
2417 * /sys/bus/rbd/...
2419 static int rbd_sysfs_init(void)
2421 int ret;
2423 rbd_bus_type.bus_attrs = rbd_bus_attrs;
2425 ret = bus_register(&rbd_bus_type);
2426 if (ret < 0)
2427 return ret;
2429 ret = device_register(&rbd_root_dev);
2431 return ret;
2434 static void rbd_sysfs_cleanup(void)
2436 device_unregister(&rbd_root_dev);
2437 bus_unregister(&rbd_bus_type);
2440 int __init rbd_init(void)
2442 int rc;
2444 rc = rbd_sysfs_init();
2445 if (rc)
2446 return rc;
2447 spin_lock_init(&node_lock);
2448 pr_info("loaded " DRV_NAME_LONG "\n");
2449 return 0;
2452 void __exit rbd_exit(void)
2454 rbd_sysfs_cleanup();
2457 module_init(rbd_init);
2458 module_exit(rbd_exit);
2460 MODULE_AUTHOR("Sage Weil <sage@newdream.net>");
2461 MODULE_AUTHOR("Yehuda Sadeh <yehuda@hq.newdream.net>");
2462 MODULE_DESCRIPTION("rados block device");
2464 /* following authorship retained from original osdblk.c */
2465 MODULE_AUTHOR("Jeff Garzik <jeff@garzik.org>");
2467 MODULE_LICENSE("GPL");