1 /* $NetBSD: dm_table.c,v 1.5 2010/01/04 00:19:08 haad Exp $ */
4 * Copyright (c) 2010-2011 Alex Hornung <alex@alexhornung.com>
5 * Copyright (c) 2008 The NetBSD Foundation, Inc.
8 * This code is derived from software contributed to The NetBSD Foundation
11 * Redistribution and use in source and binary forms, with or without
12 * modification, are permitted provided that the following conditions
14 * 1. Redistributions of source code must retain the above copyright
15 * notice, this list of conditions and the following disclaimer.
16 * 2. Redistributions in binary form must reproduce the above copyright
17 * notice, this list of conditions and the following disclaimer in the
18 * documentation and/or other materials provided with the distribution.
20 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
21 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
22 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
23 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
24 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30 * POSSIBILITY OF SUCH DAMAGE.
33 #include <sys/param.h>
34 #include <sys/malloc.h>
35 #include <cpu/atomic.h>
36 #include <dev/disk/dm/dm.h>
39 * There are two types of users of this interface:
42 * dmstrategy, dmgetdisklabel, dmsize, dm_dev_status_ioctl,
43 * dm_table_deps_ioctl, dm_table_status_ioctl, dm_table_reload_ioctl
46 * dm_dev_remove_ioctl, dm_dev_resume_ioctl, dm_table_clear_ioctl
48 * Writers can work with table_head only when there are no readers. We
49 * simply use shared/exclusive locking to ensure this.
53 * Function to increment table user reference counter. Return id
55 * DM_TABLE_ACTIVE will return active table id.
56 * DM_TABLE_INACTIVE will return inactive table id.
59 dm_table_busy(dm_table_head_t
*head
, uint8_t table_id
)
65 lockmgr(&head
->table_mtx
, LK_SHARED
);
67 if (table_id
== DM_TABLE_ACTIVE
)
68 id
= head
->cur_active_table
;
70 id
= 1 - head
->cur_active_table
;
72 atomic_add_int(&head
->io_cnt
, 1);
78 * Function release table lock and eventually wakeup all waiters.
81 dm_table_unbusy(dm_table_head_t
*head
)
83 KKASSERT(head
->io_cnt
!= 0);
85 atomic_subtract_int(&head
->io_cnt
, 1);
87 lockmgr(&head
->table_mtx
, LK_RELEASE
);
91 * Return current active table to caller, increment io_cnt reference counter.
94 dm_table_get_entry(dm_table_head_t
*head
, uint8_t table_id
)
98 id
= dm_table_busy(head
, table_id
);
100 return &head
->tables
[id
];
104 * Decrement io reference counter and release shared lock.
107 dm_table_release(dm_table_head_t
*head
, uint8_t table_id
)
109 dm_table_unbusy(head
);
113 * Switch table from inactive to active mode. Have to wait until io_cnt is 0.
116 dm_table_switch_tables(dm_table_head_t
*head
)
118 lockmgr(&head
->table_mtx
, LK_EXCLUSIVE
);
120 head
->cur_active_table
= 1 - head
->cur_active_table
;
122 lockmgr(&head
->table_mtx
, LK_RELEASE
);
126 * Destroy all table data. This function can run when there are no
127 * readers on table lists.
130 dm_table_destroy(dm_table_head_t
*head
, uint8_t table_id
)
133 dm_table_entry_t
*table_en
;
136 lockmgr(&head
->table_mtx
, LK_EXCLUSIVE
);
138 dmdebug("table_id=%d io_cnt=%d\n", table_id
, head
->io_cnt
);
140 if (table_id
== DM_TABLE_ACTIVE
)
141 id
= head
->cur_active_table
;
143 id
= 1 - head
->cur_active_table
;
145 tbl
= &head
->tables
[id
];
147 while ((table_en
= TAILQ_FIRST(tbl
)) != NULL
) {
148 TAILQ_REMOVE(tbl
, table_en
, next
);
150 if (table_en
->target
->destroy
)
151 table_en
->target
->destroy(table_en
);
152 table_en
->target_config
= NULL
;
154 dm_table_free_deps(table_en
);
156 /* decrement the refcount for the target */
157 dm_target_unbusy(table_en
->target
);
159 kfree(table_en
, M_DM
);
161 KKASSERT(TAILQ_EMPTY(tbl
));
163 lockmgr(&head
->table_mtx
, LK_RELEASE
);
169 * Return length of active or inactive table in device.
172 _dm_table_size(dm_table_head_t
*head
, int table
)
175 dm_table_entry_t
*table_en
;
180 /* Select active table */
181 tbl
= dm_table_get_entry(head
, table
);
184 * Find out what tables I want to select.
185 * if length => rawblkno then we should used that table.
187 TAILQ_FOREACH(table_en
, tbl
, next
) {
188 length
+= table_en
->length
;
191 dm_table_unbusy(head
);
197 dm_table_size(dm_table_head_t
*head
)
199 return _dm_table_size(head
, DM_TABLE_ACTIVE
);
203 dm_inactive_table_size(dm_table_head_t
*head
)
205 return _dm_table_size(head
, DM_TABLE_INACTIVE
);
209 * Return > 0 if table is at least one table entry (returns number of entries)
210 * and return 0 if there is not. Target count returned from this function
211 * doesn't need to be true when userspace user receive it (after return
212 * there can be dm_dev_resume_ioctl), therefore this is only informative.
215 dm_table_get_target_count(dm_table_head_t
*head
, uint8_t table_id
)
217 dm_table_entry_t
*table_en
;
219 uint32_t target_count
;
223 tbl
= dm_table_get_entry(head
, table_id
);
225 TAILQ_FOREACH(table_en
, tbl
, next
)
228 dm_table_unbusy(head
);
234 * Initialize dm_table_head_t structures, I'm trying to keep this structure as
235 * opaque as possible.
238 dm_table_head_init(dm_table_head_t
*head
)
240 head
->cur_active_table
= 0;
243 /* Initialize tables. */
244 TAILQ_INIT(&head
->tables
[0]);
245 TAILQ_INIT(&head
->tables
[1]);
247 lockinit(&head
->table_mtx
, "dmtbl", 0, LK_CANRECURSE
);
251 * Destroy all variables in table_head
254 dm_table_head_destroy(dm_table_head_t
*head
)
256 KKASSERT(!lockinuse(&head
->table_mtx
));
258 /* tables don't exist when I call this routine, therefore it
259 * doesn't make sense to have io_cnt != 0 */
260 KKASSERT(head
->io_cnt
== 0);
262 lockuninit(&head
->table_mtx
);
266 dm_table_init_target(dm_table_entry_t
*table_en
, void *cfg
)
268 table_en
->target_config
= cfg
;
272 dm_table_add_deps(dm_table_entry_t
*table_en
, dm_pdev_t
*pdev
)
274 dm_table_head_t
*head
;
279 head
= &table_en
->dev
->table_head
;
280 lockmgr(&head
->table_mtx
, LK_SHARED
);
282 TAILQ_FOREACH(map
, &table_en
->pdev_maps
, next
) {
283 if (map
->data
.pdev
->udev
== pdev
->udev
) {
284 lockmgr(&head
->table_mtx
, LK_RELEASE
);
289 map
= kmalloc(sizeof(*map
), M_DM
, M_WAITOK
| M_ZERO
);
290 map
->data
.pdev
= pdev
;
291 TAILQ_INSERT_TAIL(&table_en
->pdev_maps
, map
, next
);
293 lockmgr(&head
->table_mtx
, LK_RELEASE
);
299 dm_table_free_deps(dm_table_entry_t
*table_en
)
301 dm_table_head_t
*head
;
304 head
= &table_en
->dev
->table_head
;
305 lockmgr(&head
->table_mtx
, LK_SHARED
);
307 while ((map
= TAILQ_FIRST(&table_en
->pdev_maps
)) != NULL
) {
308 TAILQ_REMOVE(&table_en
->pdev_maps
, map
, next
);
311 KKASSERT(TAILQ_EMPTY(&table_en
->pdev_maps
));
313 lockmgr(&head
->table_mtx
, LK_RELEASE
);