2 * mdt.c - meta data file for NILFS
4 * Copyright (C) 2005-2008 Nippon Telegraph and Telephone Corporation.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
20 * Written by Ryusuke Konishi <ryusuke@osrg.net>
23 #include <linux/buffer_head.h>
24 #include <linux/mpage.h>
26 #include <linux/writeback.h>
27 #include <linux/backing-dev.h>
28 #include <linux/swap.h>
29 #include <linux/slab.h>
37 #define NILFS_MDT_MAX_RA_BLOCKS (16 - 1)
41 nilfs_mdt_insert_new_block(struct inode
*inode
, unsigned long block
,
42 struct buffer_head
*bh
,
43 void (*init_block
)(struct inode
*,
44 struct buffer_head
*, void *))
46 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
50 /* Caller exclude read accesses using page lock */
52 /* set_buffer_new(bh); */
55 ret
= nilfs_bmap_insert(ii
->i_bmap
, block
, (unsigned long)bh
);
59 set_buffer_mapped(bh
);
61 kaddr
= kmap_atomic(bh
->b_page
, KM_USER0
);
62 memset(kaddr
+ bh_offset(bh
), 0, 1 << inode
->i_blkbits
);
64 init_block(inode
, bh
, kaddr
);
65 flush_dcache_page(bh
->b_page
);
66 kunmap_atomic(kaddr
, KM_USER0
);
68 set_buffer_uptodate(bh
);
69 nilfs_mark_buffer_dirty(bh
);
70 nilfs_mdt_mark_dirty(inode
);
74 static int nilfs_mdt_create_block(struct inode
*inode
, unsigned long block
,
75 struct buffer_head
**out_bh
,
76 void (*init_block
)(struct inode
*,
80 struct super_block
*sb
= inode
->i_sb
;
81 struct nilfs_transaction_info ti
;
82 struct buffer_head
*bh
;
85 nilfs_transaction_begin(sb
, &ti
, 0);
88 bh
= nilfs_grab_buffer(inode
, inode
->i_mapping
, block
, 0);
93 if (buffer_uptodate(bh
))
97 if (buffer_uptodate(bh
))
100 bh
->b_bdev
= sb
->s_bdev
;
101 err
= nilfs_mdt_insert_new_block(inode
, block
, bh
, init_block
);
108 unlock_page(bh
->b_page
);
109 page_cache_release(bh
->b_page
);
114 err
= nilfs_transaction_commit(sb
);
116 nilfs_transaction_abort(sb
);
122 nilfs_mdt_submit_block(struct inode
*inode
, unsigned long blkoff
,
123 int mode
, struct buffer_head
**out_bh
)
125 struct buffer_head
*bh
;
129 bh
= nilfs_grab_buffer(inode
, inode
->i_mapping
, blkoff
, 0);
133 ret
= -EEXIST
; /* internal code */
134 if (buffer_uptodate(bh
))
138 if (!trylock_buffer(bh
)) {
142 } else /* mode == READ */
145 if (buffer_uptodate(bh
)) {
150 ret
= nilfs_bmap_lookup(NILFS_I(inode
)->i_bmap
, blkoff
, &blknum
);
155 map_bh(bh
, inode
->i_sb
, (sector_t
)blknum
);
157 bh
->b_end_io
= end_buffer_read_sync
;
166 unlock_page(bh
->b_page
);
167 page_cache_release(bh
->b_page
);
173 static int nilfs_mdt_read_block(struct inode
*inode
, unsigned long block
,
174 int readahead
, struct buffer_head
**out_bh
)
176 struct buffer_head
*first_bh
, *bh
;
177 unsigned long blkoff
;
178 int i
, nr_ra_blocks
= NILFS_MDT_MAX_RA_BLOCKS
;
181 err
= nilfs_mdt_submit_block(inode
, block
, READ
, &first_bh
);
182 if (err
== -EEXIST
) /* internal code */
190 for (i
= 0; i
< nr_ra_blocks
; i
++, blkoff
++) {
191 err
= nilfs_mdt_submit_block(inode
, blkoff
, READA
, &bh
);
192 if (likely(!err
|| err
== -EEXIST
))
194 else if (err
!= -EBUSY
)
196 /* abort readahead if bmap lookup failed */
197 if (!buffer_locked(first_bh
))
202 wait_on_buffer(first_bh
);
206 if (!buffer_uptodate(first_bh
))
219 * nilfs_mdt_get_block - read or create a buffer on meta data file.
220 * @inode: inode of the meta data file
221 * @blkoff: block offset
222 * @create: create flag
223 * @init_block: initializer used for newly allocated block
224 * @out_bh: output of a pointer to the buffer_head
226 * nilfs_mdt_get_block() looks up the specified buffer and tries to create
227 * a new buffer if @create is not zero. On success, the returned buffer is
228 * assured to be either existing or formatted using a buffer lock on success.
229 * @out_bh is substituted only when zero is returned.
231 * Return Value: On success, it returns 0. On error, the following negative
232 * error code is returned.
234 * %-ENOMEM - Insufficient memory available.
238 * %-ENOENT - the specified block does not exist (hole block)
240 * %-EROFS - Read only filesystem (for create mode)
242 int nilfs_mdt_get_block(struct inode
*inode
, unsigned long blkoff
, int create
,
243 void (*init_block
)(struct inode
*,
244 struct buffer_head
*, void *),
245 struct buffer_head
**out_bh
)
249 /* Should be rewritten with merging nilfs_mdt_read_block() */
251 ret
= nilfs_mdt_read_block(inode
, blkoff
, !create
, out_bh
);
252 if (!create
|| ret
!= -ENOENT
)
255 ret
= nilfs_mdt_create_block(inode
, blkoff
, out_bh
, init_block
);
256 if (unlikely(ret
== -EEXIST
)) {
257 /* create = 0; */ /* limit read-create loop retries */
264 * nilfs_mdt_delete_block - make a hole on the meta data file.
265 * @inode: inode of the meta data file
266 * @block: block offset
268 * Return Value: On success, zero is returned.
269 * On error, one of the following negative error code is returned.
271 * %-ENOMEM - Insufficient memory available.
275 int nilfs_mdt_delete_block(struct inode
*inode
, unsigned long block
)
277 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
280 err
= nilfs_bmap_delete(ii
->i_bmap
, block
);
281 if (!err
|| err
== -ENOENT
) {
282 nilfs_mdt_mark_dirty(inode
);
283 nilfs_mdt_forget_block(inode
, block
);
289 * nilfs_mdt_forget_block - discard dirty state and try to remove the page
290 * @inode: inode of the meta data file
291 * @block: block offset
293 * nilfs_mdt_forget_block() clears a dirty flag of the specified buffer, and
294 * tries to release the page including the buffer from a page cache.
296 * Return Value: On success, 0 is returned. On error, one of the following
297 * negative error code is returned.
299 * %-EBUSY - page has an active buffer.
301 * %-ENOENT - page cache has no page addressed by the offset.
303 int nilfs_mdt_forget_block(struct inode
*inode
, unsigned long block
)
305 pgoff_t index
= (pgoff_t
)block
>>
306 (PAGE_CACHE_SHIFT
- inode
->i_blkbits
);
308 unsigned long first_block
;
312 page
= find_lock_page(inode
->i_mapping
, index
);
316 wait_on_page_writeback(page
);
318 first_block
= (unsigned long)index
<<
319 (PAGE_CACHE_SHIFT
- inode
->i_blkbits
);
320 if (page_has_buffers(page
)) {
321 struct buffer_head
*bh
;
323 bh
= nilfs_page_get_nth_block(page
, block
- first_block
);
324 nilfs_forget_buffer(bh
);
326 still_dirty
= PageDirty(page
);
328 page_cache_release(page
);
331 invalidate_inode_pages2_range(inode
->i_mapping
, index
, index
) != 0)
337 * nilfs_mdt_mark_block_dirty - mark a block on the meta data file dirty.
338 * @inode: inode of the meta data file
339 * @block: block offset
341 * Return Value: On success, it returns 0. On error, the following negative
342 * error code is returned.
344 * %-ENOMEM - Insufficient memory available.
348 * %-ENOENT - the specified block does not exist (hole block)
350 int nilfs_mdt_mark_block_dirty(struct inode
*inode
, unsigned long block
)
352 struct buffer_head
*bh
;
355 err
= nilfs_mdt_read_block(inode
, block
, 0, &bh
);
358 nilfs_mark_buffer_dirty(bh
);
359 nilfs_mdt_mark_dirty(inode
);
364 int nilfs_mdt_fetch_dirty(struct inode
*inode
)
366 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
368 if (nilfs_bmap_test_and_clear_dirty(ii
->i_bmap
)) {
369 set_bit(NILFS_I_DIRTY
, &ii
->i_state
);
372 return test_bit(NILFS_I_DIRTY
, &ii
->i_state
);
376 nilfs_mdt_write_page(struct page
*page
, struct writeback_control
*wbc
)
379 struct super_block
*sb
;
382 redirty_page_for_writepage(wbc
, page
);
385 inode
= page
->mapping
->host
;
391 if (wbc
->sync_mode
== WB_SYNC_ALL
)
392 err
= nilfs_construct_segment(sb
);
393 else if (wbc
->for_reclaim
)
394 nilfs_flush_segment(sb
, inode
->i_ino
);
400 static const struct address_space_operations def_mdt_aops
= {
401 .writepage
= nilfs_mdt_write_page
,
402 .sync_page
= block_sync_page
,
405 static const struct inode_operations def_mdt_iops
;
406 static const struct file_operations def_mdt_fops
;
409 int nilfs_mdt_init(struct inode
*inode
, gfp_t gfp_mask
, size_t objsz
)
411 struct nilfs_mdt_info
*mi
;
413 mi
= kzalloc(max(sizeof(*mi
), objsz
), GFP_NOFS
);
417 init_rwsem(&mi
->mi_sem
);
418 inode
->i_private
= mi
;
420 inode
->i_mode
= S_IFREG
;
421 mapping_set_gfp_mask(inode
->i_mapping
, gfp_mask
);
422 inode
->i_mapping
->backing_dev_info
= inode
->i_sb
->s_bdi
;
424 inode
->i_op
= &def_mdt_iops
;
425 inode
->i_fop
= &def_mdt_fops
;
426 inode
->i_mapping
->a_ops
= &def_mdt_aops
;
431 void nilfs_mdt_set_entry_size(struct inode
*inode
, unsigned entry_size
,
432 unsigned header_size
)
434 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
436 mi
->mi_entry_size
= entry_size
;
437 mi
->mi_entries_per_block
= (1 << inode
->i_blkbits
) / entry_size
;
438 mi
->mi_first_entry_offset
= DIV_ROUND_UP(header_size
, entry_size
);
441 static const struct address_space_operations shadow_map_aops
= {
442 .sync_page
= block_sync_page
,
446 * nilfs_mdt_setup_shadow_map - setup shadow map and bind it to metadata file
447 * @inode: inode of the metadata file
448 * @shadow: shadow mapping
450 int nilfs_mdt_setup_shadow_map(struct inode
*inode
,
451 struct nilfs_shadow_map
*shadow
)
453 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
454 struct backing_dev_info
*bdi
= inode
->i_sb
->s_bdi
;
456 INIT_LIST_HEAD(&shadow
->frozen_buffers
);
457 address_space_init_once(&shadow
->frozen_data
);
458 nilfs_mapping_init(&shadow
->frozen_data
, bdi
, &shadow_map_aops
);
459 address_space_init_once(&shadow
->frozen_btnodes
);
460 nilfs_mapping_init(&shadow
->frozen_btnodes
, bdi
, &shadow_map_aops
);
461 mi
->mi_shadow
= shadow
;
466 * nilfs_mdt_save_to_shadow_map - copy bmap and dirty pages to shadow map
467 * @inode: inode of the metadata file
469 int nilfs_mdt_save_to_shadow_map(struct inode
*inode
)
471 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
472 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
473 struct nilfs_shadow_map
*shadow
= mi
->mi_shadow
;
476 ret
= nilfs_copy_dirty_pages(&shadow
->frozen_data
, inode
->i_mapping
);
480 ret
= nilfs_copy_dirty_pages(&shadow
->frozen_btnodes
,
481 &ii
->i_btnode_cache
);
485 nilfs_bmap_save(ii
->i_bmap
, &shadow
->bmap_store
);
490 int nilfs_mdt_freeze_buffer(struct inode
*inode
, struct buffer_head
*bh
)
492 struct nilfs_shadow_map
*shadow
= NILFS_MDT(inode
)->mi_shadow
;
493 struct buffer_head
*bh_frozen
;
495 int blkbits
= inode
->i_blkbits
;
497 page
= grab_cache_page(&shadow
->frozen_data
, bh
->b_page
->index
);
501 if (!page_has_buffers(page
))
502 create_empty_buffers(page
, 1 << blkbits
, 0);
504 bh_frozen
= nilfs_page_get_nth_block(page
, bh_offset(bh
) >> blkbits
);
506 if (!buffer_uptodate(bh_frozen
))
507 nilfs_copy_buffer(bh_frozen
, bh
);
508 if (list_empty(&bh_frozen
->b_assoc_buffers
)) {
509 list_add_tail(&bh_frozen
->b_assoc_buffers
,
510 &shadow
->frozen_buffers
);
511 set_buffer_nilfs_redirected(bh
);
513 brelse(bh_frozen
); /* already frozen */
517 page_cache_release(page
);
522 nilfs_mdt_get_frozen_buffer(struct inode
*inode
, struct buffer_head
*bh
)
524 struct nilfs_shadow_map
*shadow
= NILFS_MDT(inode
)->mi_shadow
;
525 struct buffer_head
*bh_frozen
= NULL
;
529 page
= find_lock_page(&shadow
->frozen_data
, bh
->b_page
->index
);
531 if (page_has_buffers(page
)) {
532 n
= bh_offset(bh
) >> inode
->i_blkbits
;
533 bh_frozen
= nilfs_page_get_nth_block(page
, n
);
536 page_cache_release(page
);
541 static void nilfs_release_frozen_buffers(struct nilfs_shadow_map
*shadow
)
543 struct list_head
*head
= &shadow
->frozen_buffers
;
544 struct buffer_head
*bh
;
546 while (!list_empty(head
)) {
547 bh
= list_first_entry(head
, struct buffer_head
,
549 list_del_init(&bh
->b_assoc_buffers
);
550 brelse(bh
); /* drop ref-count to make it releasable */
555 * nilfs_mdt_restore_from_shadow_map - restore dirty pages and bmap state
556 * @inode: inode of the metadata file
558 void nilfs_mdt_restore_from_shadow_map(struct inode
*inode
)
560 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
561 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
562 struct nilfs_shadow_map
*shadow
= mi
->mi_shadow
;
564 down_write(&mi
->mi_sem
);
566 if (mi
->mi_palloc_cache
)
567 nilfs_palloc_clear_cache(inode
);
569 nilfs_clear_dirty_pages(inode
->i_mapping
);
570 nilfs_copy_back_pages(inode
->i_mapping
, &shadow
->frozen_data
);
572 nilfs_clear_dirty_pages(&ii
->i_btnode_cache
);
573 nilfs_copy_back_pages(&ii
->i_btnode_cache
, &shadow
->frozen_btnodes
);
575 nilfs_bmap_restore(ii
->i_bmap
, &shadow
->bmap_store
);
577 up_write(&mi
->mi_sem
);
581 * nilfs_mdt_clear_shadow_map - truncate pages in shadow map caches
582 * @inode: inode of the metadata file
584 void nilfs_mdt_clear_shadow_map(struct inode
*inode
)
586 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
587 struct nilfs_shadow_map
*shadow
= mi
->mi_shadow
;
589 down_write(&mi
->mi_sem
);
590 nilfs_release_frozen_buffers(shadow
);
591 truncate_inode_pages(&shadow
->frozen_data
, 0);
592 truncate_inode_pages(&shadow
->frozen_btnodes
, 0);
593 up_write(&mi
->mi_sem
);