2 * mdt.c - meta data file for NILFS
4 * Copyright (C) 2005-2008 Nippon Telegraph and Telephone Corporation.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
20 * Written by Ryusuke Konishi <ryusuke@osrg.net>
23 #include <linux/buffer_head.h>
24 #include <linux/mpage.h>
26 #include <linux/writeback.h>
27 #include <linux/backing-dev.h>
28 #include <linux/swap.h>
29 #include <linux/slab.h>
37 #define NILFS_MDT_MAX_RA_BLOCKS (16 - 1)
41 nilfs_mdt_insert_new_block(struct inode
*inode
, unsigned long block
,
42 struct buffer_head
*bh
,
43 void (*init_block
)(struct inode
*,
44 struct buffer_head
*, void *))
46 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
50 /* Caller exclude read accesses using page lock */
52 /* set_buffer_new(bh); */
55 ret
= nilfs_bmap_insert(ii
->i_bmap
, block
, (unsigned long)bh
);
59 set_buffer_mapped(bh
);
61 kaddr
= kmap_atomic(bh
->b_page
);
62 memset(kaddr
+ bh_offset(bh
), 0, 1 << inode
->i_blkbits
);
64 init_block(inode
, bh
, kaddr
);
65 flush_dcache_page(bh
->b_page
);
68 set_buffer_uptodate(bh
);
69 mark_buffer_dirty(bh
);
70 nilfs_mdt_mark_dirty(inode
);
74 static int nilfs_mdt_create_block(struct inode
*inode
, unsigned long block
,
75 struct buffer_head
**out_bh
,
76 void (*init_block
)(struct inode
*,
80 struct super_block
*sb
= inode
->i_sb
;
81 struct nilfs_transaction_info ti
;
82 struct buffer_head
*bh
;
85 nilfs_transaction_begin(sb
, &ti
, 0);
88 bh
= nilfs_grab_buffer(inode
, inode
->i_mapping
, block
, 0);
93 if (buffer_uptodate(bh
))
97 if (buffer_uptodate(bh
))
100 bh
->b_bdev
= sb
->s_bdev
;
101 err
= nilfs_mdt_insert_new_block(inode
, block
, bh
, init_block
);
108 unlock_page(bh
->b_page
);
109 page_cache_release(bh
->b_page
);
114 err
= nilfs_transaction_commit(sb
);
116 nilfs_transaction_abort(sb
);
122 nilfs_mdt_submit_block(struct inode
*inode
, unsigned long blkoff
,
123 int mode
, struct buffer_head
**out_bh
)
125 struct buffer_head
*bh
;
129 bh
= nilfs_grab_buffer(inode
, inode
->i_mapping
, blkoff
, 0);
133 ret
= -EEXIST
; /* internal code */
134 if (buffer_uptodate(bh
))
138 if (!trylock_buffer(bh
)) {
142 } else /* mode == READ */
145 if (buffer_uptodate(bh
)) {
150 ret
= nilfs_bmap_lookup(NILFS_I(inode
)->i_bmap
, blkoff
, &blknum
);
155 map_bh(bh
, inode
->i_sb
, (sector_t
)blknum
);
157 bh
->b_end_io
= end_buffer_read_sync
;
166 unlock_page(bh
->b_page
);
167 page_cache_release(bh
->b_page
);
173 static int nilfs_mdt_read_block(struct inode
*inode
, unsigned long block
,
174 int readahead
, struct buffer_head
**out_bh
)
176 struct buffer_head
*first_bh
, *bh
;
177 unsigned long blkoff
;
178 int i
, nr_ra_blocks
= NILFS_MDT_MAX_RA_BLOCKS
;
181 err
= nilfs_mdt_submit_block(inode
, block
, READ
, &first_bh
);
182 if (err
== -EEXIST
) /* internal code */
190 for (i
= 0; i
< nr_ra_blocks
; i
++, blkoff
++) {
191 err
= nilfs_mdt_submit_block(inode
, blkoff
, READA
, &bh
);
192 if (likely(!err
|| err
== -EEXIST
))
194 else if (err
!= -EBUSY
)
196 /* abort readahead if bmap lookup failed */
197 if (!buffer_locked(first_bh
))
202 wait_on_buffer(first_bh
);
206 if (!buffer_uptodate(first_bh
))
219 * nilfs_mdt_get_block - read or create a buffer on meta data file.
220 * @inode: inode of the meta data file
221 * @blkoff: block offset
222 * @create: create flag
223 * @init_block: initializer used for newly allocated block
224 * @out_bh: output of a pointer to the buffer_head
226 * nilfs_mdt_get_block() looks up the specified buffer and tries to create
227 * a new buffer if @create is not zero. On success, the returned buffer is
228 * assured to be either existing or formatted using a buffer lock on success.
229 * @out_bh is substituted only when zero is returned.
231 * Return Value: On success, it returns 0. On error, the following negative
232 * error code is returned.
234 * %-ENOMEM - Insufficient memory available.
238 * %-ENOENT - the specified block does not exist (hole block)
240 * %-EROFS - Read only filesystem (for create mode)
242 int nilfs_mdt_get_block(struct inode
*inode
, unsigned long blkoff
, int create
,
243 void (*init_block
)(struct inode
*,
244 struct buffer_head
*, void *),
245 struct buffer_head
**out_bh
)
249 /* Should be rewritten with merging nilfs_mdt_read_block() */
251 ret
= nilfs_mdt_read_block(inode
, blkoff
, !create
, out_bh
);
252 if (!create
|| ret
!= -ENOENT
)
255 ret
= nilfs_mdt_create_block(inode
, blkoff
, out_bh
, init_block
);
256 if (unlikely(ret
== -EEXIST
)) {
257 /* create = 0; */ /* limit read-create loop retries */
264 * nilfs_mdt_delete_block - make a hole on the meta data file.
265 * @inode: inode of the meta data file
266 * @block: block offset
268 * Return Value: On success, zero is returned.
269 * On error, one of the following negative error code is returned.
271 * %-ENOMEM - Insufficient memory available.
275 int nilfs_mdt_delete_block(struct inode
*inode
, unsigned long block
)
277 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
280 err
= nilfs_bmap_delete(ii
->i_bmap
, block
);
281 if (!err
|| err
== -ENOENT
) {
282 nilfs_mdt_mark_dirty(inode
);
283 nilfs_mdt_forget_block(inode
, block
);
289 * nilfs_mdt_forget_block - discard dirty state and try to remove the page
290 * @inode: inode of the meta data file
291 * @block: block offset
293 * nilfs_mdt_forget_block() clears a dirty flag of the specified buffer, and
294 * tries to release the page including the buffer from a page cache.
296 * Return Value: On success, 0 is returned. On error, one of the following
297 * negative error code is returned.
299 * %-EBUSY - page has an active buffer.
301 * %-ENOENT - page cache has no page addressed by the offset.
303 int nilfs_mdt_forget_block(struct inode
*inode
, unsigned long block
)
305 pgoff_t index
= (pgoff_t
)block
>>
306 (PAGE_CACHE_SHIFT
- inode
->i_blkbits
);
308 unsigned long first_block
;
312 page
= find_lock_page(inode
->i_mapping
, index
);
316 wait_on_page_writeback(page
);
318 first_block
= (unsigned long)index
<<
319 (PAGE_CACHE_SHIFT
- inode
->i_blkbits
);
320 if (page_has_buffers(page
)) {
321 struct buffer_head
*bh
;
323 bh
= nilfs_page_get_nth_block(page
, block
- first_block
);
324 nilfs_forget_buffer(bh
);
326 still_dirty
= PageDirty(page
);
328 page_cache_release(page
);
331 invalidate_inode_pages2_range(inode
->i_mapping
, index
, index
) != 0)
337 * nilfs_mdt_mark_block_dirty - mark a block on the meta data file dirty.
338 * @inode: inode of the meta data file
339 * @block: block offset
341 * Return Value: On success, it returns 0. On error, the following negative
342 * error code is returned.
344 * %-ENOMEM - Insufficient memory available.
348 * %-ENOENT - the specified block does not exist (hole block)
350 int nilfs_mdt_mark_block_dirty(struct inode
*inode
, unsigned long block
)
352 struct buffer_head
*bh
;
355 err
= nilfs_mdt_read_block(inode
, block
, 0, &bh
);
358 mark_buffer_dirty(bh
);
359 nilfs_mdt_mark_dirty(inode
);
364 int nilfs_mdt_fetch_dirty(struct inode
*inode
)
366 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
368 if (nilfs_bmap_test_and_clear_dirty(ii
->i_bmap
)) {
369 set_bit(NILFS_I_DIRTY
, &ii
->i_state
);
372 return test_bit(NILFS_I_DIRTY
, &ii
->i_state
);
376 nilfs_mdt_write_page(struct page
*page
, struct writeback_control
*wbc
)
378 struct inode
*inode
= page
->mapping
->host
;
379 struct super_block
*sb
;
382 if (inode
&& (inode
->i_sb
->s_flags
& MS_RDONLY
)) {
384 * It means that filesystem was remounted in read-only
385 * mode because of error or metadata corruption. But we
386 * have dirty pages that try to be flushed in background.
387 * So, here we simply discard this dirty page.
389 nilfs_clear_dirty_page(page
, false);
394 redirty_page_for_writepage(wbc
, page
);
402 if (wbc
->sync_mode
== WB_SYNC_ALL
)
403 err
= nilfs_construct_segment(sb
);
404 else if (wbc
->for_reclaim
)
405 nilfs_flush_segment(sb
, inode
->i_ino
);
411 static const struct address_space_operations def_mdt_aops
= {
412 .writepage
= nilfs_mdt_write_page
,
415 static const struct inode_operations def_mdt_iops
;
416 static const struct file_operations def_mdt_fops
;
419 int nilfs_mdt_init(struct inode
*inode
, gfp_t gfp_mask
, size_t objsz
)
421 struct nilfs_mdt_info
*mi
;
423 mi
= kzalloc(max(sizeof(*mi
), objsz
), GFP_NOFS
);
427 init_rwsem(&mi
->mi_sem
);
428 inode
->i_private
= mi
;
430 inode
->i_mode
= S_IFREG
;
431 mapping_set_gfp_mask(inode
->i_mapping
, gfp_mask
);
432 inode
->i_mapping
->backing_dev_info
= inode
->i_sb
->s_bdi
;
434 inode
->i_op
= &def_mdt_iops
;
435 inode
->i_fop
= &def_mdt_fops
;
436 inode
->i_mapping
->a_ops
= &def_mdt_aops
;
441 void nilfs_mdt_set_entry_size(struct inode
*inode
, unsigned entry_size
,
442 unsigned header_size
)
444 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
446 mi
->mi_entry_size
= entry_size
;
447 mi
->mi_entries_per_block
= (1 << inode
->i_blkbits
) / entry_size
;
448 mi
->mi_first_entry_offset
= DIV_ROUND_UP(header_size
, entry_size
);
452 * nilfs_mdt_setup_shadow_map - setup shadow map and bind it to metadata file
453 * @inode: inode of the metadata file
454 * @shadow: shadow mapping
456 int nilfs_mdt_setup_shadow_map(struct inode
*inode
,
457 struct nilfs_shadow_map
*shadow
)
459 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
460 struct backing_dev_info
*bdi
= inode
->i_sb
->s_bdi
;
462 INIT_LIST_HEAD(&shadow
->frozen_buffers
);
463 address_space_init_once(&shadow
->frozen_data
);
464 nilfs_mapping_init(&shadow
->frozen_data
, inode
, bdi
);
465 address_space_init_once(&shadow
->frozen_btnodes
);
466 nilfs_mapping_init(&shadow
->frozen_btnodes
, inode
, bdi
);
467 mi
->mi_shadow
= shadow
;
472 * nilfs_mdt_save_to_shadow_map - copy bmap and dirty pages to shadow map
473 * @inode: inode of the metadata file
475 int nilfs_mdt_save_to_shadow_map(struct inode
*inode
)
477 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
478 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
479 struct nilfs_shadow_map
*shadow
= mi
->mi_shadow
;
482 ret
= nilfs_copy_dirty_pages(&shadow
->frozen_data
, inode
->i_mapping
);
486 ret
= nilfs_copy_dirty_pages(&shadow
->frozen_btnodes
,
487 &ii
->i_btnode_cache
);
491 nilfs_bmap_save(ii
->i_bmap
, &shadow
->bmap_store
);
496 int nilfs_mdt_freeze_buffer(struct inode
*inode
, struct buffer_head
*bh
)
498 struct nilfs_shadow_map
*shadow
= NILFS_MDT(inode
)->mi_shadow
;
499 struct buffer_head
*bh_frozen
;
501 int blkbits
= inode
->i_blkbits
;
503 page
= grab_cache_page(&shadow
->frozen_data
, bh
->b_page
->index
);
507 if (!page_has_buffers(page
))
508 create_empty_buffers(page
, 1 << blkbits
, 0);
510 bh_frozen
= nilfs_page_get_nth_block(page
, bh_offset(bh
) >> blkbits
);
512 if (!buffer_uptodate(bh_frozen
))
513 nilfs_copy_buffer(bh_frozen
, bh
);
514 if (list_empty(&bh_frozen
->b_assoc_buffers
)) {
515 list_add_tail(&bh_frozen
->b_assoc_buffers
,
516 &shadow
->frozen_buffers
);
517 set_buffer_nilfs_redirected(bh
);
519 brelse(bh_frozen
); /* already frozen */
523 page_cache_release(page
);
528 nilfs_mdt_get_frozen_buffer(struct inode
*inode
, struct buffer_head
*bh
)
530 struct nilfs_shadow_map
*shadow
= NILFS_MDT(inode
)->mi_shadow
;
531 struct buffer_head
*bh_frozen
= NULL
;
535 page
= find_lock_page(&shadow
->frozen_data
, bh
->b_page
->index
);
537 if (page_has_buffers(page
)) {
538 n
= bh_offset(bh
) >> inode
->i_blkbits
;
539 bh_frozen
= nilfs_page_get_nth_block(page
, n
);
542 page_cache_release(page
);
547 static void nilfs_release_frozen_buffers(struct nilfs_shadow_map
*shadow
)
549 struct list_head
*head
= &shadow
->frozen_buffers
;
550 struct buffer_head
*bh
;
552 while (!list_empty(head
)) {
553 bh
= list_first_entry(head
, struct buffer_head
,
555 list_del_init(&bh
->b_assoc_buffers
);
556 brelse(bh
); /* drop ref-count to make it releasable */
561 * nilfs_mdt_restore_from_shadow_map - restore dirty pages and bmap state
562 * @inode: inode of the metadata file
564 void nilfs_mdt_restore_from_shadow_map(struct inode
*inode
)
566 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
567 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
568 struct nilfs_shadow_map
*shadow
= mi
->mi_shadow
;
570 down_write(&mi
->mi_sem
);
572 if (mi
->mi_palloc_cache
)
573 nilfs_palloc_clear_cache(inode
);
575 nilfs_clear_dirty_pages(inode
->i_mapping
, true);
576 nilfs_copy_back_pages(inode
->i_mapping
, &shadow
->frozen_data
);
578 nilfs_clear_dirty_pages(&ii
->i_btnode_cache
, true);
579 nilfs_copy_back_pages(&ii
->i_btnode_cache
, &shadow
->frozen_btnodes
);
581 nilfs_bmap_restore(ii
->i_bmap
, &shadow
->bmap_store
);
583 up_write(&mi
->mi_sem
);
587 * nilfs_mdt_clear_shadow_map - truncate pages in shadow map caches
588 * @inode: inode of the metadata file
590 void nilfs_mdt_clear_shadow_map(struct inode
*inode
)
592 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
593 struct nilfs_shadow_map
*shadow
= mi
->mi_shadow
;
595 down_write(&mi
->mi_sem
);
596 nilfs_release_frozen_buffers(shadow
);
597 truncate_inode_pages(&shadow
->frozen_data
, 0);
598 truncate_inode_pages(&shadow
->frozen_btnodes
, 0);
599 up_write(&mi
->mi_sem
);