提交 6451e041 编写于 作者: J Jaegeuk Kim

f2fs: add infra for ino management

This patch changes the naming of orphan-related data structures to use as
inode numbers managed globally.
Later, we can use this facility for managing any inode number lists.
Reviewed-by: NChao Yu <chao2.yu@samsung.com>
Signed-off-by: NJaegeuk Kim <jaegeuk@kernel.org>
上级 953e6cc6
...@@ -22,7 +22,7 @@ ...@@ -22,7 +22,7 @@
#include "segment.h" #include "segment.h"
#include <trace/events/f2fs.h> #include <trace/events/f2fs.h>
static struct kmem_cache *orphan_entry_slab; static struct kmem_cache *ino_entry_slab;
static struct kmem_cache *inode_entry_slab; static struct kmem_cache *inode_entry_slab;
/* /*
...@@ -282,19 +282,18 @@ const struct address_space_operations f2fs_meta_aops = { ...@@ -282,19 +282,18 @@ const struct address_space_operations f2fs_meta_aops = {
.set_page_dirty = f2fs_set_meta_page_dirty, .set_page_dirty = f2fs_set_meta_page_dirty,
}; };
static void __add_ino_entry(struct f2fs_sb_info *sbi, nid_t ino) static void __add_ino_entry(struct f2fs_sb_info *sbi, nid_t ino, int type)
{ {
struct list_head *head; struct ino_entry *new, *e;
struct orphan_inode_entry *new, *e;
new = f2fs_kmem_cache_alloc(orphan_entry_slab, GFP_ATOMIC); new = f2fs_kmem_cache_alloc(ino_entry_slab, GFP_ATOMIC);
new->ino = ino; new->ino = ino;
spin_lock(&sbi->orphan_inode_lock); spin_lock(&sbi->ino_lock[type]);
list_for_each_entry(e, &sbi->orphan_inode_list, list) { list_for_each_entry(e, &sbi->ino_list[type], list) {
if (e->ino == ino) { if (e->ino == ino) {
spin_unlock(&sbi->orphan_inode_lock); spin_unlock(&sbi->ino_lock[type]);
kmem_cache_free(orphan_entry_slab, new); kmem_cache_free(ino_entry_slab, new);
return; return;
} }
if (e->ino > ino) if (e->ino > ino)
...@@ -303,58 +302,58 @@ static void __add_ino_entry(struct f2fs_sb_info *sbi, nid_t ino) ...@@ -303,58 +302,58 @@ static void __add_ino_entry(struct f2fs_sb_info *sbi, nid_t ino)
/* add new entry into list which is sorted by inode number */ /* add new entry into list which is sorted by inode number */
list_add_tail(&new->list, &e->list); list_add_tail(&new->list, &e->list);
spin_unlock(&sbi->orphan_inode_lock); spin_unlock(&sbi->ino_lock[type]);
} }
static void __remove_ino_entry(struct f2fs_sb_info *sbi, nid_t ino) static void __remove_ino_entry(struct f2fs_sb_info *sbi, nid_t ino, int type)
{ {
struct orphan_inode_entry *e; struct ino_entry *e;
spin_lock(&sbi->orphan_inode_lock); spin_lock(&sbi->ino_lock[type]);
list_for_each_entry(e, &sbi->orphan_inode_list, list) { list_for_each_entry(e, &sbi->ino_list[type], list) {
if (e->ino == ino) { if (e->ino == ino) {
list_del(&e->list); list_del(&e->list);
sbi->n_orphans--; sbi->n_orphans--;
spin_unlock(&sbi->orphan_inode_lock); spin_unlock(&sbi->ino_lock[type]);
kmem_cache_free(orphan_entry_slab, e); kmem_cache_free(ino_entry_slab, e);
return; return;
} }
} }
spin_unlock(&sbi->orphan_inode_lock); spin_unlock(&sbi->ino_lock[type]);
} }
int acquire_orphan_inode(struct f2fs_sb_info *sbi) int acquire_orphan_inode(struct f2fs_sb_info *sbi)
{ {
int err = 0; int err = 0;
spin_lock(&sbi->orphan_inode_lock); spin_lock(&sbi->ino_lock[ORPHAN_INO]);
if (unlikely(sbi->n_orphans >= sbi->max_orphans)) if (unlikely(sbi->n_orphans >= sbi->max_orphans))
err = -ENOSPC; err = -ENOSPC;
else else
sbi->n_orphans++; sbi->n_orphans++;
spin_unlock(&sbi->orphan_inode_lock); spin_unlock(&sbi->ino_lock[ORPHAN_INO]);
return err; return err;
} }
void release_orphan_inode(struct f2fs_sb_info *sbi) void release_orphan_inode(struct f2fs_sb_info *sbi)
{ {
spin_lock(&sbi->orphan_inode_lock); spin_lock(&sbi->ino_lock[ORPHAN_INO]);
f2fs_bug_on(sbi->n_orphans == 0); f2fs_bug_on(sbi->n_orphans == 0);
sbi->n_orphans--; sbi->n_orphans--;
spin_unlock(&sbi->orphan_inode_lock); spin_unlock(&sbi->ino_lock[ORPHAN_INO]);
} }
void add_orphan_inode(struct f2fs_sb_info *sbi, nid_t ino) void add_orphan_inode(struct f2fs_sb_info *sbi, nid_t ino)
{ {
/* add new orphan entry into list which is sorted by inode number */ /* add new orphan entry into list which is sorted by inode number */
__add_ino_entry(sbi, ino); __add_ino_entry(sbi, ino, ORPHAN_INO);
} }
void remove_orphan_inode(struct f2fs_sb_info *sbi, nid_t ino) void remove_orphan_inode(struct f2fs_sb_info *sbi, nid_t ino)
{ {
/* remove orphan entry from orphan list */ /* remove orphan entry from orphan list */
__remove_ino_entry(sbi, ino); __remove_ino_entry(sbi, ino, ORPHAN_INO);
} }
static void recover_orphan_inode(struct f2fs_sb_info *sbi, nid_t ino) static void recover_orphan_inode(struct f2fs_sb_info *sbi, nid_t ino)
...@@ -408,14 +407,14 @@ static void write_orphan_inodes(struct f2fs_sb_info *sbi, block_t start_blk) ...@@ -408,14 +407,14 @@ static void write_orphan_inodes(struct f2fs_sb_info *sbi, block_t start_blk)
unsigned short orphan_blocks = (unsigned short)((sbi->n_orphans + unsigned short orphan_blocks = (unsigned short)((sbi->n_orphans +
(F2FS_ORPHANS_PER_BLOCK - 1)) / F2FS_ORPHANS_PER_BLOCK); (F2FS_ORPHANS_PER_BLOCK - 1)) / F2FS_ORPHANS_PER_BLOCK);
struct page *page = NULL; struct page *page = NULL;
struct orphan_inode_entry *orphan = NULL; struct ino_entry *orphan = NULL;
for (index = 0; index < orphan_blocks; index++) for (index = 0; index < orphan_blocks; index++)
grab_meta_page(sbi, start_blk + index); grab_meta_page(sbi, start_blk + index);
index = 1; index = 1;
spin_lock(&sbi->orphan_inode_lock); spin_lock(&sbi->ino_lock[ORPHAN_INO]);
head = &sbi->orphan_inode_list; head = &sbi->ino_list[ORPHAN_INO];
/* loop for each orphan inode entry and write them in Jornal block */ /* loop for each orphan inode entry and write them in Jornal block */
list_for_each_entry(orphan, head, list) { list_for_each_entry(orphan, head, list) {
...@@ -455,7 +454,7 @@ static void write_orphan_inodes(struct f2fs_sb_info *sbi, block_t start_blk) ...@@ -455,7 +454,7 @@ static void write_orphan_inodes(struct f2fs_sb_info *sbi, block_t start_blk)
f2fs_put_page(page, 1); f2fs_put_page(page, 1);
} }
spin_unlock(&sbi->orphan_inode_lock); spin_unlock(&sbi->ino_lock[ORPHAN_INO]);
} }
static struct page *validate_checkpoint(struct f2fs_sb_info *sbi, static struct page *validate_checkpoint(struct f2fs_sb_info *sbi,
...@@ -939,31 +938,36 @@ void write_checkpoint(struct f2fs_sb_info *sbi, bool is_umount) ...@@ -939,31 +938,36 @@ void write_checkpoint(struct f2fs_sb_info *sbi, bool is_umount)
trace_f2fs_write_checkpoint(sbi->sb, is_umount, "finish checkpoint"); trace_f2fs_write_checkpoint(sbi->sb, is_umount, "finish checkpoint");
} }
void init_orphan_info(struct f2fs_sb_info *sbi) void init_ino_entry_info(struct f2fs_sb_info *sbi)
{ {
spin_lock_init(&sbi->orphan_inode_lock); int i;
INIT_LIST_HEAD(&sbi->orphan_inode_list);
sbi->n_orphans = 0; for (i = 0; i < MAX_INO_ENTRY; i++) {
spin_lock_init(&sbi->ino_lock[i]);
INIT_LIST_HEAD(&sbi->ino_list[i]);
}
/* /*
* considering 512 blocks in a segment 8 blocks are needed for cp * considering 512 blocks in a segment 8 blocks are needed for cp
* and log segment summaries. Remaining blocks are used to keep * and log segment summaries. Remaining blocks are used to keep
* orphan entries with the limitation one reserved segment * orphan entries with the limitation one reserved segment
* for cp pack we can have max 1020*504 orphan entries * for cp pack we can have max 1020*504 orphan entries
*/ */
sbi->n_orphans = 0;
sbi->max_orphans = (sbi->blocks_per_seg - 2 - NR_CURSEG_TYPE) sbi->max_orphans = (sbi->blocks_per_seg - 2 - NR_CURSEG_TYPE)
* F2FS_ORPHANS_PER_BLOCK; * F2FS_ORPHANS_PER_BLOCK;
} }
int __init create_checkpoint_caches(void) int __init create_checkpoint_caches(void)
{ {
orphan_entry_slab = f2fs_kmem_cache_create("f2fs_orphan_entry", ino_entry_slab = f2fs_kmem_cache_create("f2fs_ino_entry",
sizeof(struct orphan_inode_entry)); sizeof(struct ino_entry));
if (!orphan_entry_slab) if (!ino_entry_slab)
return -ENOMEM; return -ENOMEM;
inode_entry_slab = f2fs_kmem_cache_create("f2fs_dirty_dir_entry", inode_entry_slab = f2fs_kmem_cache_create("f2fs_dirty_dir_entry",
sizeof(struct dir_inode_entry)); sizeof(struct dir_inode_entry));
if (!inode_entry_slab) { if (!inode_entry_slab) {
kmem_cache_destroy(orphan_entry_slab); kmem_cache_destroy(ino_entry_slab);
return -ENOMEM; return -ENOMEM;
} }
return 0; return 0;
...@@ -971,6 +975,6 @@ int __init create_checkpoint_caches(void) ...@@ -971,6 +975,6 @@ int __init create_checkpoint_caches(void)
void destroy_checkpoint_caches(void) void destroy_checkpoint_caches(void)
{ {
kmem_cache_destroy(orphan_entry_slab); kmem_cache_destroy(ino_entry_slab);
kmem_cache_destroy(inode_entry_slab); kmem_cache_destroy(inode_entry_slab);
} }
...@@ -167,7 +167,7 @@ static void update_mem_info(struct f2fs_sb_info *sbi) ...@@ -167,7 +167,7 @@ static void update_mem_info(struct f2fs_sb_info *sbi)
si->cache_mem += npages << PAGE_CACHE_SHIFT; si->cache_mem += npages << PAGE_CACHE_SHIFT;
npages = META_MAPPING(sbi)->nrpages; npages = META_MAPPING(sbi)->nrpages;
si->cache_mem += npages << PAGE_CACHE_SHIFT; si->cache_mem += npages << PAGE_CACHE_SHIFT;
si->cache_mem += sbi->n_orphans * sizeof(struct orphan_inode_entry); si->cache_mem += sbi->n_orphans * sizeof(struct ino_entry);
si->cache_mem += sbi->n_dirty_dirs * sizeof(struct dir_inode_entry); si->cache_mem += sbi->n_dirty_dirs * sizeof(struct dir_inode_entry);
} }
......
...@@ -100,8 +100,13 @@ enum { ...@@ -100,8 +100,13 @@ enum {
META_SSA META_SSA
}; };
/* for the list of orphan inodes */ /* for the list of ino */
struct orphan_inode_entry { enum {
ORPHAN_INO, /* for orphan ino list */
MAX_INO_ENTRY, /* max. list */
};
struct ino_entry {
struct list_head list; /* list head */ struct list_head list; /* list head */
nid_t ino; /* inode number */ nid_t ino; /* inode number */
}; };
...@@ -450,9 +455,11 @@ struct f2fs_sb_info { ...@@ -450,9 +455,11 @@ struct f2fs_sb_info {
bool por_doing; /* recovery is doing or not */ bool por_doing; /* recovery is doing or not */
wait_queue_head_t cp_wait; wait_queue_head_t cp_wait;
/* for orphan inode management */ /* for inode management */
struct list_head orphan_inode_list; /* orphan inode list */ spinlock_t ino_lock[MAX_INO_ENTRY]; /* for ino entry lock */
spinlock_t orphan_inode_lock; /* for orphan inode list */ struct list_head ino_list[MAX_INO_ENTRY]; /* inode list head */
/* for orphan inode, use 0'th array */
unsigned int n_orphans; /* # of orphan inodes */ unsigned int n_orphans; /* # of orphan inodes */
unsigned int max_orphans; /* max orphan inodes */ unsigned int max_orphans; /* max orphan inodes */
...@@ -1255,7 +1262,7 @@ void add_dirty_dir_inode(struct inode *); ...@@ -1255,7 +1262,7 @@ void add_dirty_dir_inode(struct inode *);
void remove_dirty_dir_inode(struct inode *); void remove_dirty_dir_inode(struct inode *);
void sync_dirty_dir_inodes(struct f2fs_sb_info *); void sync_dirty_dir_inodes(struct f2fs_sb_info *);
void write_checkpoint(struct f2fs_sb_info *, bool); void write_checkpoint(struct f2fs_sb_info *, bool);
void init_orphan_info(struct f2fs_sb_info *); void init_ino_entry_info(struct f2fs_sb_info *);
int __init create_checkpoint_caches(void); int __init create_checkpoint_caches(void);
void destroy_checkpoint_caches(void); void destroy_checkpoint_caches(void);
......
...@@ -1003,7 +1003,7 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent) ...@@ -1003,7 +1003,7 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent)
INIT_LIST_HEAD(&sbi->dir_inode_list); INIT_LIST_HEAD(&sbi->dir_inode_list);
spin_lock_init(&sbi->dir_inode_lock); spin_lock_init(&sbi->dir_inode_lock);
init_orphan_info(sbi); init_ino_entry_info(sbi);
/* setup f2fs internal modules */ /* setup f2fs internal modules */
err = build_segment_manager(sbi); err = build_segment_manager(sbi);
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册