Commit c07dfcb4 authored by Tahsin Erdogan's avatar Tahsin Erdogan Committed by Theodore Ts'o

mbcache: make mbcache naming more generic

Make names more generic so that mbcache usage is not limited to
block sharing. In a subsequent patch in the series
("ext4: xattr inode deduplication"), we start using the mbcache code
for sharing xattr inodes. With that patch, old mb_cache_entry.e_block
field could be holding either a block number or an inode number.
Signed-off-by: default avatarTahsin Erdogan <tahsin@google.com>
Signed-off-by: default avatarTheodore Ts'o <tytso@mit.edu>
parent b6d9029d
...@@ -493,8 +493,8 @@ bad_block: ext2_error(sb, "ext2_xattr_set", ...@@ -493,8 +493,8 @@ bad_block: ext2_error(sb, "ext2_xattr_set",
* This must happen under buffer lock for * This must happen under buffer lock for
* ext2_xattr_set2() to reliably detect modified block * ext2_xattr_set2() to reliably detect modified block
*/ */
mb_cache_entry_delete_block(EXT2_SB(sb)->s_mb_cache, mb_cache_entry_delete(EXT2_SB(sb)->s_mb_cache, hash,
hash, bh->b_blocknr); bh->b_blocknr);
/* keep the buffer locked while modifying it. */ /* keep the buffer locked while modifying it. */
} else { } else {
...@@ -721,8 +721,8 @@ ext2_xattr_set2(struct inode *inode, struct buffer_head *old_bh, ...@@ -721,8 +721,8 @@ ext2_xattr_set2(struct inode *inode, struct buffer_head *old_bh,
* This must happen under buffer lock for * This must happen under buffer lock for
* ext2_xattr_set2() to reliably detect freed block * ext2_xattr_set2() to reliably detect freed block
*/ */
mb_cache_entry_delete_block(ext2_mb_cache, mb_cache_entry_delete(ext2_mb_cache, hash,
hash, old_bh->b_blocknr); old_bh->b_blocknr);
/* Free the old block. */ /* Free the old block. */
ea_bdebug(old_bh, "freeing"); ea_bdebug(old_bh, "freeing");
ext2_free_blocks(inode, old_bh->b_blocknr, 1); ext2_free_blocks(inode, old_bh->b_blocknr, 1);
...@@ -795,8 +795,8 @@ ext2_xattr_delete_inode(struct inode *inode) ...@@ -795,8 +795,8 @@ ext2_xattr_delete_inode(struct inode *inode)
* This must happen under buffer lock for ext2_xattr_set2() to * This must happen under buffer lock for ext2_xattr_set2() to
* reliably detect freed block * reliably detect freed block
*/ */
mb_cache_entry_delete_block(EXT2_SB(inode->i_sb)->s_mb_cache, mb_cache_entry_delete(EXT2_SB(inode->i_sb)->s_mb_cache, hash,
hash, bh->b_blocknr); bh->b_blocknr);
ext2_free_blocks(inode, EXT2_I(inode)->i_file_acl, 1); ext2_free_blocks(inode, EXT2_I(inode)->i_file_acl, 1);
get_bh(bh); get_bh(bh);
bforget(bh); bforget(bh);
...@@ -907,11 +907,11 @@ ext2_xattr_cache_find(struct inode *inode, struct ext2_xattr_header *header) ...@@ -907,11 +907,11 @@ ext2_xattr_cache_find(struct inode *inode, struct ext2_xattr_header *header)
while (ce) { while (ce) {
struct buffer_head *bh; struct buffer_head *bh;
bh = sb_bread(inode->i_sb, ce->e_block); bh = sb_bread(inode->i_sb, ce->e_value);
if (!bh) { if (!bh) {
ext2_error(inode->i_sb, "ext2_xattr_cache_find", ext2_error(inode->i_sb, "ext2_xattr_cache_find",
"inode %ld: block %ld read error", "inode %ld: block %ld read error",
inode->i_ino, (unsigned long) ce->e_block); inode->i_ino, (unsigned long) ce->e_value);
} else { } else {
lock_buffer(bh); lock_buffer(bh);
/* /*
...@@ -931,7 +931,7 @@ ext2_xattr_cache_find(struct inode *inode, struct ext2_xattr_header *header) ...@@ -931,7 +931,7 @@ ext2_xattr_cache_find(struct inode *inode, struct ext2_xattr_header *header)
} else if (le32_to_cpu(HDR(bh)->h_refcount) > } else if (le32_to_cpu(HDR(bh)->h_refcount) >
EXT2_XATTR_REFCOUNT_MAX) { EXT2_XATTR_REFCOUNT_MAX) {
ea_idebug(inode, "block %ld refcount %d>%d", ea_idebug(inode, "block %ld refcount %d>%d",
(unsigned long) ce->e_block, (unsigned long) ce->e_value,
le32_to_cpu(HDR(bh)->h_refcount), le32_to_cpu(HDR(bh)->h_refcount),
EXT2_XATTR_REFCOUNT_MAX); EXT2_XATTR_REFCOUNT_MAX);
} else if (!ext2_xattr_cmp(header, HDR(bh))) { } else if (!ext2_xattr_cmp(header, HDR(bh))) {
......
...@@ -678,7 +678,7 @@ ext4_xattr_release_block(handle_t *handle, struct inode *inode, ...@@ -678,7 +678,7 @@ ext4_xattr_release_block(handle_t *handle, struct inode *inode,
* This must happen under buffer lock for * This must happen under buffer lock for
* ext4_xattr_block_set() to reliably detect freed block * ext4_xattr_block_set() to reliably detect freed block
*/ */
mb_cache_entry_delete_block(ext4_mb_cache, hash, bh->b_blocknr); mb_cache_entry_delete(ext4_mb_cache, hash, bh->b_blocknr);
get_bh(bh); get_bh(bh);
unlock_buffer(bh); unlock_buffer(bh);
ext4_free_blocks(handle, inode, bh, 0, 1, ext4_free_blocks(handle, inode, bh, 0, 1,
...@@ -1113,8 +1113,8 @@ ext4_xattr_block_set(handle_t *handle, struct inode *inode, ...@@ -1113,8 +1113,8 @@ ext4_xattr_block_set(handle_t *handle, struct inode *inode,
* ext4_xattr_block_set() to reliably detect modified * ext4_xattr_block_set() to reliably detect modified
* block * block
*/ */
mb_cache_entry_delete_block(ext4_mb_cache, hash, mb_cache_entry_delete(ext4_mb_cache, hash,
bs->bh->b_blocknr); bs->bh->b_blocknr);
ea_bdebug(bs->bh, "modifying in-place"); ea_bdebug(bs->bh, "modifying in-place");
error = ext4_xattr_set_entry(i, s, handle, inode); error = ext4_xattr_set_entry(i, s, handle, inode);
if (!error) { if (!error) {
...@@ -2236,10 +2236,10 @@ ext4_xattr_cache_find(struct inode *inode, struct ext4_xattr_header *header, ...@@ -2236,10 +2236,10 @@ ext4_xattr_cache_find(struct inode *inode, struct ext4_xattr_header *header,
while (ce) { while (ce) {
struct buffer_head *bh; struct buffer_head *bh;
bh = sb_bread(inode->i_sb, ce->e_block); bh = sb_bread(inode->i_sb, ce->e_value);
if (!bh) { if (!bh) {
EXT4_ERROR_INODE(inode, "block %lu read error", EXT4_ERROR_INODE(inode, "block %lu read error",
(unsigned long) ce->e_block); (unsigned long)ce->e_value);
} else if (ext4_xattr_cmp(header, BHDR(bh)) == 0) { } else if (ext4_xattr_cmp(header, BHDR(bh)) == 0) {
*pce = ce; *pce = ce;
return bh; return bh;
......
...@@ -10,7 +10,7 @@ ...@@ -10,7 +10,7 @@
/* /*
* Mbcache is a simple key-value store. Keys need not be unique, however * Mbcache is a simple key-value store. Keys need not be unique, however
* key-value pairs are expected to be unique (we use this fact in * key-value pairs are expected to be unique (we use this fact in
* mb_cache_entry_delete_block()). * mb_cache_entry_delete()).
* *
* Ext2 and ext4 use this cache for deduplication of extended attribute blocks. * Ext2 and ext4 use this cache for deduplication of extended attribute blocks.
* They use hash of a block contents as a key and block number as a value. * They use hash of a block contents as a key and block number as a value.
...@@ -62,15 +62,15 @@ static inline struct hlist_bl_head *mb_cache_entry_head(struct mb_cache *cache, ...@@ -62,15 +62,15 @@ static inline struct hlist_bl_head *mb_cache_entry_head(struct mb_cache *cache,
* @cache - cache where the entry should be created * @cache - cache where the entry should be created
* @mask - gfp mask with which the entry should be allocated * @mask - gfp mask with which the entry should be allocated
* @key - key of the entry * @key - key of the entry
* @block - block that contains data * @value - value of the entry
* @reusable - is the block reusable by other inodes? * @reusable - is the entry reusable by others?
* *
* Creates entry in @cache with key @key and records that data is stored in * Creates entry in @cache with key @key and value @value. The function returns
* block @block. The function returns -EBUSY if entry with the same key * -EBUSY if entry with the same key and value already exists in cache.
* and for the same block already exists in cache. Otherwise 0 is returned. * Otherwise 0 is returned.
*/ */
int mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, int mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key,
sector_t block, bool reusable) u64 value, bool reusable)
{ {
struct mb_cache_entry *entry, *dup; struct mb_cache_entry *entry, *dup;
struct hlist_bl_node *dup_node; struct hlist_bl_node *dup_node;
...@@ -91,12 +91,12 @@ int mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, ...@@ -91,12 +91,12 @@ int mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key,
/* One ref for hash, one ref returned */ /* One ref for hash, one ref returned */
atomic_set(&entry->e_refcnt, 1); atomic_set(&entry->e_refcnt, 1);
entry->e_key = key; entry->e_key = key;
entry->e_block = block; entry->e_value = value;
entry->e_reusable = reusable; entry->e_reusable = reusable;
head = mb_cache_entry_head(cache, key); head = mb_cache_entry_head(cache, key);
hlist_bl_lock(head); hlist_bl_lock(head);
hlist_bl_for_each_entry(dup, dup_node, head, e_hash_list) { hlist_bl_for_each_entry(dup, dup_node, head, e_hash_list) {
if (dup->e_key == key && dup->e_block == block) { if (dup->e_key == key && dup->e_value == value) {
hlist_bl_unlock(head); hlist_bl_unlock(head);
kmem_cache_free(mb_entry_cache, entry); kmem_cache_free(mb_entry_cache, entry);
return -EBUSY; return -EBUSY;
...@@ -187,13 +187,13 @@ struct mb_cache_entry *mb_cache_entry_find_next(struct mb_cache *cache, ...@@ -187,13 +187,13 @@ struct mb_cache_entry *mb_cache_entry_find_next(struct mb_cache *cache,
EXPORT_SYMBOL(mb_cache_entry_find_next); EXPORT_SYMBOL(mb_cache_entry_find_next);
/* /*
* mb_cache_entry_get - get a cache entry by block number (and key) * mb_cache_entry_get - get a cache entry by value (and key)
* @cache - cache we work with * @cache - cache we work with
* @key - key of block number @block * @key - key
* @block - block number * @value - value
*/ */
struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key, struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key,
sector_t block) u64 value)
{ {
struct hlist_bl_node *node; struct hlist_bl_node *node;
struct hlist_bl_head *head; struct hlist_bl_head *head;
...@@ -202,7 +202,7 @@ struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key, ...@@ -202,7 +202,7 @@ struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key,
head = mb_cache_entry_head(cache, key); head = mb_cache_entry_head(cache, key);
hlist_bl_lock(head); hlist_bl_lock(head);
hlist_bl_for_each_entry(entry, node, head, e_hash_list) { hlist_bl_for_each_entry(entry, node, head, e_hash_list) {
if (entry->e_key == key && entry->e_block == block) { if (entry->e_key == key && entry->e_value == value) {
atomic_inc(&entry->e_refcnt); atomic_inc(&entry->e_refcnt);
goto out; goto out;
} }
...@@ -214,15 +214,14 @@ struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key, ...@@ -214,15 +214,14 @@ struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key,
} }
EXPORT_SYMBOL(mb_cache_entry_get); EXPORT_SYMBOL(mb_cache_entry_get);
/* mb_cache_entry_delete_block - remove information about block from cache /* mb_cache_entry_delete - remove a cache entry
* @cache - cache we work with * @cache - cache we work with
* @key - key of block @block * @key - key
* @block - block number * @value - value
* *
* Remove entry from cache @cache with key @key with data stored in @block. * Remove entry from cache @cache with key @key and value @value.
*/ */
void mb_cache_entry_delete_block(struct mb_cache *cache, u32 key, void mb_cache_entry_delete(struct mb_cache *cache, u32 key, u64 value)
sector_t block)
{ {
struct hlist_bl_node *node; struct hlist_bl_node *node;
struct hlist_bl_head *head; struct hlist_bl_head *head;
...@@ -231,7 +230,7 @@ void mb_cache_entry_delete_block(struct mb_cache *cache, u32 key, ...@@ -231,7 +230,7 @@ void mb_cache_entry_delete_block(struct mb_cache *cache, u32 key,
head = mb_cache_entry_head(cache, key); head = mb_cache_entry_head(cache, key);
hlist_bl_lock(head); hlist_bl_lock(head);
hlist_bl_for_each_entry(entry, node, head, e_hash_list) { hlist_bl_for_each_entry(entry, node, head, e_hash_list) {
if (entry->e_key == key && entry->e_block == block) { if (entry->e_key == key && entry->e_value == value) {
/* We keep hash list reference to keep entry alive */ /* We keep hash list reference to keep entry alive */
hlist_bl_del_init(&entry->e_hash_list); hlist_bl_del_init(&entry->e_hash_list);
hlist_bl_unlock(head); hlist_bl_unlock(head);
...@@ -248,7 +247,7 @@ void mb_cache_entry_delete_block(struct mb_cache *cache, u32 key, ...@@ -248,7 +247,7 @@ void mb_cache_entry_delete_block(struct mb_cache *cache, u32 key,
} }
hlist_bl_unlock(head); hlist_bl_unlock(head);
} }
EXPORT_SYMBOL(mb_cache_entry_delete_block); EXPORT_SYMBOL(mb_cache_entry_delete);
/* mb_cache_entry_touch - cache entry got used /* mb_cache_entry_touch - cache entry got used
* @cache - cache the entry belongs to * @cache - cache the entry belongs to
......
...@@ -19,15 +19,15 @@ struct mb_cache_entry { ...@@ -19,15 +19,15 @@ struct mb_cache_entry {
u32 e_key; u32 e_key;
u32 e_referenced:1; u32 e_referenced:1;
u32 e_reusable:1; u32 e_reusable:1;
/* Block number of hashed block - stable during lifetime of the entry */ /* User provided value - stable during lifetime of the entry */
sector_t e_block; u64 e_value;
}; };
struct mb_cache *mb_cache_create(int bucket_bits); struct mb_cache *mb_cache_create(int bucket_bits);
void mb_cache_destroy(struct mb_cache *cache); void mb_cache_destroy(struct mb_cache *cache);
int mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, int mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key,
sector_t block, bool reusable); u64 value, bool reusable);
void __mb_cache_entry_free(struct mb_cache_entry *entry); void __mb_cache_entry_free(struct mb_cache_entry *entry);
static inline int mb_cache_entry_put(struct mb_cache *cache, static inline int mb_cache_entry_put(struct mb_cache *cache,
struct mb_cache_entry *entry) struct mb_cache_entry *entry)
...@@ -38,10 +38,9 @@ static inline int mb_cache_entry_put(struct mb_cache *cache, ...@@ -38,10 +38,9 @@ static inline int mb_cache_entry_put(struct mb_cache *cache,
return 1; return 1;
} }
void mb_cache_entry_delete_block(struct mb_cache *cache, u32 key, void mb_cache_entry_delete(struct mb_cache *cache, u32 key, u64 value);
sector_t block);
struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key, struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key,
sector_t block); u64 value);
struct mb_cache_entry *mb_cache_entry_find_first(struct mb_cache *cache, struct mb_cache_entry *mb_cache_entry_find_first(struct mb_cache *cache,
u32 key); u32 key);
struct mb_cache_entry *mb_cache_entry_find_next(struct mb_cache *cache, struct mb_cache_entry *mb_cache_entry_find_next(struct mb_cache *cache,
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment