Commit be62c519 authored by Jingbo Xu's avatar Jingbo Xu Committed by Gao Xiang

erofs: support large folios for fscache mode

When large folios supported, one folio can be split into several slices,
each of which may be mapped to META/UNMAPPED/MAPPED, and the folio can
be unlocked as a whole only when all slices have completed.

Thus always allocate erofs_fscache_request for each .read_folio() or
.readahead(), in which case the allocated request is responsible for
unlocking folios when all slices have completed.

As described above, each folio or folio range can be mapped into several
slices, while these slices may be mapped to different cookies, and thus
each slice needs its own netfs_cache_resources.  Here we introduce
chained requests to support this, where each .read_folio() or
.readahead() calling can correspond to multiple requests.  Each request
has its own netfs_cache_resources and thus is used to access one cookie.
Among these requests, there's a primary request, with the others
pointing to the primary request.
Signed-off-by: default avatarJingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: default avatarJia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20221201074256.16639-2-jefflexu@linux.alibaba.comSigned-off-by: default avatarGao Xiang <hsiangkao@linux.alibaba.com>
parent 709fe09e
...@@ -12,6 +12,7 @@ static LIST_HEAD(erofs_domain_list); ...@@ -12,6 +12,7 @@ static LIST_HEAD(erofs_domain_list);
static struct vfsmount *erofs_pseudo_mnt; static struct vfsmount *erofs_pseudo_mnt;
struct erofs_fscache_request { struct erofs_fscache_request {
struct erofs_fscache_request *primary;
struct netfs_cache_resources cache_resources; struct netfs_cache_resources cache_resources;
struct address_space *mapping; /* The mapping being accessed */ struct address_space *mapping; /* The mapping being accessed */
loff_t start; /* Start position */ loff_t start; /* Start position */
...@@ -38,6 +39,26 @@ static struct erofs_fscache_request *erofs_fscache_req_alloc(struct address_spac ...@@ -38,6 +39,26 @@ static struct erofs_fscache_request *erofs_fscache_req_alloc(struct address_spac
return req; return req;
} }
static struct erofs_fscache_request *erofs_fscache_req_chain(struct erofs_fscache_request *primary,
size_t len)
{
struct erofs_fscache_request *req;
/* use primary request for the first submission */
if (!primary->submitted) {
refcount_inc(&primary->ref);
return primary;
}
req = erofs_fscache_req_alloc(primary->mapping,
primary->start + primary->submitted, len);
if (!IS_ERR(req)) {
req->primary = primary;
refcount_inc(&primary->ref);
}
return req;
}
static void erofs_fscache_req_complete(struct erofs_fscache_request *req) static void erofs_fscache_req_complete(struct erofs_fscache_request *req)
{ {
struct folio *folio; struct folio *folio;
...@@ -56,17 +77,19 @@ static void erofs_fscache_req_complete(struct erofs_fscache_request *req) ...@@ -56,17 +77,19 @@ static void erofs_fscache_req_complete(struct erofs_fscache_request *req)
folio_unlock(folio); folio_unlock(folio);
} }
rcu_read_unlock(); rcu_read_unlock();
if (req->cache_resources.ops)
req->cache_resources.ops->end_operation(&req->cache_resources);
kfree(req);
} }
static void erofs_fscache_req_put(struct erofs_fscache_request *req) static void erofs_fscache_req_put(struct erofs_fscache_request *req)
{ {
if (refcount_dec_and_test(&req->ref)) if (refcount_dec_and_test(&req->ref)) {
erofs_fscache_req_complete(req); if (req->cache_resources.ops)
req->cache_resources.ops->end_operation(&req->cache_resources);
if (!req->primary)
erofs_fscache_req_complete(req);
else
erofs_fscache_req_put(req->primary);
kfree(req);
}
} }
static void erofs_fscache_subreq_complete(void *priv, static void erofs_fscache_subreq_complete(void *priv,
...@@ -74,8 +97,12 @@ static void erofs_fscache_subreq_complete(void *priv, ...@@ -74,8 +97,12 @@ static void erofs_fscache_subreq_complete(void *priv,
{ {
struct erofs_fscache_request *req = priv; struct erofs_fscache_request *req = priv;
if (IS_ERR_VALUE(transferred_or_error)) if (IS_ERR_VALUE(transferred_or_error)) {
req->error = transferred_or_error; if (req->primary)
req->primary->error = transferred_or_error;
else
req->error = transferred_or_error;
}
erofs_fscache_req_put(req); erofs_fscache_req_put(req);
} }
...@@ -131,7 +158,6 @@ static int erofs_fscache_read_folios_async(struct fscache_cookie *cookie, ...@@ -131,7 +158,6 @@ static int erofs_fscache_read_folios_async(struct fscache_cookie *cookie,
done += slen; done += slen;
} }
DBG_BUGON(done != len); DBG_BUGON(done != len);
req->submitted += len;
return 0; return 0;
} }
...@@ -167,32 +193,19 @@ static int erofs_fscache_meta_read_folio(struct file *data, struct folio *folio) ...@@ -167,32 +193,19 @@ static int erofs_fscache_meta_read_folio(struct file *data, struct folio *folio)
return ret; return ret;
} }
/* static int erofs_fscache_data_read_slice(struct erofs_fscache_request *primary)
* Read into page cache in the range described by (@pos, @len).
*
* On return, if the output @unlock is true, the caller is responsible for page
* unlocking; otherwise the callee will take this responsibility through request
* completion.
*
* The return value is the number of bytes successfully handled, or negative
* error code on failure. The only exception is that, the length of the range
* instead of the error code is returned on failure after request is allocated,
* so that .readahead() could advance rac accordingly.
*/
static int erofs_fscache_data_read(struct address_space *mapping,
loff_t pos, size_t len, bool *unlock)
{ {
struct address_space *mapping = primary->mapping;
struct inode *inode = mapping->host; struct inode *inode = mapping->host;
struct super_block *sb = inode->i_sb; struct super_block *sb = inode->i_sb;
struct erofs_fscache_request *req; struct erofs_fscache_request *req;
struct erofs_map_blocks map; struct erofs_map_blocks map;
struct erofs_map_dev mdev; struct erofs_map_dev mdev;
struct iov_iter iter; struct iov_iter iter;
loff_t pos = primary->start + primary->submitted;
size_t count; size_t count;
int ret; int ret;
*unlock = true;
map.m_la = pos; map.m_la = pos;
ret = erofs_map_blocks(inode, &map, EROFS_GET_BLOCKS_RAW); ret = erofs_map_blocks(inode, &map, EROFS_GET_BLOCKS_RAW);
if (ret) if (ret)
...@@ -220,17 +233,19 @@ static int erofs_fscache_data_read(struct address_space *mapping, ...@@ -220,17 +233,19 @@ static int erofs_fscache_data_read(struct address_space *mapping,
} }
iov_iter_zero(PAGE_SIZE - size, &iter); iov_iter_zero(PAGE_SIZE - size, &iter);
erofs_put_metabuf(&buf); erofs_put_metabuf(&buf);
return PAGE_SIZE; primary->submitted += PAGE_SIZE;
return 0;
} }
count = primary->len - primary->submitted;
if (!(map.m_flags & EROFS_MAP_MAPPED)) { if (!(map.m_flags & EROFS_MAP_MAPPED)) {
count = len;
iov_iter_xarray(&iter, READ, &mapping->i_pages, pos, count); iov_iter_xarray(&iter, READ, &mapping->i_pages, pos, count);
iov_iter_zero(count, &iter); iov_iter_zero(count, &iter);
return count; primary->submitted += count;
return 0;
} }
count = min_t(size_t, map.m_llen - (pos - map.m_la), len); count = min_t(size_t, map.m_llen - (pos - map.m_la), count);
DBG_BUGON(!count || count % PAGE_SIZE); DBG_BUGON(!count || count % PAGE_SIZE);
mdev = (struct erofs_map_dev) { mdev = (struct erofs_map_dev) {
...@@ -241,68 +256,65 @@ static int erofs_fscache_data_read(struct address_space *mapping, ...@@ -241,68 +256,65 @@ static int erofs_fscache_data_read(struct address_space *mapping,
if (ret) if (ret)
return ret; return ret;
req = erofs_fscache_req_alloc(mapping, pos, count); req = erofs_fscache_req_chain(primary, count);
if (IS_ERR(req)) if (IS_ERR(req))
return PTR_ERR(req); return PTR_ERR(req);
*unlock = false;
ret = erofs_fscache_read_folios_async(mdev.m_fscache->cookie, ret = erofs_fscache_read_folios_async(mdev.m_fscache->cookie,
req, mdev.m_pa + (pos - map.m_la), count); req, mdev.m_pa + (pos - map.m_la), count);
if (ret)
req->error = ret;
erofs_fscache_req_put(req); erofs_fscache_req_put(req);
return count; primary->submitted += count;
return ret;
} }
static int erofs_fscache_read_folio(struct file *file, struct folio *folio) static int erofs_fscache_data_read(struct erofs_fscache_request *req)
{ {
bool unlock;
int ret; int ret;
DBG_BUGON(folio_size(folio) != EROFS_BLKSIZ); do {
ret = erofs_fscache_data_read_slice(req);
if (ret)
req->error = ret;
} while (!ret && req->submitted < req->len);
ret = erofs_fscache_data_read(folio_mapping(folio), folio_pos(folio), return ret;
folio_size(folio), &unlock); }
if (unlock) {
if (ret > 0) static int erofs_fscache_read_folio(struct file *file, struct folio *folio)
folio_mark_uptodate(folio); {
struct erofs_fscache_request *req;
int ret;
req = erofs_fscache_req_alloc(folio_mapping(folio),
folio_pos(folio), folio_size(folio));
if (IS_ERR(req)) {
folio_unlock(folio); folio_unlock(folio);
return PTR_ERR(req);
} }
return ret < 0 ? ret : 0;
ret = erofs_fscache_data_read(req);
erofs_fscache_req_put(req);
return ret;
} }
static void erofs_fscache_readahead(struct readahead_control *rac) static void erofs_fscache_readahead(struct readahead_control *rac)
{ {
struct folio *folio; struct erofs_fscache_request *req;
size_t len, done = 0;
loff_t start, pos;
bool unlock;
int ret, size;
if (!readahead_count(rac)) if (!readahead_count(rac))
return; return;
start = readahead_pos(rac); req = erofs_fscache_req_alloc(rac->mapping,
len = readahead_length(rac); readahead_pos(rac), readahead_length(rac));
if (IS_ERR(req))
return;
do { /* The request completion will drop refs on the folios. */
pos = start + done; while (readahead_folio(rac))
ret = erofs_fscache_data_read(rac->mapping, pos, ;
len - done, &unlock);
if (ret <= 0)
return;
size = ret; erofs_fscache_data_read(req);
while (size) { erofs_fscache_req_put(req);
folio = readahead_folio(rac);
size -= folio_size(folio);
if (unlock) {
folio_mark_uptodate(folio);
folio_unlock(folio);
}
}
} while ((done += ret) < len);
} }
static const struct address_space_operations erofs_fscache_meta_aops = { static const struct address_space_operations erofs_fscache_meta_aops = {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment