[RFC PATCH 4/6] erofs: implement .read_iter for page cache sharing
From: Jingbo Xu
Date: Fri Jan 06 2023 - 07:54:18 EST
When page cache sharing enabled, page caches are managed in the address
space of blobs rather than erofs inodes. All erofs inodes sharing one
chunk will refer to and share the page cache in the blob's address
space.
Signed-off-by: Jingbo Xu <jefflexu@xxxxxxxxxxxxxxxxx>
---
fs/erofs/fscache.c | 64 +++++++++++++++++++++++++++++++++++++++++++++
fs/erofs/internal.h | 1 +
2 files changed, 65 insertions(+)
diff --git a/fs/erofs/fscache.c b/fs/erofs/fscache.c
index ea276884f043..1f2a42dd1590 100644
--- a/fs/erofs/fscache.c
+++ b/fs/erofs/fscache.c
@@ -320,6 +320,70 @@ const struct address_space_operations erofs_fscache_access_aops = {
static const struct file_operations erofs_fscache_meta_fops = {};
+static ssize_t erofs_fscache_share_file_read_iter(struct kiocb *iocb,
+ struct iov_iter *to)
+{
+ struct file *filp = iocb->ki_filp;
+ struct inode *inode = file_inode(filp);
+ /* since page cache sharing is enabled only when i_size <= chunk_size */
+ struct erofs_map_blocks map = {}; /* .m_la = 0 */
+ struct erofs_map_dev mdev;
+ struct folio *folio;
+ ssize_t already_read = 0;
+ int ret = 0;
+
+ /* no need taking (shared) inode lock since it's a ro filesystem */
+ if (!iov_iter_count(to))
+ return 0;
+
+ if (IS_DAX(inode) || iocb->ki_flags & IOCB_DIRECT)
+ return -EOPNOTSUPP;
+
+ ret = erofs_map_blocks(inode, &map, EROFS_GET_BLOCKS_RAW);
+ if (ret)
+ return ret;
+
+ mdev = (struct erofs_map_dev) {
+ .m_deviceid = map.m_deviceid,
+ .m_pa = map.m_pa,
+ };
+ ret = erofs_map_dev(inode->i_sb, &mdev);
+ if (ret)
+ return ret;
+
+ do {
+ size_t bytes, copied, offset, fsize;
+ pgoff_t index = (mdev.m_pa >> PAGE_SHIFT) + (iocb->ki_pos >> PAGE_SHIFT);
+
+ folio = read_cache_folio(mdev.m_fscache->inode->i_mapping, index, NULL, NULL);
+ if (IS_ERR(folio)) {
+ ret = PTR_ERR(folio);
+ break;
+ }
+
+ fsize = folio_size(folio);
+ offset = iocb->ki_pos & (fsize - 1);
+ bytes = min_t(size_t, inode->i_size - iocb->ki_pos, iov_iter_count(to));
+ bytes = min_t(size_t, bytes, fsize - offset);
+ copied = copy_folio_to_iter(folio, offset, bytes, to);
+ folio_put(folio);
+ iocb->ki_pos += copied;
+ already_read += copied;
+ if (copied < bytes) {
+ ret = -EFAULT;
+ break;
+ }
+ } while (iov_iter_count(to) && iocb->ki_pos < inode->i_size);
+
+ file_accessed(filp);
+ return already_read ? already_read : ret;
+}
+
+const struct file_operations erofs_fscache_share_file_fops = {
+ .llseek = generic_file_llseek,
+ .read_iter = erofs_fscache_share_file_read_iter,
+};
+
static void erofs_fscache_domain_put(struct erofs_domain *domain)
{
mutex_lock(&erofs_domain_list_lock);
diff --git a/fs/erofs/internal.h b/fs/erofs/internal.h
index 24d471fe2fa4..386e2fd4c025 100644
--- a/fs/erofs/internal.h
+++ b/fs/erofs/internal.h
@@ -617,6 +617,7 @@ struct erofs_fscache *erofs_fscache_register_cookie(struct super_block *sb,
void erofs_fscache_unregister_cookie(struct erofs_fscache *fscache);
extern const struct address_space_operations erofs_fscache_access_aops;
+extern const struct file_operations erofs_fscache_share_file_fops;
#else
static inline int erofs_fscache_register_fs(struct super_block *sb)
{
--
2.19.1.6.gb485710b