Re: [PATCH v11 4/7] fs/proc/task_mmu: Implement IOCTL to get and optionally clear info about PTEs
From: Michał Mirosław
Date: Mon Mar 13 2023 - 12:02:56 EST
On Thu, 9 Mar 2023 at 14:58, Muhammad Usama Anjum
<usama.anjum@xxxxxxxxxxxxx> wrote:
>
> This IOCTL, PAGEMAP_SCAN on pagemap file can be used to get and/or clear
> the info about page table entries. The following operations are supported
> in this ioctl:
> - Get the information if the pages have been written-to (PAGE_IS_WRITTEN),
> file mapped (PAGE_IS_FILE), present (PAGE_IS_PRESENT) or swapped
> (PAGE_IS_SWAPPED).
> - Find pages which have been written-to and write protect the pages
> (atomic PAGE_IS_WRITTEN + PAGEMAP_WP_ENGAGE)
[...]
> --- a/fs/proc/task_mmu.c
> +++ b/fs/proc/task_mmu.c
> @@ -19,6 +19,7 @@
> #include <linux/shmem_fs.h>
> #include <linux/uaccess.h>
> #include <linux/pkeys.h>
> +#include <linux/minmax.h>
>
> #include <asm/elf.h>
> #include <asm/tlb.h>
> @@ -1132,6 +1133,18 @@ static inline void clear_soft_dirty(struct vm_area_struct *vma,
> }
> #endif
>
> +static inline bool is_pte_uffd_wp(pte_t pte)
> +{
> + return ((pte_present(pte) && pte_uffd_wp(pte)) ||
> + (pte_swp_uffd_wp_any(pte)));
Parentheses around pte_swp_uffd_wp_any() are redundant. Please remove
here and in all following if()s. (Nit: those extra parentheses are
used inconsistently in the patch anyway.)
[...]
> +static inline bool pagemap_scan_is_wt_required(struct pagemap_scan_private *p)
This seems to check if the PAGE_IS_WRITTEN flag is tested, so
"pagemap_scan_needs_wp_checks()"? Or maybe document/expand the "wt"
acronym as it seems used also on following code.
> +{
> + return ((p->required_mask & PAGE_IS_WRITTEN) ||
> + (p->anyof_mask & PAGE_IS_WRITTEN) ||
> + (p->excluded_mask & PAGE_IS_WRITTEN));
Nit: It looks like it should answer "do any of the masks contain
PAGE_IS_WRITTEN?" so maybe:
return (p->required_mask | p->anyof_mask | p->excluded_mask) & PAGE_IS_WRITTEN;
[...]
> +static int pagemap_scan_output(bool wt, bool file, bool pres, bool swap,
> + struct pagemap_scan_private *p,
> + unsigned long addr, unsigned int n_pages)
> +{
> + unsigned long bitmap = PM_SCAN_BITMAP(wt, file, pres, swap);
> + struct page_region *cur = &p->cur;
> + bool cpy = true;
> +
> + if (p->max_pages && (p->found_pages == p->max_pages))
> + return -ENOSPC;
> +
> + if (!n_pages)
> + return -EINVAL;
> +
> + if (p->required_mask)
> + cpy = ((p->required_mask & bitmap) == p->required_mask);
> + if (cpy && p->anyof_mask)
> + cpy = (p->anyof_mask & bitmap);
> + if (cpy && p->excluded_mask)
> + cpy = !(p->excluded_mask & bitmap);
Since the rest of the code is executed only when `cpy` is true, this
could just return early for easier understanding.
BTW, some of the tests are redundant. Eg: if required_mask == 0, then
`required_mask & x == required_mask` will always hold. Same for
`excluded_mask & x == 0`.
> +
> + bitmap = bitmap & p->return_mask;
Nit: bitmap &= p->return_mask;
> + if (cpy && bitmap) {
Assuming early returns on `!cpy` are done earlier:
if (!bitmap)
return 0;
> + if ((cur->len) && (cur->bitmap == bitmap) &&
> + (cur->start + cur->len * PAGE_SIZE == addr)) {
I'd recommend removing the extra parentheses as they make the code
less readable for me (too many parentheses to match visually).
The `cur->len` test seems redundant: is it possible to have
`cur->start == addr` in that case (I guess it would have to get
`n_pages == 0` in an earlier invocation)?
> +
> + cur->len += n_pages;
> + p->found_pages += n_pages;
Please add an early return so that 'else' chaining won't be necessary.
> + } else if ((!p->vec_index) ||
> + ((p->vec_index + 1) < p->vec_len)) {
Can you explain this test? Why not just `p->vec_index < p->vec_len`? Or better:
if (vec_index >= p->vec_len)
return -ENOSPC;
> + if (cur->len) {
> + memcpy(&p->vec[p->vec_index], cur,
> + sizeof(struct page_region));
> + p->vec_index++;
> + }
> +
> + cur->start = addr;
> + cur->len = n_pages;
> + cur->bitmap = bitmap;
> + p->found_pages += n_pages;
> + } else {
> + return -ENOSPC;
> + }
> + }
> +
> + return 0;
> +}
[...]
> +static int pagemap_scan_deposit(struct pagemap_scan_private *p,
> + struct page_region __user *vec,
> + unsigned long *vec_index)
> +{
> + struct page_region *cur = &p->cur;
> +
> + if (cur->len) {
if (!cur->len)
return 0;
> + if (copy_to_user(&vec[*vec_index], cur,
> + sizeof(struct page_region)))
> + return -EFAULT;
> +
> + p->vec_index++;
> + (*vec_index)++;
> + }
> +
> + return 0;
> +}
> +static int pagemap_scan_pmd_entry(pmd_t *pmd, unsigned long start,
> + unsigned long end, struct mm_walk *walk)
> +{
> + struct pagemap_scan_private *p = walk->private;
> + struct vm_area_struct *vma = walk->vma;
> + bool is_writ, is_file, is_pres, is_swap;
> + unsigned long addr = end;
> + spinlock_t *ptl;
> + int ret = 0;
> + pte_t *pte;
> +
> +#ifdef CONFIG_TRANSPARENT_HUGEPAGE
Is the `#ifdef` needed? `pmd_trans_huge_lock()` will always return
NULL if transparent hugepages are not compiled in. OTOH I see
BUILD_BUG() is possible in HPAGE_SIZE definition (irrelevant in this
case), so that would need to be worked around first.
> + ptl = pmd_trans_huge_lock(pmd, vma);
> + if (ptl) {
> + unsigned long n_pages;
> +
> + is_writ = !is_pmd_uffd_wp(*pmd);
`is_written`?
> + /*
> + * Break huge page into small pages if operation needs to be
> + * performed is on a portion of the huge page.
> + */
> + if (is_writ && PM_SCAN_OP_IS_WP(p) &&
> + (end - start < HPAGE_SIZE)) {
> + spin_unlock(ptl);
> +
> + split_huge_pmd(vma, pmd, start);
> + goto process_smaller_pages;
> + }
> +
> + n_pages = (end - start)/PAGE_SIZE;
> + if (p->max_pages &&
> + p->found_pages + n_pages >= p->max_pages)
Nit: greater-than is also correct and avoids no-op assignment.
> + n_pages = p->max_pages - p->found_pages;
> +
> + ret = pagemap_scan_output(is_writ, vma->vm_file,
> + pmd_present(*pmd), is_swap_pmd(*pmd),
> + p, start, n_pages);
> + spin_unlock(ptl);
if (ret || !is_written)
return ret;
This will avoid those tests in the following if().
> +
> + if (!ret && is_writ && PM_SCAN_OP_IS_WP(p) &&
> + uffd_wp_range(walk->mm, vma, start, HPAGE_SIZE, true) < 0)
> + ret = -EINVAL;
> +
> + return ret;
After above early returns, this will be always `return 0;`.
> + }
> +process_smaller_pages:
> + if (pmd_trans_unstable(pmd))
> + return 0;
> +#endif /* CONFIG_TRANSPARENT_HUGEPAGE */
> +
> + for (addr = start; !ret && addr < end; pte++, addr += PAGE_SIZE) {
The `!ret` can be removed if the EINVAL case was to `break` by itself.
> + pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl);
> +
> + is_writ = !is_pte_uffd_wp(*pte);
> + is_file = vma->vm_file;
> + is_pres = pte_present(*pte);
> + is_swap = is_swap_pte(*pte);
> +
> + pte_unmap_unlock(pte, ptl);
> +
> + ret = pagemap_scan_output(is_writ, is_file, is_pres, is_swap,
> + p, addr, 1);
> + if (ret)
> + break;
> +
> + if (PM_SCAN_OP_IS_WP(p) && is_writ &&
> + uffd_wp_range(walk->mm, vma, addr, PAGE_SIZE, true) < 0)
> + ret = -EINVAL;
> + }
> +
> + cond_resched();
> + return ret;
> +}
> +
> +static int pagemap_scan_pte_hole(unsigned long addr, unsigned long end,
> + int depth, struct mm_walk *walk)
> +{
> + struct pagemap_scan_private *p = walk->private;
> + struct vm_area_struct *vma = walk->vma;
> + unsigned long n_pages;
> + int ret = 0;
> +
> + if (vma) {
if (!vma) return 0;
> + n_pages = (end - addr)/PAGE_SIZE;
> + if (p->max_pages &&
> + p->found_pages + n_pages >= p->max_pages)
> + n_pages = p->max_pages - p->found_pages;
> +
> + ret = pagemap_scan_output(false, vma->vm_file, false, false, p,
> + addr, n_pages);
> + }
> +
> + return ret;
> +}
> +/* No hugetlb support is present. */
"FIXME: hugetlb support is not implemented."? (There seems to be no
#ifdef CONFIG_HUGETLB or similar, so I guess the comment is about the
current implementation.)
> +static const struct mm_walk_ops pagemap_scan_ops = {
> + .test_walk = pagemap_scan_test_walk,
> + .pmd_entry = pagemap_scan_pmd_entry,
> + .pte_hole = pagemap_scan_pte_hole,
> +};
> +
> +static bool pagemap_scan_args_valid(struct pm_scan_arg *arg,
> + struct page_region __user *vec,
> + unsigned long start)
> +{
> + /* Detect illegal size, flags and masks */
> + if (arg->size != sizeof(struct pm_scan_arg))
> + return false;
> + if (arg->flags & ~PM_SCAN_OPS)
> + return false;
> + if ((arg->required_mask | arg->anyof_mask | arg->excluded_mask |
> + arg->return_mask) & ~PM_SCAN_BITS_ALL)
> + return false;
> + if (!arg->required_mask && !arg->anyof_mask &&
> + !arg->excluded_mask)
> + return false;
Is there an assumption in the code that those checks are needed? I'd
expect that no selection criteria makes a valid page set?
> + if (!arg->return_mask)
> + return false;
> +
> + /* Validate memory ranges */
> + if (!(arg->flags & PM_SCAN_OP_GET))
> + return false;
> + if (!arg->vec)
> + return false;
> + if (arg->vec_len == 0)
> + return false;
> + if (!access_ok((void __user *)vec,
> + arg->vec_len * sizeof(struct page_region)))
> + return false;
Is there a provision that userspace threads are all blocked from
manipulating mmaps during this ioctl()? If not, this is a TOCTOU bug
and the writes should be checked each time as another userspace thread
could remap the memory while the ioctl() is working. Anyway, the
return should be EFAULT for this case.
> + if (!IS_ALIGNED(start, PAGE_SIZE))
> + return false;
> + if (!access_ok((void __user *)start, arg->len))
> + return false;
This I guess want's to check if the range to be scanned is mapped -
but isn't this what the ioctl() should do during the scan? (But, also
see above.)
> + if (PM_SCAN_OP_IS_WP(arg)) {
if (!...IS_WP) return true;
> + if (arg->required_mask & PM_SCAN_NON_WT_BITS)
> + return false;
> + if (arg->anyof_mask & PM_SCAN_NON_WT_BITS)
> + return false;
> + if (arg->excluded_mask & PM_SCAN_NON_WT_BITS)
> + return false;
Please see: pagemap_scan_is_wt_required comment. Also, it seems this
constant is used only here, so ~PAGE_IS_WRITTEN might be enough?
[...]
> +static long do_pagemap_cmd(struct mm_struct *mm, struct pm_scan_arg *arg)
> +{
> + unsigned long start, end, walk_start, walk_end;
> + unsigned long empty_slots, vec_index = 0;
> + struct page_region __user *vec;
> + struct pagemap_scan_private p;
> + int ret = 0;
> +
> + start = (unsigned long)untagged_addr(arg->start);
> + vec = (struct page_region *)(unsigned long)untagged_addr(arg->vec);
> +
> + if (!pagemap_scan_args_valid(arg, vec, start))
> + return -EINVAL;
> +
> + end = start + arg->len;
> + p.max_pages = arg->max_pages;
> + p.found_pages = 0;
> + p.flags = arg->flags;
> + p.required_mask = arg->required_mask;
> + p.anyof_mask = arg->anyof_mask;
> + p.excluded_mask = arg->excluded_mask;
> + p.return_mask = arg->return_mask;
> + p.cur.len = 0;
> + p.vec = NULL;
> + p.vec_len = (PAGEMAP_WALK_SIZE >> PAGE_SHIFT);
> +
> + /*
> + * Allocate smaller buffer to get output from inside the page walk
> + * functions and walk page range in PAGEMAP_WALK_SIZE size chunks. As
> + * we want to return output to user in compact form where no two
> + * consecutive regions should be continuous and have the same flags.
> + * So store the latest element in p.cur between different walks and
> + * store the p.cur at the end of the walk to the user buffer.
> + */
> + p.vec = kmalloc_array(p.vec_len, sizeof(struct page_region),
> + GFP_KERNEL);
> + if (!p.vec)
> + return -ENOMEM;
> +
> + walk_start = walk_end = start;
> + while (walk_end < end) {
> + p.vec_index = 0;
> +
> + empty_slots = arg->vec_len - vec_index;
> + p.vec_len = min(p.vec_len, empty_slots);
> +
> + walk_end = (walk_start + PAGEMAP_WALK_SIZE) & PAGEMAP_WALK_MASK;
> + if (walk_end > end)
> + walk_end = end;
> +
> + mmap_read_lock(mm);
> + ret = walk_page_range(mm, walk_start, walk_end,
> + &pagemap_scan_ops, &p);
> + mmap_read_unlock(mm);
> +
> + if (!(!ret || ret == -ENOSPC))
if (ret && ret != -ENOSPC)
> + goto free_data;
> +
> + walk_start = walk_end;
> + if (p.vec_index) {
> + if (copy_to_user(&vec[vec_index], p.vec,
> + p.vec_index *
> + sizeof(struct page_region))) {
> + ret = -EFAULT;
> + goto free_data;
> + }
> + vec_index += p.vec_index;
> + }
> + }
> + ret = pagemap_scan_deposit(&p, vec, &vec_index);
> + if (!ret)
> + ret = vec_index;
> +free_data:
> + kfree(p.vec);
> +
> + return ret;
> +}
> +
> +static long pagemap_scan_ioctl(struct file *file, unsigned int cmd,
> + unsigned long arg)
> +{
> + struct pm_scan_arg __user *uarg = (struct pm_scan_arg __user *)arg;
> + struct mm_struct *mm = file->private_data;
> + struct pm_scan_arg argument;
> +
> + if (cmd == PAGEMAP_SCAN) {
switch() for easier expansion later?
> + if (copy_from_user(&argument, uarg,
> + sizeof(struct pm_scan_arg)))
sizeof(*argument);
Could you push this to do_pagemap_cmd()? In case this file gets more
ioctl() commands there won't be need to add more command-specific
structures in this function.
> + return -EFAULT;
> + return do_pagemap_cmd(mm, &argument);
> + }
> +
> + return -EINVAL;
> +}
> +
> const struct file_operations proc_pagemap_operations = {
> .llseek = mem_lseek, /* borrow this */
> .read = pagemap_read,
> .open = pagemap_open,
> .release = pagemap_release,
> + .unlocked_ioctl = pagemap_scan_ioctl,
> + .compat_ioctl = pagemap_scan_ioctl,
Is this correct? Would the code need a different userspace pointer
handling for 32-bit userspace on 64-bit kernel?
> };
> #endif /* CONFIG_PROC_PAGE_MONITOR */