Re: [PATCH v1 14/16] mm: rename zap_page_range_single() to zap_vma_range()

From: Alice Ryhl

Date: Sat Feb 28 2026 - 07:44:58 EST


On Fri, Feb 27, 2026 at 09:08:45PM +0100, David Hildenbrand (Arm) wrote:
> diff --git a/drivers/android/binder/page_range.rs b/drivers/android/binder/page_range.rs
> index fdd97112ef5c..2fddd4ed8d4c 100644
> --- a/drivers/android/binder/page_range.rs
> +++ b/drivers/android/binder/page_range.rs
> @@ -130,7 +130,7 @@ pub(crate) struct ShrinkablePageRange {
> pid: Pid,
> /// The mm for the relevant process.
> mm: ARef<Mm>,
> - /// Used to synchronize calls to `vm_insert_page` and `zap_page_range_single`.
> + /// Used to synchronize calls to `vm_insert_page` and `zap_vma_range`.
> #[pin]
> mm_lock: Mutex<()>,
> /// Spinlock protecting changes to pages.
> @@ -719,7 +719,7 @@ fn drop(self: Pin<&mut Self>) {
>
> if let Some(vma) = mmap_read.vma_lookup(vma_addr) {
> let user_page_addr = vma_addr + (page_index << PAGE_SHIFT);
> - vma.zap_page_range_single(user_page_addr, PAGE_SIZE);
> + vma.zap_vma_range(user_page_addr, PAGE_SIZE);
> }

LGTM. Be aware that this will have a merge conflict with patches
currently in char-misc-linus that are scheduled to land in an -rc.

> diff --git a/drivers/android/binder_alloc.c b/drivers/android/binder_alloc.c
> index dd2046bd5cde..e4488ad86a65 100644
> --- a/drivers/android/binder_alloc.c
> +++ b/drivers/android/binder_alloc.c
> @@ -1185,7 +1185,7 @@ enum lru_status binder_alloc_free_page(struct list_head *item,
> if (vma) {
> trace_binder_unmap_user_start(alloc, index);
>
> - zap_page_range_single(vma, page_addr, PAGE_SIZE);
> + zap_vma_range(vma, page_addr, PAGE_SIZE);
>
> trace_binder_unmap_user_end(alloc, index);

LGTM.

> diff --git a/rust/kernel/mm/virt.rs b/rust/kernel/mm/virt.rs
> index b8e59e4420f3..04b3cc925d67 100644
> --- a/rust/kernel/mm/virt.rs
> +++ b/rust/kernel/mm/virt.rs
> @@ -113,7 +113,7 @@ pub fn end(&self) -> usize {
> /// kernel goes further in freeing unused page tables, but for the purposes of this operation
> /// we must only assume that the leaf level is cleared.
> #[inline]
> - pub fn zap_page_range_single(&self, address: usize, size: usize) {
> + pub fn zap_vma_range(&self, address: usize, size: usize) {
> let (end, did_overflow) = address.overflowing_add(size);
> if did_overflow || address < self.start() || self.end() < end {
> // TODO: call WARN_ONCE once Rust version of it is added
> @@ -124,7 +124,7 @@ pub fn zap_page_range_single(&self, address: usize, size: usize) {
> // sufficient for this method call. This method has no requirements on the vma flags. The
> // address range is checked to be within the vma.
> unsafe {
> - bindings::zap_page_range_single(self.as_ptr(), address, size)
> + bindings::zap_vma_range(self.as_ptr(), address, size)
> };
> }

Same as previous patch: please run rustfmt. It will format on a single
line, like this:

unsafe { bindings::zap_vma_range(self.as_ptr(), address, size) };

with the above change applied:

Acked-by: Alice Ryhl <aliceryhl@xxxxxxxxxx> # Rust and Binder

Alice