[PATCH v2 -next 2/3] x86/mm/pat: Cleanup unused parameter in follow_phys

From: Wupeng Ma
Date: Mon Mar 06 2023 - 21:26:35 EST


From: Ma Wupeng <mawupeng1@xxxxxxxxxx>

Parameter flags is always zero in caller untrack_pfn() and
track_pfn_copy(). let's drop it.

Signed-off-by: Ma Wupeng <mawupeng1@xxxxxxxxxx>
---
arch/x86/mm/pat/memtype.c | 10 +++-------
1 file changed, 3 insertions(+), 7 deletions(-)

diff --git a/arch/x86/mm/pat/memtype.c b/arch/x86/mm/pat/memtype.c
index 4fb1e89f32e4..3ec025011c90 100644
--- a/arch/x86/mm/pat/memtype.c
+++ b/arch/x86/mm/pat/memtype.c
@@ -951,8 +951,7 @@ static void free_pfn_range(u64 paddr, unsigned long size)
}

static int follow_phys(struct vm_area_struct *vma, unsigned long address,
- unsigned int flags, unsigned long *prot,
- resource_size_t *phys)
+ unsigned long *prot, resource_size_t *phys)
{
int ret = -EINVAL;
pte_t *ptep, pte;
@@ -965,9 +964,6 @@ static int follow_phys(struct vm_area_struct *vma, unsigned long address,
goto out;
pte = *ptep;

- if ((flags & FOLL_WRITE) && !pte_write(pte))
- goto unlock;
-
*prot = pgprot_val(pte_pgprot(pte));
*phys = (resource_size_t)pte_pfn(pte) << PAGE_SHIFT;

@@ -997,7 +993,7 @@ int track_pfn_copy(struct vm_area_struct *vma)
* reserve the whole chunk covered by vma. We need the
* starting address and protection from pte.
*/
- if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) {
+ if (follow_phys(vma, vma->vm_start, &prot, &paddr)) {
WARN_ON_ONCE(1);
return -EINVAL;
}
@@ -1084,7 +1080,7 @@ void untrack_pfn(struct vm_area_struct *vma, unsigned long pfn,
/* free the chunk starting from pfn or the whole chunk */
paddr = (resource_size_t)pfn << PAGE_SHIFT;
if (!paddr && !size) {
- if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) {
+ if (follow_phys(vma, vma->vm_start, &prot, &paddr)) {
WARN_ON_ONCE(1);
return;
}
--
2.25.1