Re: [PATCH] mm, soft-offline: convert parameter to pfn

From: Naoya Horiguchi
Date: Wed Oct 16 2019 - 04:54:41 EST


On Wed, Oct 16, 2019 at 10:34:52AM +0200, David Hildenbrand wrote:
> On 16.10.19 10:27, Naoya Horiguchi wrote:
> > On Wed, Oct 16, 2019 at 09:56:19AM +0200, David Hildenbrand wrote:
> > > On 16.10.19 09:09, Naoya Horiguchi wrote:
> > > > Hi,
> > > >
> > > > I wrote a simple cleanup for parameter of soft_offline_page(),
> > > > based on thread https://lkml.org/lkml/2019/10/11/57.
> > > >
> > > > I know that we need more cleanup on hwpoison-inject, but I think
> > > > that will be mentioned in re-write patchset Oscar is preparing now.
> > > > So let me shared only this part as a separate one now.
> > ...
> > >
> > > I think you should rebase that patch on linux-next (where the
> > > pfn_to_online_page() check is in place). I assume you'll want to move the
> > > pfn_to_online_page() check into soft_offline_page() then as well?
> >
> > I rebased to next-20191016. And yes, we will move pfn_to_online_page()
> > into soft offline code. It seems that we can also move pfn_valid(),
> > but is simply moving like below good enough for you?
>
> At least I can't am the patch to current next/master (due to
> pfn_to_online_page()).
>
> >
> > @@ -1877,11 +1877,17 @@ static int soft_offline_free_page(struct page *page)
> > * This is not a 100% solution for all memory, but tries to be
> > * ``good enough'' for the majority of memory.
> > */
> > -int soft_offline_page(struct page *page, int flags)
> > +int soft_offline_page(unsigned long pfn, int flags)
> > {
> > int ret;
> > - unsigned long pfn = page_to_pfn(page);
> > + struct page *page;
> > + if (!pfn_valid(pfn))
> > + return -ENXIO;
> > + /* Only online pages can be soft-offlined (esp., not ZONE_DEVICE). */
> > + if (!pfn_to_online_page(pfn))
> > + return -EIO;
> > + page = pfn_to_page(pfn);
> > if (is_zone_device_page(page)) {
> > pr_debug_ratelimited("soft_offline: %#lx page is device page\n",
> > pfn);
> > --
> >
> > Or we might have an option to do as memory_failure() does like below:
>
> In contrast to soft offlining, memory failure can deal with devmem. So I
> think the above makes sense.

OK, so here's the revised one.

Thanks,
Naoya Horiguchi
---
From: Naoya Horiguchi <n-horiguchi@xxxxxxxxxxxxx>
Date: Wed, 16 Oct 2019 17:00:33 +0900
Subject: [PATCH] mm, soft-offline: convert parameter to pfn

Currently soft_offline_page() receives struct page, and its sibling
memory_failure() receives pfn. This discrepancy looks weird and makes
precheck on pfn validity tricky. So let's align them.

Signed-off-by: Naoya Horiguchi <n-horiguchi@xxxxxxxxxxxxx>
---
drivers/base/memory.c | 7 +------
include/linux/mm.h | 2 +-
mm/madvise.c | 2 +-
mm/memory-failure.c | 14 ++++++++++----
4 files changed, 13 insertions(+), 12 deletions(-)

diff --git a/drivers/base/memory.c b/drivers/base/memory.c
index 55907c27075b..a757d9ed88a7 100644
--- a/drivers/base/memory.c
+++ b/drivers/base/memory.c
@@ -538,12 +538,7 @@ static ssize_t soft_offline_page_store(struct device *dev,
if (kstrtoull(buf, 0, &pfn) < 0)
return -EINVAL;
pfn >>= PAGE_SHIFT;
- if (!pfn_valid(pfn))
- return -ENXIO;
- /* Only online pages can be soft-offlined (esp., not ZONE_DEVICE). */
- if (!pfn_to_online_page(pfn))
- return -EIO;
- ret = soft_offline_page(pfn_to_page(pfn), 0);
+ ret = soft_offline_page(pfn, 0);
return ret == 0 ? count : ret;
}

diff --git a/include/linux/mm.h b/include/linux/mm.h
index 44d058723db9..fd360d208346 100644
--- a/include/linux/mm.h
+++ b/include/linux/mm.h
@@ -2794,7 +2794,7 @@ extern int sysctl_memory_failure_early_kill;
extern int sysctl_memory_failure_recovery;
extern void shake_page(struct page *p, int access);
extern atomic_long_t num_poisoned_pages __read_mostly;
-extern int soft_offline_page(struct page *page, int flags);
+extern int soft_offline_page(unsigned long pfn, int flags);


/*
diff --git a/mm/madvise.c b/mm/madvise.c
index 2be9f3fdb05e..99dd06fecfa9 100644
--- a/mm/madvise.c
+++ b/mm/madvise.c
@@ -887,7 +887,7 @@ static int madvise_inject_error(int behavior,
pr_info("Soft offlining pfn %#lx at process virtual address %#lx\n",
pfn, start);

- ret = soft_offline_page(page, MF_COUNT_INCREASED);
+ ret = soft_offline_page(pfn, MF_COUNT_INCREASED);
if (ret)
return ret;
continue;
diff --git a/mm/memory-failure.c b/mm/memory-failure.c
index 05c8c6df25e6..bdf408d7f65c 100644
--- a/mm/memory-failure.c
+++ b/mm/memory-failure.c
@@ -1476,7 +1476,7 @@ static void memory_failure_work_func(struct work_struct *work)
if (!gotten)
break;
if (entry.flags & MF_SOFT_OFFLINE)
- soft_offline_page(pfn_to_page(entry.pfn), entry.flags);
+ soft_offline_page(entry.pfn, entry.flags);
else
memory_failure(entry.pfn, entry.flags);
}
@@ -1857,7 +1857,7 @@ static int soft_offline_free_page(struct page *page)

/**
* soft_offline_page - Soft offline a page.
- * @page: page to offline
+ * @pfn: pfn to soft-offline
* @flags: flags. Same as memory_failure().
*
* Returns 0 on success, otherwise negated errno.
@@ -1877,11 +1877,17 @@ static int soft_offline_free_page(struct page *page)
* This is not a 100% solution for all memory, but tries to be
* ``good enough'' for the majority of memory.
*/
-int soft_offline_page(struct page *page, int flags)
+int soft_offline_page(unsigned long pfn, int flags)
{
int ret;
- unsigned long pfn = page_to_pfn(page);
+ struct page *page;

+ if (!pfn_valid(pfn))
+ return -ENXIO;
+ /* Only online pages can be soft-offlined (esp., not ZONE_DEVICE). */
+ page = pfn_to_online_page(pfn);
+ if (!page)
+ return -EIO;
if (is_zone_device_page(page)) {
pr_debug_ratelimited("soft_offline: %#lx page is device page\n",
pfn);
--
2.17.1