Re: [RFC PATCH] madvise: make madvise_cold_or_pageout_pte_range() support large folio

From: Yu Zhao
Date: Fri Jul 14 2023 - 11:42:35 EST


On Thu, Jul 13, 2023 at 11:57 PM Yin, Fengwei <fengwei.yin@xxxxxxxxx> wrote:
>
>
> >> - if (pageout_anon_only_filter && !folio_test_anon(folio))
> >> + /* Do not interfere with other mappings of this folio */
> >> + if (folio_mapcount(folio) != 1)
> >> continue;
> >>
> >> - VM_BUG_ON_FOLIO(folio_test_large(folio), folio);
> >> -
> >> - if (pte_young(ptent)) {
> >> - ptent = ptep_get_and_clear_full(mm, addr, pte,
> >> - tlb->fullmm);
> >> - ptent = pte_mkold(ptent);
> >> - set_pte_at(mm, addr, pte, ptent);
> >> - tlb_remove_tlb_entry(tlb, pte, addr);
> >> - }
> >> -
> >> - /*
> >> - * We are deactivating a folio for accelerating reclaiming.
> >> - * VM couldn't reclaim the folio unless we clear PG_young.
> >> - * As a side effect, it makes confuse idle-page tracking
> >> - * because they will miss recent referenced history.
> >> - */
> >> - folio_clear_referenced(folio);
> >> - folio_test_clear_young(folio);
> >> - if (folio_test_active(folio))
> >> - folio_set_workingset(folio);
> >> +pageout_cold_folio:
> >> if (pageout) {
> >> if (folio_isolate_lru(folio)) {
> >> if (folio_test_unevictable(folio))
> >> @@ -529,8 +542,30 @@ static int madvise_cold_or_pageout_pte_range(pmd_t *pmd,
> >> arch_leave_lazy_mmu_mode();
> >> pte_unmap_unlock(start_pte, ptl);
> >> }
> >> - if (pageout)
> >> - reclaim_pages(&folio_list);
> >> +
> >> + if (pageout) {
> >> + LIST_HEAD(reclaim_list);
> >> +
> >> + while (!list_empty(&folio_list)) {
> >> + int refs;
> >> + unsigned long flags;
> >> + struct mem_cgroup *memcg = folio_memcg(folio);
> >> +
> >> + folio = lru_to_folio(&folio_list);
> >> + list_del(&folio->lru);
> >> +
> >> + refs = folio_referenced(folio, 0, memcg, &flags);
> >> +
> >> + if ((flags & VM_LOCKED) || (refs == -1)) {
> >> + folio_putback_lru(folio);
> >> + continue;
> >> + }
> >> +
> >> + folio_test_clear_referenced(folio);
> >> + list_add(&folio->lru, &reclaim_list);
> >> + }
> >> + reclaim_pages(&reclaim_list);
> >> + }
> >
> > i overlooked the chunk above -- it's unnecessary: after we split the
> > large folio (and splice the base folios onto the same LRU list), we
> > continue at the position of the first base folio because of:
> >
> > pte--;
> > addr -= PAGE_SIZE;
> > continue;
> >
> > And then we do pte_mkold(), which takes care of the A-bit.
> This patch moves the A-bit clear out of the folio isolation loop. So
> even the folio is split and loop restarts from the first base folio,
> the A-bit is not cleared. A-bit is only cleared in reclaim loop.
>
> There is one option for A-bit clearing:
> - clear A-bit of base 4K page in isolation loop and leave large folio
> A-bit clearing to reclaim loop.
>
> This patch didn't use it because don't want to introduce A-bit clearing
> in two places. But I am open about clearing base 4K page A-bit cleared in
> isolation loop. Thanks.

Sorry but why are we trying to do multiple things in one patch that I
assumed is supposed to simply fix madvise() for large anon folios? And
none of those things seems to have a clear rationale behind it.

The only patch that makes sense at the moment (or the first patch of a
series) is what I said before:

- if (folio_mapcount(folio) != 1)
+ if (folio_estimated_sharers(folio) != 1)

And probably clarify (preferrably in the comments above) this is an
estimate because we think it's a better tradeoff if we do so (less
code/overhead from checking the mapcounts of the rest of folios within
the range).