On 04.03.2020 05:34, Shile Zhang wrote:
Hi Kirill,Here is problem: prev_nr_pages must be updated.
Thanks for your quickly reply!
On 2020/3/4 00:52, Kirill Tkhai wrote:
On 03.03.2020 19:15, Shile Zhang wrote:Yes, you're right! I missed this point.
When 'CONFIG_DEFERRED_STRUCT_PAGE_INIT' is set, 'pgdatinit' kthread willThe way I understand the original problem, that Pavel fixed:
initialise the deferred pages with local interrupts disabled. It is
introduced by commit 3a2d7fa8a3d5 ("mm: disable interrupts while
initializing deferred pages").
The local interrupt will be disabled long time inside
deferred_init_memmap(), depends on memory size.
On machine with NCPUS <= 2, the 'pgdatinit' kthread could be pined on
boot CPU, then the tick timer will stuck long time, which caused the
system wall time inaccuracy.
For example, the dmesg shown that:
ÂÂ [ÂÂÂ 0.197975] node 0 initialised, 32170688 pages in 1ms
Obviously, 1ms is unreasonable.
Now, fix it by restore in the pending interrupts inside the while loop.
The reasonable demsg shown likes:
[ÂÂÂ 1.069306] node 0 initialised, 32203456 pages in 894ms
we need disable irqs in deferred_init_memmap() since this function may be called
in parallel with deferred_grow_zone() called from interrupt handler. So, Pavel
added lock to fix the race.
In case of we temporary unlock the lock, interrupt still be possible,
so my previous proposition returns the problem back.
Now thought again, I think we have to just add:
ÂÂÂÂpgdat_resize_unlock();
ÂÂÂÂpgdat_resize_lock();
instead of releasing interrupts, since in case of we just release them with lock held,
a call of interrupt->deferred_grow_zone() bring us to a deadlock.
So, unlock the lock is must.
Thanks for your comment!
Yeah, your patch is better!Signed-off-by: Shile Zhang<shile.zhang@xxxxxxxxxxxxxxxxx>I think we need here something like below (untested):
---
 mm/page_alloc.c | 6 +++++-
 1 file changed, 5 insertions(+), 1 deletion(-)
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 3c4eb750a199..d3f337f2e089 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -1809,8 +1809,12 @@ static int __init deferred_init_memmap(void *data)
ÂÂÂÂÂÂ * that we can avoid introducing any issues with the buddy
ÂÂÂÂÂÂ * allocator.
ÂÂÂÂÂÂ */
-ÂÂÂ while (spfn < epfn)
+ÂÂÂ while (spfn < epfn) {
ÂÂÂÂÂÂÂÂÂ nr_pages += deferred_init_maxorder(&i, zone, &spfn, &epfn);
+ÂÂÂÂÂÂÂ /* let in any pending interrupts */
+ÂÂÂÂÂÂÂ local_irq_restore(flags);
+ÂÂÂÂÂÂÂ local_irq_save(flags);
+ÂÂÂ }
 zone_empty:
ÂÂÂÂÂ pgdat_resize_unlock(pgdat, &flags);
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 79e950d76ffc..323afa9a4db5 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -1828,7 +1828,7 @@ static int __init deferred_init_memmap(void *data)
 {
ÂÂÂÂÂ pg_data_t *pgdat = data;
ÂÂÂÂÂ const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id);
-ÂÂÂ unsigned long spfn = 0, epfn = 0, nr_pages = 0;
+ÂÂÂ unsigned long spfn = 0, epfn = 0, nr_pages = 0, prev_nr_pages = 0;
ÂÂÂÂÂ unsigned long first_init_pfn, flags;
ÂÂÂÂÂ unsigned long start = jiffies;
ÂÂÂÂÂ struct zone *zone;
@@ -1869,8 +1869,18 @@ static int __init deferred_init_memmap(void *data)
ÂÂÂÂÂÂ * that we can avoid introducing any issues with the buddy
ÂÂÂÂÂÂ * allocator.
ÂÂÂÂÂÂ */
-ÂÂÂ while (spfn < epfn)
+ÂÂÂ while (spfn < epfn) {
ÂÂÂÂÂÂÂÂÂ nr_pages += deferred_init_maxorder(&i, zone, &spfn, &epfn);
+ÂÂÂÂÂÂÂ /*
+ÂÂÂÂÂÂÂÂ * Release interrupts every 1Gb to give a possibility
+ÂÂÂÂÂÂÂÂ * a timer to advance jiffies.
+ÂÂÂÂÂÂÂÂ */
+ÂÂÂÂÂÂÂ if (nr_pages - prev_nr_pages > (1UL << (30 - PAGE_SHIFT))) {
+ÂÂÂÂÂÂÂÂÂÂÂ prev_nr_pages = nr_pages;
+ÂÂÂÂÂÂÂÂÂÂÂ pgdat_resize_unlock(pgdat, &flags);
+ÂÂÂÂÂÂÂÂÂÂÂ pgdat_resize_lock(pgdat, &flags);
+ÂÂÂÂÂÂÂ }
+ÂÂÂ }
 zone_empty:
ÂÂÂÂÂ pgdat_resize_unlock(pgdat, &flags);
(I believe the comment may be improved more).
I test your code and it works!
But it seems that 1G is still hold the interrupts too long, about 40ms in my env
with Intel(R) Xeon(R) 2.5GHz). I tried other size, it is OK to use 1024 pages (4MB),
which suggested by Andrew's before.
Could you please help to review it again?
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 3c4eb750a199..5def66d3ffcd 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -1768,7 +1768,7 @@ static int __init deferred_init_memmap(void *data)
Â{
ÂÂÂÂÂÂÂ pg_data_t *pgdat = data;
ÂÂÂÂÂÂÂ const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id);
-ÂÂÂÂÂÂ unsigned long spfn = 0, epfn = 0, nr_pages = 0;
+ÂÂÂÂÂÂ unsigned long spfn = 0, epfn = 0, nr_pages = 0, prev_nr_pages = 0;
ÂÂÂÂÂÂÂ unsigned long first_init_pfn, flags;
ÂÂÂÂÂÂÂ unsigned long start = jiffies;
ÂÂÂÂÂÂÂ struct zone *zone;
@@ -1809,8 +1809,17 @@ static int __init deferred_init_memmap(void *data)
ÂÂÂÂÂÂÂÂ * that we can avoid introducing any issues with the buddy
ÂÂÂÂÂÂÂÂ * allocator.
ÂÂÂÂÂÂÂÂ */
-ÂÂÂÂÂÂ while (spfn < epfn)
+ÂÂÂÂÂÂ while (spfn < epfn) {
ÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂ nr_pages += deferred_init_maxorder(&i, zone, &spfn, &epfn);
+ÂÂÂÂÂÂÂÂÂÂÂÂÂÂ /*
+ÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂ * Restore pending interrupts every 1024 pages to give
+ÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂ * the chance tick timer to advance jiffies.
+ÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂ */
+ÂÂÂÂÂÂÂÂÂÂÂÂÂÂ if (nr_pages - prev_nr_pages > 1024) {
+ÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂ pgdat_resize_unlock(&flags);
+ÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂÂ pgdat_resize_lock(&flags);
Anyway, releasing every 4M looks wrong for me, since you removes the fix that Pavel introduced.
He protected against big allocations made from interrupt content. But in case of we unlock
the lock after 4Mb, only 4Mb will be available for allocations from interrupts. pgdat->first_deferred_pfn
is updated at the start of function, so interrupt allocations won't be able to initialize
mode for themselve.
In case of you want unlock interrupts very often, you should make some creativity with first_deferred_pfn.
We should update it sequentially. Something like below (untested):
---
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 79e950d76ffc..be09d158baeb 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -1828,7 +1828,7 @@ static int __init deferred_init_memmap(void *data)
{
pg_data_t *pgdat = data;
const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id);
- unsigned long spfn = 0, epfn = 0, nr_pages = 0;
+ unsigned long spfn = 0, epfn = 0, nr_pages;
unsigned long first_init_pfn, flags;
unsigned long start = jiffies;
struct zone *zone;
@@ -1838,7 +1838,7 @@ static int __init deferred_init_memmap(void *data)
/* Bind memory initialisation thread to a local node if possible */
if (!cpumask_empty(cpumask))
set_cpus_allowed_ptr(current, cpumask);
-
+again:
pgdat_resize_lock(pgdat, &flags);
first_init_pfn = pgdat->first_deferred_pfn;
if (first_init_pfn == ULONG_MAX) {
@@ -1850,7 +1850,6 @@ static int __init deferred_init_memmap(void *data)
/* Sanity check boundaries */
BUG_ON(pgdat->first_deferred_pfn < pgdat->node_start_pfn);
BUG_ON(pgdat->first_deferred_pfn > pgdat_end_pfn(pgdat));
- pgdat->first_deferred_pfn = ULONG_MAX;
/* Only the highest zone is deferred so find it */
for (zid = 0; zid < MAX_NR_ZONES; zid++) {
@@ -1864,14 +1863,30 @@ static int __init deferred_init_memmap(void *data)
first_init_pfn))
goto zone_empty;
+ nr_pages = 0;
+
/*
* Initialize and free pages in MAX_ORDER sized increments so
* that we can avoid introducing any issues with the buddy
* allocator.
+ * Final iteration marker is: spfn=ULONG_MAX and epfn=0.
*/
- while (spfn < epfn)
+ while (spfn < epfn) {
nr_pages += deferred_init_maxorder(&i, zone, &spfn, &epfn);
+ if (!epfn)
+ break;
+ pgdat->first_deferred_pfn = epfn;
+ /*I update the patch based on your comment, it passed the test.
+ * Restore pending interrupts every 128Mb to give
+ * the chance tick timer to advance jiffies.
+ */
+ if (nr_pages > (1UL << 27 - PAGE_SHIFT)) {
+ pgdat_resize_unlock(pgdat, &flags);
+ goto again;
+ }
+ }
zone_empty:
+ pgdat->first_deferred_pfn = ULONG_MAX;
pgdat_resize_unlock(pgdat, &flags);
/* Sanity check that the next zone really is unpopulated */