After updating the lru drain sequence, new comers avoid waiting for
the current drainer, because he is flushing works on each online CPU,
by trying to lock the mutex; the drainer OTOH tries to do works for
those who fail to acquire the lock by checking the lru drain sequence
after releasing lock.
See eef1a429f234 ("mm/swap.c: piggyback lru_add_drain_all() calls")
for reasons why we can skip waiting for the lock.
The memory barriers around the sequence and the lock come together
to remove waiters without their drain works bandoned.
Cc: Sebastian Andrzej Siewior <bigeasy@xxxxxxxxxxxxx>
Cc: Konstantin Khlebnikov <khlebnikov@xxxxxxxxxxxxxx>
Signed-off-by: Hillf Danton <hdanton@xxxxxxxx>
---
This is inspired by one of the works from Sebastian.
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -714,10 +714,11 @@ static void lru_add_drain_per_cpu(struct
*/
void lru_add_drain_all(void)
{
- static seqcount_t seqcount = SEQCNT_ZERO(seqcount);
+ static unsigned int lru_drain_seq;
static DEFINE_MUTEX(lock);
static struct cpumask has_work;
- int cpu, seq;
+ int cpu;
+ unsigned int seq;
/*
* Make sure nobody triggers this path before mm_percpu_wq is fully
@@ -726,18 +727,16 @@ void lru_add_drain_all(void)
if (WARN_ON(!mm_percpu_wq))
return;
- seq = raw_read_seqcount_latch(&seqcount);
+ lru_drain_seq++;
+ smp_mb();
- mutex_lock(&lock);
+more_work:
- /*
- * Piggyback on drain started and finished while we waited for lock:
- * all pages pended at the time of our enter were drained from vectors.
- */
- if (__read_seqcount_retry(&seqcount, seq))
- goto done;
+ if (!mutex_trylock(&lock))
+ return;
- raw_write_seqcount_latch(&seqcount);
+ smp_mb();
+ seq = lru_drain_seq;
cpumask_clear(&has_work);
@@ -759,8 +758,11 @@ void lru_add_drain_all(void)
for_each_cpu(cpu, &has_work)
flush_work(&per_cpu(lru_add_drain_work, cpu));
-done:
mutex_unlock(&lock);
+
+ smp_mb();
+ if (seq != lru_drain_seq)
+ goto more_work;
}
#else
void lru_add_drain_all(void)
--