[PATCH 2/2] mm: swap_cgroup: get rid of __lookup_swap_cgroup()
From: Roman Gushchin
Date: Fri Nov 15 2024 - 14:10:26 EST
Because swap_cgroup map is now virtually contiguous,
swap_cgroup_record() can be simplified, which eliminates
a need to use __lookup_swap_cgroup().
Now as __lookup_swap_cgroup() is really trivial and is used only once,
it can be inlined.
Signed-off-by: Roman Gushchin <roman.gushchin@xxxxxxxxx>
---
mm/swap_cgroup.c | 18 ++----------------
1 file changed, 2 insertions(+), 16 deletions(-)
diff --git a/mm/swap_cgroup.c b/mm/swap_cgroup.c
index 18de498c84a4..0db907308c94 100644
--- a/mm/swap_cgroup.c
+++ b/mm/swap_cgroup.c
@@ -33,13 +33,6 @@ static struct swap_cgroup_ctrl swap_cgroup_ctrl[MAX_SWAPFILES];
*
* TODO: we can push these buffers out to HIGHMEM.
*/
-
-static struct swap_cgroup *__lookup_swap_cgroup(struct swap_cgroup_ctrl *ctrl,
- pgoff_t offset)
-{
- return &ctrl->map[offset];
-}
-
static struct swap_cgroup *lookup_swap_cgroup(swp_entry_t ent,
struct swap_cgroup_ctrl **ctrlp)
{
@@ -49,7 +42,7 @@ static struct swap_cgroup *lookup_swap_cgroup(swp_entry_t ent,
ctrl = &swap_cgroup_ctrl[swp_type(ent)];
if (ctrlp)
*ctrlp = ctrl;
- return __lookup_swap_cgroup(ctrl, offset);
+ return &ctrl->map[offset];
}
/**
@@ -104,16 +97,9 @@ unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id,
spin_lock_irqsave(&ctrl->lock, flags);
old = sc->id;
- for (;;) {
+ for (; offset < end; offset++, sc++) {
VM_BUG_ON(sc->id != old);
sc->id = id;
- offset++;
- if (offset == end)
- break;
- if (offset % SC_PER_PAGE)
- sc++;
- else
- sc = __lookup_swap_cgroup(ctrl, offset);
}
spin_unlock_irqrestore(&ctrl->lock, flags);
--
2.47.0.338.g60cca15819-goog