[RFC PATCH 00/16] Core scheduling v6
From: Vineeth Remanan Pillai
Date: Tue Jun 30 2020 - 17:33:16 EST
Sixth iteration of the Core-Scheduling feature.
Core scheduling is a feature that allows only trusted tasks to run
concurrently on cpus sharing compute resources (eg: hyperthreads on a
core). The goal is to mitigate the core-level side-channel attacks
without requiring to disable SMT (which has a significant impact on
performance in some situations). Core scheduling (as of v6) mitigates
user-space to user-space attacks and user to kernel attack when one of
the siblings enters the kernel via interrupts. It is still possible to
have a task attack the sibling thread when it enters the kernel via
syscalls.
By default, the feature doesn't change any of the current scheduler
behavior. The user decides which tasks can run simultaneously on the
same core (for now by having them in the same tagged cgroup). When a
tag is enabled in a cgroup and a task from that cgroup is running on a
hardware thread, the scheduler ensures that only idle or trusted tasks
run on the other sibling(s). Besides security concerns, this feature
can also be beneficial for RT and performance applications where we
want to control how tasks make use of SMT dynamically.
This iteration is mostly a cleanup of v5 except for a major feature of
pausing sibling when a cpu enters kernel via nmi/irq/softirq. Also
introducing documentation and includes minor crash fixes.
One major cleanup was removing the hotplug support and related code.
The hotplug related crashes were not documented and the fixes piled up
over time leading to complex code. We were not able to reproduce the
crashes in the limited testing done. But if they are reroducable, we
don't want to hide them. We should document them and design better
fixes if any.
In terms of performance, the results in this release are similar to
v5. On a x86 system with N hardware threads:
- if only N/2 hardware threads are busy, the performance is similar
between baseline, corescheduling and nosmt
- if N hardware threads are busy with N different corescheduling
groups, the impact of corescheduling is similar to nosmt
- if N hardware threads are busy and multiple active threads share the
same corescheduling cookie, they gain a performance improvement over
nosmt.
The specific performance impact depends on the workload, but for a
really busy database 12-vcpu VM (1 coresched tag) running on a 36
hardware threads NUMA node with 96 mostly idle neighbor VMs (each in
their own coresched tag), the performance drops by 54% with
corescheduling and drops by 90% with nosmt.
v6 is rebased on 5.7.6(a06eb423367e)
https://github.com/digitalocean/linux-coresched/tree/coresched/v6-v5.7.y
Changes in v6
-------------
- Documentation
- Joel
- Pause siblings on entering nmi/irq/softirq
- Joel, Vineeth
- Fix for RCU crash
- Joel
- Fix for a crash in pick_next_task
- Yu Chen, Vineeth
- Minor re-write of core-wide vruntime comparison
- Aaron Lu
- Cleanup: Address Review comments
- Cleanup: Remove hotplug support (for now)
- Build fixes: 32 bit, SMT=n, AUTOGROUP=n etc
- Joel, Vineeth
Changes in v5
-------------
- Fixes for cgroup/process tagging during corner cases like cgroup
destroy, task moving across cgroups etc
- Tim Chen
- Coresched aware task migrations
- Aubrey Li
- Other minor stability fixes.
Changes in v4
-------------
- Implement a core wide min_vruntime for vruntime comparison of tasks
across cpus in a core.
- Aaron Lu
- Fixes a typo bug in setting the forced_idle cpu.
- Aaron Lu
Changes in v3
-------------
- Fixes the issue of sibling picking up an incompatible task
- Aaron Lu
- Vineeth Pillai
- Julien Desfossez
- Fixes the issue of starving threads due to forced idle
- Peter Zijlstra
- Fixes the refcounting issue when deleting a cgroup with tag
- Julien Desfossez
- Fixes a crash during cpu offline/online with coresched enabled
- Vineeth Pillai
- Fixes a comparison logic issue in sched_core_find
- Aaron Lu
Changes in v2
-------------
- Fixes for couple of NULL pointer dereference crashes
- Subhra Mazumdar
- Tim Chen
- Improves priority comparison logic for process in different cpus
- Peter Zijlstra
- Aaron Lu
- Fixes a hard lockup in rq locking
- Vineeth Pillai
- Julien Desfossez
- Fixes a performance issue seen on IO heavy workloads
- Vineeth Pillai
- Julien Desfossez
- Fix for 32bit build
- Aubrey Li
ISSUES
------
- Aaron(Intel) found an issue with load balancing when the tasks have
different weights(nice or cgroup shares). Task weight is not considered
in coresched aware load balancing and causes those higher weights task
to starve. This issue was in v5 as well and is carried over.
- Joel(ChromeOS) found an issue where RT task may be preempted by a
lower class task. This was also in v6 and is carried over.
- Coresched RB-tree doesn't get updated when task priority changes
- Potential starvation of untagged tasks(0 cookie) - a side effect of
0 cookie tasks not in the coresched RB-Tree
TODO
----
- MAJOR: Core wide vruntime comparison re-work
https://lwn.net/ml/linux-kernel/20200506143506.GH5298@xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx/
https://lwn.net/ml/linux-kernel/20200514130248.GD2940@xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx/
- MAJOR: Decide on the interfaces/API for exposing the feature to userland.
- prctl/set_schedattr
- A new coresched cgroup?
- fork/clone behaviour: whether child to inherit the cookie,
process vs thread behaviour etc
- Auto tagging based on user/session/process etc.
- procfs/sysfs interface to core scheduling.
- MAJOR: Load balancing/Migration fixes for core scheduling
With v6, Load balancing is partially coresched aware, but has some
issues w.r.t process/taskgroup weights:
https://lwn.net/ml/linux-kernel/20200225034438.GA617271@xxxxxxxxxxxxxxxxxxxxxxxxxxxx/
- System wide trusted cookie:
As of now, cookie value of 0 is considered special - default for tasks
that are not explicitly tagged. Kernel threads are not tagged by
default and hence tasks with 0 cookie is currently considered as a
system wide trusted trusted group. IRQ pause feature assumes this.
Discussions regarding this needs to continue and converge on a system
wide trusted cookie.
- Investigate the source of the overhead even when no tasks are tagged:
https://lkml.org/lkml/2019/10/29/242
- Core scheduling test framework: kselftests, torture tests etc
---
Aaron Lu (2):
sched/fair: wrapper for cfs_rq->min_vruntime
sched/fair: core wide cfs task priority comparison
Aubrey Li (1):
sched: migration changes for core scheduling
Chen Yu (1):
sched: Fix pick_next_task() race condition in core scheduling
Joel Fernandes (Google) (2):
irq: Add support for core-wide protection of IRQ and softirq
Documentation: Add documentation on core scheduling
Peter Zijlstra (9):
sched: Wrap rq::lock access
sched: Introduce sched_class::pick_task()
sched: Core-wide rq->lock
sched/fair: Add a few assertions
sched: Basic tracking of matching tasks
sched: Add core wide task selection and scheduling.
sched: Trivial forced-newidle balancer
sched: cgroup tagging interface for core scheduling
sched: Debug bits...
vpillai (1):
sched/fair: Fix forced idle sibling starvation corner case
Documentation/admin-guide/hw-vuln/core-scheduling.rst | 241 ++++
Documentation/admin-guide/hw-vuln/index.rst | 1 +
Documentation/admin-guide/kernel-parameters.txt | 9 +
include/linux/sched.h | 14 +-
kernel/Kconfig.preempt | 19 +
kernel/sched/core.c | 1076 ++++++++++++++++-
kernel/sched/cpuacct.c | 12 +-
kernel/sched/deadline.c | 34 +-
kernel/sched/debug.c | 4 +-
kernel/sched/fair.c | 409 +++++--
kernel/sched/idle.c | 13 +-
kernel/sched/pelt.h | 2 +-
kernel/sched/rt.c | 22 +-
kernel/sched/sched.h | 256 +++-
kernel/sched/stop_task.c | 13 +-
kernel/sched/topology.c | 4 +-
kernel/softirq.c | 46 +
17 files changed, 1953 insertions(+), 222 deletions(-)
create mode 100644 Documentation/admin-guide/hw-vuln/core-scheduling.rst
--
2.17.1