Re: [PATCH] cpuset: Make rebuild_sched_domains() usable from anycontext (take 2)
From: Max Krasnyansky
Date: Wed Jul 16 2008 - 14:27:53 EST
Paul Menage wrote:
> On Tue, Jul 15, 2008 at 11:21 PM, Max Krasnyansky <maxk@xxxxxxxxxxxx> wrote:
>> From: Max Krasnyanskiy <maxk@xxxxxxxxxxxx>
>>
>> Basically as Paul J. pointed out rebuild_sched_domains() is the
>> only way to rebuild sched domains correctly based on the current
>> cpuset settings. What this means is that we need to be able to
>> call it from different contexts, like cpuhotplug for example.
>> Also if you look at the cpu_active_map patches, the scheduler now
>> calls rebuild_sched_domains() directly from places like
>> arch_init_sched_domains().
>>
>> In order to support that properly we need to change cpuset locking
>> a bit, to avoid circular locking dependencies.
>> This patch makes rebuild_sched_domains() callable from any context.
>> The only requirement now is the the caller has to hold cpu_hotplug.lock
>> (ie get_online_cpus()).
>
> Calling get_online_cpus() doesn't technically result in you holding
> cpu_hotplug.lock - it ensures that either you're the active
> cpu_hotplug writer or else no-one else holds cpu_hotplug.lock and
> cpu_hotplug.refcount > 0. Can you specify this more precisely? Maybe
> say "the caller has to hold cpu_hotplug for read or write"?
Very good point.
How about we just say
"has to be called under get_online_cpus()"
I think that reflects what the calling requirement is without diving into
to much detail as to how get_online_cpus() works internally.
> A useful
> patch might be to rename "struct {} cpu_hotplug" to "struct {}
> cpu_hotplug_recursive_rw_mutex", since that's exactly what it is.
Hmm, I guess the idea was that struct cpu_hotplug would hold more stuff
then just the lock. But maybe I'm wrong.
> Then this patch could say "caller has to hold
> cpu_hotplug_recursive_rw_mutex". Yes, it's a bit ugly, but at least it
> exposes it properly.
I think at this level we should just say
"call me inside get_online_cpus() ... put_online_cpus()"
and leave the details. I mean we should just stop using "hold
cpu_hotplug.lock" notation because it's something internal. All we really care
about is that we called under get_online_cpus() to make sure that cpus are not
unplugged at the wrong time.
>> + /* We have to iterate cgroup hierarchy, make sure nobody is messing
>> + * with it. */
>> + cgroup_lock();
>> +
>> /* Special case for the 99% of systems with one, full, sched domain */
>> if (is_sched_load_balance(&top_cpuset)) {
>> ndoms = 1;
>> @@ -598,10 +606,10 @@ void rebuild_sched_domains(void)
>>
>> q = kfifo_alloc(number_of_cpusets * sizeof(cp), GFP_KERNEL, NULL);
>> if (IS_ERR(q))
>> - goto done;
>> + goto unlock;
>> csa = kmalloc(number_of_cpusets * sizeof(cp), GFP_KERNEL);
>> if (!csa)
>> - goto done;
>> + goto unlock;
>> csn = 0;
>>
>> cp = &top_cpuset;
>> @@ -688,10 +696,16 @@ restart:
>> BUG_ON(nslot != ndoms);
>>
>> rebuild:
>> + /* Drop cgroup lock before calling the scheduler.
>> + * This is not strictly necesseary but simplifies lock nesting. */
>> + cgroup_unlock();
>> +
>> /* Have scheduler rebuild sched domains */
>> - get_online_cpus();
>> partition_sched_domains(ndoms, doms, dattr);
>> - put_online_cpus();
>> + goto done;
>> +
>> +unlock:
>> + cgroup_unlock();
>
> This goto ordering's a bit ugly. rebuild_sched_domains() is ripe for a
> refactoring, but I guess that can wait.
Yeah. I tried a couple of option and settled on the current one for now.
If we did not have to drop cgroup_lock() before calling the scheduler some of
the gotos would go away but I'd rather drop it to avoid more nesting in the
future.
Sounds like you're ok with the current version logic wise. I'll take another
cleanup pass some time late.
>> +/*
>> + * Internal helper for rebuilding sched domains when something changes.
>> + * rebuild_sched_domains() must be called under get_online_cpus() and
>> + * it needs to take cgroup_lock(). But most of the cpuset code is already
>> + * holding cgroup_lock() while calling __rebuild_sched_domains().
>> + * In order to avoid incorrect lock nesting we delegate the actual domain
>> + * rebuilding to the workqueue.
>> + */
>> +static void __rebuild_sched_domains(void)
>> +{
>> + queue_work(cpuset_wq, &rebuild_domains_work);
>> +}
>> +
>
> The __ prefix is normally used to indicate a lower-level or pre-locked
> version of a function. In this case it's a higher-level function so I
> don't think that prefix is appropriate. How about
> async_rebuild_sched_domains() ?
Yep. Much better name.
>> -static void common_cpu_mem_hotplug_unplug(int rebuild_sd)
>> +static void common_cpu_mem_hotplug_unplug(void)
>> {
>> cgroup_lock();
>>
>> @@ -1902,13 +1934,6 @@ static void common_cpu_mem_hotplug_unplug(int rebuild_sd)
>> top_cpuset.mems_allowed = node_states[N_HIGH_MEMORY];
>> scan_for_empty_cpusets(&top_cpuset);
>
> This is still unsafe because it accesses cpu_online_map without
> get_online_cpus() (when called as the memory hotplug handler)?
>
> Maybe scan_for_empty_cpusets() should take a parameter indicating
> whether we're interested in cpus or mems?
Ah, I thought I separated them out, but you're right it still accesses
cpu_online_map from mem hotplug events. I think instead of the argument
we can just split them up completely. The common part is
scan_for_empty_cpusets() the rest is not. I'll fix that.
>> hotcpu_notifier(cpuset_handle_cpuhp, 0);
>> +
>> + cpuset_wq = create_singlethread_workqueue("cpuset");
>> + BUG_ON(!cpuset_wq);
>
> Seems a bit of a shame to waste a kthread on this. Is there no generic
> single-threaded workqueue that we could piggyback on? Maybe create one
> in workqueue.c that can be used by anyone who specifically needs a
> singlethreaded workqueue for occasional work?
Agree
$ git grep create_singlethread_workqueue | wc
104 412 9037
Granted, it's unlikely that a single machine would have all that stuff enabled
but 104 kthreads just for workqueues does seems like an overkill.
I was going to take a pass at the workqueue users and replaces
flush_scheduled_work() with flush_work()/cancel_work_sync() so I might take a
crack at this too.
Max
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/