Re: [PATCH v2 2/2] x86/xen: allow privcmd hypercalls to be preempted
From: Luis R. Rodriguez
Date: Wed Dec 10 2014 - 20:03:51 EST
On Wed, Dec 10, 2014 at 04:29:06PM -0800, H. Peter Anvin wrote:
> On 12/10/2014 03:34 PM, Luis R. Rodriguez wrote:
> > diff --git a/arch/x86/kernel/entry_32.S b/arch/x86/kernel/entry_32.S
> > index 344b63f..40b5c0c 100644
> > --- a/arch/x86/kernel/entry_32.S
> > +++ b/arch/x86/kernel/entry_32.S
> > @@ -982,7 +982,28 @@ ENTRY(xen_hypervisor_callback)
> > ENTRY(xen_do_upcall)
> > 1: mov %esp, %eax
> > call xen_evtchn_do_upcall
> > +#ifdef CONFIG_PREEMPT
> > jmp ret_from_intr
> > +#else
> > + GET_THREAD_INFO(%ebp)
> > +#ifdef CONFIG_VM86
> > + movl PT_EFLAGS(%esp), %eax # mix EFLAGS and CS
> > + movb PT_CS(%esp), %al
> > + andl $(X86_EFLAGS_VM | SEGMENT_RPL_MASK), %eax
> > +#else
> > + movl PT_CS(%esp), %eax
> > + andl $SEGMENT_RPL_MASK, %eax
> > +#endif
> > + cmpl $USER_RPL, %eax
> > + jae resume_userspace # returning to v8086 or userspace
> > + DISABLE_INTERRUPTS(CLBR_ANY)
> > + cmpb $0,PER_CPU_VAR(xen_in_preemptible_hcall)
> > + jz resume_kernel
> > + movb $0,PER_CPU_VAR(xen_in_preemptible_hcall)
> > + call cond_resched_irq
> > + movb $1,PER_CPU_VAR(xen_in_preemptible_hcall)
> > + jmp resume_kernel
> > +#endif /* CONFIG_PREEMPT */
> > CFI_ENDPROC
> > ENDPROC(xen_hypervisor_callback)
> >
> > diff --git a/arch/x86/kernel/entry_64.S b/arch/x86/kernel/entry_64.S
> > index c0226ab..0ccdd06 100644
> > --- a/arch/x86/kernel/entry_64.S
> > +++ b/arch/x86/kernel/entry_64.S
> > @@ -1170,7 +1170,23 @@ ENTRY(xen_do_hypervisor_callback) # do_hypervisor_callback(struct *pt_regs)
> > popq %rsp
> > CFI_DEF_CFA_REGISTER rsp
> > decl PER_CPU_VAR(irq_count)
> > +#ifdef CONFIG_PREEMPT
> > jmp error_exit
> > +#else
> > + movl %ebx, %eax
> > + RESTORE_REST
> > + DISABLE_INTERRUPTS(CLBR_NONE)
> > + TRACE_IRQS_OFF
> > + GET_THREAD_INFO(%rcx)
> > + testl %eax, %eax
> > + je error_exit_user
> > + cmpb $0,PER_CPU_VAR(xen_in_preemptible_hcall)
> > + jz retint_kernel
> > + movb $0,PER_CPU_VAR(xen_in_preemptible_hcall)
> > + call cond_resched_irq
> > + movb $1,PER_CPU_VAR(xen_in_preemptible_hcall)
> > + jmp retint_kernel
> > +#endif /* CONFIG_PREEMPT */
> > CFI_ENDPROC
> > END(xen_do_hypervisor_callback)
> >
> > @@ -1398,6 +1414,7 @@ ENTRY(error_exit)
> > GET_THREAD_INFO(%rcx)
> > testl %eax,%eax
> > jne retint_kernel
> > +error_exit_user:
> > LOCKDEP_SYS_EXIT_IRQ
> > movl TI_flags(%rcx),%edx
> > movl $_TIF_WORK_MASK,%edi
>
> You're adding a bunch of code for the *non*-preemptive case here... why?
This is an issue onloy for for non*-preemptive kernels.
Some of Xen's hypercalls can take a long time and unfortunately for
*non*-preemptive kernels this can be quite a bit of an issue.
We've handled situations like this with cond_resched() before which will
push even *non*-preemptive kernels to behave as voluntarily preemptive,
I was not aware to what extent this was done and precedents set but
its pretety widespread now... this then just addresses once particular
case where this is also an issuefor but now in IRQ context.
I agree its a hack but so are all the other cond_reshed() calls then.
I don't think its a good idea to be spreading use of something like
this everywhere but after careful review and trying toa void this
exact code for a while I have not been able to find any other reasonable
alternative.
Luis
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/