Re: KASAN: null-ptr-deref Write in io_uring_cancel_sqpoll

From: Pavel Begunkov
Date: Tue Apr 27 2021 - 07:20:31 EST


On 4/27/21 11:39 AM, Palash Oswal wrote:
> On Tue, Apr 27, 2021 at 2:07 PM Pavel Begunkov <asml.silence@xxxxxxxxx> wrote:
>>
>> io_sq_offload_create() {
>> ...
>> ret = io_uring_alloc_task_context(tsk, ctx);
>> wake_up_new_task(tsk);
>> if (ret)
>> goto err;
>> }
>>
>> Shouldn't happen unless offload create has failed. Just add
>> a return in *cancel_sqpoll() for this case. It's failing
>> so no requests has been submitted and no cancellation is needed.
>
> io_uring_cancel_sqpoll can be called by two flows:
> 1. io_uring_task_cancel() -> io_sqpoll_cancel_sync() ->
> io_uring_cancel_sqpoll ; which properly sanitises current->io_uring
> to be non NULL. (
> https://elixir.bootlin.com/linux/v5.12/source/include/linux/io_uring.h#L21
> )
> 2. io_sq_offload_create -> io_sq_thread -> io_uring_cancel_sqpoll ;
> which does not check the value of current->io_uring
>
> In the second flow,
> https://elixir.bootlin.com/linux/v5.12/source/fs/io_uring.c#L7970
> The initialization of current->io_uring (i.e
> io_uring_alloc_task_context() ) happens after calling io_sq_thread.
> And, therefore io_uring_cancel_sqpoll receives a NULL value for
> current->io_uring.

Right, exactly as I wrote in the previous message. And still placing
the check in io_uring_cancel_sqpoll() is a better (safer) option.

Just send a patch for 5.13 and mark it stable

>
> The backtrace from the crash confirms the second scenario:
> [ 70.661551] ==================================================================
> [ 70.662764] BUG: KASAN: null-ptr-deref in io_uring_cancel_sqpoll+0x203/0x350
> [ 70.663834] Write of size 4 at addr 0000000000000060 by task iou-sqp-750/755
> [ 70.664025]
> [ 70.664025] CPU: 1 PID: 755 Comm: iou-sqp-750 Not tainted 5.12.0 #101
> [ 70.664025] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996),
> BIOS 1.14.0-1 04/01/2014
> [ 70.664025] Call Trace:
> [ 70.664025] dump_stack+0xe9/0x168
> [ 70.664025] ? io_uring_cancel_sqpoll+0x203/0x350
> [ 70.664025] __kasan_report+0x166/0x1c0
> [ 70.664025] ? io_uring_cancel_sqpoll+0x203/0x350
> [ 70.664025] kasan_report+0x4f/0x70
> [ 70.664025] kasan_check_range+0x2f3/0x340
> [ 70.664025] __kasan_check_write+0x14/0x20
> [ 70.664025] io_uring_cancel_sqpoll+0x203/0x350
> [ 70.664025] ? io_sq_thread_unpark+0xd0/0xd0
> [ 70.664025] ? mutex_lock+0xbb/0x130
> [ 70.664025] ? init_wait_entry+0xe0/0xe0
> [ 70.664025] ? wait_for_completion_killable_timeout+0x20/0x20
> [ 70.664025] io_sq_thread+0x174c/0x18c0
> [ 70.664025] ? io_rsrc_put_work+0x380/0x380
> [ 70.664025] ? init_wait_entry+0xe0/0xe0
> [ 70.664025] ? _raw_spin_lock_irq+0xa5/0x180
> [ 70.664025] ? _raw_spin_lock_irqsave+0x190/0x190
> [ 70.664025] ? calculate_sigpending+0x6b/0xa0
> [ 70.664025] ? io_rsrc_put_work+0x380/0x380
> [ 70.664025] ret_from_fork+0x22/0x30
>
> We might want to add additional validation before calling
> io_uring_cancel_sqpoll. I did verify that the reproducer stopped
> producing the bug after the following change.
> ---
> diff --git a/fs/io_uring.c b/fs/io_uring.c
> index dff34975d86b..36fc9abe8022 100644
> --- a/fs/io_uring.c
> +++ b/fs/io_uring.c
> @@ -6832,8 +6832,10 @@ static int io_sq_thread(void *data)
> timeout = jiffies + sqd->sq_thread_idle;
> }
>
> - list_for_each_entry(ctx, &sqd->ctx_list, sqd_list)
> - io_uring_cancel_sqpoll(ctx);
> + list_for_each_entry(ctx, &sqd->ctx_list, sqd_list) {
> + if (current->io_uring)
> + io_uring_cancel_sqpoll(ctx);
> + }
> sqd->thread = NULL;
> list_for_each_entry(ctx, &sqd->ctx_list, sqd_list)
> io_ring_set_wakeup_flag(ctx);
>

--
Pavel Begunkov