Re: [PATCH v3] vmw_balloon: fixing double free when batching mode is off
From: Greg Kroah-Hartman
Date: Fri Jun 01 2018 - 04:00:04 EST
On Thu, May 31, 2018 at 08:56:52PM +0000, Nadav Amit wrote:
> Nadav Amit <namit@xxxxxxxxxx> wrote:
>
> > Nadav Amit <namit@xxxxxxxxxx> wrote:
> >
> >> Ping. Please consider it for inclusion for rc4.
> >>
> >> Nadav Amit <namit@xxxxxxxxxx> wrote:
> >>
> >>> From: Gil Kupfer <gilkup@xxxxxxxxx>
> >>>
> >>> The balloon.page field is used for two different purposes if batching is
> >>> on or off. If batching is on, the field point to the page which is used
> >>> to communicate with with the hypervisor. If it is off, balloon.page
> >>> points to the page that is about to be (un)locked.
> >>>
> >>> Unfortunately, this dual-purpose of the field introduced a bug: when the
> >>> balloon is popped (e.g., when the machine is reset or the balloon driver
> >>> is explicitly removed), the balloon driver frees, unconditionally, the
> >>> page that is held in balloon.page. As a result, if batching is
> >>> disabled, this leads to double freeing the last page that is sent to the
> >>> hypervisor.
> >>>
> >>> The following error occurs during rmmod when kernel checkers are on, and
> >>> the balloon is not empty:
> >>>
> >>> [ 42.307653] ------------[ cut here ]------------
> >>> [ 42.307657] Kernel BUG at ffffffffba1e4b28 [verbose debug info unavailable]
> >>> [ 42.307720] invalid opcode: 0000 [#1] SMP DEBUG_PAGEALLOC
> >>> [ 42.312512] Modules linked in: vmw_vsock_vmci_transport vsock ppdev joydev vmw_balloon(-) input_leds serio_raw vmw_vmci parport_pc shpchp parport i2c_piix4 nfit mac_hid autofs4 vmwgfx drm_kms_helper hid_generic syscopyarea sysfillrect usbhid sysimgblt fb_sys_fops hid ttm mptspi scsi_transport_spi ahci mptscsih drm psmouse vmxnet3 libahci mptbase pata_acpi
> >>> [ 42.312766] CPU: 10 PID: 1527 Comm: rmmod Not tainted 4.12.0+ #5
> >>> [ 42.312803] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 09/30/2016
> >>> [ 42.313042] task: ffff9bf9680f8000 task.stack: ffffbfefc1638000
> >>> [ 42.313290] RIP: 0010:__free_pages+0x38/0x40
> >>> [ 42.313510] RSP: 0018:ffffbfefc163be98 EFLAGS: 00010246
> >>> [ 42.313731] RAX: 000000000000003e RBX: ffffffffc02b9720 RCX: 0000000000000006
> >>> [ 42.313972] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff9bf97e08e0a0
> >>> [ 42.314201] RBP: ffffbfefc163be98 R08: 0000000000000000 R09: 0000000000000000
> >>> [ 42.314435] R10: 0000000000000000 R11: 0000000000000000 R12: ffffffffc02b97e4
> >>> [ 42.314505] R13: ffffffffc02b9748 R14: ffffffffc02b9728 R15: 0000000000000200
> >>> [ 42.314550] FS: 00007f3af5fec700(0000) GS:ffff9bf97e080000(0000) knlGS:0000000000000000
> >>> [ 42.314599] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> >>> [ 42.314635] CR2: 00007f44f6f4ab24 CR3: 00000003a7d12000 CR4: 00000000000006e0
> >>> [ 42.314864] Call Trace:
> >>> [ 42.315774] vmballoon_pop+0x102/0x130 [vmw_balloon]
> >>> [ 42.315816] vmballoon_exit+0x42/0xd64 [vmw_balloon]
> >>> [ 42.315853] SyS_delete_module+0x1e2/0x250
> >>> [ 42.315891] entry_SYSCALL_64_fastpath+0x23/0xc2
> >>> [ 42.315924] RIP: 0033:0x7f3af5b0e8e7
> >>> [ 42.315949] RSP: 002b:00007fffe6ce0148 EFLAGS: 00000206 ORIG_RAX: 00000000000000b0
> >>> [ 42.315996] RAX: ffffffffffffffda RBX: 000055be676401e0 RCX: 00007f3af5b0e8e7
> >>> [ 42.316951] RDX: 000000000000000a RSI: 0000000000000800 RDI: 000055be67640248
> >>> [ 42.317887] RBP: 0000000000000003 R08: 0000000000000000 R09: 1999999999999999
> >>> [ 42.318845] R10: 0000000000000883 R11: 0000000000000206 R12: 00007fffe6cdf130
> >>> [ 42.319755] R13: 0000000000000000 R14: 0000000000000000 R15: 000055be676401e0
> >>> [ 42.320606] Code: c0 74 1c f0 ff 4f 1c 74 02 5d c3 85 f6 74 07 e8 0f d8 ff ff 5d c3 31 f6 e8 c6 fb ff ff 5d c3 48 c7 c6 c8 0f c5 ba e8 58 be 02 00 <0f> 0b 66 0f 1f 44 00 00 66 66 66 66 90 48 85 ff 75 01 c3 55 48
> >>> [ 42.323462] RIP: __free_pages+0x38/0x40 RSP: ffffbfefc163be98
> >>> [ 42.325735] ---[ end trace 872e008e33f81508 ]---
> >>>
> >>> To solve the bug, we eliminate the dual purpose of balloon.page.
> >>>
> >>> Fixes: f220a80f0c2e ("VMware balloon: add batching to the vmw_balloon.")
> >>> Cc: stable@xxxxxxxxxxxxxxx
> >>> Reported-by: Oleksandr Natalenko <onatalen@xxxxxxxxxx>
> >>> Signed-off-by: Gil Kupfer <gilkup@xxxxxxxxx>
> >>> Signed-off-by: Nadav Amit <namit@xxxxxxxxxx>
> >>> Reviewed-by: Xavier Deguillard <xdeguillard@xxxxxxxxxx>
> >>> ---
> >>> drivers/misc/vmw_balloon.c | 23 +++++++----------------
> >>> 1 file changed, 7 insertions(+), 16 deletions(-)
> >>>
> >>> diff --git a/drivers/misc/vmw_balloon.c b/drivers/misc/vmw_balloon.c
> >>> index 9047c0a529b2..efd733472a35 100644
> >>> --- a/drivers/misc/vmw_balloon.c
> >>> +++ b/drivers/misc/vmw_balloon.c
> >>> @@ -576,15 +576,9 @@ static void vmballoon_pop(struct vmballoon *b)
> >>> }
> >>> }
> >>>
> >>> - if (b->batch_page) {
> >>> - vunmap(b->batch_page);
> >>> - b->batch_page = NULL;
> >>> - }
> >>> -
> >>> - if (b->page) {
> >>> - __free_page(b->page);
> >>> - b->page = NULL;
> >>> - }
> >>> + /* Clearing the batch_page unconditionally has no adverse effect */
> >>> + free_page((unsigned long)b->batch_page);
> >>> + b->batch_page = NULL;
> >>> }
> >>>
> >>> /*
> >>> @@ -991,16 +985,13 @@ static const struct vmballoon_ops vmballoon_batched_ops = {
> >>>
> >>> static bool vmballoon_init_batching(struct vmballoon *b)
> >>> {
> >>> - b->page = alloc_page(VMW_PAGE_ALLOC_NOSLEEP);
> >>> - if (!b->page)
> >>> - return false;
> >>> + struct page *page;
> >>>
> >>> - b->batch_page = vmap(&b->page, 1, VM_MAP, PAGE_KERNEL);
> >>> - if (!b->batch_page) {
> >>> - __free_page(b->page);
> >>> + page = alloc_page(GFP_KERNEL | __GFP_ZERO);
> >>> + if (!page)
> >>> return false;
> >>> - }
> >>>
> >>> + b->batch_page = page_address(page);
> >>> return true;
> >>> }
> >>>
> >>> --
> >>> 2.14.1
> >
> > Greg, can you please include this patch?? 4.17 is almost out of the door,
> > and the last version was sent a month ago.
> >
> > If you have any reservations, please let me know immediately.
>
> Arnd,
>
> Perhaps you can - the very least - respond (or just include the patch)?
>
> The last version of this patch was sent over a month ago.
It's too late for 4.17 now, and as this touches core code, I'll wait for
the next merge window cycle. It's also not even in my patch review
queue anymore, I guess the long "does this solve the problem" delays
that this went through pushed it out.
So I would need it resent no matter what.
thanks,
greg k-h