RE: [PATCH V3 08/13] HV/Vmbus: Initialize VMbus ring buffer for Isolation VM
From: Michael Kelley
Date: Mon Aug 16 2021 - 13:28:48 EST
From: Tianyu Lan <ltykernel@xxxxxxxxx> Sent: Monday, August 9, 2021 10:56 AM
>
> VMbus ring buffer are shared with host and it's need to
s/it's need/it needs/
> be accessed via extra address space of Isolation VM with
> SNP support. This patch is to map the ring buffer
> address in extra address space via ioremap(). HV host
It's actually using vmap_pfn(), not ioremap().
> visibility hvcall smears data in the ring buffer and
> so reset the ring buffer memory to zero after calling
> visibility hvcall.
>
> Signed-off-by: Tianyu Lan <Tianyu.Lan@xxxxxxxxxxxxx>
> ---
> drivers/hv/Kconfig | 1 +
> drivers/hv/channel.c | 10 +++++
> drivers/hv/hyperv_vmbus.h | 2 +
> drivers/hv/ring_buffer.c | 84 ++++++++++++++++++++++++++++++---------
> 4 files changed, 79 insertions(+), 18 deletions(-)
>
> diff --git a/drivers/hv/Kconfig b/drivers/hv/Kconfig
> index d1123ceb38f3..dd12af20e467 100644
> --- a/drivers/hv/Kconfig
> +++ b/drivers/hv/Kconfig
> @@ -8,6 +8,7 @@ config HYPERV
> || (ARM64 && !CPU_BIG_ENDIAN))
> select PARAVIRT
> select X86_HV_CALLBACK_VECTOR if X86
> + select VMAP_PFN
> help
> Select this option to run Linux as a Hyper-V client operating
> system.
> diff --git a/drivers/hv/channel.c b/drivers/hv/channel.c
> index 4c4717c26240..60ef881a700c 100644
> --- a/drivers/hv/channel.c
> +++ b/drivers/hv/channel.c
> @@ -712,6 +712,16 @@ static int __vmbus_open(struct vmbus_channel *newchannel,
> if (err)
> goto error_clean_ring;
>
> + err = hv_ringbuffer_post_init(&newchannel->outbound,
> + page, send_pages);
> + if (err)
> + goto error_free_gpadl;
> +
> + err = hv_ringbuffer_post_init(&newchannel->inbound,
> + &page[send_pages], recv_pages);
> + if (err)
> + goto error_free_gpadl;
> +
> /* Create and init the channel open message */
> open_info = kzalloc(sizeof(*open_info) +
> sizeof(struct vmbus_channel_open_channel),
> diff --git a/drivers/hv/hyperv_vmbus.h b/drivers/hv/hyperv_vmbus.h
> index 40bc0eff6665..15cd23a561f3 100644
> --- a/drivers/hv/hyperv_vmbus.h
> +++ b/drivers/hv/hyperv_vmbus.h
> @@ -172,6 +172,8 @@ extern int hv_synic_cleanup(unsigned int cpu);
> /* Interface */
>
> void hv_ringbuffer_pre_init(struct vmbus_channel *channel);
> +int hv_ringbuffer_post_init(struct hv_ring_buffer_info *ring_info,
> + struct page *pages, u32 page_cnt);
>
> int hv_ringbuffer_init(struct hv_ring_buffer_info *ring_info,
> struct page *pages, u32 pagecnt, u32 max_pkt_size);
> diff --git a/drivers/hv/ring_buffer.c b/drivers/hv/ring_buffer.c
> index 2aee356840a2..d4f93fca1108 100644
> --- a/drivers/hv/ring_buffer.c
> +++ b/drivers/hv/ring_buffer.c
> @@ -17,6 +17,8 @@
> #include <linux/vmalloc.h>
> #include <linux/slab.h>
> #include <linux/prefetch.h>
> +#include <linux/io.h>
> +#include <asm/mshyperv.h>
>
> #include "hyperv_vmbus.h"
>
> @@ -179,43 +181,89 @@ void hv_ringbuffer_pre_init(struct vmbus_channel *channel)
> mutex_init(&channel->outbound.ring_buffer_mutex);
> }
>
> -/* Initialize the ring buffer. */
> -int hv_ringbuffer_init(struct hv_ring_buffer_info *ring_info,
> - struct page *pages, u32 page_cnt, u32 max_pkt_size)
> +int hv_ringbuffer_post_init(struct hv_ring_buffer_info *ring_info,
> + struct page *pages, u32 page_cnt)
> {
> + u64 physic_addr = page_to_pfn(pages) << PAGE_SHIFT;
> + unsigned long *pfns_wraparound;
> + void *vaddr;
> int i;
> - struct page **pages_wraparound;
>
> - BUILD_BUG_ON((sizeof(struct hv_ring_buffer) != PAGE_SIZE));
> + if (!hv_isolation_type_snp())
> + return 0;
> +
> + physic_addr += ms_hyperv.shared_gpa_boundary;
>
> /*
> * First page holds struct hv_ring_buffer, do wraparound mapping for
> * the rest.
> */
> - pages_wraparound = kcalloc(page_cnt * 2 - 1, sizeof(struct page *),
> + pfns_wraparound = kcalloc(page_cnt * 2 - 1, sizeof(unsigned long),
> GFP_KERNEL);
> - if (!pages_wraparound)
> + if (!pfns_wraparound)
> return -ENOMEM;
>
> - pages_wraparound[0] = pages;
> + pfns_wraparound[0] = physic_addr >> PAGE_SHIFT;
> for (i = 0; i < 2 * (page_cnt - 1); i++)
> - pages_wraparound[i + 1] = &pages[i % (page_cnt - 1) + 1];
> -
> - ring_info->ring_buffer = (struct hv_ring_buffer *)
> - vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, PAGE_KERNEL);
> -
> - kfree(pages_wraparound);
> + pfns_wraparound[i + 1] = (physic_addr >> PAGE_SHIFT) +
> + i % (page_cnt - 1) + 1;
>
> -
> - if (!ring_info->ring_buffer)
> + vaddr = vmap_pfn(pfns_wraparound, page_cnt * 2 - 1, PAGE_KERNEL_IO);
> + kfree(pfns_wraparound);
> + if (!vaddr)
> return -ENOMEM;
>
> - ring_info->ring_buffer->read_index =
> - ring_info->ring_buffer->write_index = 0;
> + /* Clean memory after setting host visibility. */
> + memset((void *)vaddr, 0x00, page_cnt * PAGE_SIZE);
> +
> + ring_info->ring_buffer = (struct hv_ring_buffer *)vaddr;
> + ring_info->ring_buffer->read_index = 0;
> + ring_info->ring_buffer->write_index = 0;
>
> /* Set the feature bit for enabling flow control. */
> ring_info->ring_buffer->feature_bits.value = 1;
>
> + return 0;
> +}
> +
> +/* Initialize the ring buffer. */
> +int hv_ringbuffer_init(struct hv_ring_buffer_info *ring_info,
> + struct page *pages, u32 page_cnt, u32 max_pkt_size)
> +{
> + int i;
> + struct page **pages_wraparound;
> +
> + BUILD_BUG_ON((sizeof(struct hv_ring_buffer) != PAGE_SIZE));
> +
> + if (!hv_isolation_type_snp()) {
> + /*
> + * First page holds struct hv_ring_buffer, do wraparound mapping for
> + * the rest.
> + */
> + pages_wraparound = kcalloc(page_cnt * 2 - 1, sizeof(struct page *),
> + GFP_KERNEL);
> + if (!pages_wraparound)
> + return -ENOMEM;
> +
> + pages_wraparound[0] = pages;
> + for (i = 0; i < 2 * (page_cnt - 1); i++)
> + pages_wraparound[i + 1] = &pages[i % (page_cnt - 1) + 1];
> +
> + ring_info->ring_buffer = (struct hv_ring_buffer *)
> + vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, PAGE_KERNEL);
> +
> + kfree(pages_wraparound);
> +
> + if (!ring_info->ring_buffer)
> + return -ENOMEM;
> +
> + ring_info->ring_buffer->read_index =
> + ring_info->ring_buffer->write_index = 0;
> +
> + /* Set the feature bit for enabling flow control. */
> + ring_info->ring_buffer->feature_bits.value = 1;
> + }
> +
> ring_info->ring_size = page_cnt << PAGE_SHIFT;
> ring_info->ring_size_div10_reciprocal =
> reciprocal_value(ring_info->ring_size / 10);
> --
> 2.25.1
This patch does the following:
1) The existing ring buffer wrap-around mapping functionality is still
executed in hv_ringbuffer_init() when not doing SNP isolation.
This mapping is based on an array of struct page's that describe the
contiguous physical memory.
2) New ring buffer wrap-around mapping functionality is added in
hv_ringbuffer_post_init() for the SNP isolation case. The case is
handled in hv_ringbuffer_post_init() because it must be done after
the GPADL is established, since that's where the host visibility
is set. What's interesting is that this case is exactly the same
as #1 above, except that the mapping is based on physical
memory addresses instead of struct page's. We have to use physical
addresses because of applying the GPA boundary, and there are no
struct page's for those physical addresses.
Unfortunately, this duplicates a lot of logic in #1 and #2, except
for the struct page vs. physical address difference.
Proposal: Couldn't we always do #2, even for the normal case
where SNP isolation is not being used? The difference would
only be in whether the GPA boundary is added. And it looks like
the normal case could be done after the GPADL is established,
as setting up the GPADL doesn't have any dependencies on
having the ring buffer mapped. This approach would remove
a lot of duplication. Just move the calls to hv_ringbuffer_init()
to after the GPADL is established, and do all the work there for
both cases.
Michael