Re: [RFC 0/8] Define coherent device memory node
From: Jerome Glisse
Date: Mon Oct 24 2016 - 13:09:27 EST
On Mon, Oct 24, 2016 at 10:01:49AM +0530, Anshuman Khandual wrote:
> [...]
> Core kernel memory features like reclamation, evictions etc. might
> need to be restricted or modified on the coherent device memory node as
> they can be performance limiting. The RFC does not propose anything on this
> yet but it can be looked into later on. For now it just disables Auto NUMA
> for any VMA which has coherent device memory.
>
> Seamless integration of coherent device memory with system memory
> will enable various other features, some of which can be listed as follows.
>
> a. Seamless migrations between system RAM and the coherent memory
> b. Will have asynchronous and high throughput migrations
> c. Be able to allocate huge order pages from these memory regions
> d. Restrict allocations to a large extent to the tasks using the
> device for workload acceleration
>
> Before concluding, will look into the reasons why the existing
> solutions don't work. There are two basic requirements which have to be
> satisfies before the coherent device memory can be integrated with core
> kernel seamlessly.
>
> a. PFN must have struct page
> b. Struct page must able to be inside standard LRU lists
>
> The above two basic requirements discard the existing method of
> device memory representation approaches like these which then requires the
> need of creating a new framework.
I do not believe the LRU list is a hard requirement, yes when faulting in
a page inside the page cache it assumes it needs to be added to lru list.
But i think this can easily be work around.
In HMM i am using ZONE_DEVICE and because memory is not accessible from CPU
(not everyone is bless with decent system bus like CAPI, CCIX, Gen-Z, ...)
so in my case a file back page must always be spawn first from a regular
page and once read from disk then i can migrate to GPU page.
So if you accept this intermediary step you can easily use ZONE_DEVICE for
device memory. This way no lru, no complex dance to make the memory out of
reach from regular memory allocator.
I think we would have much to gain if we pool our effort on a single common
solution for device memory. In my case the device memory is not accessible
by the CPU (because PCIE restrictions), in your case it is. Thus the only
difference is that in my case it can not be map inside the CPU page table
while in yours it can.
>
> (1) Traditional ioremap
>
> a. Memory is mapped into kernel (linear and virtual) and user space
> b. These PFNs do not have struct pages associated with it
> c. These special PFNs are marked with special flags inside the PTE
> d. Cannot participate in core VM functions much because of this
> e. Cannot do easy user space migrations
>
> (2) Zone ZONE_DEVICE
>
> a. Memory is mapped into kernel and user space
> b. PFNs do have struct pages associated with it
> c. These struct pages are allocated inside it's own memory range
> d. Unfortunately the struct page's union containing LRU has been
> used for struct dev_pagemap pointer
> e. Hence it cannot be part of any LRU (like Page cache)
> f. Hence file cached mapping cannot reside on these PFNs
> g. Cannot do easy migrations
>
> I had also explored non LRU representation of this coherent device
> memory where the integration with system RAM in the core VM is limited only
> to the following functions. Not being inside LRU is definitely going to
> reduce the scope of tight integration with system RAM.
>
> (1) Migration support between system RAM and coherent memory
> (2) Migration support between various coherent memory nodes
> (3) Isolation of the coherent memory
> (4) Mapping the coherent memory into user space through driver's
> struct vm_operations
> (5) HW poisoning of the coherent memory
>
> Allocating the entire memory of the coherent device node right
> after hot plug into ZONE_MOVABLE (where the memory is already inside the
> buddy system) will still expose a time window where other user space
> allocations can come into the coherent device memory node and prevent the
> intended isolation. So traditional hot plug is not the solution. Hence
> started looking into CMA based non LRU solution but then hit the following
> roadblocks.
>
> (1) CMA does not support hot plugging of new memory node
> a. CMA area needs to be marked during boot before buddy is
> initialized
> b. cma_alloc()/cma_release() can happen on the marked area
> c. Should be able to mark the CMA areas just after memory hot plug
> d. cma_alloc()/cma_release() can happen later after the hot plug
> e. This is not currently supported right now
>
> (2) Mapped non LRU migration of pages
> a. Recent work from Michan Kim makes non LRU page migratable
> b. But it still does not support migration of mapped non LRU pages
> c. With non LRU CMA reserved, again there are some additional
> challenges
>
> With hot pluggable CMA and non LRU mapped migration support there
> may be an alternate approach to represent coherent device memory. Please
> do review this RFC proposal and let me know your comments or suggestions.
> Thank you.
You can take a look at hmm-v13 if you want to see how i do non LRU page
migration. While i put most of the migration code inside hmm_migrate.c it
could easily be move to migrate.c without hmm_ prefix.
There is 2 missing piece with existing migrate code. First is to put memory
allocation for destination under control of who call the migrate code. Second
is to allow offloading the copy operation to device (ie not use the CPU to
copy data).
I believe same requirement also make sense for platform you are targeting.
Thus same code can be use.
hmm-v13 https://cgit.freedesktop.org/~glisse/linux/log/?h=hmm-v13
I haven't posted this patchset yet because we are doing some modifications
to the device driver API to accomodate some new features. But the ZONE_DEVICE
changes and the overall migration code will stay the same more or less (i have
patches that move it to migrate.c and share more code with existing migrate
code).
If you think i missed anything about lru and page cache please point it to
me. Because when i audited code for that i didn't see any road block with
the few fs i was looking at (ext4, xfs and core page cache code).
> [...]
Cheers,
Jérôme