Because the swapping API doesn't adapt well to dynamic changes inEach page is either in frontswap OR on the normal swap device,I see. So why not implement this as an ordinary swap device, with a
never both. So, yes, both reads and writes are avoided if memory
is available and there is no write issued to the io subsystem if
memory is available. The is_memory_available decision is determined
by the hypervisor dynamically for each page when the guest attempts
a "frontswap_put". So, yes, you are indeed "swapping to the
hypervisor" but, at least in the case of Xen, the hypervisor
never swaps any memory to disk so there is never double swapping.
higher priority than the disk device? this way we reuse an API and
keep
things asynchronous, instead of introducing a special purpose API.
the size and availability of the underlying "swap" device, which
is very useful for swap to (bare-metal) hypervisor.
Doesn't this commit the hypervisor to retain this memory? If so, isn'tYes the hypervisor is committed to retain the memory. In
it simpler to give the page to the guest (so now it doesn't need to
swap at all)?
some ways, giving a page of memory to a guest (via ballooning)
is simpler and in some ways not. When a guest "owns" a page,
it can do whatever it wants with it, independent of what is best
for the "whole" virtualized system. When the hypervisor
"owns" the page on behalf of the guest but the guest can't
directly address it, the hypervisor has more flexibility.
For example, tmem optionally compresses all frontswap pages,
effectively doubling the size of its available memory.
In the future, knowing that a guest application can never
access the pages directly, it might store all frontswap pages in
(slower but still synchronous) phase change memory or "far NUMA"
memory.
I think it will be true in an overwhelming number of cases. Flash isI wasn't referring to hardware capability but to the availability
new enough that most devices support scatter/gather.
and timing constraints of the pages that need to be swapped.