On 08/13/2018 09:29 PM, Austin S. Hemmelgarn wrote:The easy short term solution is to just disallow the replace command (with the intent of getting it working in the future), but ensure that the older style add/remove method works. That uses the balance code internally, so it should honor any restrictions on block placement for the new device, and therefore should be pretty easy to get working.
On 2018-08-13 15:20, Hannes Reinecke wrote:[ .. ]
On 08/13/2018 08:42 PM, David Sterba wrote:
On Fri, Aug 10, 2018 at 03:04:33AM +0900, Naohiro Aota wrote:
This series adds zoned block device support to btrfs.
Yay, thanks!
Device replace is implemented in largely the same manner as most otherDevice replace is disabled, but the changlog suggests there's a way toHow would a device replace work in general?
make it work, so it's a matter of implementation. And this should be
implemented at the time of merge.
While I do understand that device replace is possible with RAID
thingies, I somewhat fail to see how could do a device replacement
without RAID functionality.
Is it even possible?
If so, how would it be different from a simple umount?
live data migration tools (for example, LVM2's pvmove command).
In short, when you issue a replace command for a given device, all
writes that would go to that device are instead sent to the new device.
While this is happening, old data is copied over from the old device to
the new one. Once all the data is copied, the old device is released
(and it's BTRFS signature wiped), and the new device has it's device ID
updated to that of the old device.
This is possible largely because of the COW infrastructure, but it's
implemented in a way that doesn't entirely depend on it (otherwise it
wouldn't work for NOCOW files).
Handling this on zoned devices is not likely to be easy though, you
would functionally have to freeze I/O that would hit the device being
replaced so that you don't accidentally write to a sequential zone out
of order.
Ah. Oh. Hmm.
It would be possible in principle if we freeze accesses to any partially
filled zones on the original device. Then all new writes will be going
into new/empty zones on the new disks, and we can copy over the old data
with no issue at all.
We end up with some partially filled zones on the new disk, but they
really should be cleaned up eventually either by the allocator filling
up the partially filled zones or once garbage collection clears out
stale zones.
However, I fear the required changes to the btrfs allocator are beyond
my btrfs knowledge :-(