Re: [PATCH 2/4] lightnvm: move metadata mapping to lower level driver

From: Javier GonzÃlez
Date: Thu Aug 01 2019 - 03:46:08 EST


> On 1 Aug 2019, at 09.15, Hans Holmberg <hans@xxxxxxxxxxxxx> wrote:
>
> On Wed, Jul 31, 2019 at 3:59 PM Javier GonzÃlez <javier@xxxxxxxxxxx> wrote:
>>> On 31 Jul 2019, at 11.41, Hans Holmberg <hans@xxxxxxxxxxxxx> wrote:
>>>
>>> Now that blk_rq_map_kern can map both kmem and vmem, move
>>> internal metadata mapping down to the lower level driver.
>>>
>>> Signed-off-by: Hans Holmberg <hans@xxxxxxxxxxxxx>
>>> ---
>>> drivers/lightnvm/core.c | 16 +++---
>>> drivers/lightnvm/pblk-core.c | 113 +++++----------------------------------
>>> drivers/lightnvm/pblk-read.c | 22 ++------
>>> drivers/lightnvm/pblk-recovery.c | 39 ++------------
>>> drivers/lightnvm/pblk-write.c | 20 ++-----
>>> drivers/lightnvm/pblk.h | 8 +--
>>> drivers/nvme/host/lightnvm.c | 20 +++++--
>>> include/linux/lightnvm.h | 6 +--
>>> 8 files changed, 54 insertions(+), 190 deletions(-)
>>>
>>> diff --git a/drivers/lightnvm/core.c b/drivers/lightnvm/core.c
>>> index 01d098fb96ac..3cd03582a2ed 100644
>>> --- a/drivers/lightnvm/core.c
>>> +++ b/drivers/lightnvm/core.c
>>> @@ -731,7 +731,7 @@ static int nvm_set_flags(struct nvm_geo *geo, struct nvm_rq *rqd)
>>> return flags;
>>> }
>>>
>>> -int nvm_submit_io(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd)
>>> +int nvm_submit_io(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd, void *buf)
>>> {
>>> struct nvm_dev *dev = tgt_dev->parent;
>>> int ret;
>>> @@ -745,7 +745,7 @@ int nvm_submit_io(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd)
>>> rqd->flags = nvm_set_flags(&tgt_dev->geo, rqd);
>>>
>>> /* In case of error, fail with right address format */
>>> - ret = dev->ops->submit_io(dev, rqd);
>>> + ret = dev->ops->submit_io(dev, rqd, buf);
>>> if (ret)
>>> nvm_rq_dev_to_tgt(tgt_dev, rqd);
>>> return ret;
>>> @@ -759,7 +759,8 @@ static void nvm_sync_end_io(struct nvm_rq *rqd)
>>> complete(waiting);
>>> }
>>>
>>> -static int nvm_submit_io_wait(struct nvm_dev *dev, struct nvm_rq *rqd)
>>> +static int nvm_submit_io_wait(struct nvm_dev *dev, struct nvm_rq *rqd,
>>> + void *buf)
>>> {
>>> DECLARE_COMPLETION_ONSTACK(wait);
>>> int ret = 0;
>>> @@ -767,7 +768,7 @@ static int nvm_submit_io_wait(struct nvm_dev *dev, struct nvm_rq *rqd)
>>> rqd->end_io = nvm_sync_end_io;
>>> rqd->private = &wait;
>>>
>>> - ret = dev->ops->submit_io(dev, rqd);
>>> + ret = dev->ops->submit_io(dev, rqd, buf);
>>> if (ret)
>>> return ret;
>>>
>>> @@ -776,7 +777,8 @@ static int nvm_submit_io_wait(struct nvm_dev *dev, struct nvm_rq *rqd)
>>> return 0;
>>> }
>>>
>>> -int nvm_submit_io_sync(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd)
>>> +int nvm_submit_io_sync(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd,
>>> + void *buf)
>>> {
>>> struct nvm_dev *dev = tgt_dev->parent;
>>> int ret;
>>> @@ -789,7 +791,7 @@ int nvm_submit_io_sync(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd)
>>> rqd->dev = tgt_dev;
>>> rqd->flags = nvm_set_flags(&tgt_dev->geo, rqd);
>>>
>>> - ret = nvm_submit_io_wait(dev, rqd);
>>> + ret = nvm_submit_io_wait(dev, rqd, buf);
>>>
>>> return ret;
>>> }
>>> @@ -816,7 +818,7 @@ static int nvm_submit_io_sync_raw(struct nvm_dev *dev, struct nvm_rq *rqd)
>>> rqd->dev = NULL;
>>> rqd->flags = nvm_set_flags(&dev->geo, rqd);
>>>
>>> - return nvm_submit_io_wait(dev, rqd);
>>> + return nvm_submit_io_wait(dev, rqd, NULL);
>>> }
>>>
>>> static int nvm_bb_chunk_sense(struct nvm_dev *dev, struct ppa_addr ppa)
>>> diff --git a/drivers/lightnvm/pblk-core.c b/drivers/lightnvm/pblk-core.c
>>> index f546e6f28b8a..a58d3c84a3f2 100644
>>> --- a/drivers/lightnvm/pblk-core.c
>>> +++ b/drivers/lightnvm/pblk-core.c
>>> @@ -507,7 +507,7 @@ void pblk_set_sec_per_write(struct pblk *pblk, int sec_per_write)
>>> pblk->sec_per_write = sec_per_write;
>>> }
>>>
>>> -int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd)
>>> +int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd, void *buf)
>>> {
>>> struct nvm_tgt_dev *dev = pblk->dev;
>>>
>>> @@ -518,7 +518,7 @@ int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd)
>>> return NVM_IO_ERR;
>>> #endif
>>>
>>> - return nvm_submit_io(dev, rqd);
>>> + return nvm_submit_io(dev, rqd, buf);
>>> }
>>>
>>> void pblk_check_chunk_state_update(struct pblk *pblk, struct nvm_rq *rqd)
>>> @@ -541,7 +541,7 @@ void pblk_check_chunk_state_update(struct pblk *pblk, struct nvm_rq *rqd)
>>> }
>>> }
>>>
>>> -int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd)
>>> +int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd, void *buf)
>>> {
>>> struct nvm_tgt_dev *dev = pblk->dev;
>>> int ret;
>>> @@ -553,7 +553,7 @@ int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd)
>>> return NVM_IO_ERR;
>>> #endif
>>>
>>> - ret = nvm_submit_io_sync(dev, rqd);
>>> + ret = nvm_submit_io_sync(dev, rqd, buf);
>>>
>>> if (trace_pblk_chunk_state_enabled() && !ret &&
>>> rqd->opcode == NVM_OP_PWRITE)
>>> @@ -562,65 +562,19 @@ int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd)
>>> return ret;
>>> }
>>>
>>> -int pblk_submit_io_sync_sem(struct pblk *pblk, struct nvm_rq *rqd)
>>> +static int pblk_submit_io_sync_sem(struct pblk *pblk, struct nvm_rq *rqd,
>>> + void *buf)
>>> {
>>> struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd);
>>> int ret;
>>>
>>> pblk_down_chunk(pblk, ppa_list[0]);
>>> - ret = pblk_submit_io_sync(pblk, rqd);
>>> + ret = pblk_submit_io_sync(pblk, rqd, buf);
>>> pblk_up_chunk(pblk, ppa_list[0]);
>>>
>>> return ret;
>>> }
>>>
>>> -static void pblk_bio_map_addr_endio(struct bio *bio)
>>> -{
>>> - bio_put(bio);
>>> -}
>>> -
>>> -struct bio *pblk_bio_map_addr(struct pblk *pblk, void *data,
>>> - unsigned int nr_secs, unsigned int len,
>>> - int alloc_type, gfp_t gfp_mask)
>>> -{
>>> - struct nvm_tgt_dev *dev = pblk->dev;
>>> - void *kaddr = data;
>>> - struct page *page;
>>> - struct bio *bio;
>>> - int i, ret;
>>> -
>>> - if (alloc_type == PBLK_KMALLOC_META)
>>> - return bio_map_kern(dev->q, kaddr, len, gfp_mask);
>>> -
>>> - bio = bio_kmalloc(gfp_mask, nr_secs);
>>> - if (!bio)
>>> - return ERR_PTR(-ENOMEM);
>>> -
>>> - for (i = 0; i < nr_secs; i++) {
>>> - page = vmalloc_to_page(kaddr);
>>> - if (!page) {
>>> - pblk_err(pblk, "could not map vmalloc bio\n");
>>> - bio_put(bio);
>>> - bio = ERR_PTR(-ENOMEM);
>>> - goto out;
>>> - }
>>> -
>>> - ret = bio_add_pc_page(dev->q, bio, page, PAGE_SIZE, 0);
>>> - if (ret != PAGE_SIZE) {
>>> - pblk_err(pblk, "could not add page to bio\n");
>>> - bio_put(bio);
>>> - bio = ERR_PTR(-ENOMEM);
>>> - goto out;
>>> - }
>>> -
>>> - kaddr += PAGE_SIZE;
>>> - }
>>> -
>>> - bio->bi_end_io = pblk_bio_map_addr_endio;
>>> -out:
>>> - return bio;
>>> -}
>>> -
>>> int pblk_calc_secs(struct pblk *pblk, unsigned long secs_avail,
>>> unsigned long secs_to_flush, bool skip_meta)
>>> {
>>> @@ -722,9 +676,7 @@ u64 pblk_line_smeta_start(struct pblk *pblk, struct pblk_line *line)
>>>
>>> int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line)
>>> {
>>> - struct nvm_tgt_dev *dev = pblk->dev;
>>> struct pblk_line_meta *lm = &pblk->lm;
>>> - struct bio *bio;
>>> struct ppa_addr *ppa_list;
>>> struct nvm_rq rqd;
>>> u64 paddr = pblk_line_smeta_start(pblk, line);
>>> @@ -736,16 +688,6 @@ int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line)
>>> if (ret)
>>> return ret;
>>>
>>> - bio = bio_map_kern(dev->q, line->smeta, lm->smeta_len, GFP_KERNEL);
>>> - if (IS_ERR(bio)) {
>>> - ret = PTR_ERR(bio);
>>> - goto clear_rqd;
>>> - }
>>> -
>>> - bio->bi_iter.bi_sector = 0; /* internal bio */
>>> - bio_set_op_attrs(bio, REQ_OP_READ, 0);
>>> -
>>> - rqd.bio = bio;
>>> rqd.opcode = NVM_OP_PREAD;
>>> rqd.nr_ppas = lm->smeta_sec;
>>> rqd.is_seq = 1;
>>> @@ -754,10 +696,9 @@ int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line)
>>> for (i = 0; i < lm->smeta_sec; i++, paddr++)
>>> ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line->id);
>>>
>>> - ret = pblk_submit_io_sync(pblk, &rqd);
>>> + ret = pblk_submit_io_sync(pblk, &rqd, line->smeta);
>>> if (ret) {
>>> pblk_err(pblk, "smeta I/O submission failed: %d\n", ret);
>>> - bio_put(bio);
>>> goto clear_rqd;
>>> }
>>>
>>> @@ -776,9 +717,7 @@ int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line)
>>> static int pblk_line_smeta_write(struct pblk *pblk, struct pblk_line *line,
>>> u64 paddr)
>>> {
>>> - struct nvm_tgt_dev *dev = pblk->dev;
>>> struct pblk_line_meta *lm = &pblk->lm;
>>> - struct bio *bio;
>>> struct ppa_addr *ppa_list;
>>> struct nvm_rq rqd;
>>> __le64 *lba_list = emeta_to_lbas(pblk, line->emeta->buf);
>>> @@ -791,16 +730,6 @@ static int pblk_line_smeta_write(struct pblk *pblk, struct pblk_line *line,
>>> if (ret)
>>> return ret;
>>>
>>> - bio = bio_map_kern(dev->q, line->smeta, lm->smeta_len, GFP_KERNEL);
>>> - if (IS_ERR(bio)) {
>>> - ret = PTR_ERR(bio);
>>> - goto clear_rqd;
>>> - }
>>> -
>>> - bio->bi_iter.bi_sector = 0; /* internal bio */
>>> - bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
>>> -
>>> - rqd.bio = bio;
>>> rqd.opcode = NVM_OP_PWRITE;
>>> rqd.nr_ppas = lm->smeta_sec;
>>> rqd.is_seq = 1;
>>> @@ -814,10 +743,9 @@ static int pblk_line_smeta_write(struct pblk *pblk, struct pblk_line *line,
>>> meta->lba = lba_list[paddr] = addr_empty;
>>> }
>>>
>>> - ret = pblk_submit_io_sync_sem(pblk, &rqd);
>>> + ret = pblk_submit_io_sync_sem(pblk, &rqd, line->smeta);
>>> if (ret) {
>>> pblk_err(pblk, "smeta I/O submission failed: %d\n", ret);
>>> - bio_put(bio);
>>> goto clear_rqd;
>>> }
>>>
>>> @@ -838,10 +766,8 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
>>> {
>>> struct nvm_tgt_dev *dev = pblk->dev;
>>> struct nvm_geo *geo = &dev->geo;
>>> - struct pblk_line_mgmt *l_mg = &pblk->l_mg;
>>> struct pblk_line_meta *lm = &pblk->lm;
>>> void *ppa_list_buf, *meta_list;
>>> - struct bio *bio;
>>> struct ppa_addr *ppa_list;
>>> struct nvm_rq rqd;
>>> u64 paddr = line->emeta_ssec;
>>> @@ -867,17 +793,6 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
>>> rq_ppas = pblk_calc_secs(pblk, left_ppas, 0, false);
>>> rq_len = rq_ppas * geo->csecs;
>>>
>>> - bio = pblk_bio_map_addr(pblk, emeta_buf, rq_ppas, rq_len,
>>> - l_mg->emeta_alloc_type, GFP_KERNEL);
>>> - if (IS_ERR(bio)) {
>>> - ret = PTR_ERR(bio);
>>> - goto free_rqd_dma;
>>> - }
>>> -
>>> - bio->bi_iter.bi_sector = 0; /* internal bio */
>>> - bio_set_op_attrs(bio, REQ_OP_READ, 0);
>>> -
>>> - rqd.bio = bio;
>>> rqd.meta_list = meta_list;
>>> rqd.ppa_list = ppa_list_buf;
>>> rqd.dma_meta_list = dma_meta_list;
>>> @@ -896,7 +811,6 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
>>> while (test_bit(pos, line->blk_bitmap)) {
>>> paddr += min;
>>> if (pblk_boundary_paddr_checks(pblk, paddr)) {
>>> - bio_put(bio);
>>> ret = -EINTR;
>>> goto free_rqd_dma;
>>> }
>>> @@ -906,7 +820,6 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
>>> }
>>>
>>> if (pblk_boundary_paddr_checks(pblk, paddr + min)) {
>>> - bio_put(bio);
>>> ret = -EINTR;
>>> goto free_rqd_dma;
>>> }
>>> @@ -915,10 +828,9 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
>>> ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line_id);
>>> }
>>>
>>> - ret = pblk_submit_io_sync(pblk, &rqd);
>>> + ret = pblk_submit_io_sync(pblk, &rqd, emeta_buf);
>>> if (ret) {
>>> pblk_err(pblk, "emeta I/O submission failed: %d\n", ret);
>>> - bio_put(bio);
>>> goto free_rqd_dma;
>>> }
>>>
>>> @@ -963,7 +875,7 @@ static int pblk_blk_erase_sync(struct pblk *pblk, struct ppa_addr ppa)
>>> /* The write thread schedules erases so that it minimizes disturbances
>>> * with writes. Thus, there is no need to take the LUN semaphore.
>>> */
>>> - ret = pblk_submit_io_sync(pblk, &rqd);
>>> + ret = pblk_submit_io_sync(pblk, &rqd, NULL);
>>> rqd.private = pblk;
>>> __pblk_end_io_erase(pblk, &rqd);
>>>
>>> @@ -1792,7 +1704,7 @@ int pblk_blk_erase_async(struct pblk *pblk, struct ppa_addr ppa)
>>> /* The write thread schedules erases so that it minimizes disturbances
>>> * with writes. Thus, there is no need to take the LUN semaphore.
>>> */
>>> - err = pblk_submit_io(pblk, rqd);
>>> + err = pblk_submit_io(pblk, rqd, NULL);
>>> if (err) {
>>> struct nvm_tgt_dev *dev = pblk->dev;
>>> struct nvm_geo *geo = &dev->geo;
>>> @@ -1923,7 +1835,6 @@ void pblk_line_close_meta(struct pblk *pblk, struct pblk_line *line)
>>> static void pblk_save_lba_list(struct pblk *pblk, struct pblk_line *line)
>>> {
>>> struct pblk_line_meta *lm = &pblk->lm;
>>> - struct pblk_line_mgmt *l_mg = &pblk->l_mg;
>>> unsigned int lba_list_size = lm->emeta_len[2];
>>> struct pblk_w_err_gc *w_err_gc = line->w_err_gc;
>>> struct pblk_emeta *emeta = line->emeta;
>>> diff --git a/drivers/lightnvm/pblk-read.c b/drivers/lightnvm/pblk-read.c
>>> index d98ea392fe33..d572d4559e4e 100644
>>> --- a/drivers/lightnvm/pblk-read.c
>>> +++ b/drivers/lightnvm/pblk-read.c
>>> @@ -342,7 +342,7 @@ void pblk_submit_read(struct pblk *pblk, struct bio *bio)
>>> bio_put(int_bio);
>>> int_bio = bio_clone_fast(bio, GFP_KERNEL, &pblk_bio_set);
>>> goto split_retry;
>>> - } else if (pblk_submit_io(pblk, rqd)) {
>>> + } else if (pblk_submit_io(pblk, rqd, NULL)) {
>>> /* Submitting IO to drive failed, let's report an error */
>>> rqd->error = -ENODEV;
>>> pblk_end_io_read(rqd);
>>> @@ -419,7 +419,6 @@ int pblk_submit_read_gc(struct pblk *pblk, struct pblk_gc_rq *gc_rq)
>>> {
>>> struct nvm_tgt_dev *dev = pblk->dev;
>>> struct nvm_geo *geo = &dev->geo;
>>> - struct bio *bio;
>>> struct nvm_rq rqd;
>>> int data_len;
>>> int ret = NVM_IO_OK;
>>> @@ -447,25 +446,12 @@ int pblk_submit_read_gc(struct pblk *pblk, struct pblk_gc_rq *gc_rq)
>>> goto out;
>>>
>>> data_len = (gc_rq->secs_to_gc) * geo->csecs;
>>> - bio = pblk_bio_map_addr(pblk, gc_rq->data, gc_rq->secs_to_gc, data_len,
>>> - PBLK_VMALLOC_META, GFP_KERNEL);
>>> - if (IS_ERR(bio)) {
>>> - pblk_err(pblk, "could not allocate GC bio (%lu)\n",
>>> - PTR_ERR(bio));
>>> - ret = PTR_ERR(bio);
>>> - goto err_free_dma;
>>> - }
>>> -
>>> - bio->bi_iter.bi_sector = 0; /* internal bio */
>>> - bio_set_op_attrs(bio, REQ_OP_READ, 0);
>>> -
>>> rqd.opcode = NVM_OP_PREAD;
>>> rqd.nr_ppas = gc_rq->secs_to_gc;
>>> - rqd.bio = bio;
>>>
>>> - if (pblk_submit_io_sync(pblk, &rqd)) {
>>> + if (pblk_submit_io_sync(pblk, &rqd, gc_rq->data)) {
>>> ret = -EIO;
>>> - goto err_free_bio;
>>> + goto err_free_dma;
>>> }
>>>
>>> pblk_read_check_rand(pblk, &rqd, gc_rq->lba_list, gc_rq->nr_secs);
>>> @@ -489,8 +475,6 @@ int pblk_submit_read_gc(struct pblk *pblk, struct pblk_gc_rq *gc_rq)
>>> pblk_free_rqd_meta(pblk, &rqd);
>>> return ret;
>>>
>>> -err_free_bio:
>>> - bio_put(bio);
>>> err_free_dma:
>>> pblk_free_rqd_meta(pblk, &rqd);
>>> return ret;
>>> diff --git a/drivers/lightnvm/pblk-recovery.c b/drivers/lightnvm/pblk-recovery.c
>>> index e6dda04de144..d5e210c3c5b7 100644
>>> --- a/drivers/lightnvm/pblk-recovery.c
>>> +++ b/drivers/lightnvm/pblk-recovery.c
>>> @@ -178,12 +178,11 @@ static int pblk_recov_pad_line(struct pblk *pblk, struct pblk_line *line,
>>> void *meta_list;
>>> struct pblk_pad_rq *pad_rq;
>>> struct nvm_rq *rqd;
>>> - struct bio *bio;
>>> struct ppa_addr *ppa_list;
>>> void *data;
>>> __le64 *lba_list = emeta_to_lbas(pblk, line->emeta->buf);
>>> u64 w_ptr = line->cur_sec;
>>> - int left_line_ppas, rq_ppas, rq_len;
>>> + int left_line_ppas, rq_ppas;
>>> int i, j;
>>> int ret = 0;
>>>
>>> @@ -212,28 +211,15 @@ static int pblk_recov_pad_line(struct pblk *pblk, struct pblk_line *line,
>>> goto fail_complete;
>>> }
>>>
>>> - rq_len = rq_ppas * geo->csecs;
>>> -
>>> - bio = pblk_bio_map_addr(pblk, data, rq_ppas, rq_len,
>>> - PBLK_VMALLOC_META, GFP_KERNEL);
>>> - if (IS_ERR(bio)) {
>>> - ret = PTR_ERR(bio);
>>> - goto fail_complete;
>>> - }
>>> -
>>> - bio->bi_iter.bi_sector = 0; /* internal bio */
>>> - bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
>>> -
>>> rqd = pblk_alloc_rqd(pblk, PBLK_WRITE_INT);
>>>
>>> ret = pblk_alloc_rqd_meta(pblk, rqd);
>>> if (ret) {
>>> pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT);
>>> - bio_put(bio);
>>> goto fail_complete;
>>> }
>>>
>>> - rqd->bio = bio;
>>> + rqd->bio = NULL;
>>> rqd->opcode = NVM_OP_PWRITE;
>>> rqd->is_seq = 1;
>>> rqd->nr_ppas = rq_ppas;
>>> @@ -275,13 +261,12 @@ static int pblk_recov_pad_line(struct pblk *pblk, struct pblk_line *line,
>>> kref_get(&pad_rq->ref);
>>> pblk_down_chunk(pblk, ppa_list[0]);
>>>
>>> - ret = pblk_submit_io(pblk, rqd);
>>> + ret = pblk_submit_io(pblk, rqd, data);
>>> if (ret) {
>>> pblk_err(pblk, "I/O submission failed: %d\n", ret);
>>> pblk_up_chunk(pblk, ppa_list[0]);
>>> kref_put(&pad_rq->ref, pblk_recov_complete);
>>> pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT);
>>> - bio_put(bio);
>>> goto fail_complete;
>>> }
>>>
>>> @@ -375,7 +360,6 @@ static int pblk_recov_scan_oob(struct pblk *pblk, struct pblk_line *line,
>>> struct ppa_addr *ppa_list;
>>> void *meta_list;
>>> struct nvm_rq *rqd;
>>> - struct bio *bio;
>>> void *data;
>>> dma_addr_t dma_ppa_list, dma_meta_list;
>>> __le64 *lba_list;
>>> @@ -407,15 +391,7 @@ static int pblk_recov_scan_oob(struct pblk *pblk, struct pblk_line *line,
>>> rq_len = rq_ppas * geo->csecs;
>>>
>>> retry_rq:
>>> - bio = bio_map_kern(dev->q, data, rq_len, GFP_KERNEL);
>>> - if (IS_ERR(bio))
>>> - return PTR_ERR(bio);
>>> -
>>> - bio->bi_iter.bi_sector = 0; /* internal bio */
>>> - bio_set_op_attrs(bio, REQ_OP_READ, 0);
>>> - bio_get(bio);
>>> -
>>> - rqd->bio = bio;
>>> + rqd->bio = NULL;
>>> rqd->opcode = NVM_OP_PREAD;
>>> rqd->meta_list = meta_list;
>>> rqd->nr_ppas = rq_ppas;
>>> @@ -445,10 +421,9 @@ static int pblk_recov_scan_oob(struct pblk *pblk, struct pblk_line *line,
>>> addr_to_gen_ppa(pblk, paddr + j, line->id);
>>> }
>>>
>>> - ret = pblk_submit_io_sync(pblk, rqd);
>>> + ret = pblk_submit_io_sync(pblk, rqd, data);
>>> if (ret) {
>>> pblk_err(pblk, "I/O submission failed: %d\n", ret);
>>> - bio_put(bio);
>>> return ret;
>>> }
>>>
>>> @@ -460,24 +435,20 @@ static int pblk_recov_scan_oob(struct pblk *pblk, struct pblk_line *line,
>>>
>>> if (padded) {
>>> pblk_log_read_err(pblk, rqd);
>>> - bio_put(bio);
>>> return -EINTR;
>>> }
>>>
>>> pad_distance = pblk_pad_distance(pblk, line);
>>> ret = pblk_recov_pad_line(pblk, line, pad_distance);
>>> if (ret) {
>>> - bio_put(bio);
>>> return ret;
>>> }
>>>
>>> padded = true;
>>> - bio_put(bio);
>>> goto retry_rq;
>>> }
>>>
>>> pblk_get_packed_meta(pblk, rqd);
>>> - bio_put(bio);
>>>
>>> for (i = 0; i < rqd->nr_ppas; i++) {
>>> struct pblk_sec_meta *meta = pblk_get_meta(pblk, meta_list, i);
>>> diff --git a/drivers/lightnvm/pblk-write.c b/drivers/lightnvm/pblk-write.c
>>> index 4e63f9b5954c..b9a2aeba95ab 100644
>>> --- a/drivers/lightnvm/pblk-write.c
>>> +++ b/drivers/lightnvm/pblk-write.c
>>> @@ -373,7 +373,6 @@ int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line)
>>> struct pblk_emeta *emeta = meta_line->emeta;
>>> struct ppa_addr *ppa_list;
>>> struct pblk_g_ctx *m_ctx;
>>> - struct bio *bio;
>>> struct nvm_rq *rqd;
>>> void *data;
>>> u64 paddr;
>>> @@ -391,20 +390,9 @@ int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line)
>>> rq_len = rq_ppas * geo->csecs;
>>> data = ((void *)emeta->buf) + emeta->mem;
>>>
>>> - bio = pblk_bio_map_addr(pblk, data, rq_ppas, rq_len,
>>> - l_mg->emeta_alloc_type, GFP_KERNEL);
>>> - if (IS_ERR(bio)) {
>>> - pblk_err(pblk, "failed to map emeta io");
>>> - ret = PTR_ERR(bio);
>>> - goto fail_free_rqd;
>>> - }
>>> - bio->bi_iter.bi_sector = 0; /* internal bio */
>>> - bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
>>> - rqd->bio = bio;
>>> -
>>> ret = pblk_alloc_w_rq(pblk, rqd, rq_ppas, pblk_end_io_write_meta);
>>> if (ret)
>>> - goto fail_free_bio;
>>> + goto fail_free_rqd;
>>>
>>> ppa_list = nvm_rq_to_ppa_list(rqd);
>>> for (i = 0; i < rqd->nr_ppas; ) {
>>> @@ -423,7 +411,7 @@ int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line)
>>>
>>> pblk_down_chunk(pblk, ppa_list[0]);
>>>
>>> - ret = pblk_submit_io(pblk, rqd);
>>> + ret = pblk_submit_io(pblk, rqd, data);
>>> if (ret) {
>>> pblk_err(pblk, "emeta I/O submission failed: %d\n", ret);
>>> goto fail_rollback;
>>> @@ -437,8 +425,6 @@ int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line)
>>> pblk_dealloc_page(pblk, meta_line, rq_ppas);
>>> list_add(&meta_line->list, &meta_line->list);
>>> spin_unlock(&l_mg->close_lock);
>>> -fail_free_bio:
>>> - bio_put(bio);
>>> fail_free_rqd:
>>> pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT);
>>> return ret;
>>> @@ -523,7 +509,7 @@ static int pblk_submit_io_set(struct pblk *pblk, struct nvm_rq *rqd)
>>> meta_line = pblk_should_submit_meta_io(pblk, rqd);
>>>
>>> /* Submit data write for current data line */
>>> - err = pblk_submit_io(pblk, rqd);
>>> + err = pblk_submit_io(pblk, rqd, NULL);
>>> if (err) {
>>> pblk_err(pblk, "data I/O submission failed: %d\n", err);
>>> return NVM_IO_ERR;
>>> diff --git a/drivers/lightnvm/pblk.h b/drivers/lightnvm/pblk.h
>>> index a67855387f53..d515d3409a74 100644
>>> --- a/drivers/lightnvm/pblk.h
>>> +++ b/drivers/lightnvm/pblk.h
>>> @@ -783,14 +783,10 @@ struct nvm_chk_meta *pblk_chunk_get_off(struct pblk *pblk,
>>> struct ppa_addr ppa);
>>> void pblk_log_write_err(struct pblk *pblk, struct nvm_rq *rqd);
>>> void pblk_log_read_err(struct pblk *pblk, struct nvm_rq *rqd);
>>> -int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd);
>>> -int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd);
>>> -int pblk_submit_io_sync_sem(struct pblk *pblk, struct nvm_rq *rqd);
>>> +int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd, void *buf);
>>> +int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd, void *buf);
>>> int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line);
>>> void pblk_check_chunk_state_update(struct pblk *pblk, struct nvm_rq *rqd);
>>> -struct bio *pblk_bio_map_addr(struct pblk *pblk, void *data,
>>> - unsigned int nr_secs, unsigned int len,
>>> - int alloc_type, gfp_t gfp_mask);
>>> struct pblk_line *pblk_line_get(struct pblk *pblk);
>>> struct pblk_line *pblk_line_get_first_data(struct pblk *pblk);
>>> struct pblk_line *pblk_line_replace_data(struct pblk *pblk);
>>> diff --git a/drivers/nvme/host/lightnvm.c b/drivers/nvme/host/lightnvm.c
>>> index d6f121452d5d..ec46693f6b64 100644
>>> --- a/drivers/nvme/host/lightnvm.c
>>> +++ b/drivers/nvme/host/lightnvm.c
>>> @@ -667,11 +667,14 @@ static struct request *nvme_nvm_alloc_request(struct request_queue *q,
>>> return rq;
>>> }
>>>
>>> -static int nvme_nvm_submit_io(struct nvm_dev *dev, struct nvm_rq *rqd)
>>> +static int nvme_nvm_submit_io(struct nvm_dev *dev, struct nvm_rq *rqd,
>>> + void *buf)
>>> {
>>> + struct nvm_geo *geo = &dev->geo;
>>> struct request_queue *q = dev->q;
>>> struct nvme_nvm_command *cmd;
>>> struct request *rq;
>>> + int ret;
>>>
>>> cmd = kzalloc(sizeof(struct nvme_nvm_command), GFP_KERNEL);
>>> if (!cmd)
>>> @@ -679,8 +682,15 @@ static int nvme_nvm_submit_io(struct nvm_dev *dev, struct nvm_rq *rqd)
>>>
>>> rq = nvme_nvm_alloc_request(q, rqd, cmd);
>>> if (IS_ERR(rq)) {
>>> - kfree(cmd);
>>> - return PTR_ERR(rq);
>>> + ret = PTR_ERR(rq);
>>> + goto err_free_cmd;
>>> + }
>>> +
>>> + if (buf) {
>>> + ret = blk_rq_map_kern(q, rq, buf, geo->csecs * rqd->nr_ppas,
>>> + GFP_KERNEL);
>>> + if (ret)
>>> + goto err_free_cmd;
>>> }
>>>
>>> rq->end_io_data = rqd;
>>> @@ -688,6 +698,10 @@ static int nvme_nvm_submit_io(struct nvm_dev *dev, struct nvm_rq *rqd)
>>> blk_execute_rq_nowait(q, NULL, rq, 0, nvme_nvm_end_io);
>>>
>>> return 0;
>>> +
>>> +err_free_cmd:
>>> + kfree(cmd);
>>> + return ret;
>>> }
>>>
>>> static void *nvme_nvm_create_dma_pool(struct nvm_dev *nvmdev, char *name,
>>> diff --git a/include/linux/lightnvm.h b/include/linux/lightnvm.h
>>> index 8891647b24b1..ee8ec2e68055 100644
>>> --- a/include/linux/lightnvm.h
>>> +++ b/include/linux/lightnvm.h
>>> @@ -88,7 +88,7 @@ typedef int (nvm_op_bb_tbl_fn)(struct nvm_dev *, struct ppa_addr, u8 *);
>>> typedef int (nvm_op_set_bb_fn)(struct nvm_dev *, struct ppa_addr *, int, int);
>>> typedef int (nvm_get_chk_meta_fn)(struct nvm_dev *, sector_t, int,
>>> struct nvm_chk_meta *);
>>> -typedef int (nvm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *);
>>> +typedef int (nvm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *, void *);
>>> typedef void *(nvm_create_dma_pool_fn)(struct nvm_dev *, char *, int);
>>> typedef void (nvm_destroy_dma_pool_fn)(void *);
>>> typedef void *(nvm_dev_dma_alloc_fn)(struct nvm_dev *, void *, gfp_t,
>>> @@ -680,8 +680,8 @@ extern int nvm_get_chunk_meta(struct nvm_tgt_dev *, struct ppa_addr,
>>> int, struct nvm_chk_meta *);
>>> extern int nvm_set_chunk_meta(struct nvm_tgt_dev *, struct ppa_addr *,
>>> int, int);
>>> -extern int nvm_submit_io(struct nvm_tgt_dev *, struct nvm_rq *);
>>> -extern int nvm_submit_io_sync(struct nvm_tgt_dev *, struct nvm_rq *);
>>> +extern int nvm_submit_io(struct nvm_tgt_dev *, struct nvm_rq *, void *);
>>> +extern int nvm_submit_io_sync(struct nvm_tgt_dev *, struct nvm_rq *, void *);
>>> extern void nvm_end_io(struct nvm_rq *);
>>>
>>> #else /* CONFIG_NVM */
>>> --
>>> 2.7.4
>>
>> Itâs very good to get rid of pblk_bio_map_addr(). Need to look at this
>> closer because I remember a number of corner cases due to the metadata
>> allocation. Have you tested the vmalloc allocation path? Note that you
>> need big lines for this to happen as it will try to do kmalloc if
>> possible
>
> Yeah, i checked that vmalloc worked for the emeta buffers(by switching
> over manually), so we should be good.
>

Cool. Code looks good otherwise.

Reviewed-by: Javier GonzÃlez <javier@xxxxxxxxxxx>


Attachment: signature.asc
Description: Message signed with OpenPGP