Re: [PATCH v2 5/9] ASoC: qcom: Add support for codec dma driver

From: Srinivasa Rao Mandadapu
Date: Tue Oct 12 2021 - 08:25:51 EST



On 10/7/2021 5:30 PM, Srinivas Kandagatla wrote:
Thanks for Your Time Srini!!!

On 05/10/2021 14:52, Srinivasa Rao Mandadapu wrote:
Upadate lpass cpu and platform driver to support audio over codec dma
in ADSP bypass use case.

Signed-off-by: Venkata Prasad Potturu <potturu@xxxxxxxxxxxxxx>
Signed-off-by: Srinivasa Rao Mandadapu <srivasam@xxxxxxxxxxxxxx>
---
  sound/soc/qcom/lpass-cpu.c      | 252 +++++++++++++++++++-
  sound/soc/qcom/lpass-platform.c | 492 +++++++++++++++++++++++++++++++++++++---
  2 files changed, 706 insertions(+), 38 deletions(-)


Firstly, am unable to apply this patch on top of sound-next for testing purposes.

I have 3 major comments in general

1> split the patch one for regmap and other for dma handling, its hard to review this long patch.
Okay. will create New patch for regmap.

2> in general we should deal with range of dais supported for va, tx and rx macros, this patch seems to be just handling only 3 dai ids.
Yes. Actually this is ADSP bypass use-case. Here we are currently handling only 3 dai id's in snd_soc_dai_driver.

3> Also some details on why dma is handled differently in tx/rx, va dais when compared to other dais. Probably dt-bindings for that too.
Yes. Here The code DMA LPM memory is different for both RX/TX Path and VA Macro Path.


diff --git a/sound/soc/qcom/lpass-cpu.c b/sound/soc/qcom/lpass-cpu.c
index 3bd9eb3..6664d03 100644
--- a/sound/soc/qcom/lpass-cpu.c
+++ b/sound/soc/qcom/lpass-cpu.c
@@ -28,6 +28,8 @@
  #define LPASS_CPU_I2S_SD2_3_MASK    GENMASK(3, 2)
  #define LPASS_CPU_I2S_SD0_1_2_MASK    GENMASK(2, 0)
  #define LPASS_CPU_I2S_SD0_1_2_3_MASK    GENMASK(3, 0)
+#define LPASS_REG_READ 1
+#define LPASS_REG_WRITE 0
    /*
   * Channel maps for Quad channel playbacks on MI2S Secondary
@@ -798,6 +800,189 @@ static struct regmap_config lpass_hdmi_regmap_config = {
      .cache_type = REGCACHE_FLAT,
  };
  +static bool __lpass_rxtx_regmap_accessible(struct device *dev, unsigned int reg, bool rw)
+{
+    struct lpass_data *drvdata = dev_get_drvdata(dev);
+    struct lpass_variant *v = drvdata->variant;
+    int i;
+
+    for (i = 0; i < v->rxtx_irq_ports; ++i) {
+        if (reg == LPAIF_RXTX_IRQCLEAR_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+        if (reg == LPAIF_RXTX_IRQEN_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+        if (reg == LPAIF_RXTX_IRQSTAT_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+    }
+
+    for (i = 0; i < v->rxtx_rdma_channels; ++i) {
+        if (reg == LPAIF_CDC_RDMACTL_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+        if (reg == LPAIF_CDC_RDMABASE_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+        if (reg == LPAIF_CDC_RDMABUFF_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+        if (rw == LPASS_REG_READ) {
+            if (reg == LPAIF_CDC_RDMACURR_REG(v, i, LPASS_CDC_DMA_RX0))
+                return true;
+        }
+        if (reg == LPAIF_CDC_RDMAPER_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+        if (reg == LPAIF_CDC_RDMA_INTF_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+    }
+
+    for (i = 0; i < v->rxtx_wrdma_channels; ++i) {
+        if (reg == LPAIF_CDC_WRDMACTL_REG(v, i + v->rxtx_wrdma_channel_start,
+                            LPASS_CDC_DMA_TX3))
+            return true;
+        if (reg == LPAIF_CDC_WRDMABASE_REG(v, i + v->rxtx_wrdma_channel_start,
+                            LPASS_CDC_DMA_TX3))
+            return true;
+        if (reg == LPAIF_CDC_WRDMABUFF_REG(v, i + v->rxtx_wrdma_channel_start,
+                            LPASS_CDC_DMA_TX3))
+            return true;
+        if (rw == LPASS_REG_READ) {
+            if (reg == LPAIF_CDC_WRDMACURR_REG(v, i, LPASS_CDC_DMA_RX0))
+                return true;
+        }
+        if (reg == LPAIF_CDC_WRDMAPER_REG(v, i + v->rxtx_wrdma_channel_start,
+                            LPASS_CDC_DMA_TX3))
+            return true;
+        if (reg == LPAIF_CDC_WRDMA_INTF_REG(v, i + v->rxtx_wrdma_channel_start,
+                            LPASS_CDC_DMA_TX3))
+            return true;
+    }
+    return false;
+}
+
+static bool lpass_rxtx_regmap_writeable(struct device *dev, unsigned int reg)
+{
+    return __lpass_rxtx_regmap_accessible(dev, reg, LPASS_REG_WRITE);
+}
+
+static bool lpass_rxtx_regmap_readable(struct device *dev, unsigned int reg)
+{
+    return __lpass_rxtx_regmap_accessible(dev, reg, LPASS_REG_READ);
+}
+
+static bool lpass_rxtx_regmap_volatile(struct device *dev, unsigned int reg)
+{
+    struct lpass_data *drvdata = dev_get_drvdata(dev);
+    struct lpass_variant *v = drvdata->variant;
+    int i;
+
+    for (i = 0; i < v->rxtx_irq_ports; ++i) {
+        if (reg == LPAIF_RXTX_IRQCLEAR_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+        if (reg == LPAIF_RXTX_IRQSTAT_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+    }
+
+    for (i = 0; i < v->rxtx_rdma_channels; ++i)
+        if (reg == LPAIF_CDC_RDMACURR_REG(v, i, LPASS_CDC_DMA_RX0))
+            return true;
+
+    for (i = 0; i < v->rxtx_wrdma_channels; ++i)
+        if (reg == LPAIF_CDC_WRDMACURR_REG(v, i + v->rxtx_wrdma_channel_start,
+                            LPASS_CDC_DMA_TX3))
+            return true;
+
+    return false;
+}
+
+static bool __lpass_va_regmap_accessible(struct device *dev, unsigned int reg, bool rw)
+{
+    struct lpass_data *drvdata = dev_get_drvdata(dev);
+    struct lpass_variant *v = drvdata->variant;
+    int i;
+
+    for (i = 0; i < v->va_irq_ports; ++i) {
+        if (reg == LPAIF_RXTX_IRQCLEAR_REG(v, i, LPASS_CDC_DMA_VA_TX0))
+            return true;
+        if (reg == LPAIF_RXTX_IRQEN_REG(v, i, LPASS_CDC_DMA_VA_TX0))
+            return true;
+        if (reg == LPAIF_RXTX_IRQSTAT_REG(v, i, LPASS_CDC_DMA_VA_TX0))
+            return true;
+    }
+
+    for (i = 0; i < v->va_wrdma_channels; ++i) {
+        if (reg == LPAIF_CDC_WRDMACTL_REG(v, i + v->va_wrdma_channel_start,
+                            LPASS_CDC_DMA_VA_TX0))
+            return true;
+        if (reg == LPAIF_CDC_WRDMABASE_REG(v, i + v->va_wrdma_channel_start,
+                            LPASS_CDC_DMA_VA_TX0))
+            return true;
+        if (reg == LPAIF_CDC_WRDMABUFF_REG(v, i + v->va_wrdma_channel_start,
+                            LPASS_CDC_DMA_VA_TX0))
+            return true;
+        if (rw == LPASS_REG_READ) {
+            if (reg == LPAIF_CDC_WRDMACURR_REG(v, i + v->va_wrdma_channel_start,
+                            LPASS_CDC_DMA_VA_TX0))
+                return true;
+        }
+        if (reg == LPAIF_CDC_WRDMAPER_REG(v, i + v->va_wrdma_channel_start,
+                            LPASS_CDC_DMA_VA_TX0))
+            return true;
+        if (reg == LPAIF_CDC_WRDMA_INTF_REG(v, i + v->va_wrdma_channel_start,
+                            LPASS_CDC_DMA_VA_TX0))
+            return true;
+    }
+    return false;
+}
+
+static bool lpass_va_regmap_writeable(struct device *dev, unsigned int reg)
+{
+    return __lpass_va_regmap_accessible(dev, reg, LPASS_REG_WRITE);
+}
+
+static bool lpass_va_regmap_readable(struct device *dev, unsigned int reg)
+{
+    return __lpass_va_regmap_accessible(dev, reg, LPASS_REG_READ);
+}
+
+static bool lpass_va_regmap_volatile(struct device *dev, unsigned int reg)
+{
+    struct lpass_data *drvdata = dev_get_drvdata(dev);
+    struct lpass_variant *v = drvdata->variant;
+    int i;
+
+    for (i = 0; i < v->va_irq_ports; ++i) {
+        if (reg == LPAIF_RXTX_IRQCLEAR_REG(v, i, LPASS_CDC_DMA_VA_TX0))
+            return true;
+        if (reg == LPAIF_RXTX_IRQSTAT_REG(v, i, LPASS_CDC_DMA_VA_TX0))
+            return true;
+    }
+
+    for (i = 0; i < v->va_wrdma_channels; ++i) {
+        if (reg == LPAIF_CDC_WRDMACURR_REG(v, i + v->va_wrdma_channel_start,
+                            LPASS_CDC_DMA_VA_TX0))
+            return true;
+    }
+
+    return false;
+}
+
+static struct regmap_config lpass_rxtx_regmap_config = {
+    .reg_bits = 32,
+    .reg_stride = 4,
+    .val_bits = 32,
+    .writeable_reg = lpass_rxtx_regmap_writeable,
+    .readable_reg = lpass_rxtx_regmap_readable,
+    .volatile_reg = lpass_rxtx_regmap_volatile,
+    .cache_type = REGCACHE_FLAT,
+};
+
+static struct regmap_config lpass_va_regmap_config = {
+    .reg_bits = 32,
+    .reg_stride = 4,
+    .val_bits = 32,
+    .writeable_reg = lpass_va_regmap_writeable,
+    .readable_reg = lpass_va_regmap_readable,
+    .volatile_reg = lpass_va_regmap_volatile,
+    .cache_type = REGCACHE_FLAT,
+};
+
  static unsigned int of_lpass_cpu_parse_sd_lines(struct device *dev,
                          struct device_node *node,
                          const char *name)
@@ -837,6 +1022,17 @@ static unsigned int of_lpass_cpu_parse_sd_lines(struct device *dev,
      }
  }
  +static bool is_cdc_dma_port(int dai_id)
+{
+    switch (dai_id) {
+    case LPASS_CDC_DMA_RX0:
+    case LPASS_CDC_DMA_TX3:
+    case LPASS_CDC_DMA_VA_TX0:

What happens to the other VA/TX/RX dai ids?

You should probably add them here.
As Explained above in ADSP Bypass use case currently handling only 3 paths. i.e. RX0, TX3, VA_TX0.

+        return true;
+    default:
+        return false;
+    }
+}
  static void of_lpass_cpu_parse_dai_data(struct device *dev,
                      struct lpass_data *data)
  {
@@ -857,7 +1053,9 @@ static void of_lpass_cpu_parse_dai_data(struct device *dev,
          }
          if (id == LPASS_DP_RX) {
              data->hdmi_port_enable = 1;
-        } else {
+        } else if (is_cdc_dma_port(id))
+            data->wcd_codec_enable = 1;
naming of wcd is bit misleading, as VA has nothing to do with WCD. May be a better naming here would help.
Okay will change accordingly.

+        else {
              data->mi2s_playback_sd_mode[id] =
                  of_lpass_cpu_parse_sd_lines(dev, node,
                                  "qcom,playback-sd-lines");

On the other hand may be you can convert this to a swith case and include other dai-ids here, instead of adding is_cdc_dma_port().
Okay. In my opinion as port number checking many times, Helper function will do better, in readability.

@@ -897,6 +1095,53 @@ int asoc_qcom_lpass_cpu_platform_probe(struct platform_device *pdev)
        of_lpass_cpu_parse_dai_data(dev, drvdata);
  +    drvdata->num_clks =  variant->num_clks;
+    if (drvdata->wcd_codec_enable) {
+        drvdata->rxtx_lpaif =
+                devm_platform_ioremap_resource_byname(pdev, "lpass-rxtx-lpaif");
+        if (IS_ERR(drvdata->rxtx_lpaif))
+            return PTR_ERR(drvdata->rxtx_lpaif);
+
+        drvdata->va_lpaif = devm_platform_ioremap_resource_byname(pdev, "lpass-va-lpaif");
+        if (IS_ERR(drvdata->va_lpaif))
+            return PTR_ERR(drvdata->va_lpaif);
+
+        lpass_rxtx_regmap_config.max_register = LPAIF_CDC_WRDMAPER_REG(variant,
+                    variant->rxtx_wrdma_channels +
+                    variant->rxtx_wrdma_channel_start, LPASS_CDC_DMA_TX3);
+
+        drvdata->rxtx_lpaif_map = devm_regmap_init_mmio(dev, drvdata->rxtx_lpaif,
+                    &lpass_rxtx_regmap_config);
+        if (IS_ERR(drvdata->rxtx_lpaif_map)) {
+            dev_err(dev, "error initializing rxtx regmap: %ld\n",
+                PTR_ERR(drvdata->rxtx_lpaif_map));
+            return PTR_ERR(drvdata->rxtx_lpaif_map);
+        }
+        lpass_va_regmap_config.max_register = LPAIF_CDC_WRDMAPER_REG(variant,
+                    variant->va_wrdma_channels +
+                    variant->va_wrdma_channel_start, LPASS_CDC_DMA_VA_TX0);
+
+        drvdata->va_lpaif_map = devm_regmap_init_mmio(dev, drvdata->va_lpaif,
+                    &lpass_va_regmap_config);
+        if (IS_ERR(drvdata->va_lpaif_map)) {
+            dev_err(dev, "error initializing va regmap: %ld\n",
+                PTR_ERR(drvdata->va_lpaif_map));
+            return PTR_ERR(drvdata->va_lpaif_map);
+        }
+
+        drvdata->cdc_clks = devm_kcalloc(dev, variant->cdc_dma_num_clks,
+                    sizeof(*drvdata->cdc_clks), GFP_KERNEL);
+        drvdata->cdc_num_clks = variant->cdc_dma_num_clks;
+        for (i = 0; i < drvdata->cdc_num_clks; i++)
+            drvdata->cdc_clks[i].id = variant->cdc_dma_clk_names[i];
+
+        ret = devm_clk_bulk_get(dev, drvdata->cdc_num_clks, drvdata->cdc_clks);
+        if (ret) {
+            dev_err(dev, "Failed to get clocks %d\n", ret);
+            return ret;
+        }
+    }
+
      drvdata->lpaif = devm_platform_ioremap_resource_byname(pdev, "lpass-lpaif");
      if (IS_ERR(drvdata->lpaif))
          return PTR_ERR(drvdata->lpaif);
@@ -939,7 +1184,10 @@ int asoc_qcom_lpass_cpu_platform_probe(struct platform_device *pdev)
        for (i = 0; i < variant->num_dai; i++) {
          dai_id = variant->dai_driver[i].id;
-        if (dai_id == LPASS_DP_RX)
+        if (dai_id == LPASS_DP_RX ||
+            dai_id == LPASS_CDC_DMA_RX0 ||
+            dai_id == LPASS_CDC_DMA_TX3 ||
+            dai_id == LPASS_CDC_DMA_VA_TX0)

same here, we should deal with other dai ids too.
Explained above. please suggest your opinion.


              continue;
            drvdata->mi2s_osr_clk[dai_id] = devm_clk_get_optional(dev,
diff --git a/sound/soc/qcom/lpass-platform.c b/sound/soc/qcom/lpass-platform.c
index a000041..75dff5e 100644
--- a/sound/soc/qcom/lpass-platform.c
+++ b/sound/soc/qcom/lpass-platform.c
@@ -20,6 +20,10 @@
    #define LPASS_PLATFORM_BUFFER_SIZE    (24 *  2 * 1024)
  #define LPASS_PLATFORM_PERIODS        2
+#define LPSAS_RXTX_CDC_DMA_LPM_ADDR 0x3260000

What is this? This should come from DT.
Okay. Will handle it from DT.

+#define LPSAS_RXTX_CDC_DMA_LPM_BUFF_SIZE (24 * 1024)
+#define LPSAS_VA_CDC_DMA_LPM_ADDR 0x336C000


+#define LPSAS_VA_CDC_DMA_LPM_BUFF_SIZE (12 * 1024)
    static const struct snd_pcm_hardware lpass_platform_pcm_hardware = {
      .info            =    SNDRV_PCM_INFO_MMAP |
@@ -45,6 +49,103 @@ static const struct snd_pcm_hardware lpass_platform_pcm_hardware = {
      .fifo_size        =    0,
  };
  +static const struct snd_pcm_hardware lpass_platform_rxtx_hardware = {
+    .info            =    SNDRV_PCM_INFO_MMAP |
+                    SNDRV_PCM_INFO_MMAP_VALID |
+                    SNDRV_PCM_INFO_INTERLEAVED |
+                    SNDRV_PCM_INFO_PAUSE |
+                    SNDRV_PCM_INFO_RESUME,
+    .formats        =    SNDRV_PCM_FMTBIT_S16 |
+                    SNDRV_PCM_FMTBIT_S24 |
+                    SNDRV_PCM_FMTBIT_S32,
+    .rates            =    SNDRV_PCM_RATE_8000_192000,
+    .rate_min        =    8000,
+    .rate_max        =    192000,
+    .channels_min        =    1,
+    .channels_max        =    8,
+    .buffer_bytes_max    =    LPSAS_RXTX_CDC_DMA_LPM_BUFF_SIZE,
+    .period_bytes_max    =    LPSAS_RXTX_CDC_DMA_LPM_BUFF_SIZE /
+                        LPASS_PLATFORM_PERIODS,
+    .period_bytes_min    =    LPSAS_RXTX_CDC_DMA_LPM_BUFF_SIZE /
+                        LPASS_PLATFORM_PERIODS,
+    .periods_min        =    LPASS_PLATFORM_PERIODS,
+    .periods_max        =    LPASS_PLATFORM_PERIODS,
+    .fifo_size        =    0,
+};
+
+static const struct snd_pcm_hardware lpass_platform_va_hardware = {
+    .info            =    SNDRV_PCM_INFO_MMAP |
+                    SNDRV_PCM_INFO_MMAP_VALID |
+                    SNDRV_PCM_INFO_INTERLEAVED |
+                    SNDRV_PCM_INFO_PAUSE |
+                    SNDRV_PCM_INFO_RESUME,
+    .formats        =    SNDRV_PCM_FMTBIT_S16 |
+                    SNDRV_PCM_FMTBIT_S24 |
+                    SNDRV_PCM_FMTBIT_S32,
+    .rates            =    SNDRV_PCM_RATE_8000_192000,
+    .rate_min        =    8000,
+    .rate_max        =    192000,
+    .channels_min        =    1,
+    .channels_max        =    8,
+    .buffer_bytes_max    =    LPSAS_VA_CDC_DMA_LPM_BUFF_SIZE,
+    .period_bytes_max    =    LPSAS_VA_CDC_DMA_LPM_BUFF_SIZE /
+                        LPASS_PLATFORM_PERIODS,
+    .period_bytes_min    =    LPSAS_VA_CDC_DMA_LPM_BUFF_SIZE /
+                        LPASS_PLATFORM_PERIODS,
+    .periods_min        =    LPASS_PLATFORM_PERIODS,
+    .periods_max        =    LPASS_PLATFORM_PERIODS,
+    .fifo_size        =    0,
+};
+
+static int lpass_platform_alloc_rxtx_dmactl_fields(struct device *dev,
+                     struct regmap *map)
+{
+    struct lpass_data *drvdata = dev_get_drvdata(dev);
+    struct lpass_variant *v = drvdata->variant;
+    struct lpaif_dmactl *rd_dmactl, *wr_dmactl;
+    int rval;
+
+    drvdata->rxtx_rd_dmactl = devm_kzalloc(dev, sizeof(struct lpaif_dmactl),
+                      GFP_KERNEL);
+    if (drvdata->rxtx_rd_dmactl == NULL)
+        return -ENOMEM;
+
+    drvdata->rxtx_wr_dmactl = devm_kzalloc(dev, sizeof(struct lpaif_dmactl),
+                      GFP_KERNEL);
+    if (drvdata->rxtx_wr_dmactl == NULL)
+        return -ENOMEM;
+
+    rd_dmactl = drvdata->rxtx_rd_dmactl;
+    wr_dmactl = drvdata->rxtx_wr_dmactl;
+
+    rval = devm_regmap_field_bulk_alloc(dev, map, &rd_dmactl->intf,
+                        &v->rxtx_rdma_intf, 15);
+    if (rval)
+        return rval;
+
+    return devm_regmap_field_bulk_alloc(dev, map, &wr_dmactl->intf,
+                        &v->rxtx_wrdma_intf, 15);
+}
+
+static int lpass_platform_alloc_va_dmactl_fields(struct device *dev,
+                     struct regmap *map)
+{
+    struct lpass_data *drvdata = dev_get_drvdata(dev);
+    struct lpass_variant *v = drvdata->variant;
+    struct lpaif_dmactl *wr_dmactl;
+
+    drvdata->va_wr_dmactl = devm_kzalloc(dev, sizeof(struct lpaif_dmactl),
+                      GFP_KERNEL);
+    if (drvdata->va_wr_dmactl == NULL)
+        return -ENOMEM;
+
+    wr_dmactl = drvdata->va_wr_dmactl;
+
+    return devm_regmap_field_bulk_alloc(dev, map, &wr_dmactl->intf,
+                        &v->va_wrdma_intf, 15);
+}
+
+
  static int lpass_platform_alloc_dmactl_fields(struct device *dev,
                       struct regmap *map)
  {
@@ -126,24 +227,44 @@ static int lpass_platform_pcmops_open(struct snd_soc_component *component,
      if (cpu_dai->driver->id == LPASS_DP_RX) {
          map = drvdata->hdmiif_map;
          drvdata->hdmi_substream[dma_ch] = substream;
+    } else if (dai_id == LPASS_CDC_DMA_RX0 || dai_id == LPASS_CDC_DMA_TX3) {
+        map = drvdata->rxtx_lpaif_map;
+        drvdata->rxtx_substream[dma_ch] = substream;
+    } else if (dai_id == LPASS_CDC_DMA_VA_TX0) {
+        map = drvdata->va_lpaif_map;
+        drvdata->va_substream[dma_ch] = substream;
      } else {
          map = drvdata->lpaif_map;
          drvdata->substream[dma_ch] = substream;
      }
      data->dma_ch = dma_ch;
-    ret = regmap_write(map,
-            LPAIF_DMACTL_REG(v, dma_ch, dir, data->i2s_port), 0);
-    if (ret) {
-        dev_err(soc_runtime->dev,
-            "error writing to rdmactl reg: %d\n", ret);
-        return ret;
+    if (!(dai_id == LPASS_CDC_DMA_RX0 ||
+        dai_id == LPASS_CDC_DMA_TX3 ||
+        dai_id == LPASS_CDC_DMA_VA_TX0)) {
+        ret = regmap_write(map, LPAIF_DMACTL_REG(v, dma_ch, dir, data->i2s_port), 0);
+        if (ret) {
+            dev_err(soc_runtime->dev,
+                "error writing to rdmactl reg: %d\n", ret);
+            return ret;
+        }
      }
-    snd_soc_set_runtime_hwparams(substream, &lpass_platform_pcm_hardware);
-
-    runtime->dma_bytes = lpass_platform_pcm_hardware.buffer_bytes_max;
  -    ret = snd_pcm_hw_constraint_integer(runtime,
-            SNDRV_PCM_HW_PARAM_PERIODS);
+    switch (dai_id) {
+    case LPASS_CDC_DMA_RX0:
+    case LPASS_CDC_DMA_TX3:
+        snd_soc_set_runtime_hwparams(substream, &lpass_platform_rxtx_hardware);
+        runtime->dma_bytes = lpass_platform_rxtx_hardware.buffer_bytes_max;
+        break;
+    case LPASS_CDC_DMA_VA_TX0:
+        snd_soc_set_runtime_hwparams(substream, &lpass_platform_va_hardware);
+        runtime->dma_bytes = lpass_platform_va_hardware.buffer_bytes_max;
+        break;
+    default:
+        snd_soc_set_runtime_hwparams(substream, &lpass_platform_pcm_hardware);
+        runtime->dma_bytes = lpass_platform_pcm_hardware.buffer_bytes_max;
+        break;
+    }
+    ret = snd_pcm_hw_constraint_integer(runtime, SNDRV_PCM_HW_PARAM_PERIODS);
      if (ret < 0) {
          kfree(data);
          dev_err(soc_runtime->dev, "setting constraints failed: %d\n",
@@ -170,6 +291,10 @@ static int lpass_platform_pcmops_close(struct snd_soc_component *component,
      data = runtime->private_data;
      if (dai_id == LPASS_DP_RX)
          drvdata->hdmi_substream[data->dma_ch] = NULL;
+    else if (dai_id == LPASS_CDC_DMA_RX0 || dai_id == LPASS_CDC_DMA_TX3)
simillar comment about dai ids here.
+ drvdata->rxtx_substream[data->dma_ch] = NULL;
+    else if (dai_id == LPASS_CDC_DMA_VA_TX0)
+        drvdata->va_substream[data->dma_ch] = NULL;
      else
          drvdata->substream[data->dma_ch] = NULL;
      if (v->free_dma_channel)
@@ -202,12 +327,22 @@ static int lpass_platform_pcmops_hw_params(struct snd_soc_component *component,
          id = pcm_data->dma_ch;
          if (dai_id == LPASS_DP_RX)
              dmactl = drvdata->hdmi_rd_dmactl;
+        else if (dai_id == LPASS_CDC_DMA_RX0)
+            dmactl = drvdata->rxtx_rd_dmactl;
          else
              dmactl = drvdata->rd_dmactl;
        } else {
-        dmactl = drvdata->wr_dmactl;
-        id = pcm_data->dma_ch - v->wrdma_channel_start;
+        if (dai_id == LPASS_CDC_DMA_TX3) {
+            dmactl = drvdata->rxtx_wr_dmactl;
+            id = pcm_data->dma_ch - v->rxtx_wrdma_channel_start;
+        } else if (dai_id == LPASS_CDC_DMA_VA_TX0) {
+            dmactl = drvdata->va_wr_dmactl;
+            id = pcm_data->dma_ch - v->va_wrdma_channel_start;
+        } else {
+            dmactl = drvdata->wr_dmactl;
+            id = pcm_data->dma_ch - v->wrdma_channel_start;
+        }
      }
        bitwidth = snd_pcm_format_width(format);
@@ -230,6 +365,10 @@ static int lpass_platform_pcmops_hw_params(struct snd_soc_component *component,
      }
        switch (dai_id) {
+    case LPASS_CDC_DMA_RX0:
+    case LPASS_CDC_DMA_TX3:
+    case LPASS_CDC_DMA_VA_TX0:
+        break;
      case LPASS_DP_RX:
          ret = regmap_fields_write(dmactl->burst8, id,
                              LPAIF_DMACTL_BURSTEN_INCR4);
@@ -354,9 +493,12 @@ static int lpass_platform_pcmops_hw_free(struct snd_soc_component *component,
        if (dai_id == LPASS_DP_RX)
          map = drvdata->hdmiif_map;
+    else if (dai_id == LPASS_CDC_DMA_RX0 ||
+        dai_id == LPASS_CDC_DMA_TX3 ||
+        dai_id == LPASS_CDC_DMA_VA_TX0)
+        return 0;
      else
          map = drvdata->lpaif_map;
-
      reg = LPAIF_DMACTL_REG(v, pcm_data->dma_ch, substream->stream, dai_id);
      ret = regmap_write(map, reg, 0);
      if (ret)
@@ -387,6 +529,9 @@ static int lpass_platform_pcmops_prepare(struct snd_soc_component *component,
          if (dai_id == LPASS_DP_RX) {
              dmactl = drvdata->hdmi_rd_dmactl;
              map = drvdata->hdmiif_map;
+        } else if (dai_id == LPASS_CDC_DMA_RX0) {
+            map = drvdata->rxtx_lpaif_map;
+            dmactl = drvdata->rxtx_rd_dmactl;
          } else {
              dmactl = drvdata->rd_dmactl;
              map = drvdata->lpaif_map;
@@ -394,9 +539,19 @@ static int lpass_platform_pcmops_prepare(struct snd_soc_component *component,
            id = pcm_data->dma_ch;
      } else {
-        dmactl = drvdata->wr_dmactl;
-        id = pcm_data->dma_ch - v->wrdma_channel_start;
-        map = drvdata->lpaif_map;
+        if (dai_id == LPASS_CDC_DMA_TX3) {
+            dmactl = drvdata->rxtx_wr_dmactl;
+            map = drvdata->rxtx_lpaif_map;
+            id = pcm_data->dma_ch - v->rxtx_wrdma_channel_start;
+        } else if (dai_id == LPASS_CDC_DMA_VA_TX0) {
+            dmactl = drvdata->va_wr_dmactl;
+            map = drvdata->va_lpaif_map;
+            id = pcm_data->dma_ch - v->va_wrdma_channel_start;
+        } else {
+            dmactl = drvdata->wr_dmactl;
+            id = pcm_data->dma_ch - v->wrdma_channel_start;
+            map = drvdata->lpaif_map;
+        }
      }
        ret = regmap_write(map, LPAIF_DMABASE_REG(v, ch, dir, dai_id),
@@ -423,6 +578,16 @@ static int lpass_platform_pcmops_prepare(struct snd_soc_component *component,
          return ret;
      }
  +    if (dai_id == LPASS_CDC_DMA_RX0 ||
+        dai_id == LPASS_CDC_DMA_TX3 ||
+        dai_id == LPASS_CDC_DMA_VA_TX0) {
+        ret = regmap_fields_write(dmactl->fifowm, id, LPAIF_DMACTL_FIFOWM_8);
+        if (ret) {
+            dev_err(soc_runtime->dev, "error writing fifowm field to dmactl reg: %d, id: %d\n",
+                ret, id);
+            return ret;
+        }
+    }
      ret = regmap_fields_write(dmactl->enable, id, LPAIF_DMACTL_ENABLE_ON);
      if (ret) {
          dev_err(soc_runtime->dev, "error writing to rdmactl reg: %d\n",
@@ -457,14 +622,27 @@ static int lpass_platform_pcmops_trigger(struct snd_soc_component *component,
          if (dai_id == LPASS_DP_RX) {
              dmactl = drvdata->hdmi_rd_dmactl;
              map = drvdata->hdmiif_map;
+        } else if (dai_id == LPASS_CDC_DMA_RX0) {
+            map = drvdata->rxtx_lpaif_map;
+            dmactl = drvdata->rxtx_rd_dmactl;
          } else {
              dmactl = drvdata->rd_dmactl;
              map = drvdata->lpaif_map;
          }
      } else {
-        dmactl = drvdata->wr_dmactl;
-        id = pcm_data->dma_ch - v->wrdma_channel_start;
-        map = drvdata->lpaif_map;
+        if (dai_id == LPASS_CDC_DMA_TX3) {
+            dmactl = drvdata->rxtx_wr_dmactl;
+            map = drvdata->rxtx_lpaif_map;
+            id = pcm_data->dma_ch - v->rxtx_wrdma_channel_start;
+        } else if (dai_id == LPASS_CDC_DMA_VA_TX0) {
+            dmactl = drvdata->va_wr_dmactl;
+            map = drvdata->va_lpaif_map;
+            id = pcm_data->dma_ch - v->va_wrdma_channel_start;
+        } else {
+            dmactl = drvdata->wr_dmactl;
+            id = pcm_data->dma_ch - v->wrdma_channel_start;
+            map = drvdata->lpaif_map;
+        }
      }
        switch (cmd) {
@@ -479,6 +657,22 @@ static int lpass_platform_pcmops_trigger(struct snd_soc_component *component,
              return ret;
          }
          switch (dai_id) {
+        case LPASS_CDC_DMA_RX0:
+        case LPASS_CDC_DMA_TX3:
+        case LPASS_CDC_DMA_VA_TX0:
+            ret = regmap_fields_write(dmactl->dyncclk, id, LPAIF_DMACTL_DYNCLK_ON);
+            if (ret) {
+                dev_err(soc_runtime->dev,
+                    "error writing to rdmactl reg field: %d\n", ret);
+                return ret;
+            }
+            reg_irqclr = LPAIF_RXTX_IRQCLEAR_REG(v, LPAIF_IRQ_PORT_HOST, dai_id);
+            val_irqclr = LPAIF_IRQ_ALL(ch);
+
+            reg_irqen = LPAIF_RXTX_IRQEN_REG(v, LPAIF_IRQ_PORT_HOST, dai_id);
+            val_mask = LPAIF_IRQ_ALL(ch);
+            val_irqen = LPAIF_IRQ_ALL(ch);
+            break;
          case LPASS_DP_RX:
              ret = regmap_fields_write(dmactl->dyncclk, id,
                       LPAIF_DMACTL_DYNCLK_ON);
@@ -543,6 +737,24 @@ static int lpass_platform_pcmops_trigger(struct snd_soc_component *component,
              return ret;
          }
          switch (dai_id) {
+        case LPASS_CDC_DMA_RX0:
+        case LPASS_CDC_DMA_TX3:
+        case LPASS_CDC_DMA_VA_TX0:
+            ret = regmap_fields_write(dmactl->dyncclk, id, LPAIF_DMACTL_DYNCLK_OFF);
+            if (ret) {
+                dev_err(soc_runtime->dev,
+                    "error writing to rdmactl reg field: %d\n", ret);
+                return ret;
+            }
+
+            reg_irqclr = LPAIF_RXTX_IRQCLEAR_REG(v, LPAIF_IRQ_PORT_HOST, dai_id);
+            val_irqclr = LPAIF_IRQ_ALL(ch);
+
+            reg_irqen = LPAIF_RXTX_IRQEN_REG(v, LPAIF_IRQ_PORT_HOST, dai_id);
+            val_mask = LPAIF_IRQ_ALL(ch);
+            val_irqen = LPAIF_IRQ_ALL(ch);
+
+        break;
          case LPASS_DP_RX:
              ret = regmap_fields_write(dmactl->dyncclk, id,
                       LPAIF_DMACTL_DYNCLK_OFF);
@@ -601,6 +813,12 @@ static snd_pcm_uframes_t lpass_platform_pcmops_pointer(
        if (dai_id == LPASS_DP_RX)
          map = drvdata->hdmiif_map;
+    else if (dai_id == LPASS_CDC_DMA_RX0 ||
+        dai_id == LPASS_CDC_DMA_TX3 ||
+        dai_id == LPASS_CDC_DMA_VA_TX0)
+        map = (dai_id == LPASS_CDC_DMA_VA_TX0) ?
+            drvdata->va_lpaif_map :
+            drvdata->rxtx_lpaif_map;
      else
          map = drvdata->lpaif_map;
  @@ -625,12 +843,37 @@ static snd_pcm_uframes_t lpass_platform_pcmops_pointer(
      return bytes_to_frames(substream->runtime, curr_addr - base_addr);
  }
  +static int lpass_platform_cdc_dma_mmap(struct snd_soc_component *component,
+            struct snd_pcm_substream *substream,
+            struct vm_area_struct *vma)
+{
+    struct snd_pcm_runtime *runtime = substream->runtime;
+    unsigned long size, offset;
+
+    vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
+    size = vma->vm_end - vma->vm_start;
+    offset = vma->vm_pgoff << PAGE_SHIFT;
+    return io_remap_pfn_range(vma, vma->vm_start,
+            (runtime->dma_addr + offset) >> PAGE_SHIFT,
+            size, vma->vm_page_prot);
+
+}
+
  static int lpass_platform_pcmops_mmap(struct snd_soc_component *component,
                        struct snd_pcm_substream *substream,
                        struct vm_area_struct *vma)
  {
      struct snd_pcm_runtime *runtime = substream->runtime;
+    struct snd_soc_pcm_runtime *soc_runtime = asoc_substream_to_rtd(substream);
+    struct snd_soc_dai *cpu_dai = asoc_rtd_to_cpu(soc_runtime, 0);
+    unsigned int dai_id = cpu_dai->driver->id;
  +    if (dai_id == LPASS_CDC_DMA_RX0 ||
+        dai_id == LPASS_CDC_DMA_TX3 ||
+        dai_id == LPASS_CDC_DMA_VA_TX0) {
+        lpass_platform_cdc_dma_mmap(component, substream, vma);
+        return 0;
+    }
      return dma_mmap_coherent(component->dev, vma, runtime->dma_area,
                   runtime->dma_addr, runtime->dma_bytes);
  }
@@ -651,6 +894,14 @@ static irqreturn_t lpass_dma_interrupt_handler(
        mask = LPAIF_IRQ_ALL(chan);
      switch (dai_id) {
+    case LPASS_CDC_DMA_RX0:
+    case LPASS_CDC_DMA_TX3:
+    case LPASS_CDC_DMA_VA_TX0:
+        map = (dai_id == LPASS_CDC_DMA_VA_TX0) ?
+                drvdata->va_lpaif_map : drvdata->rxtx_lpaif_map;
+        reg = LPAIF_RXTX_IRQCLEAR_REG(v, LPAIF_IRQ_PORT_HOST, dai_id);
+        val = 0;
+        break;
      case LPASS_DP_RX:
          map = drvdata->hdmiif_map;
          reg = LPASS_HDMITX_APP_IRQCLEAR_REG(v);
@@ -774,41 +1025,131 @@ static irqreturn_t lpass_platform_hdmiif_irq(int irq, void *data)
                  return rv;
          }
      }
+    return IRQ_HANDLED;
+}
+
+static irqreturn_t lpass_platform_rxtxif_irq(int irq, void *data)
+{
+    struct lpass_data *drvdata = data;
+    struct lpass_variant *v = drvdata->variant;
+    unsigned int irqs;
+    int rv, chan;
+
+    rv = regmap_read(drvdata->rxtx_lpaif_map,
+            LPAIF_RXTX_IRQSTAT_REG(v, LPAIF_IRQ_PORT_HOST, LPASS_CDC_DMA_RX0), &irqs);
+    if (rv)
+        return IRQ_NONE;
+    /* Handle per channel interrupts */
+    for (chan = 0; chan < LPASS_MAX_CDC_DMA_CHANNELS; chan++) {
+        if (irqs & LPAIF_IRQ_ALL(chan) && drvdata->rxtx_substream[chan]) {
+            rv = lpass_dma_interrupt_handler(
+                        drvdata->rxtx_substream[chan],
+                        drvdata, chan, irqs);
+            if (rv != IRQ_HANDLED)
+                return rv;
+        }
+    }
+
+    return IRQ_HANDLED;
+}
+
+static irqreturn_t lpass_platform_vaif_irq(int irq, void *data)
+{
+    struct lpass_data *drvdata = data;
+    struct lpass_variant *v = drvdata->variant;
+    unsigned int irqs;
+    int rv, chan;
  +    rv = regmap_read(drvdata->va_lpaif_map,
+            LPAIF_RXTX_IRQSTAT_REG(v, LPAIF_IRQ_PORT_HOST,
+            LPASS_CDC_DMA_VA_TX0), &irqs);
+    if (rv)
+        return IRQ_NONE;
+    /* Handle per channel interrupts */
+    for (chan = 0; chan < LPASS_MAX_VA_CDC_DMA_CHANNELS; chan++) {
+        if (irqs & LPAIF_IRQ_ALL(chan) && drvdata->va_substream[chan]) {
+            rv = lpass_dma_interrupt_handler(
+                        drvdata->va_substream[chan],
+                        drvdata, chan, irqs);
+            if (rv != IRQ_HANDLED)
+                return rv;
+        }
+    }
      return IRQ_HANDLED;
  }
  +static int lpass_platform_prealloc_cdc_dma_buffer(struct snd_pcm *pcm,
+            struct snd_pcm_substream *substream, int dai_id)
+{
+    struct snd_dma_buffer *buf = &substream->dma_buffer;
+    int ret;
+
+    ret = dma_coerce_mask_and_coherent(pcm->card->dev, DMA_BIT_MASK(64));
+    if (ret)
+        return ret;
+
+    buf->dev.dev = pcm->card->dev;
+    buf->private_data = NULL;
+
+    /* Assign Codec DMA buffer pointers */
+    buf->dev.type = SNDRV_DMA_TYPE_CONTINUOUS;
+    if (dai_id == LPASS_CDC_DMA_RX0) {
+        buf->bytes = lpass_platform_rxtx_hardware.buffer_bytes_max;
+        buf->addr = LPSAS_RXTX_CDC_DMA_LPM_ADDR;

This is totally broken, and will not work on any other platform other than the one you tested on.
This function get's called only for codec dma paths.
+    } else if (dai_id == LPASS_CDC_DMA_TX3) {
+        buf->bytes = lpass_platform_rxtx_hardware.buffer_bytes_max;
+        buf->addr = LPSAS_RXTX_CDC_DMA_LPM_ADDR + LPSAS_RXTX_CDC_DMA_LPM_BUFF_SIZE;
+    } else if (dai_id == LPASS_CDC_DMA_VA_TX0) {
+        buf->bytes = lpass_platform_va_hardware.buffer_bytes_max;
+        buf->addr = LPSAS_VA_CDC_DMA_LPM_ADDR;
+    }
+
+    buf->area = (unsigned char * __force)ioremap(buf->addr, buf->bytes);
+
+    return 0;
+}
+
  static int lpass_platform_pcm_new(struct snd_soc_component *component,
                    struct snd_soc_pcm_runtime *soc_runtime)
  {
      struct snd_pcm *pcm = soc_runtime->pcm;
      struct snd_pcm_substream *psubstream, *csubstream;
      int ret;
+    struct snd_soc_dai *cpu_dai = asoc_rtd_to_cpu(soc_runtime, 0);
+    unsigned int dai_id = cpu_dai->driver->id;
+
This code does not exist in upstream version, please refer to latest code.
Okay. Will check and update.

      size_t size = lpass_platform_pcm_hardware.buffer_bytes_max;
        psubstream = pcm->streams[SNDRV_PCM_STREAM_PLAYBACK].substream;
      if (psubstream) {
-        ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV,
-                    component->dev,
-                    size, &psubstream->dma_buffer);
-        if (ret) {
-            dev_err(soc_runtime->dev, "Cannot allocate buffer(s)\n");
-            return ret;
+        if (dai_id == LPASS_CDC_DMA_RX0) {
+            lpass_platform_prealloc_cdc_dma_buffer(pcm, psubstream, dai_id);
+        } else {
+            ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV,
+                        component->dev,
+                        size, &psubstream->dma_buffer);
+            if (ret) {
+                dev_err(soc_runtime->dev, "Cannot allocate buffer(s)\n");
+                return ret;
+            }
          }
      }
        csubstream = pcm->streams[SNDRV_PCM_STREAM_CAPTURE].substream;
      if (csubstream) {
-        ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV,
-                    component->dev,
-                    size, &csubstream->dma_buffer);
-        if (ret) {
-            dev_err(soc_runtime->dev, "Cannot allocate buffer(s)\n");
-            if (psubstream)
- snd_dma_free_pages(&psubstream->dma_buffer);
-            return ret;
+        if (dai_id == LPASS_CDC_DMA_TX3 || dai_id == LPASS_CDC_DMA_VA_TX0) {
+            lpass_platform_prealloc_cdc_dma_buffer(pcm, csubstream, dai_id);
+        } else {
+            ret = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV,
+                        component->dev,
+                        size, &csubstream->dma_buffer);
+            if (ret) {
+                dev_err(soc_runtime->dev, "Cannot allocate buffer(s)\n");
+                if (psubstream)
+ snd_dma_free_pages(&psubstream->dma_buffer);
+                return ret;
+            }
          }
-
      }
        return 0;
@@ -818,18 +1159,42 @@ static void lpass_platform_pcm_free(struct snd_soc_component *component,
                      struct snd_pcm *pcm)
  {
      struct snd_pcm_substream *substream;
+    unsigned int dai_id = component->id;
      int i;
        for_each_pcm_streams(i) {
          substream = pcm->streams[i].substream;
          if (substream) {
-            snd_dma_free_pages(&substream->dma_buffer);
+            if (dai_id == LPASS_CDC_DMA_RX0 ||
+                dai_id == LPASS_CDC_DMA_TX3 ||
+                dai_id == LPASS_CDC_DMA_VA_TX0) {
+                if (substream->dma_buffer.area)
+                    iounmap((void __iomem *)substream->dma_buffer.area);
+            } else {
+ snd_dma_free_pages(&substream->dma_buffer);
+            }
              substream->dma_buffer.area = NULL;
              substream->dma_buffer.addr = 0;
          }
      }
  }
  +int lpass_platform_copy(struct snd_soc_component *component,
+             struct snd_pcm_substream *substream, int channel,
+             unsigned long pos, void __user *buf, unsigned long bytes)
+{
+    struct snd_pcm_runtime *rt = substream->runtime;
+    unsigned char *dma_buf = rt->dma_area + pos +
+                channel * (rt->dma_bytes / rt->channels);
+
+    if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK)
+        return copy_from_user_toio(dma_buf, buf, bytes);
+    else if (substream->stream == SNDRV_PCM_STREAM_CAPTURE)
+        return copy_to_user_fromio(buf, dma_buf, bytes);
+
+    return -EINVAL;
+}
+
  static int lpass_platform_pcmops_suspend(struct snd_soc_component *component)
  {
      struct lpass_data *drvdata = snd_soc_component_get_drvdata(component);
@@ -875,6 +1240,7 @@ static const struct snd_soc_component_driver lpass_component_driver = {
      .mmap        = lpass_platform_pcmops_mmap,
      .pcm_construct    = lpass_platform_pcm_new,
      .pcm_destruct    = lpass_platform_pcm_free,
+    .copy_user        = lpass_platform_copy,

Have you tested other dais like I2S with this patch?

Yes. Verified for other dais.
      .suspend        = lpass_platform_pcmops_suspend,
      .resume            = lpass_platform_pcmops_resume,
  @@ -914,6 +1280,60 @@ int asoc_qcom_lpass_platform_register(struct platform_device *pdev)
          return ret;
      }
  +    if (drvdata->wcd_codec_enable) {
+        ret = regmap_write(drvdata->rxtx_lpaif_map,
+            LPAIF_RXTX_IRQEN_REG(v, LPAIF_IRQ_PORT_HOST, LPASS_CDC_DMA_TX3), 0x0);
+        if (ret) {
+            dev_err(&pdev->dev, "error writing to rxtx irqen reg: %d\n", ret);
+            return ret;
+        }
+        ret = regmap_write(drvdata->va_lpaif_map,
+            LPAIF_RXTX_IRQEN_REG(v, LPAIF_IRQ_PORT_HOST, LPASS_CDC_DMA_VA_TX0), 0x0);
+        if (ret) {
+            dev_err(&pdev->dev, "error writing to rxtx irqen reg: %d\n", ret);
+            return ret;
+        }
+        drvdata->rxtxif_irq = platform_get_irq_byname(pdev, "lpass-irq-rxtxif");
+        if (drvdata->rxtxif_irq < 0)
+            return -ENODEV;
+
+        ret = devm_request_irq(&pdev->dev, drvdata->rxtxif_irq,
+                lpass_platform_rxtxif_irq, IRQF_TRIGGER_RISING,
+                "lpass-irq-rxtxif", drvdata);
+        if (ret) {
+            dev_err(&pdev->dev, "rxtx irq request failed: %d\n", ret);
+            return ret;
+        }
+
+        ret = lpass_platform_alloc_rxtx_dmactl_fields(&pdev->dev,
+                         drvdata->rxtx_lpaif_map);
+        if (ret) {
+            dev_err(&pdev->dev,
+                "error initializing rxtx dmactl fields: %d\n", ret);
+            return ret;
+        }
+
+        drvdata->vaif_irq = platform_get_irq_byname(pdev, "lpass-irq-vaif");
+        if (drvdata->vaif_irq < 0)
+            return -ENODEV;
+
+        ret = devm_request_irq(&pdev->dev, drvdata->vaif_irq,
+                lpass_platform_vaif_irq, IRQF_TRIGGER_RISING,
+                "lpass-irq-vaif", drvdata);
+        if (ret) {
+            dev_err(&pdev->dev, "va irq request failed: %d\n", ret);
+            return ret;
+        }
+
+        ret = lpass_platform_alloc_va_dmactl_fields(&pdev->dev,
+                         drvdata->va_lpaif_map);
+        if (ret) {
+            dev_err(&pdev->dev,
+                "error initializing va dmactl fields: %d\n", ret);
+            return ret;
+        }
+    }
+
      if (drvdata->hdmi_port_enable) {
          drvdata->hdmiif_irq = platform_get_irq_byname(pdev, "lpass-irq-hdmi");
          if (drvdata->hdmiif_irq < 0)

--
Qualcomm India Private Limited, on behalf of Qualcomm Innovation Center, Inc.,
is a member of Code Aurora Forum, a Linux Foundation Collaborative Project.