RE: [EXT] Re: [PATCH v9 04/13] media: amphion: add vpu core driver

From: Ming Qian
Date: Thu Sep 23 2021 - 06:32:38 EST


> -----Original Message-----
> From: Nicolas Dufresne [mailto:nicolas@xxxxxxxxxxxx]
> Sent: Wednesday, September 22, 2021 9:10 PM
> To: Ming Qian <ming.qian@xxxxxxx>; mchehab@xxxxxxxxxx;
> shawnguo@xxxxxxxxxx; robh+dt@xxxxxxxxxx; s.hauer@xxxxxxxxxxxxxx
> Cc: hverkuil-cisco@xxxxxxxxx; kernel@xxxxxxxxxxxxxx; festevam@xxxxxxxxx;
> dl-linux-imx <linux-imx@xxxxxxx>; Aisheng Dong <aisheng.dong@xxxxxxx>;
> linux-media@xxxxxxxxxxxxxxx; linux-kernel@xxxxxxxxxxxxxxx;
> devicetree@xxxxxxxxxxxxxxx; linux-arm-kernel@xxxxxxxxxxxxxxxxxxx
> Subject: [EXT] Re: [PATCH v9 04/13] media: amphion: add vpu core driver
>
> Caution: EXT Email
>
> Hi Ming,
>
> seem comment below.
>
> gLe lundi 13 septembre 2021 à 17:11 +0800, Ming Qian a écrit :
> > The vpu supports encoder and decoder.
> > it needs mu core to handle it.
> > core will run either encoder or decoder firmware.
> >
> > This driver is for support the vpu core.
> >
> > Signed-off-by: Ming Qian <ming.qian@xxxxxxx>
> > Signed-off-by: Shijie Qin <shijie.qin@xxxxxxx>
> > Signed-off-by: Zhou Peng <eagle.zhou@xxxxxxx>
> > ---
> > drivers/media/platform/amphion/vpu_codec.h | 68 ++
> > drivers/media/platform/amphion/vpu_core.c | 901
> +++++++++++++++++++++
> > drivers/media/platform/amphion/vpu_core.h | 16 +
> > drivers/media/platform/amphion/vpu_dbg.c | 496 ++++++++++++
> > drivers/media/platform/amphion/vpu_log.h | 44 +
>
> When loading this driver, I immediately stumbled across non-standard logging.
>
> [ 29.617994] [VPU CORE]core vpu_core probe
> [ 29.666447] [VPU CORE][0] decoder boot-region :
> <0x0000000084000000,
> 0x2000000>
> [ 29.738152] [VPU IMX8Q]imx8q fuse = 0x0
> [ 29.799421] [VPU CORE]core vpu_core probe<
> [ 30.111805] [VPU CORE][0] decoder boot
>
> I remember a thread related to Venus driver, and it seems that its best to
> remove this custom logging infrastructure and use the existing infrastructure.

OK, I'll make a change to improve the log

>
> > drivers/media/platform/amphion/vpu_rpc.c | 263 ++++++
> > drivers/media/platform/amphion/vpu_rpc.h | 463 +++++++++++
> > 7 files changed, 2251 insertions(+)
> > create mode 100644 drivers/media/platform/amphion/vpu_codec.h
> > create mode 100644 drivers/media/platform/amphion/vpu_core.c
> > create mode 100644 drivers/media/platform/amphion/vpu_core.h
> > create mode 100644 drivers/media/platform/amphion/vpu_dbg.c
> > create mode 100644 drivers/media/platform/amphion/vpu_log.h
> > create mode 100644 drivers/media/platform/amphion/vpu_rpc.c
> > create mode 100644 drivers/media/platform/amphion/vpu_rpc.h
> >
> > diff --git a/drivers/media/platform/amphion/vpu_codec.h
> b/drivers/media/platform/amphion/vpu_codec.h
> > new file mode 100644
> > index 000000000000..4dc41d6808b1
> > --- /dev/null
> > +++ b/drivers/media/platform/amphion/vpu_codec.h
> > @@ -0,0 +1,68 @@
> > +/* SPDX-License-Identifier: GPL-2.0 */
> > +/*
> > + * Copyright 2020-2021 NXP
> > + */
> > +
> > +#ifndef _AMPHION_VPU_CODEC_H
> > +#define _AMPHION_VPU_CODEC_H
> > +
> > +struct vpu_encode_params {
> > + u32 input_format;
> > + u32 codec_format;
> > + u32 profile;
> > + u32 tier;
> > + u32 level;
> > + u32 frame_rate_num;
> > + u32 frame_rate_den;
> > + u32 src_stride;
> > + u32 src_width;
> > + u32 src_height;
> > + struct v4l2_rect crop;
> > + u32 out_width;
> > + u32 out_height;
> > +
> > + u32 gop_length;
> > + u32 bframes;
> > +
> > + u32 rc_mode;
> > + u32 bitrate;
> > + u32 bitrate_min;
> > + u32 bitrate_max;
> > +
> > + u32 i_frame_qp;
> > + u32 p_frame_qp;
> > + u32 b_frame_qp;
> > + u32 qp_min;
> > + u32 qp_max;
> > + u32 qp_min_i;
> > + u32 qp_max_i;
> > +
> > + struct {
> > + u32 enable;
> > + u32 idc;
> > + u32 width;
> > + u32 height;
> > + } sar;
> > +
> > + struct {
> > + u32 primaries;
> > + u32 transfer;
> > + u32 matrix;
> > + u32 full_range;
> > + } color;
> > +};
> > +
> > +struct vpu_decode_params {
> > + u32 codec_format;
> > + u32 output_format;
> > + u32 b_dis_reorder;
> > + u32 b_non_frame;
> > + u32 frame_count;
> > + u32 end_flag;
> > + struct {
> > + u32 base;
> > + u32 size;
> > + } udata;
> > +};
> > +
> > +#endif
> > diff --git a/drivers/media/platform/amphion/vpu_core.c
> b/drivers/media/platform/amphion/vpu_core.c
> > new file mode 100644
> > index 000000000000..332ad3710abe
> > --- /dev/null
> > +++ b/drivers/media/platform/amphion/vpu_core.c
> > @@ -0,0 +1,901 @@
> > +// SPDX-License-Identifier: GPL-2.0
> > +/*
> > + * Copyright 2020-2021 NXP
> > + */
> > +
> > +#define TAG "CORE"
> > +
> > +#include <linux/init.h>
> > +#include <linux/interconnect.h>
> > +#include <linux/ioctl.h>
> > +#include <linux/list.h>
> > +#include <linux/kernel.h>
> > +#include <linux/module.h>
> > +#include <linux/of_device.h>
> > +#include <linux/of_address.h>
> > +#include <linux/platform_device.h>
> > +#include <linux/slab.h>
> > +#include <linux/types.h>
> > +#include <linux/pm_runtime.h>
> > +#include <linux/pm_domain.h>
> > +#include <linux/firmware.h>
> > +#include "vpu.h"
> > +#include "vpu_defs.h"
> > +#include "vpu_core.h"
> > +#include "vpu_mbox.h"
> > +#include "vpu_msgs.h"
> > +#include "vpu_rpc.h"
> > +#include "vpu_log.h"
> > +#include "vpu_cmds.h"
> > +
> > +unsigned int vpu_dbg_level = LVL_ERR | LVL_WARN | LVL_INFO;
> > +module_param(vpu_dbg_level, uint, 0644);
> > +
> > +void csr_writel(struct vpu_core *core, u32 reg, u32 val)
> > +{
> > + writel(val, core->base + reg);
> > +}
> > +
> > +u32 csr_readl(struct vpu_core *core, u32 reg)
> > +{
> > + return readl(core->base + reg);
> > +}
> > +
> > +static int vpu_core_load_firmware(struct vpu_core *core)
> > +{
> > + const struct firmware *pfw = NULL;
> > + int ret = 0;
> > +
> > + WARN_ON(!core || !core->res || !core->res->fwname);
> > + if (!core->fw.virt) {
> > + core_err(core, "firmware buffer is not ready\n");
> > + return -EINVAL;
> > + }
> > +
> > + ret = request_firmware(&pfw, core->res->fwname, core->dev);
> > + core_dbg(core, LVL_DEBUG, "request_firmware %s : %d\n",
> core->res->fwname, ret);
> > + if (ret) {
> > + core_err(core, "request firmware %s failed, ret = %d\n",
> > + core->res->fwname, ret);
> > + return ret;
> > + }
> > +
> > + if (core->fw.length < pfw->size) {
> > + core_err(core, "firmware buffer size want %zu, but %d\n",
> > + pfw->size, core->fw.length);
> > + ret = -EINVAL;
> > + goto exit;
> > + }
> > +
> > + memset_io(core->fw.virt, 0, core->fw.length);
> > + memcpy(core->fw.virt, pfw->data, pfw->size);
> > + core->fw.bytesused = pfw->size;
> > + ret = vpu_iface_on_firmware_loaded(core);
> > +exit:
> > + release_firmware(pfw);
> > + pfw = NULL;
> > +
> > + return ret;
> > +}
> > +
> > +static int vpu_core_wait_boot_done(struct vpu_core *core)
> > +{
> > + int ret;
> > + u32 fw_version;
> > +
> > + ret = wait_for_completion_timeout(&core->cmp, VPU_TIMEOUT);
> > + if (!ret) {
> > + core_err(core, "boot timeout\n");
> > + return -EINVAL;
> > + }
> > +
> > + fw_version = vpu_iface_get_version(core);
> > + core_dbg(core, LVL_WARN, "firmware version : %d.%d.%d\n",
> > + (fw_version >> 16) & 0xff,
> > + (fw_version >> 8) & 0xff,
> > + fw_version & 0xff);
> > + core->supported_instance_count =
> vpu_iface_get_max_instance_count(core);
> > + if (core->res->act_size) {
> > + u32 count = core->act.length / core->res->act_size;
> > +
> > + core->supported_instance_count =
> min(core->supported_instance_count, count);
> > + }
> > + core->fw_version = fw_version;
> > +
> > + return 0;
> > +}
> > +
> > +static int vpu_core_boot(struct vpu_core *core, bool load)
> > +{
> > + int ret;
> > +
> > + WARN_ON(!core);
> > +
> > + if (!core->res->standalone)
> > + return 0;
> > +
> > + core_dbg(core, LVL_WARN, "boot\n");
> > + reinit_completion(&core->cmp);
> > + if (load) {
> > + ret = vpu_core_load_firmware(core);
> > + if (ret)
> > + return ret;
> > + }
> > +
> > + vpu_iface_boot_core(core);
> > + return vpu_core_wait_boot_done(core);
> > +}
> > +
> > +static int vpu_core_shutdown(struct vpu_core *core)
> > +{
> > + if (!core->res->standalone)
> > + return 0;
> > + return vpu_iface_shutdown_core(core);
> > +}
> > +
> > +static int vpu_core_restore(struct vpu_core *core)
> > +{
> > + if (!core->res->standalone)
> > + return 0;
> > + return vpu_iface_restore_core(core);
> > +}
> > +
> > +static int __vpu_alloc_dma(struct device *dev, struct vpu_buffer *buf)
> > +{
> > + gfp_t gfp = GFP_KERNEL | GFP_DMA32;
> > +
> > + WARN_ON(!dev || !buf);
> > +
> > + if (!buf->length)
> > + return 0;
> > +
> > + buf->virt = dma_alloc_coherent(dev, buf->length, &buf->phys, gfp);
> > + if (!buf->virt)
> > + return -ENOMEM;
> > +
> > + buf->dev = dev;
> > +
> > + return 0;
> > +}
> > +
> > +void vpu_free_dma(struct vpu_buffer *buf)
> > +{
> > + WARN_ON(!buf);
> > +
> > + if (!buf->virt || !buf->dev)
> > + return;
> > +
> > + dma_free_coherent(buf->dev, buf->length, buf->virt, buf->phys);
> > + buf->virt = NULL;
> > + buf->phys = 0;
> > + buf->length = 0;
> > + buf->bytesused = 0;
> > + buf->dev = NULL;
> > +}
> > +
> > +int vpu_alloc_dma(struct vpu_core *core, struct vpu_buffer *buf)
> > +{
> > + WARN_ON(!core || !buf);
> > +
> > + return __vpu_alloc_dma(core->dev, buf);
> > +}
> > +
> > +static struct vpu_core *vpu_core_find_next_by_type(struct vpu_dev *vpu,
> u32 type)
> > +{
> > + struct vpu_core *c;
> > +
> > + WARN_ON(!vpu);
> > +
> > + list_for_each_entry(c, &vpu->cores, list) {
> > + if (c->type == type)
> > + return c;
> > + }
> > +
> > + return NULL;
> > +}
> > +
> > +int vpu_core_check_fmt(struct vpu_core *core, u32 pixelfmt)
> > +{
> > + if (!core)
> > + return -EINVAL;
> > +
> > + if (vpu_iface_check_format(core, pixelfmt))
> > + return 0;
> > +
> > + return -EINVAL;
> > +}
> > +
> > +static void vpu_core_check_hang(struct vpu_core *core)
> > +{
> > + if (core->hang_mask)
> > + core->state = VPU_CORE_HANG;
> > +}
> > +
> > +static struct vpu_core *vpu_core_find_proper_by_type(struct vpu_dev *vpu,
> u32 type)
> > +{
> > + struct vpu_core *core = NULL;
> > + int request_count = INT_MAX;
> > + struct vpu_core *c;
> > +
> > + WARN_ON(!vpu);
> > +
> > + list_for_each_entry(c, &vpu->cores, list) {
> > + core_dbg(c, LVL_DEBUG, "instance_mask = 0x%lx, state
> = %d\n",
> > + c->instance_mask,
> > + c->state);
> > + if (c->type != type)
> > + continue;
> > + if (c->state == VPU_CORE_DEINIT) {
> > + core = c;
> > + break;
> > + }
> > + vpu_core_check_hang(c);
> > + if (c->state != VPU_CORE_ACTIVE)
> > + continue;
> > + if (c->request_count < request_count) {
> > + request_count = c->request_count;
> > + core = c;
> > + }
> > + if (!request_count)
> > + break;
> > + }
> > +
> > + return core;
> > +}
> > +
> > +static bool vpu_core_is_exist(struct vpu_dev *vpu, struct vpu_core *core)
> > +{
> > + struct vpu_core *c;
> > +
> > + list_for_each_entry(c, &vpu->cores, list) {
> > + if (c == core)
> > + return true;
> > + }
> > +
> > + return false;
> > +}
> > +
> > +static void vpu_core_get_vpu(struct vpu_core *core)
> > +{
> > + core->vpu->get_vpu(core->vpu);
> > + if (core->type == VPU_CORE_TYPE_ENC)
> > + core->vpu->get_enc(core->vpu);
> > + if (core->type == VPU_CORE_TYPE_DEC)
> > + core->vpu->get_dec(core->vpu);
> > +}
> > +
> > +static int vpu_core_register(struct device *dev, struct vpu_core *core)
> > +{
> > + struct vpu_dev *vpu = dev_get_drvdata(dev);
> > + int ret = 0;
> > +
> > + core_dbg(core, LVL_DEBUG, "register core\n");
> > + if (vpu_core_is_exist(vpu, core))
> > + return 0;
> > +
> > + core->workqueue = alloc_workqueue("vpu", WQ_UNBOUND |
> WQ_MEM_RECLAIM, 1);
> > + if (!core->workqueue) {
> > + core_err(core, "fail to alloc workqueue\n");
> > + return -ENOMEM;
> > + }
> > + INIT_WORK(&core->msg_work, vpu_msg_run_work);
> > + INIT_DELAYED_WORK(&core->msg_delayed_work,
> vpu_msg_delayed_work);
> > + core->msg_buffer_size =
> roundup_pow_of_two(VPU_MSG_BUFFER_SIZE);
> > + core->msg_buffer = vzalloc(core->msg_buffer_size);
> > + if (!core->msg_buffer) {
> > + core_err(core, "failed allocate buffer for fifo\n");
> > + ret = -ENOMEM;
> > + goto error;
> > + }
> > + ret = kfifo_init(&core->msg_fifo, core->msg_buffer,
> core->msg_buffer_size);
> > + if (ret) {
> > + core_err(core, "failed init kfifo\n");
> > + goto error;
> > + }
> > +
> > + list_add_tail(&core->list, &vpu->cores);
> > +
> > + vpu_core_get_vpu(core);
> > +
> > + if (core->type == VPU_CORE_TYPE_ENC && !vpu->vdev_enc)
> > + venc_create_video_device(vpu);
> > + if (core->type == VPU_CORE_TYPE_DEC && !vpu->vdev_dec)
> > + vdec_create_video_device(vpu);
> > +
> > + return 0;
> > +error:
> > + if (core->msg_buffer) {
> > + vfree(core->msg_buffer);
> > + core->msg_buffer = NULL;
> > + }
> > + if (core->workqueue) {
> > + destroy_workqueue(core->workqueue);
> > + core->workqueue = NULL;
> > + }
> > + return ret;
> > +}
> > +
> > +static void vpu_core_put_vpu(struct vpu_core *core)
> > +{
> > + if (core->type == VPU_CORE_TYPE_ENC)
> > + core->vpu->put_enc(core->vpu);
> > + if (core->type == VPU_CORE_TYPE_DEC)
> > + core->vpu->put_dec(core->vpu);
> > + core->vpu->put_vpu(core->vpu);
> > +}
> > +
> > +static int vpu_core_unregister(struct device *dev, struct vpu_core *core)
> > +{
> > + struct vpu_dev *vpu = dev_get_drvdata(dev);
> > +
> > + list_del_init(&core->list);
> > +
> > + vpu_core_put_vpu(core);
> > + core->vpu = NULL;
> > + vfree(core->msg_buffer);
> > + core->msg_buffer = NULL;
> > +
> > + if (core->workqueue) {
> > + cancel_work_sync(&core->msg_work);
> > + cancel_delayed_work_sync(&core->msg_delayed_work);
> > + destroy_workqueue(core->workqueue);
> > + core->workqueue = NULL;
> > + }
> > +
> > + if (vpu_core_find_next_by_type(vpu, core->type))
> > + return 0;
> > +
> > + if (core->type == VPU_CORE_TYPE_ENC)
> > + video_unregister_device(vpu->vdev_enc);
> > + if (core->type == VPU_CORE_TYPE_DEC)
> > + video_unregister_device(vpu->vdev_dec);
> > +
> > + return 0;
> > +}
> > +
> > +static int vpu_core_acquire_instance(struct vpu_core *core)
> > +{
> > + int id;
> > +
> > + WARN_ON(!core);
> > +
> > + id = ffz(core->instance_mask);
> > + if (id >= core->supported_instance_count)
> > + return -EINVAL;
> > +
> > + set_bit(id, &core->instance_mask);
> > +
> > + return id;
> > +}
> > +
> > +static void vpu_core_release_instance(struct vpu_core *core, int id)
> > +{
> > + WARN_ON(!core);
> > +
> > + if (id < 0 || id >= core->supported_instance_count)
> > + return;
> > +
> > + clear_bit(id, &core->instance_mask);
> > +}
> > +
> > +struct vpu_inst *vpu_inst_get(struct vpu_inst *inst)
> > +{
> > + if (!inst)
> > + return NULL;
> > +
> > + atomic_inc(&inst->ref_count);
> > +
> > + return inst;
> > +}
> > +
> > +void vpu_inst_put(struct vpu_inst *inst)
> > +{
> > + if (!inst)
> > + return;
> > + if (atomic_dec_and_test(&inst->ref_count)) {
> > + if (inst->release)
> > + inst->release(inst);
> > + }
> > +}
> > +
> > +struct vpu_core *vpu_request_core(struct vpu_dev *vpu, enum
> vpu_core_type type)
> > +{
> > + struct vpu_core *core = NULL;
> > + int ret;
> > +
> > + mutex_lock(&vpu->lock);
> > +
> > + core = vpu_core_find_proper_by_type(vpu, type);
> > + if (!core)
> > + goto exit;
> > +
> > + core_dbg(core, LVL_DEBUG, "is found\n");
> > + mutex_lock(&core->lock);
> > + pm_runtime_get_sync(core->dev);
> > +
> > + if (core->state == VPU_CORE_DEINIT) {
> > + ret = vpu_core_boot(core, true);
> > + if (ret) {
> > + pm_runtime_put_sync(core->dev);
> > + mutex_unlock(&core->lock);
> > + core = NULL;
> > + goto exit;
> > + }
> > + core->state = VPU_CORE_ACTIVE;
> > + }
> > +
> > + core->request_count++;
> > +
> > + mutex_unlock(&core->lock);
> > +exit:
> > + mutex_unlock(&vpu->lock);
> > +
> > + return core;
> > +}
> > +
> > +void vpu_release_core(struct vpu_core *core)
> > +{
> > + if (!core)
> > + return;
> > +
> > + mutex_lock(&core->lock);
> > + pm_runtime_put_sync(core->dev);
> > + if (core->request_count)
> > + core->request_count--;
> > + mutex_unlock(&core->lock);
> > +}
> > +
> > +int vpu_inst_register(struct vpu_inst *inst)
> > +{
> > + struct vpu_core *core;
> > + int ret = 0;
> > +
> > + WARN_ON(!inst || !inst->core);
> > +
> > + core = inst->core;
> > + mutex_lock(&core->lock);
> > + if (inst->id >= 0 && inst->id < core->supported_instance_count)
> > + goto exit;
> > +
> > + ret = vpu_core_acquire_instance(core);
> > + if (ret < 0)
> > + goto exit;
> > +
> > + inst->id = ret;
> > + list_add_tail(&inst->list, &core->instances);
> > + ret = 0;
> > + if (core->res->act_size) {
> > + inst->act.phys = core->act.phys + core->res->act_size *
> inst->id;
> > + inst->act.virt = core->act.virt + core->res->act_size * inst->id;
> > + inst->act.length = core->res->act_size;
> > + }
> > + vpu_inst_create_dbgfs_file(inst);
> > +exit:
> > + mutex_unlock(&core->lock);
> > +
> > + if (ret)
> > + core_err(core, "register instance fail\n");
> > + return ret;
> > +}
> > +
> > +int vpu_inst_unregister(struct vpu_inst *inst)
> > +{
> > + struct vpu_core *core;
> > +
> > + WARN_ON(!inst);
> > +
> > + core = inst->core;
> > +
> > + vpu_clear_request(inst);
> > + mutex_lock(&core->lock);
> > + if (inst->id >= 0 && inst->id < core->supported_instance_count) {
> > + vpu_inst_remove_dbgfs_file(inst);
> > + list_del_init(&inst->list);
> > + vpu_core_release_instance(core, inst->id);
> > + inst->id = VPU_INST_NULL_ID;
> > + }
> > + vpu_core_check_hang(core);
> > + if (core->state == VPU_CORE_HANG && !core->instance_mask) {
> > + core_dbg(core, LVL_WARN, "reset hang core\n");
> > + if (!vpu_core_sw_reset(core)) {
> > + core->state = VPU_CORE_ACTIVE;
> > + core->hang_mask = 0;
> > + }
> > + }
> > + mutex_unlock(&core->lock);
> > +
> > + return 0;
> > +}
> > +
> > +struct vpu_inst *vpu_core_find_instance(struct vpu_core *core, u32 index)
> > +{
> > + struct vpu_inst *inst = NULL;
> > + struct vpu_inst *tmp;
> > +
> > + mutex_lock(&core->lock);
> > + if (!test_bit(index, &core->instance_mask))
> > + goto exit;
> > + list_for_each_entry(tmp, &core->instances, list) {
> > + if (tmp->id == index) {
> > + inst = vpu_inst_get(tmp);
> > + break;
> > + }
> > + }
> > +exit:
> > + mutex_unlock(&core->lock);
> > +
> > + return inst;
> > +}
> > +
> > +static int vpu_core_parse_dt(struct vpu_core *core, struct device_node *np)
> > +{
> > + struct device_node *node;
> > + struct resource res;
> > +
> > + if (of_count_phandle_with_args(np, "memory-region", NULL) < 2) {
> > + core_err(core, "need 2 memory-region for boot and rpc\n");
> > + return -ENODEV;
> > + }
> > +
> > + node = of_parse_phandle(np, "memory-region", 0);
> > + if (!node) {
> > + core_err(core, "boot-region of_parse_phandle error\n");
> > + return -ENODEV;
> > + }
> > + if (of_address_to_resource(node, 0, &res)) {
> > + core_err(core, "boot-region of_address_to_resource
> error\n");
> > + return -EINVAL;
> > + }
> > + core->fw.phys = res.start;
> > + core->fw.length = resource_size(&res);
> > + core_dbg(core, LVL_INFO, "boot-region : <%pad, 0x%x>\n",
> > + &core->fw.phys, core->fw.length);
> > +
> > + node = of_parse_phandle(np, "memory-region", 1);
> > + if (!node) {
> > + core_err(core, "rpc-region of_parse_phandle error\n");
> > + return -ENODEV;
> > + }
> > + if (of_address_to_resource(node, 0, &res)) {
> > + core_err(core, "rpc-region of_address_to_resource error\n");
> > + return -EINVAL;
> > + }
> > + core->rpc.phys = res.start;
> > + core->rpc.length = resource_size(&res);
> > + core_dbg(core, LVL_DEBUG, "rpc-region : <%pad, 0x%x>\n",
> > + &core->rpc.phys, core->rpc.length);
> > +
> > + if (core->rpc.length < core->res->rpc_size + core->res->fwlog_size) {
> > + core_err(core, "the rpc-region <%pad, 0x%x> is not
> enough\n",
> > + &core->rpc.phys, core->rpc.length);
> > + return -EINVAL;
> > + }
> > +
> > + core->fw.virt = ioremap_wc(core->fw.phys, core->fw.length);
> > + core->rpc.virt = ioremap_wc(core->rpc.phys, core->rpc.length);
> > + memset_io(core->rpc.virt, 0, core->rpc.length);
> > +
> > + if (vpu_iface_check_memory_region(core,
> > + core->rpc.phys,
> > + core->rpc.length) !=
> VPU_CORE_MEMORY_UNCACHED) {
> > + core_err(core, "rpc region<%pad, 0x%x> isn't uncached\n",
> > + &core->rpc.phys, core->rpc.length);
> > + return -EINVAL;
> > + }
> > +
> > + core->log.phys = core->rpc.phys + core->res->rpc_size;
> > + core->log.virt = core->rpc.virt + core->res->rpc_size;
> > + core->log.length = core->res->fwlog_size;
> > + core->act.phys = core->log.phys + core->log.length;
> > + core->act.virt = core->log.virt + core->log.length;
> > + core->act.length = core->rpc.length - core->res->rpc_size -
> core->log.length;
> > + core->rpc.length = core->res->rpc_size;
> > +
> > + return 0;
> > +}
> > +
> > +static int vpu_core_probe(struct platform_device *pdev)
> > +{
> > + struct device *dev = &pdev->dev;
> > + struct vpu_core *core;
> > + struct vpu_dev *vpu = dev_get_drvdata(dev->parent);
> > + struct vpu_shared_addr *iface;
> > + u32 iface_data_size;
> > + int ret;
> > +
> > + vpu_dbg(LVL_WARN, "core %s probe\n", pdev->dev.of_node->name);
> > + if (!vpu)
> > + return -EINVAL;
> > + core = devm_kzalloc(dev, sizeof(*core), GFP_KERNEL);
> > + if (!core)
> > + return -ENOMEM;
> > +
> > + core->pdev = pdev;
> > + core->dev = dev;
> > + platform_set_drvdata(pdev, core);
> > + core->vpu = vpu;
> > + INIT_LIST_HEAD(&core->instances);
> > + mutex_init(&core->lock);
> > + mutex_init(&core->cmd_lock);
> > + init_completion(&core->cmp);
> > + init_waitqueue_head(&core->ack_wq);
> > + core->state = VPU_CORE_DEINIT;
> > +
> > + core->res = of_device_get_match_data(dev);
> > + if (!core->res)
> > + return -ENODEV;
> > +
> > + core->type = core->res->type;
> > + core->id = of_alias_get_id(dev->of_node, "vpu_core");
> > + if (core->id < 0) {
> > + vpu_err("can't get vpu core id\n");
> > + return core->id;
> > + }
> > + ret = vpu_core_parse_dt(core, dev->of_node);
> > + if (ret)
> > + return ret;
> > +
> > + core->base = devm_platform_ioremap_resource(pdev, 0);
> > + if (IS_ERR(core->base))
> > + return PTR_ERR(core->base);
> > +
> > + if (!vpu_iface_check_codec(core)) {
> > + core_err(core, "is not supported\n");
> > + return -EINVAL;
> > + }
> > +
> > + ret = vpu_mbox_init(core);
> > + if (ret)
> > + return ret;
> > +
> > + iface = devm_kzalloc(dev, sizeof(*iface), GFP_KERNEL);
> > + if (!iface)
> > + return -ENOMEM;
> > +
> > + iface_data_size = vpu_iface_get_data_size(core);
> > + if (iface_data_size) {
> > + iface->priv = devm_kzalloc(dev, iface_data_size,
> GFP_KERNEL);
> > + if (!iface->priv)
> > + return -ENOMEM;
> > + }
> > +
> > + ret = vpu_iface_init(core, iface, &core->rpc, core->fw.phys);
> > + if (ret) {
> > + core_err(core, "init iface fail, ret = %d\n", ret);
> > + return ret;
> > + }
> > +
> > + vpu_iface_config_system(core, vpu->res->mreg_base, vpu->base);
> > + vpu_iface_set_log_buf(core, &core->log);
> > +
> > + pm_runtime_enable(dev);
> > + ret = pm_runtime_get_sync(dev);
> > + if (ret) {
> > + pm_runtime_put_noidle(dev);
> > + pm_runtime_set_suspended(dev);
> > + goto err_runtime_disable;
> > + }
> > +
> > + if (vpu_iface_get_power_state(core))
> > + ret = vpu_core_restore(core);
> > + if (ret)
> > + goto err_core_boot;
> > +
> > + ret = vpu_core_register(dev->parent, core);
> > + if (ret)
> > + goto err_core_register;
> > + core->parent = dev->parent;
> > +
> > + pm_runtime_put_sync(dev);
> > + vpu_core_create_dbgfs_file(core);
> > +
> > + return 0;
> > +
> > +err_core_register:
> > + vpu_core_shutdown(core);
> > +err_core_boot:
> > + pm_runtime_put_sync(dev);
> > +err_runtime_disable:
> > + pm_runtime_disable(dev);
> > +
> > + return ret;
> > +}
> > +
> > +static int vpu_core_remove(struct platform_device *pdev)
> > +{
> > + struct device *dev = &pdev->dev;
> > + struct vpu_core *core = platform_get_drvdata(pdev);
> > + int ret;
> > +
> > + vpu_core_remove_dbgfs_file(core);
> > + ret = pm_runtime_get_sync(dev);
> > + WARN_ON(ret < 0);
> > +
> > + vpu_core_shutdown(core);
> > + pm_runtime_put_sync(dev);
> > + pm_runtime_disable(dev);
> > +
> > + vpu_core_unregister(core->parent, core);
> > + iounmap(core->fw.virt);
> > + iounmap(core->rpc.virt);
> > + mutex_destroy(&core->lock);
> > + mutex_destroy(&core->cmd_lock);
> > +
> > + return 0;
> > +}
> > +
> > +static int __maybe_unused vpu_core_runtime_resume(struct device *dev)
> > +{
> > + struct vpu_core *core = dev_get_drvdata(dev);
> > +
> > + return vpu_mbox_request(core);
> > +}
> > +
> > +static int __maybe_unused vpu_core_runtime_suspend(struct device *dev)
> > +{
> > + struct vpu_core *core = dev_get_drvdata(dev);
> > +
> > + vpu_mbox_free(core);
> > + return 0;
> > +}
> > +
> > +static void vpu_core_cancel_work(struct vpu_core *core)
> > +{
> > + struct vpu_inst *inst = NULL;
> > +
> > + cancel_work_sync(&core->msg_work);
> > + cancel_delayed_work_sync(&core->msg_delayed_work);
> > +
> > + mutex_lock(&core->lock);
> > + list_for_each_entry(inst, &core->instances, list)
> > + cancel_work_sync(&inst->msg_work);
> > + mutex_unlock(&core->lock);
> > +}
> > +
> > +static void vpu_core_resume_work(struct vpu_core *core)
> > +{
> > + struct vpu_inst *inst = NULL;
> > + unsigned long delay = msecs_to_jiffies(10);
> > +
> > + queue_work(core->workqueue, &core->msg_work);
> > + queue_delayed_work(core->workqueue, &core->msg_delayed_work,
> delay);
> > +
> > + mutex_lock(&core->lock);
> > + list_for_each_entry(inst, &core->instances, list)
> > + queue_work(inst->workqueue, &inst->msg_work);
> > + mutex_unlock(&core->lock);
> > +}
> > +
> > +static int __maybe_unused vpu_core_resume(struct device *dev)
> > +{
> > + struct vpu_core *core = dev_get_drvdata(dev);
> > + int ret = 0;
> > +
> > + if (!core->res->standalone)
> > + return 0;
> > +
> > + mutex_lock(&core->lock);
> > + pm_runtime_get_sync(dev);
> > + vpu_core_get_vpu(core);
> > + if (core->state != VPU_CORE_SNAPSHOT)
> > + goto exit;
> > +
> > + if (!vpu_iface_get_power_state(core)) {
> > + if (!list_empty(&core->instances)) {
> > + ret = vpu_core_boot(core, false);
> > + if (ret) {
> > + core_err(core, "%s boot fail\n", __func__);
> > + core->state = VPU_CORE_DEINIT;
> > + goto exit;
> > + }
> > + core->state = VPU_CORE_ACTIVE;
> > + } else {
> > + core->state = VPU_CORE_DEINIT;
> > + }
> > + } else {
> > + if (!list_empty(&core->instances)) {
> > + ret = vpu_core_sw_reset(core);
> > + if (ret) {
> > + core_err(core, "%s sw_reset fail\n",
> __func__);
> > + core->state = VPU_CORE_HANG;
> > + goto exit;
> > + }
> > + }
> > + core->state = VPU_CORE_ACTIVE;
> > + }
> > +
> > +exit:
> > + pm_runtime_put_sync(dev);
> > + mutex_unlock(&core->lock);
> > +
> > + vpu_core_resume_work(core);
> > + return ret;
> > +}
> > +
> > +static int __maybe_unused vpu_core_suspend(struct device *dev)
> > +{
> > + struct vpu_core *core = dev_get_drvdata(dev);
> > + int ret = 0;
> > +
> > + if (!core->res->standalone)
> > + return 0;
> > +
> > + mutex_lock(&core->lock);
> > + if (core->state == VPU_CORE_ACTIVE) {
> > + if (!list_empty(&core->instances)) {
> > + ret = vpu_core_snapshot(core);
> > + if (ret) {
> > + mutex_unlock(&core->lock);
> > + return ret;
> > + }
> > + }
> > +
> > + core->state = VPU_CORE_SNAPSHOT;
> > + }
> > + mutex_unlock(&core->lock);
> > +
> > + vpu_core_cancel_work(core);
> > +
> > + mutex_lock(&core->lock);
> > + vpu_core_put_vpu(core);
> > + mutex_unlock(&core->lock);
> > + return ret;
> > +}
> > +
> > +static const struct dev_pm_ops vpu_core_pm_ops = {
> > + SET_RUNTIME_PM_OPS(vpu_core_runtime_suspend,
> vpu_core_runtime_resume, NULL)
> > + SET_SYSTEM_SLEEP_PM_OPS(vpu_core_suspend, vpu_core_resume)
> > +};
> > +
> > +static struct vpu_core_resources imx8q_enc = {
> > + .type = VPU_CORE_TYPE_ENC,
> > + .fwname = "vpu/vpu_fw_imx8_enc.bin",
> > + .stride = 16,
> > + .max_width = 1920,
> > + .max_height = 1920,
> > + .min_width = 64,
> > + .min_height = 48,
> > + .step_width = 2,
> > + .step_height = 2,
> > + .rpc_size = 0x80000,
> > + .fwlog_size = 0x80000,
> > + .act_size = 0xc0000,
> > + .standalone = true,
> > +};
> > +
> > +static struct vpu_core_resources imx8q_dec = {
> > + .type = VPU_CORE_TYPE_DEC,
> > + .fwname = "vpu/vpu_fw_imx8_dec.bin",
> > + .stride = 256,
> > + .max_width = 8188,
> > + .max_height = 8188,
> > + .min_width = 16,
> > + .min_height = 16,
> > + .step_width = 1,
> > + .step_height = 1,
> > + .rpc_size = 0x80000,
> > + .fwlog_size = 0x80000,
> > + .standalone = true,
> > +};
> > +
> > +static const struct of_device_id vpu_core_dt_match[] = {
> > + { .compatible = "nxp,imx8q-vpu-encoder", .data = &imx8q_enc },
> > + { .compatible = "nxp,imx8q-vpu-decoder", .data = &imx8q_dec },
> > + {}
> > +};
> > +MODULE_DEVICE_TABLE(of, vpu_core_dt_match);
> > +
> > +static struct platform_driver amphion_vpu_core_driver = {
> > + .probe = vpu_core_probe,
> > + .remove = vpu_core_remove,
> > + .driver = {
> > + .name = "amphion-vpu-core",
> > + .of_match_table = vpu_core_dt_match,
> > + .pm = &vpu_core_pm_ops,
> > + },
> > +};
> > +module_platform_driver(amphion_vpu_core_driver);
> > +
> > +MODULE_AUTHOR("Freescale Semiconductor, Inc.");
> > +MODULE_DESCRIPTION("Linux VPU driver for Freescale i.MX8Q");
> > +MODULE_LICENSE("GPL v2");
> > diff --git a/drivers/media/platform/amphion/vpu_core.h
> b/drivers/media/platform/amphion/vpu_core.h
> > new file mode 100644
> > index 000000000000..12880e2e14b2
> > --- /dev/null
> > +++ b/drivers/media/platform/amphion/vpu_core.h
> > @@ -0,0 +1,16 @@
> > +/* SPDX-License-Identifier: GPL-2.0 */
> > +/*
> > + * Copyright 2020-2021 NXP
> > + */
> > +
> > +#ifndef _AMPHION_VPU_CORE_H
> > +#define _AMPHION_VPU_CORE_H
> > +
> > +void csr_writel(struct vpu_core *core, u32 reg, u32 val);
> > +u32 csr_readl(struct vpu_core *core, u32 reg);
> > +int vpu_alloc_dma(struct vpu_core *core, struct vpu_buffer *buf);
> > +void vpu_free_dma(struct vpu_buffer *buf);
> > +struct vpu_inst *vpu_core_find_instance(struct vpu_core *core, u32 index);
> > +int vpu_core_check_fmt(struct vpu_core *core, u32 pixelfmt);
> > +
> > +#endif
> > diff --git a/drivers/media/platform/amphion/vpu_dbg.c
> b/drivers/media/platform/amphion/vpu_dbg.c
> > new file mode 100644
> > index 000000000000..4dbfbd6d2f46
> > --- /dev/null
> > +++ b/drivers/media/platform/amphion/vpu_dbg.c
> > @@ -0,0 +1,496 @@
> > +// SPDX-License-Identifier: GPL-2.0
> > +/*
> > + * Copyright 2020-2021 NXP
> > + */
> > +
> > +#define TAG "DEBUG"
> > +
> > +#include <linux/init.h>
> > +#include <linux/device.h>
> > +#include <linux/ioctl.h>
> > +#include <linux/list.h>
> > +#include <linux/module.h>
> > +#include <linux/kernel.h>
> > +#include <linux/types.h>
> > +#include <linux/pm_runtime.h>
> > +#include <media/v4l2-device.h>
> > +#include <linux/debugfs.h>
> > +#include "vpu.h"
> > +#include "vpu_defs.h"
> > +#include "vpu_helpers.h"
> > +#include "vpu_cmds.h"
> > +#include "vpu_rpc.h"
> > +#include "vpu_log.h"
> > +
> > +struct print_buf_desc {
> > + u32 start_h_phy;
> > + u32 start_h_vir;
> > + u32 start_m;
> > + u32 bytes;
> > + u32 read;
> > + u32 write;
> > + char buffer[0];
> > +};
> > +
> > +static char *vb2_stat_name[] = {
> > + [VB2_BUF_STATE_DEQUEUED] = "dequeued",
> > + [VB2_BUF_STATE_IN_REQUEST] = "in_request",
> > + [VB2_BUF_STATE_PREPARING] = "preparing",
> > + [VB2_BUF_STATE_QUEUED] = "queued",
> > + [VB2_BUF_STATE_ACTIVE] = "active",
> > + [VB2_BUF_STATE_DONE] = "done",
> > + [VB2_BUF_STATE_ERROR] = "error",
> > +};
> > +
> > +static char *vpu_stat_name[] = {
> > + [VPU_BUF_STATE_IDLE] = "idle",
> > + [VPU_BUF_STATE_INUSE] = "inuse",
> > + [VPU_BUF_STATE_DECODED] = "decoded",
> > + [VPU_BUF_STATE_READY] = "ready",
> > + [VPU_BUF_STATE_SKIP] = "skip",
> > + [VPU_BUF_STATE_ERROR] = "error",
> > +};
> > +
> > +static int vpu_dbg_instance(struct seq_file *s, void *data)
> > +{
> > + struct vpu_inst *inst = s->private;
> > + char str[128];
> > + int num;
> > + struct vb2_queue *vq;
> > + int i;
> > +
> > + num = scnprintf(str, sizeof(str), "[%s]\n",
> vpu_core_type_desc(inst->type));
> > + if (seq_write(s, str, num))
> > + return 0;
> > +
> > + num = scnprintf(str, sizeof(str), "tgig = %d,pid = %d\n", inst->tgid,
> inst->pid);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + num = scnprintf(str, sizeof(str), "state = %d\n", inst->state);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + num = scnprintf(str, sizeof(str),
> > + "min_buffer_out = %d, min_buffer_cap = %d\n",
> > + inst->min_buffer_out, inst->min_buffer_cap);
> > + if (seq_write(s, str, num))
> > + return 0;
> > +
> > +
> > + vq = v4l2_m2m_get_src_vq(inst->m2m_ctx);
> > + num = scnprintf(str, sizeof(str),
> > + "output(%d) : fmt = %c%c%c%c %d x %d, %d;",
> > + vb2_is_streaming(vq),
> > + inst->out_format.pixfmt,
> > + inst->out_format.pixfmt >> 8,
> > + inst->out_format.pixfmt >> 16,
> > + inst->out_format.pixfmt >> 24,
> > + inst->out_format.width,
> > + inst->out_format.height,
> > + vq->last_buffer_dequeued);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + for (i = 0; i < inst->out_format.num_planes; i++) {
> > + num = scnprintf(str, sizeof(str), " %d(%d)",
> > + inst->out_format.sizeimage[i],
> > + inst->out_format.bytesperline[i]);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + }
> > + if (seq_write(s, "\n", 1))
> > + return 0;
> > +
> > + vq = v4l2_m2m_get_dst_vq(inst->m2m_ctx);
> > + num = scnprintf(str, sizeof(str),
> > + "capture(%d): fmt = %c%c%c%c %d x %d, %d;",
> > + vb2_is_streaming(vq),
> > + inst->cap_format.pixfmt,
> > + inst->cap_format.pixfmt >> 8,
> > + inst->cap_format.pixfmt >> 16,
> > + inst->cap_format.pixfmt >> 24,
> > + inst->cap_format.width,
> > + inst->cap_format.height,
> > + vq->last_buffer_dequeued);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + for (i = 0; i < inst->cap_format.num_planes; i++) {
> > + num = scnprintf(str, sizeof(str), " %d(%d)",
> > + inst->cap_format.sizeimage[i],
> > + inst->cap_format.bytesperline[i]);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + }
> > + if (seq_write(s, "\n", 1))
> > + return 0;
> > + num = scnprintf(str, sizeof(str), "crop: (%d, %d) %d x %d\n",
> > + inst->crop.left,
> > + inst->crop.top,
> > + inst->crop.width,
> > + inst->crop.height);
> > + if (seq_write(s, str, num))
> > + return 0;
> > +
> > + vq = v4l2_m2m_get_src_vq(inst->m2m_ctx);
> > + for (i = 0; i < vq->num_buffers; i++) {
> > + struct vb2_buffer *vb = vq->bufs[i];
> > + struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
> > + struct vpu_vb2_buffer *vpu_buf = to_vpu_vb2_buffer(vbuf);
> > +
> > + if (vb->state == VB2_BUF_STATE_DEQUEUED)
> > + continue;
> > + num = scnprintf(str, sizeof(str),
> > + "output [%2d] state = %10s, %8s\n",
> > + i, vb2_stat_name[vb->state],
> > + vpu_stat_name[vpu_buf->state]);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + }
> > +
> > + vq = v4l2_m2m_get_dst_vq(inst->m2m_ctx);
> > + for (i = 0; i < vq->num_buffers; i++) {
> > + struct vb2_buffer *vb = vq->bufs[i];
> > + struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb);
> > + struct vpu_vb2_buffer *vpu_buf = to_vpu_vb2_buffer(vbuf);
> > +
> > + if (vb->state == VB2_BUF_STATE_DEQUEUED)
> > + continue;
> > + num = scnprintf(str, sizeof(str),
> > + "capture[%2d] state = %10s, %8s\n",
> > + i, vb2_stat_name[vb->state],
> > + vpu_stat_name[vpu_buf->state]);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + }
> > +
> > + num = scnprintf(str, sizeof(str), "sequence = %d\n", inst->sequence);
> > + if (seq_write(s, str, num))
> > + return 0;
> > +
> > + if (inst->use_stream_buffer) {
> > + num = scnprintf(str, sizeof(str), "stream_buffer = %d / %d,
> <%pad, 0x%x>\n",
> > + vpu_helper_get_used_space(inst),
> > + inst->stream_buffer.length,
> > + &inst->stream_buffer.phys,
> > + inst->stream_buffer.length);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + }
> > + num = scnprintf(str, sizeof(str), "kfifo len = 0x%x\n",
> kfifo_len(&inst->msg_fifo));
> > + if (seq_write(s, str, num))
> > + return 0;
> > +
> > + num = scnprintf(str, sizeof(str), "flow :\n");
> > + if (seq_write(s, str, num))
> > + return 0;
> > +
> > + mutex_lock(&inst->core->cmd_lock);
> > + for (i = 0; i < ARRAY_SIZE(inst->flows); i++) {
> > + u32 idx = (inst->flow_idx + i) % (ARRAY_SIZE(inst->flows));
> > +
> > + if (!inst->flows[idx])
> > + continue;
> > + num = scnprintf(str, sizeof(str), "\t[%s]0x%x\n",
> > + inst->flows[idx] >= VPU_MSG_ID_NOOP ?
> "M" : "C",
> > + inst->flows[idx]);
> > + if (seq_write(s, str, num)) {
> > + mutex_unlock(&inst->core->cmd_lock);
> > + return 0;
> > + }
> > + }
> > + mutex_unlock(&inst->core->cmd_lock);
> > +
> > + i = 0;
> > + while (true) {
> > + num = call_vop(inst, get_debug_info, str, sizeof(str), i++);
> > + if (num <= 0)
> > + break;
> > + if (seq_write(s, str, num))
> > + return 0;
> > + }
> > +
> > + return 0;
> > +}
> > +
> > +static int vpu_dbg_core(struct seq_file *s, void *data)
> > +{
> > + struct vpu_core *core = s->private;
> > + struct vpu_shared_addr *iface = core->iface;
> > + char str[128];
> > + int num;
> > +
> > + num = scnprintf(str, sizeof(str), "[%s]\n",
> vpu_core_type_desc(core->type));
> > + if (seq_write(s, str, num))
> > + return 0;
> > +
> > + num = scnprintf(str, sizeof(str), "boot_region = <%pad, 0x%x>\n",
> > + &core->fw.phys, core->fw.length);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + num = scnprintf(str, sizeof(str), "rpc_region = <%pad, 0x%x> used =
> 0x%x\n",
> > + &core->rpc.phys, core->rpc.length,
> core->rpc.bytesused);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + num = scnprintf(str, sizeof(str), "fwlog_region = <%pad, 0x%x>\n",
> > + &core->log.phys, core->log.length);
> > + if (seq_write(s, str, num))
> > + return 0;
> > +
> > + num = scnprintf(str, sizeof(str), "state = %d\n", core->state);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + if (core->state == VPU_CORE_DEINIT)
> > + return 0;
> > + num = scnprintf(str, sizeof(str), "fw version = %d.%d.%d\n",
> > + (core->fw_version >> 16) & 0xff,
> > + (core->fw_version >> 8) & 0xff,
> > + core->fw_version & 0xff);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + num = scnprintf(str, sizeof(str), "instances = %d/%d (0x%02lx), %d\n",
> > + hweight32(core->instance_mask),
> > + core->supported_instance_count,
> > + core->instance_mask,
> > + core->request_count);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + num = scnprintf(str, sizeof(str), "kfifo len = 0x%x\n",
> kfifo_len(&core->msg_fifo));
> > + if (seq_write(s, str, num))
> > + return 0;
> > + num = scnprintf(str, sizeof(str),
> > + "cmd_buf:[0x%x, 0x%x], wptr = 0x%x, rptr =
> 0x%x\n",
> > + iface->cmd_desc->start,
> > + iface->cmd_desc->end,
> > + iface->cmd_desc->wptr,
> > + iface->cmd_desc->rptr);
> > + if (seq_write(s, str, num))
> > + return 0;
> > + num = scnprintf(str, sizeof(str),
> > + "msg_buf:[0x%x, 0x%x], wptr = 0x%x, rptr =
> 0x%x\n",
> > + iface->msg_desc->start,
> > + iface->msg_desc->end,
> > + iface->msg_desc->wptr,
> > + iface->msg_desc->rptr);
> > + if (seq_write(s, str, num))
> > + return 0;
> > +
> > + return 0;
> > +}
> > +
> > +static int vpu_dbg_fwlog(struct seq_file *s, void *data)
> > +{
> > + struct vpu_core *core = s->private;
> > + struct print_buf_desc *print_buf;
> > + int length;
> > + u32 rptr;
> > + u32 wptr;
> > + int ret = 0;
> > +
> > + if (!core->log.virt || core->state == VPU_CORE_DEINIT)
> > + return 0;
> > +
> > + print_buf = core->log.virt;
> > + rptr = print_buf->read;
> > + wptr = print_buf->write;
> > +
> > + if (rptr == wptr)
> > + return 0;
> > + else if (rptr < wptr)
> > + length = wptr - rptr;
> > + else
> > + length = print_buf->bytes + wptr - rptr;
> > +
> > + if (s->count + length >= s->size) {
> > + s->count = s->size;
> > + return 0;
> > + }
> > +
> > + if (rptr + length >= print_buf->bytes) {
> > + int num = print_buf->bytes - rptr;
> > +
> > + if (seq_write(s, print_buf->buffer + rptr, num))
> > + ret = -1;
> > + length -= num;
> > + rptr = 0;
> > + }
> > +
> > + if (length) {
> > + if (seq_write(s, print_buf->buffer + rptr, length))
> > + ret = -1;
> > + rptr += length;
> > + }
> > + if (!ret)
> > + print_buf->read = rptr;
> > +
> > + return 0;
> > +}
> > +
> > +static int vpu_dbg_inst_open(struct inode *inode, struct file *filp)
> > +{
> > + return single_open(filp, vpu_dbg_instance, inode->i_private);
> > +}
> > +
> > +static ssize_t vpu_dbg_inst_write(struct file *file,
> > + const char __user *user_buf, size_t size, loff_t
> *ppos)
> > +{
> > + struct seq_file *s = file->private_data;
> > + struct vpu_inst *inst = s->private;
> > +
> > + vpu_session_debug(inst);
> > +
> > + return size;
> > +}
> > +
> > +static ssize_t vpu_dbg_core_write(struct file *file,
> > + const char __user *user_buf, size_t size, loff_t
> *ppos)
> > +{
> > + struct seq_file *s = file->private_data;
> > + struct vpu_core *core = s->private;
> > +
> > + pm_runtime_get_sync(core->dev);
> > + mutex_lock(&core->lock);
> > + if (core->state != VPU_CORE_DEINIT && !core->instance_mask) {
> > + core_dbg(core, LVL_INFO, "reset core\n");
> > + if (!vpu_core_sw_reset(core)) {
> > + core->state = VPU_CORE_ACTIVE;
> > + core->hang_mask = 0;
> > + }
> > + }
> > + mutex_unlock(&core->lock);
> > + pm_runtime_put_sync(core->dev);
> > +
> > + return size;
> > +}
> > +
> > +static int vpu_dbg_core_open(struct inode *inode, struct file *filp)
> > +{
> > + return single_open(filp, vpu_dbg_core, inode->i_private);
> > +}
> > +
> > +static int vpu_dbg_fwlog_open(struct inode *inode, struct file *filp)
> > +{
> > + return single_open(filp, vpu_dbg_fwlog, inode->i_private);
> > +}
> > +
> > +static const struct file_operations vpu_dbg_inst_fops = {
> > + .owner = THIS_MODULE,
> > + .open = vpu_dbg_inst_open,
> > + .release = single_release,
> > + .read = seq_read,
> > + .write = vpu_dbg_inst_write,
> > +};
> > +
> > +static const struct file_operations vpu_dbg_core_fops = {
> > + .owner = THIS_MODULE,
> > + .open = vpu_dbg_core_open,
> > + .release = single_release,
> > + .read = seq_read,
> > + .write = vpu_dbg_core_write,
> > +};
> > +
> > +static const struct file_operations vpu_dbg_fwlog_fops = {
> > + .owner = THIS_MODULE,
> > + .open = vpu_dbg_fwlog_open,
> > + .release = single_release,
> > + .read = seq_read,
> > +};
> > +
> > +int vpu_inst_create_dbgfs_file(struct vpu_inst *inst)
> > +{
> > + struct vpu_dev *vpu;
> > + char name[64];
> > +
> > + if (!inst || !inst->core || !inst->core->vpu)
> > + return -EINVAL;
> > +
> > + vpu = inst->core->vpu;
> > + if (!vpu->debugfs)
> > + return -EINVAL;
> > +
> > + if (inst->debugfs)
> > + return 0;
> > +
> > + scnprintf(name, sizeof(name), "instance.%d.%d",
> > + inst->core->id, inst->id);
> > + inst->debugfs = debugfs_create_file((const char *)name,
> > + VERIFY_OCTAL_PERMISSIONS(0644),
> > + vpu->debugfs,
> > + inst,
> > + &vpu_dbg_inst_fops);
> > + if (!inst->debugfs) {
> > + vpu_err("vpu create debugfs %s fail\n", name);
> > + return -EINVAL;
> > + }
> > +
> > + return 0;
> > +}
> > +
> > +int vpu_inst_remove_dbgfs_file(struct vpu_inst *inst)
> > +{
> > + if (!inst)
> > + return 0;
> > +
> > + debugfs_remove(inst->debugfs);
> > + inst->debugfs = NULL;
> > +
> > + return 0;
> > +}
> > +
> > +int vpu_core_create_dbgfs_file(struct vpu_core *core)
> > +{
> > + struct vpu_dev *vpu;
> > + char name[64];
> > +
> > + if (!core || !core->vpu)
> > + return -EINVAL;
> > +
> > + vpu = core->vpu;
> > + if (!vpu->debugfs)
> > + return -EINVAL;
> > +
> > + if (!core->debugfs) {
> > + scnprintf(name, sizeof(name), "core.%d", core->id);
> > + core->debugfs = debugfs_create_file((const char *)name,
> > +
> VERIFY_OCTAL_PERMISSIONS(0644),
> > + vpu->debugfs,
> > + core,
> > + &vpu_dbg_core_fops);
> > + if (!core->debugfs) {
> > + vpu_err("vpu create debugfs %s fail\n", name);
> > + return -EINVAL;
> > + }
> > + }
> > + if (!core->debugfs_fwlog) {
> > + scnprintf(name, sizeof(name), "fwlog.%d", core->id);
> > + core->debugfs_fwlog = debugfs_create_file((const char
> *)name,
> > +
> VERIFY_OCTAL_PERMISSIONS(0444),
> > + vpu->debugfs,
> > + core,
> > + &vpu_dbg_fwlog_fops);
> > + if (!core->debugfs_fwlog) {
> > + vpu_err("vpu create debugfs %s fail\n", name);
> > + return -EINVAL;
> > + }
> > + }
> > +
> > + return 0;
> > +}
> > +
> > +int vpu_core_remove_dbgfs_file(struct vpu_core *core)
> > +{
> > + if (!core)
> > + return 0;
> > + debugfs_remove(core->debugfs);
> > + core->debugfs = NULL;
> > + debugfs_remove(core->debugfs_fwlog);
> > + core->debugfs_fwlog = NULL;
> > +
> > + return 0;
> > +}
> > +
> > +void vpu_inst_record_flow(struct vpu_inst *inst, u32 flow)
> > +{
> > + if (!inst)
> > + return;
> > +
> > + inst->flows[inst->flow_idx] = flow;
> > + inst->flow_idx = (inst->flow_idx + 1) % (ARRAY_SIZE(inst->flows));
> > +}
> > diff --git a/drivers/media/platform/amphion/vpu_log.h
> b/drivers/media/platform/amphion/vpu_log.h
> > new file mode 100644
> > index 000000000000..fec98368fad1
> > --- /dev/null
> > +++ b/drivers/media/platform/amphion/vpu_log.h
> > @@ -0,0 +1,44 @@
> > +/* SPDX-License-Identifier: GPL-2.0 */
> > +/*
> > + * Copyright 2020-2021 NXP
> > + */
> > +
> > +#ifndef _AMPHION_VPU_LOG_H
> > +#define _AMPHION_VPU_LOG_H
> > +
> > +#define LVL_ERR (1 << 0)
> > +#define LVL_WARN (1 << 1)
> > +#define LVL_INFO (1 << 2)
> > +#define LVL_DEBUG (1 << 3)
> > +#define LVL_IRQ (1 << 4)
> > +#define LVL_CMD (1 << 5)
> > +#define LVL_EVT (1 << 6)
> > +#define LVL_CTRL (1 << 7)
> > +#define LVL_TS (1 << 8)
> > +#define LVL_FLOW (1 << 13)
> > +
> > +extern unsigned int vpu_dbg_level;
> > +
> > +#ifdef TAG
> > +#define vpu_dbg(level, fmt, arg...) \
> > + do { \
> > + if ((vpu_dbg_level & (level)) || ((level) & LVL_ERR)) \
> > + pr_info("[VPU "TAG"]"fmt, ## arg); \
> > + } while (0)
> > +#else
> > +#define vpu_dbg(level, fmt, arg...) \
> > + do { \
> > + if ((vpu_dbg_level & (level)) || ((level) & LVL_ERR)) \
> > + pr_info("[VPU]"fmt, ## arg); \
> > + } while (0)
> > +#endif
> > +
> > +#define vpu_err(fmt, arg...) vpu_dbg(LVL_ERR, fmt, ##arg)
> > +#define inst_dbg(inst, level, fmt, arg...) \
> > + vpu_dbg(level, "[%d:%d] "fmt, inst->core->id, inst->id, ## arg)
> > +#define inst_err(inst, fmt, arg...) inst_dbg(inst, LVL_ERR, fmt, ## arg)
> > +#define core_dbg(core, level, fmt, arg...) \
> > + vpu_dbg(level, "[%d] %s "fmt, core->id,
> vpu_core_type_desc(core->type), ## arg)
> > +#define core_err(core, fmt, arg...) core_dbg(core, LVL_ERR, fmt, ## arg)
> > +
> > +#endif
> > diff --git a/drivers/media/platform/amphion/vpu_rpc.c
> b/drivers/media/platform/amphion/vpu_rpc.c
> > new file mode 100644
> > index 000000000000..21c95b20d48d
> > --- /dev/null
> > +++ b/drivers/media/platform/amphion/vpu_rpc.c
> > @@ -0,0 +1,263 @@
> > +// SPDX-License-Identifier: GPL-2.0
> > +/*
> > + * Copyright 2020-2021 NXP
> > + */
> > +
> > +#define TAG "RPC"
> > +#include <linux/init.h>
> > +#include <linux/interconnect.h>
> > +#include <linux/ioctl.h>
> > +#include <linux/list.h>
> > +#include <linux/kernel.h>
> > +#include <linux/module.h>
> > +#include <linux/of_device.h>
> > +#include <linux/of_address.h>
> > +#include <linux/platform_device.h>
> > +#include <linux/firmware/imx/ipc.h>
> > +#include <linux/firmware/imx/svc/misc.h>
> > +#include "vpu.h"
> > +#include "vpu_rpc.h"
> > +#include "vpu_imx8q.h"
> > +#include "vpu_windsor.h"
> > +#include "vpu_malone.h"
> > +#include "vpu_log.h"
> > +
> > +u32 vpu_iface_check_memory_region(struct vpu_core *core, dma_addr_t
> addr, u32 size)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->check_memory_region)
> > + return VPU_CORE_MEMORY_INVALID;
> > +
> > + return ops->check_memory_region(core->fw.phys, addr, size);
> > +}
> > +
> > +static u32 vpu_rpc_check_buffer_space(struct vpu_rpc_buffer_desc *desc,
> bool write)
> > +{
> > + u32 ptr1;
> > + u32 ptr2;
> > + u32 size;
> > +
> > + WARN_ON(!desc);
> > +
> > + size = desc->end - desc->start;
> > + if (write) {
> > + ptr1 = desc->wptr;
> > + ptr2 = desc->rptr;
> > + } else {
> > + ptr1 = desc->rptr;
> > + ptr2 = desc->wptr;
> > + }
> > +
> > + if (ptr1 == ptr2) {
> > + if (!write)
> > + return 0;
> > + else
> > + return size;
> > + }
> > +
> > + return (ptr2 + size - ptr1) % size;
> > +}
> > +
> > +static int vpu_rpc_send_cmd_buf(struct vpu_shared_addr *shared,
> > + struct vpu_rpc_event *cmd)
> > +{
> > + struct vpu_rpc_buffer_desc *desc;
> > + u32 space = 0;
> > + u32 *data;
> > + u32 wptr;
> > + u32 i;
> > +
> > + WARN_ON(!shared || !shared->cmd_mem_vir || !cmd);
> > +
> > + desc = shared->cmd_desc;
> > + space = vpu_rpc_check_buffer_space(desc, true);
> > + if (space < (((cmd->hdr.num + 1) << 2) + 16)) {
> > + vpu_err("Cmd Buffer is no space for [%d] %d\n",
> > + cmd->hdr.index, cmd->hdr.id);
> > + return -EINVAL;
> > + }
> > + wptr = desc->wptr;
> > + data = (u32 *)(shared->cmd_mem_vir + desc->wptr - desc->start);
> > + *data = 0;
> > + *data |= ((cmd->hdr.index & 0xff) << 24);
> > + *data |= ((cmd->hdr.num & 0xff) << 16);
> > + *data |= (cmd->hdr.id & 0x3fff);
> > + wptr += 4;
> > + data++;
> > + if (wptr >= desc->end) {
> > + wptr = desc->start;
> > + data = shared->cmd_mem_vir;
> > + }
> > +
> > + for (i = 0; i < cmd->hdr.num; i++) {
> > + *data = cmd->data[i];
> > + wptr += 4;
> > + data++;
> > + if (wptr >= desc->end) {
> > + wptr = desc->start;
> > + data = shared->cmd_mem_vir;
> > + }
> > + }
> > +
> > + /*update wptr after data is written*/
> > + mb();
> > + desc->wptr = wptr;
> > +
> > + return 0;
> > +}
> > +
> > +static bool vpu_rpc_check_msg(struct vpu_shared_addr *shared)
> > +{
> > + struct vpu_rpc_buffer_desc *desc;
> > + u32 space = 0;
> > + u32 msgword;
> > + u32 msgnum;
> > +
> > + WARN_ON(!shared || !shared->msg_desc);
> > +
> > + desc = shared->msg_desc;
> > + space = vpu_rpc_check_buffer_space(desc, 0);
> > + space = (space >> 2);
> > +
> > + if (space) {
> > + msgword = *(u32 *)(shared->msg_mem_vir + desc->rptr -
> desc->start);
> > + msgnum = (msgword & 0xff0000) >> 16;
> > + if (msgnum <= space)
> > + return true;
> > + }
> > +
> > + return false;
> > +}
> > +
> > +static int vpu_rpc_receive_msg_buf(struct vpu_shared_addr *shared, struct
> vpu_rpc_event *msg)
> > +{
> > + struct vpu_rpc_buffer_desc *desc;
> > + u32 *data;
> > + u32 msgword;
> > + u32 rptr;
> > + u32 i;
> > +
> > + WARN_ON(!shared || !shared->msg_desc || !msg);
> > +
> > + if (!vpu_rpc_check_msg(shared))
> > + return -EINVAL;
> > +
> > + desc = shared->msg_desc;
> > + data = (u32 *)(shared->msg_mem_vir + desc->rptr - desc->start);
> > + rptr = desc->rptr;
> > + msgword = *data;
> > + data++;
> > + rptr += 4;
> > + if (rptr >= desc->end) {
> > + rptr = desc->start;
> > + data = shared->msg_mem_vir;
> > + }
> > +
> > + msg->hdr.index = (msgword >> 24) & 0xff;
> > + msg->hdr.num = (msgword >> 16) & 0xff;
> > + msg->hdr.id = msgword & 0x3fff;
> > +
> > + if (msg->hdr.num > ARRAY_SIZE(msg->data)) {
> > + vpu_err("msg(%d) data length(%d) is out of range\n",
> > + msg->hdr.id, msg->hdr.num);
> > + return -EINVAL;
> > + }
> > +
> > + for (i = 0; i < msg->hdr.num; i++) {
> > + msg->data[i] = *data;
> > + data++;
> > + rptr += 4;
> > + if (rptr >= desc->end) {
> > + rptr = desc->start;
> > + data = shared->msg_mem_vir;
> > + }
> > + }
> > +
> > + /*update rptr after data is read*/
> > + mb();
> > + desc->rptr = rptr;
> > +
> > + return 0;
> > +}
> > +
> > +struct vpu_iface_ops imx8q_rpc_ops[] = {
> > + [VPU_CORE_TYPE_ENC] = {
> > + .check_codec = vpu_imx8q_check_codec,
> > + .check_fmt = vpu_imx8q_check_fmt,
> > + .boot_core = vpu_imx8q_boot_core,
> > + .get_power_state = vpu_imx8q_get_power_state,
> > + .on_firmware_loaded = vpu_imx8q_on_firmware_loaded,
> > + .get_data_size = vpu_windsor_get_data_size,
> > + .check_memory_region =
> vpu_imx8q_check_memory_region,
> > + .init_rpc = vpu_windsor_init_rpc,
> > + .set_log_buf = vpu_windsor_set_log_buf,
> > + .set_system_cfg = vpu_windsor_set_system_cfg,
> > + .get_version = vpu_windsor_get_version,
> > + .send_cmd_buf = vpu_rpc_send_cmd_buf,
> > + .receive_msg_buf = vpu_rpc_receive_msg_buf,
> > + .pack_cmd = vpu_windsor_pack_cmd,
> > + .convert_msg_id = vpu_windsor_convert_msg_id,
> > + .unpack_msg_data = vpu_windsor_unpack_msg_data,
> > + .config_memory_resource =
> vpu_windsor_config_memory_resource,
> > + .get_stream_buffer_size =
> vpu_windsor_get_stream_buffer_size,
> > + .config_stream_buffer = vpu_windsor_config_stream_buffer,
> > + .get_stream_buffer_desc =
> vpu_windsor_get_stream_buffer_desc,
> > + .update_stream_buffer =
> vpu_windsor_update_stream_buffer,
> > + .set_encode_params = vpu_windsor_set_encode_params,
> > + .input_frame = vpu_windsor_input_frame,
> > + .get_max_instance_count =
> vpu_windsor_get_max_instance_count,
> > + },
> > + [VPU_CORE_TYPE_DEC] = {
> > + .check_codec = vpu_imx8q_check_codec,
> > + .check_fmt = vpu_imx8q_check_fmt,
> > + .boot_core = vpu_imx8q_boot_core,
> > + .get_power_state = vpu_imx8q_get_power_state,
> > + .on_firmware_loaded = vpu_imx8q_on_firmware_loaded,
> > + .get_data_size = vpu_malone_get_data_size,
> > + .check_memory_region =
> vpu_imx8q_check_memory_region,
> > + .init_rpc = vpu_malone_init_rpc,
> > + .set_log_buf = vpu_malone_set_log_buf,
> > + .set_system_cfg = vpu_malone_set_system_cfg,
> > + .get_version = vpu_malone_get_version,
> > + .send_cmd_buf = vpu_rpc_send_cmd_buf,
> > + .receive_msg_buf = vpu_rpc_receive_msg_buf,
> > + .get_stream_buffer_size =
> vpu_malone_get_stream_buffer_size,
> > + .config_stream_buffer = vpu_malone_config_stream_buffer,
> > + .set_decode_params = vpu_malone_set_decode_params,
> > + .pack_cmd = vpu_malone_pack_cmd,
> > + .convert_msg_id = vpu_malone_convert_msg_id,
> > + .unpack_msg_data = vpu_malone_unpack_msg_data,
> > + .get_stream_buffer_desc =
> vpu_malone_get_stream_buffer_desc,
> > + .update_stream_buffer =
> vpu_malone_update_stream_buffer,
> > + .add_scode = vpu_malone_add_scode,
> > + .input_frame = vpu_malone_input_frame,
> > + .pre_send_cmd = vpu_malone_pre_cmd,
> > + .post_send_cmd = vpu_malone_post_cmd,
> > + .init_instance = vpu_malone_init_instance,
> > + .get_max_instance_count =
> vpu_malone_get_max_instance_count,
> > + },
> > +};
> > +
> > +struct vpu_iface_ops *vpu_iface_get(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *rpc_ops = NULL;
> > + u32 size = 0;
> > +
> > + WARN_ON(!core || !core->vpu || !core->vpu->res);
> > +
> > + switch (core->vpu->res->plat_type) {
> > + case IMX8QXP:
> > + case IMX8QM:
> > + rpc_ops = imx8q_rpc_ops;
> > + size = ARRAY_SIZE(imx8q_rpc_ops);
> > + break;
> > + default:
> > + return NULL;
> > + }
> > +
> > + if (core->type >= size)
> > + return NULL;
> > +
> > + return &rpc_ops[core->type];
> > +}
> > diff --git a/drivers/media/platform/amphion/vpu_rpc.h
> b/drivers/media/platform/amphion/vpu_rpc.h
> > new file mode 100644
> > index 000000000000..b066a83e67d3
> > --- /dev/null
> > +++ b/drivers/media/platform/amphion/vpu_rpc.h
> > @@ -0,0 +1,463 @@
> > +/* SPDX-License-Identifier: GPL-2.0 */
> > +/*
> > + * Copyright 2020-2021 NXP
> > + */
> > +
> > +#ifndef _AMPHION_VPU_RPC_H
> > +#define _AMPHION_VPU_RPC_H
> > +
> > +#include <media/videobuf2-core.h>
> > +#include "vpu_codec.h"
> > +
> > +struct vpu_rpc_buffer_desc {
> > + u32 wptr;
> > + u32 rptr;
> > + u32 start;
> > + u32 end;
> > +};
> > +
> > +struct vpu_shared_addr {
> > + void *iface;
> > + struct vpu_rpc_buffer_desc *cmd_desc;
> > + void *cmd_mem_vir;
> > + struct vpu_rpc_buffer_desc *msg_desc;
> > + void *msg_mem_vir;
> > +
> > + unsigned long boot_addr;
> > + struct vpu_core *core;
> > + void *priv;
> > +};
> > +
> > +struct vpu_rpc_event_header {
> > + u32 index;
> > + u32 id;
> > + u32 num;
> > +};
> > +
> > +struct vpu_rpc_event {
> > + struct vpu_rpc_event_header hdr;
> > + u32 data[128];
> > +};
> > +
> > +struct vpu_iface_ops {
> > + bool (*check_codec)(enum vpu_core_type type);
> > + bool (*check_fmt)(enum vpu_core_type type, u32 pixelfmt);
> > + u32 (*get_data_size)(void);
> > + u32 (*check_memory_region)(dma_addr_t base, dma_addr_t addr,
> u32 size);
> > + int (*boot_core)(struct vpu_core *core);
> > + int (*shutdown_core)(struct vpu_core *core);
> > + int (*restore_core)(struct vpu_core *core);
> > + int (*get_power_state)(struct vpu_core *core);
> > + int (*on_firmware_loaded)(struct vpu_core *core);
> > + void (*init_rpc)(struct vpu_shared_addr *shared,
> > + struct vpu_buffer *rpc, dma_addr_t boot_addr);
> > + void (*set_log_buf)(struct vpu_shared_addr *shared,
> > + struct vpu_buffer *log);
> > + void (*set_system_cfg)(struct vpu_shared_addr *shared,
> > + u32 regs_base, void __iomem *regs, u32 index);
> > + void (*set_stream_cfg)(struct vpu_shared_addr *shared, u32 index);
> > + u32 (*get_version)(struct vpu_shared_addr *shared);
> > + u32 (*get_max_instance_count)(struct vpu_shared_addr *shared);
> > + int (*get_stream_buffer_size)(struct vpu_shared_addr *shared);
> > + int (*send_cmd_buf)(struct vpu_shared_addr *shared,
> > + struct vpu_rpc_event *cmd);
> > + int (*receive_msg_buf)(struct vpu_shared_addr *shared,
> > + struct vpu_rpc_event *msg);
> > + int (*pack_cmd)(struct vpu_rpc_event *pkt, u32 index, u32 id, void
> *data);
> > + int (*convert_msg_id)(u32 msg_id);
> > + int (*unpack_msg_data)(struct vpu_rpc_event *pkt, void *data);
> > + int (*input_frame)(struct vpu_shared_addr *shared,
> > + struct vpu_inst *inst, struct vb2_buffer *vb);
> > + int (*config_memory_resource)(struct vpu_shared_addr *shared,
> > + u32 instance,
> > + u32 type,
> > + u32 index,
> > + struct vpu_buffer *buf);
> > + int (*config_stream_buffer)(struct vpu_shared_addr *shared,
> > + u32 instance,
> > + struct vpu_buffer *buf);
> > + int (*update_stream_buffer)(struct vpu_shared_addr *shared,
> > + u32 instance, u32 ptr, bool
> write);
> > + int (*get_stream_buffer_desc)(struct vpu_shared_addr *shared,
> > + u32 instance,
> > + struct vpu_rpc_buffer_desc
> *desc);
> > + int (*set_encode_params)(struct vpu_shared_addr *shared,
> > + u32 instance,
> > + struct vpu_encode_params *params, u32 update);
> > + int (*set_decode_params)(struct vpu_shared_addr *shared,
> > + u32 instance,
> > + struct vpu_decode_params *params, u32 update);
> > + int (*add_scode)(struct vpu_shared_addr *shared,
> > + u32 instance,
> > + struct vpu_buffer *stream_buffer,
> > + u32 pixelformat,
> > + u32 scode_type);
> > + int (*pre_send_cmd)(struct vpu_shared_addr *shared, u32 instance);
> > + int (*post_send_cmd)(struct vpu_shared_addr *shared, u32 instance);
> > + int (*init_instance)(struct vpu_shared_addr *shared, u32 instance);
> > +};
> > +
> > +enum {
> > + VPU_CORE_MEMORY_INVALID = 0,
> > + VPU_CORE_MEMORY_CACHED,
> > + VPU_CORE_MEMORY_UNCACHED
> > +};
> > +
> > +struct vpu_rpc_region_t {
> > + dma_addr_t start;
> > + dma_addr_t end;
> > + dma_addr_t type;
> > +};
> > +
> > +struct vpu_iface_ops *vpu_iface_get(struct vpu_core *core);
> > +u32 vpu_iface_check_memory_region(struct vpu_core *core, dma_addr_t
> addr, u32 size);
> > +
> > +static inline bool vpu_iface_check_codec(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (ops && ops->check_codec)
> > + return ops->check_codec(core->type);
> > +
> > + return true;
> > +}
> > +
> > +static inline bool vpu_iface_check_format(struct vpu_core *core, u32
> pixelfmt)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (ops && ops->check_fmt)
> > + return ops->check_fmt(core->type, pixelfmt);
> > +
> > + return true;
> > +}
> > +
> > +static inline int vpu_iface_boot_core(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (ops && ops->boot_core)
> > + return ops->boot_core(core);
> > + return 0;
> > +}
> > +
> > +static inline int vpu_iface_get_power_state(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (ops && ops->get_power_state)
> > + return ops->get_power_state(core);
> > + return 1;
> > +}
> > +
> > +static inline int vpu_iface_shutdown_core(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (ops && ops->shutdown_core)
> > + return ops->shutdown_core(core);
> > + return 0;
> > +}
> > +
> > +static inline int vpu_iface_restore_core(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (ops && ops->restore_core)
> > + return ops->restore_core(core);
> > + return -EINVAL;
> > +}
> > +
> > +static inline int vpu_iface_on_firmware_loaded(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (ops && ops->on_firmware_loaded)
> > + return ops->on_firmware_loaded(core);
> > +
> > + return 0;
> > +}
> > +
> > +static inline u32 vpu_iface_get_data_size(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->get_data_size)
> > + return 0;
> > +
> > + return ops->get_data_size();
> > +}
> > +
> > +static inline int vpu_iface_init(struct vpu_core *core,
> > + struct vpu_shared_addr *shared,
> > + struct vpu_buffer *rpc,
> > + dma_addr_t boot_addr)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->init_rpc)
> > + return -EINVAL;
> > +
> > + ops->init_rpc(shared, rpc, boot_addr);
> > + core->iface = shared;
> > + shared->core = core;
> > + if (rpc->bytesused > rpc->length)
> > + return -ENOSPC;
> > + return 0;
> > +}
> > +
> > +static inline int vpu_iface_set_log_buf(struct vpu_core *core,
> > + struct vpu_buffer *log)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops)
> > + return -EINVAL;
> > +
> > + if (ops->set_log_buf)
> > + ops->set_log_buf(core->iface, log);
> > +
> > + return 0;
> > +}
> > +
> > +static inline int vpu_iface_config_system(struct vpu_core *core,
> > + u32 regs_base, void __iomem *regs)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops)
> > + return -EINVAL;
> > + if (ops->set_system_cfg)
> > + ops->set_system_cfg(core->iface, regs_base, regs, core->id);
> > +
> > + return 0;
> > +}
> > +
> > +static inline int vpu_iface_get_stream_buffer_size(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->get_stream_buffer_size)
> > + return 0;
> > +
> > + return ops->get_stream_buffer_size(core->iface);
> > +}
> > +
> > +static inline int vpu_iface_config_stream(struct vpu_inst *inst)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (!ops)
> > + return -EINVAL;
> > + if (ops->set_stream_cfg)
> > + ops->set_stream_cfg(inst->core->iface, inst->id);
> > + return 0;
> > +}
> > +
> > +static inline int vpu_iface_send_cmd(struct vpu_core *core, struct
> vpu_rpc_event *cmd)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->send_cmd_buf)
> > + return -EINVAL;
> > +
> > + return ops->send_cmd_buf(core->iface, cmd);
> > +}
> > +
> > +static inline int vpu_iface_receive_msg(struct vpu_core *core, struct
> vpu_rpc_event *msg)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->receive_msg_buf)
> > + return -EINVAL;
> > +
> > + return ops->receive_msg_buf(core->iface, msg);
> > +}
> > +
> > +static inline int vpu_iface_pack_cmd(struct vpu_core *core,
> > + struct vpu_rpc_event *pkt,
> > + u32 index, u32 id, void *data)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->pack_cmd)
> > + return -EINVAL;
> > + return ops->pack_cmd(pkt, index, id, data);
> > +}
> > +
> > +static inline int vpu_iface_convert_msg_id(struct vpu_core *core, u32
> msg_id)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->convert_msg_id)
> > + return -EINVAL;
> > +
> > + return ops->convert_msg_id(msg_id);
> > +}
> > +
> > +static inline int vpu_iface_unpack_msg_data(struct vpu_core *core,
> > + struct vpu_rpc_event
> *pkt, void *data)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->unpack_msg_data)
> > + return -EINVAL;
> > +
> > + return ops->unpack_msg_data(pkt, data);
> > +}
> > +
> > +static inline int vpu_iface_input_frame(struct vpu_inst *inst,
> > + struct vb2_buffer *vb)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + if (!ops || !ops->input_frame)
> > + return -EINVAL;
> > +
> > + return ops->input_frame(inst->core->iface, inst, vb);
> > +}
> > +
> > +static inline int vpu_iface_config_memory_resource(struct vpu_inst *inst,
> > + u32 type, u32 index, struct vpu_buffer *buf)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (!ops || !ops->config_memory_resource)
> > + return -EINVAL;
> > +
> > + return ops->config_memory_resource(inst->core->iface,
> > + inst->id,
> > + type, index, buf);
> > +}
> > +
> > +static inline int vpu_iface_config_stream_buffer(struct vpu_inst *inst,
> > + struct vpu_buffer *buf)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (!ops || !ops->config_stream_buffer)
> > + return -EINVAL;
> > +
> > + return ops->config_stream_buffer(inst->core->iface, inst->id, buf);
> > +}
> > +
> > +static inline int vpu_iface_update_stream_buffer(struct vpu_inst *inst,
> > + u32 ptr, bool write)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (!ops || !ops->update_stream_buffer)
> > + return -EINVAL;
> > +
> > + return ops->update_stream_buffer(inst->core->iface, inst->id, ptr,
> write);
> > +}
> > +
> > +static inline int vpu_iface_get_stream_buffer_desc(struct vpu_inst *inst,
> > + struct vpu_rpc_buffer_desc
> *desc)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (!ops || !ops->get_stream_buffer_desc)
> > + return -EINVAL;
> > +
> > + if (!desc)
> > + return 0;
> > +
> > + return ops->get_stream_buffer_desc(inst->core->iface, inst->id, desc);
> > +}
> > +
> > +static inline u32 vpu_iface_get_version(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->get_version)
> > + return 0;
> > +
> > + return ops->get_version(core->iface);
> > +}
> > +
> > +static inline u32 vpu_iface_get_max_instance_count(struct vpu_core *core)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(core);
> > +
> > + if (!ops || !ops->get_max_instance_count)
> > + return 0;
> > +
> > + return ops->get_max_instance_count(core->iface);
> > +}
> > +
> > +static inline int vpu_iface_set_encode_params(struct vpu_inst *inst,
> > + struct vpu_encode_params *params, u32
> update)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (!ops || !ops->set_encode_params)
> > + return -EINVAL;
> > +
> > + return ops->set_encode_params(inst->core->iface, inst->id, params,
> update);
> > +}
> > +
> > +static inline int vpu_iface_set_decode_params(struct vpu_inst *inst,
> > + struct vpu_decode_params *params, u32
> update)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (!ops || !ops->set_decode_params)
> > + return -EINVAL;
> > +
> > + return ops->set_decode_params(inst->core->iface, inst->id, params,
> update);
> > +}
> > +
> > +static inline int vpu_iface_add_scode(struct vpu_inst *inst, u32 scode_type)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (!ops || !ops->add_scode)
> > + return -EINVAL;
> > +
> > + return ops->add_scode(inst->core->iface, inst->id,
> > + &inst->stream_buffer,
> > + inst->out_format.pixfmt,
> > + scode_type);
> > +}
> > +
> > +static inline int vpu_iface_pre_send_cmd(struct vpu_inst *inst)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (ops && ops->pre_send_cmd)
> > + return ops->pre_send_cmd(inst->core->iface, inst->id);
> > + return 0;
> > +}
> > +
> > +static inline int vpu_iface_post_send_cmd(struct vpu_inst *inst)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (ops && ops->post_send_cmd)
> > + return ops->post_send_cmd(inst->core->iface, inst->id);
> > + return 0;
> > +}
> > +
> > +static inline int vpu_iface_init_instance(struct vpu_inst *inst)
> > +{
> > + struct vpu_iface_ops *ops = vpu_iface_get(inst->core);
> > +
> > + WARN_ON(inst->id < 0);
> > + if (ops && ops->init_instance)
> > + return ops->init_instance(inst->core->iface, inst->id);
> > +
> > + return 0;
> > +}
> > +
> > +#endif
>