Re: [PATCH v2] media: platform: add VPFE capture driver support for AM437X

From: Hans Verkuil
Date: Wed Dec 03 2014 - 05:57:52 EST


On 12/03/14 01:29, Lad, Prabhakar wrote:
> From: Benoit Parrot <bparrot@xxxxxx>
>
> This patch adds Video Processing Front End (VPFE) driver for
> AM437X family of devices
> Driver supports the following:
> - V4L2 API using MMAP buffer access based on videobuf2 api
> - Asynchronous sensor/decoder sub device registration
> - DT support
>
> Signed-off-by: Benoit Parrot <bparrot@xxxxxx>
> Signed-off-by: Darren Etheridge <detheridge@xxxxxx>
> Signed-off-by: Lad, Prabhakar <prabhakar.csengg@xxxxxxxxx>
> ---
> Changes for v2:
> a> Fixed review comments pointed by Hans.
>
> .../devicetree/bindings/media/ti-am437x-vpfe.txt | 61 +
> MAINTAINERS | 9 +
> drivers/media/platform/Kconfig | 1 +
> drivers/media/platform/Makefile | 2 +
> drivers/media/platform/am437x/Kconfig | 11 +
> drivers/media/platform/am437x/Makefile | 2 +
> drivers/media/platform/am437x/am437x-vpfe.c | 2785 ++++++++++++++++++++
> drivers/media/platform/am437x/am437x-vpfe.h | 287 ++
> drivers/media/platform/am437x/am437x-vpfe_regs.h | 140 +
> include/uapi/linux/Kbuild | 1 +
> include/uapi/linux/am437x-vpfe.h | 122 +
> 11 files changed, 3421 insertions(+)
> create mode 100644 Documentation/devicetree/bindings/media/ti-am437x-vpfe.txt
> create mode 100644 drivers/media/platform/am437x/Kconfig
> create mode 100644 drivers/media/platform/am437x/Makefile
> create mode 100644 drivers/media/platform/am437x/am437x-vpfe.c
> create mode 100644 drivers/media/platform/am437x/am437x-vpfe.h
> create mode 100644 drivers/media/platform/am437x/am437x-vpfe_regs.h
> create mode 100644 include/uapi/linux/am437x-vpfe.h
>
> diff --git a/Documentation/devicetree/bindings/media/ti-am437x-vpfe.txt b/Documentation/devicetree/bindings/media/ti-am437x-vpfe.txt
> new file mode 100644
> index 0000000..3932e76
> --- /dev/null
> +++ b/Documentation/devicetree/bindings/media/ti-am437x-vpfe.txt
> @@ -0,0 +1,61 @@
> +Texas Instruments AM437x CAMERA (VPFE)
> +--------------------------------------
> +
> +The Video Processing Front End (VPFE) is a key component for image capture
> +applications. The capture module provides the system interface and the
> +processing capability to connect RAW image-sensor modules and video decoders
> +to the AM437x device.
> +
> +Required properties:
> +- compatible: must be "ti,am437x-vpfe"
> +- reg: physical base address and length of the registers set for the device;
> +- interrupts: should contain IRQ line for the VPFE;
> +- ti,am437x-vpfe-interface: can be one of the following,
> + 0 - Raw Bayer Interface.
> + 1 - 8 Bit BT656 Interface.
> + 2 - 10 Bit BT656 Interface.
> + 3 - YCbCr 8 Bit Interface.
> + 4 - YCbCr 16 Bit Interface.
> +
> +VPFE supports a single port node with parallel bus. It should contain one
> +'port' child node with child 'endpoint' node. Please refer to the bindings
> +defined in Documentation/devicetree/bindings/media/video-interfaces.txt.
> +
> +Example:
> + vpfe: vpfe@f0034000 {
> + compatible = "ti,am437x-vpfe";
> + reg = <0x48328000 0x2000>;
> + interrupts = <GIC_SPI 50 IRQ_TYPE_LEVEL_HIGH>;
> +
> + pinctrl-names = "default", "sleep";
> + pinctrl-0 = <&vpfe_pins_default>;
> + pinctrl-1 = <&vpfe_pins_sleep>;
> +
> + port {
> + #address-cells = <1>;
> + #size-cells = <0>;
> +
> + vpfe0_ep: endpoint {
> + remote-endpoint = <&ov2659_1>;
> + ti,am437x-vpfe-interface = <0>;
> + bus-width = <8>;
> + hsync-active = <0>;
> + vsync-active = <0>;
> + };
> + };
> + };
> +
> + i2c1: i2c@4802a000 {
> +
> + ov2659@30 {
> + compatible = "ti,ov2659";
> + reg = <0x30>;
> +
> + port {
> + ov2659_1: endpoint {
> + remote-endpoint = <&vpfe0_ep>;
> + bus-width = <8>;
> + mclk-frequency = <12000000>;
> + };
> + };
> + };
> diff --git a/MAINTAINERS b/MAINTAINERS
> index a6288ca..a42d367 100644
> --- a/MAINTAINERS
> +++ b/MAINTAINERS
> @@ -8537,6 +8537,15 @@ S: Maintained
> F: drivers/media/platform/davinci/
> F: include/media/davinci/
>
> +TI AM437X VPFE DRIVER
> +M: Lad, Prabhakar <prabhakar.csengg@xxxxxxxxx>
> +L: linux-media@xxxxxxxxxxxxxxx
> +W: http://linuxtv.org/
> +Q: http://patchwork.linuxtv.org/project/linux-media/list/
> +T: git git://linuxtv.org/mhadli/v4l-dvb-davinci_devices.git
> +S: Maintained
> +F: drivers/media/platform/am437x/
> +
> SIS 190 ETHERNET DRIVER
> M: Francois Romieu <romieu@xxxxxxxxxxxxx>
> L: netdev@xxxxxxxxxxxxxxx
> diff --git a/drivers/media/platform/Kconfig b/drivers/media/platform/Kconfig
> index 0c61155..6d94045 100644
> --- a/drivers/media/platform/Kconfig
> +++ b/drivers/media/platform/Kconfig
> @@ -126,6 +126,7 @@ config VIDEO_S3C_CAMIF
> source "drivers/media/platform/soc_camera/Kconfig"
> source "drivers/media/platform/exynos4-is/Kconfig"
> source "drivers/media/platform/s5p-tv/Kconfig"
> +source "drivers/media/platform/am437x/Kconfig"
>
> endif # V4L_PLATFORM_DRIVERS
>
> diff --git a/drivers/media/platform/Makefile b/drivers/media/platform/Makefile
> index b818afb..7bb6d46 100644
> --- a/drivers/media/platform/Makefile
> +++ b/drivers/media/platform/Makefile
> @@ -49,4 +49,6 @@ obj-$(CONFIG_VIDEO_RENESAS_VSP1) += vsp1/
>
> obj-y += omap/
>
> +obj-$(CONFIG_VIDEO_AM437X_VPFE) += am437x/
> +
> ccflags-y += -I$(srctree)/drivers/media/i2c
> diff --git a/drivers/media/platform/am437x/Kconfig b/drivers/media/platform/am437x/Kconfig
> new file mode 100644
> index 0000000..97dea72
> --- /dev/null
> +++ b/drivers/media/platform/am437x/Kconfig
> @@ -0,0 +1,11 @@
> +config VIDEO_AM437X_VPFE
> + tristate "TI AM437x VPFE video capture driver"
> + depends on VIDEO_V4L2 && VIDEO_V4L2_SUBDEV_API
> + depends on SOC_AM43XX || COMPILE_TEST
> + select VIDEOBUF2_DMA_CONTIG
> + help
> + Support for AM437x Video Processing Front End based Video
> + Capture Driver.
> +
> + To compile this driver as a module, choose M here. The module
> + will be called ti_vpfe.
> diff --git a/drivers/media/platform/am437x/Makefile b/drivers/media/platform/am437x/Makefile
> new file mode 100644
> index 0000000..3c6b7bd
> --- /dev/null
> +++ b/drivers/media/platform/am437x/Makefile
> @@ -0,0 +1,2 @@
> +ti-vpfe-y := am437x-vpfe.o
> +obj-$(CONFIG_VIDEO_AM437X_VPFE) += ti-vpfe.o
> diff --git a/drivers/media/platform/am437x/am437x-vpfe.c b/drivers/media/platform/am437x/am437x-vpfe.c
> new file mode 100644
> index 0000000..f3db23d
> --- /dev/null
> +++ b/drivers/media/platform/am437x/am437x-vpfe.c
> @@ -0,0 +1,2785 @@
> +/*
> + * TI VPFE capture Driver
> + *
> + * Copyright (C) 2013 - 2014 Texas Instruments, Inc.
> + *
> + * Benoit Parrot <bparrot@xxxxxx>
> + * Lad, Prabhakar <prabhakar.csengg@xxxxxxxxx>
> + *
> + * This program is free software; you may redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation; version 2 of the License.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
> + * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
> + * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
> + * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
> + * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
> + * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
> + * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
> + * SOFTWARE.
> + */
> +
> +#include <linux/delay.h>
> +#include <linux/err.h>
> +#include <linux/init.h>
> +#include <linux/interrupt.h>
> +#include <linux/io.h>
> +#include <linux/module.h>
> +#include <linux/platform_device.h>
> +#include <linux/pm_runtime.h>
> +#include <linux/slab.h>
> +#include <linux/uaccess.h>
> +#include <linux/videodev2.h>
> +
> +#include <media/v4l2-common.h>
> +#include <media/v4l2-ctrls.h>
> +#include <media/v4l2-event.h>
> +#include <media/v4l2-of.h>
> +
> +#include "am437x-vpfe.h"
> +
> +#define VPFE_MODULE_NAME "vpfe"
> +#define VPFE_VERSION "0.1.0"
> +
> +static int debug;
> +module_param(debug, int, 0644);
> +MODULE_PARM_DESC(debug, "Debug level 0-8");
> +
> +#define vpfe_dbg(level, dev, fmt, arg...) \
> + v4l2_dbg(level, debug, &dev->v4l2_dev, fmt, ##arg)
> +#define vpfe_info(dev, fmt, arg...) \
> + v4l2_info(&dev->v4l2_dev, fmt, ##arg)
> +#define vpfe_err(dev, fmt, arg...) \
> + v4l2_err(&dev->v4l2_dev, fmt, ##arg)
> +
> +/* standard information */
> +struct vpfe_standard {
> + v4l2_std_id std_id;
> + unsigned int width;
> + unsigned int height;
> + struct v4l2_fract pixelaspect;
> + int frame_format;
> +};
> +
> +const struct vpfe_standard vpfe_standards[] = {
> + {V4L2_STD_525_60, 720, 480, {11, 10}, 1},
> + {V4L2_STD_625_50, 720, 576, {54, 59}, 1},
> +};
> +
> +struct bus_format {
> + unsigned int width;
> + unsigned int bpp;
> +};
> +
> +/*
> + * struct vpfe_fmt - VPFE media bus format information
> + * @name: V4L2 format description
> + * @code: V4L2 media bus format code
> + * @shifted: V4L2 media bus format code for the same pixel layout but
> + * shifted to be 8 bits per pixel. =0 if format is not shiftable.
> + * @pixelformat: V4L2 pixel format FCC identifier
> + * @width: Bits per pixel (when transferred over a bus)
> + * @bpp: Bytes per pixel (when stored in memory)
> + * @supported: Indicates format supported by subdev
> + */
> +struct vpfe_fmt {
> + const char *name;
> + u32 fourcc;
> + u32 code;
> + struct bus_format l;
> + struct bus_format s;
> + bool supported;
> + u32 index;
> +};
> +
> +static struct vpfe_fmt formats[] = {
> + {
> + .name = "YUV 4:2:2 packed, YCbYCr",
> + .fourcc = V4L2_PIX_FMT_YUYV,
> + .code = MEDIA_BUS_FMT_YUYV8_2X8,
> + .l.width = 10,
> + .l.bpp = 4,
> + .s.width = 8,
> + .s.bpp = 2,
> + .supported = false,
> + }, {
> + .name = "YUV 4:2:2 packed, CbYCrY",
> + .fourcc = V4L2_PIX_FMT_UYVY,
> + .code = MEDIA_BUS_FMT_UYVY8_2X8,
> + .l.width = 10,
> + .l.bpp = 4,
> + .s.width = 8,
> + .s.bpp = 2,
> + .supported = false,
> + }, {
> + .name = "YUV 4:2:2 packed, YCrYCb",
> + .fourcc = V4L2_PIX_FMT_YVYU,
> + .code = MEDIA_BUS_FMT_YVYU8_2X8,
> + .l.width = 10,
> + .l.bpp = 4,
> + .s.width = 8,
> + .s.bpp = 2,
> + .supported = false,
> + }, {
> + .name = "YUV 4:2:2 packed, CrYCbY",
> + .fourcc = V4L2_PIX_FMT_VYUY,
> + .code = MEDIA_BUS_FMT_VYUY8_2X8,
> + .l.width = 10,
> + .l.bpp = 4,
> + .s.width = 8,
> + .s.bpp = 2,
> + .supported = false,
> + }, {
> + .name = "RAW8 BGGR",
> + .fourcc = V4L2_PIX_FMT_SBGGR8,
> + .code = MEDIA_BUS_FMT_SBGGR8_1X8,
> + .l.width = 10,
> + .l.bpp = 2,
> + .s.width = 8,
> + .s.bpp = 1,
> + .supported = false,
> + }, {
> + .name = "RAW8 GBRG",
> + .fourcc = V4L2_PIX_FMT_SGBRG8,
> + .code = MEDIA_BUS_FMT_SGBRG8_1X8,
> + .l.width = 10,
> + .l.bpp = 2,
> + .s.width = 8,
> + .s.bpp = 1,
> + .supported = false,
> + }, {
> + .name = "RAW8 GRBG",
> + .fourcc = V4L2_PIX_FMT_SGRBG8,
> + .code = MEDIA_BUS_FMT_SGRBG8_1X8,
> + .l.width = 10,
> + .l.bpp = 2,
> + .s.width = 8,
> + .s.bpp = 1,
> + .supported = false,
> + }, {
> + .name = "RAW8 RGGB",
> + .fourcc = V4L2_PIX_FMT_SRGGB8,
> + .code = MEDIA_BUS_FMT_SRGGB8_1X8,
> + .l.width = 10,
> + .l.bpp = 2,
> + .s.width = 8,
> + .s.bpp = 1,
> + .supported = false,
> + }, {
> + .name = "RGB565 (LE)",
> + .fourcc = V4L2_PIX_FMT_RGB565,
> + .code = MEDIA_BUS_FMT_RGB565_2X8_LE,
> + .l.width = 10,
> + .l.bpp = 4,
> + .s.width = 8,
> + .s.bpp = 2,
> + .supported = false,
> + }, {
> + .name = "RGB565 (BE)",
> + .fourcc = V4L2_PIX_FMT_RGB565X,
> + .code = MEDIA_BUS_FMT_RGB565_2X8_BE,
> + .l.width = 10,
> + .l.bpp = 4,
> + .s.width = 8,
> + .s.bpp = 2,
> + .supported = false,
> + },
> +};
> +
> +static int
> +__vpfe_get_format(struct vpfe_device *vpfe,
> + struct v4l2_format *format, unsigned int *bpp);
> +
> +static struct vpfe_fmt *find_format_by_code(unsigned int code, bool set)
> +{
> + struct vpfe_fmt *fmt;
> + static u32 index;
> + unsigned int k;
> +
> + for (k = 0; k < ARRAY_SIZE(formats); k++) {
> + fmt = &formats[k];
> + if (fmt->code == code) {
> + if (set) {
> + fmt->supported = true;
> + fmt->index = index++;
> + }

It's ugly and unexpected to set these fields here.
Why not do that where you actually need it?

E.g. instead of calling find_format_by_code(code, true) do:

fmt = find_format_by_code(code);
if (fmt) {
fmt->supported = true;
fmt->index = index++;
}

That's much cleaner.

> + return fmt;
> + }
> + }
> +
> + return NULL;
> +}
> +
> +static struct vpfe_fmt *find_format_by_pix(unsigned int pixelformat)
> +{
> + struct vpfe_fmt *fmt;
> + unsigned int k;
> +
> + for (k = 0; k < ARRAY_SIZE(formats); k++) {
> + fmt = &formats[k];
> + if (fmt->fourcc == pixelformat)
> + return fmt;
> + }
> +
> + return NULL;
> +}
> +
> +static void
> +mbus_to_pix(struct vpfe_device *vpfe,
> + const struct v4l2_mbus_framefmt *mbus,
> + struct v4l2_pix_format *pix, unsigned int *bpp)
> +{
> + struct vpfe_subdev_info *sdinfo = vpfe->current_subdev;
> + unsigned int bus_width = sdinfo->vpfe_param.bus_width;
> + struct vpfe_fmt *fmt;
> +
> + memset(pix, 0, sizeof(*pix));
> + pix->width = mbus->width;
> + pix->height = mbus->height;
> +
> + fmt = find_format_by_code(mbus->code, false);
> + if (WARN_ON(fmt == NULL)) {
> + pr_err("Invalid mbus code set\n");
> + *bpp = 1;
> + return;
> + }
> +
> + pix->colorspace = mbus->colorspace;
> + pix->field = mbus->field;

Use the v4l2_fill_pix_format helper (media/v4l2-mediabus.h).
That will also fill in the new ycbcr_enc and quantization fields.

> + pix->pixelformat = fmt->fourcc;
> + *bpp = (bus_width == 10) ? fmt->l.bpp : fmt->s.bpp;
> +
> + pix->bytesperline = pix->width * *bpp;
> + /* pitch should be 32 bytes aligned */
> + pix->bytesperline = ALIGN(pix->bytesperline, 32);
> + pix->sizeimage = pix->bytesperline * pix->height;
> +}
> +
> +static int
> +pix_to_mbus(struct vpfe_device *vpfe,
> + struct v4l2_pix_format *pix,
> + struct v4l2_mbus_framefmt *mbus)
> +{
> + struct vpfe_fmt *fmt;
> +
> + memset(mbus, 0, sizeof(*mbus));
> + mbus->width = pix->width;
> + mbus->height = pix->height;
> +
> + fmt = find_format_by_pix(pix->pixelformat);
> + if (!fmt) {
> + /* default to first entry */
> + vpfe_dbg(3, vpfe, "Invalid pixel code: %x, default used instead\n",
> + pix->pixelformat);
> + fmt = &formats[0];
> + }
> +
> + mbus->code = fmt->code;
> + mbus->colorspace = pix->colorspace;
> + mbus->field = pix->field;

Use v4l2_fill_mbus_format().

> +
> + return 0;
> +}
> +
> +/* Print Four-character-code (FOURCC) */
> +static char *print_fourcc(u32 fmt)
> +{
> + static char code[5];
> +
> + code[0] = (unsigned char)(fmt & 0xff);
> + code[1] = (unsigned char)((fmt >> 8) & 0xff);
> + code[2] = (unsigned char)((fmt >> 16) & 0xff);
> + code[3] = (unsigned char)((fmt >> 24) & 0xff);
> + code[4] = '\0';
> +
> + return code;
> +}
> +
> +static int
> +cmp_v4l2_format(const struct v4l2_format *lhs, const struct v4l2_format *rhs)
> +{
> + return lhs->type == rhs->type &&
> + lhs->fmt.pix.width == rhs->fmt.pix.width &&
> + lhs->fmt.pix.height == rhs->fmt.pix.height &&
> + lhs->fmt.pix.pixelformat == rhs->fmt.pix.pixelformat &&
> + lhs->fmt.pix.field == rhs->fmt.pix.field &&
> + lhs->fmt.pix.colorspace == rhs->fmt.pix.colorspace;
> +}
> +
> +static inline u32 vpfe_reg_read(struct vpfe_ccdc *ccdc, u32 offset)
> +{
> + return ioread32(ccdc->ccdc_cfg.base_addr + offset);
> +}
> +
> +static inline void vpfe_reg_write(struct vpfe_ccdc *ccdc, u32 val, u32 offset)
> +{
> + iowrite32(val, ccdc->ccdc_cfg.base_addr + offset);
> +}
> +
> +static inline struct vpfe_device *to_vpfe(struct vpfe_ccdc *ccdc)
> +{
> + return container_of(ccdc, struct vpfe_device, ccdc);
> +}
> +
> +static inline struct vpfe_cap_buffer *to_vpfe_buffer(struct vb2_buffer *vb)
> +{
> + return container_of(vb, struct vpfe_cap_buffer, vb);
> +}
> +
> +static inline void vpfe_pcr_enable(struct vpfe_ccdc *ccdc, int flag)
> +{
> + vpfe_reg_write(ccdc, !!flag, VPFE_PCR);
> +}
> +
> +static void vpfe_config_enable(struct vpfe_ccdc *ccdc, int flag)
> +{
> + unsigned int cfg;
> +
> + if (!flag) {
> + cfg = vpfe_reg_read(ccdc, VPFE_CONFIG);
> + cfg &= ~(VPFE_CONFIG_EN_ENABLE << VPFE_CONFIG_EN_SHIFT);
> + } else {
> + cfg = VPFE_CONFIG_EN_ENABLE << VPFE_CONFIG_EN_SHIFT;
> + }
> +
> + vpfe_reg_write(ccdc, cfg, VPFE_CONFIG);
> +}
> +
> +static void vpfe_ccdc_setwin(struct vpfe_ccdc *ccdc,
> + struct v4l2_rect *image_win,
> + enum ccdc_frmfmt frm_fmt,
> + int bpp)
> +{
> + int horz_start, horz_nr_pixels;
> + int vert_start, vert_nr_lines;
> + int val, mid_img;
> +
> + /*
> + * ppc - per pixel count. indicates how many pixels per cell
> + * output to SDRAM. example, for ycbcr, it is one y and one c, so 2.
> + * raw capture this is 1
> + */
> + horz_start = image_win->left * bpp;
> + horz_nr_pixels = (image_win->width * bpp) - 1;
> + vpfe_reg_write(ccdc, (horz_start << VPFE_HORZ_INFO_SPH_SHIFT) |
> + horz_nr_pixels, VPFE_HORZ_INFO);
> +
> + vert_start = image_win->top;
> +
> + if (frm_fmt == CCDC_FRMFMT_INTERLACED) {
> + vert_nr_lines = (image_win->height >> 1) - 1;
> + vert_start >>= 1;
> + /* Since first line doesn't have any data */
> + vert_start += 1;
> + /* configure VDINT0 */
> + val = (vert_start << VPFE_VDINT_VDINT0_SHIFT);
> + } else {
> + /* Since first line doesn't have any data */
> + vert_start += 1;
> + vert_nr_lines = image_win->height - 1;
> + /*
> + * configure VDINT0 and VDINT1. VDINT1 will be at half
> + * of image height
> + */
> + mid_img = vert_start + (image_win->height / 2);
> + val = (vert_start << VPFE_VDINT_VDINT0_SHIFT) |
> + (mid_img & VPFE_VDINT_VDINT1_MASK);
> + }
> +
> + vpfe_reg_write(ccdc, val, VPFE_VDINT);
> +
> + vpfe_reg_write(ccdc, (vert_start << VPFE_VERT_START_SLV0_SHIFT) |
> + vert_start, VPFE_VERT_START);
> + vpfe_reg_write(ccdc, vert_nr_lines, VPFE_VERT_LINES);
> +}
> +
> +static void vpfe_reg_dump(struct vpfe_ccdc *ccdc)
> +{
> + struct vpfe_device *vpfe = to_vpfe(ccdc);
> +
> + vpfe_dbg(3, vpfe, "ALAW: 0x%x\n", vpfe_reg_read(ccdc, VPFE_ALAW));
> + vpfe_dbg(3, vpfe, "CLAMP: 0x%x\n", vpfe_reg_read(ccdc, VPFE_CLAMP));
> + vpfe_dbg(3, vpfe, "DCSUB: 0x%x\n", vpfe_reg_read(ccdc, VPFE_DCSUB));
> + vpfe_dbg(3, vpfe, "BLKCMP: 0x%x\n", vpfe_reg_read(ccdc, VPFE_BLKCMP));
> + vpfe_dbg(3, vpfe, "COLPTN: 0x%x\n", vpfe_reg_read(ccdc, VPFE_COLPTN));
> + vpfe_dbg(3, vpfe, "SDOFST: 0x%x\n", vpfe_reg_read(ccdc, VPFE_SDOFST));
> + vpfe_dbg(3, vpfe, "SYN_MODE: 0x%x\n",
> + vpfe_reg_read(ccdc, VPFE_SYNMODE));
> + vpfe_dbg(3, vpfe, "HSIZE_OFF: 0x%x\n",
> + vpfe_reg_read(ccdc, VPFE_HSIZE_OFF));
> + vpfe_dbg(3, vpfe, "HORZ_INFO: 0x%x\n",
> + vpfe_reg_read(ccdc, VPFE_HORZ_INFO));
> + vpfe_dbg(3, vpfe, "VERT_START: 0x%x\n",
> + vpfe_reg_read(ccdc, VPFE_VERT_START));
> + vpfe_dbg(3, vpfe, "VERT_LINES: 0x%x\n",
> + vpfe_reg_read(ccdc, VPFE_VERT_LINES));
> +}
> +
> +static int
> +vpfe_ccdc_validate_param(struct vpfe_ccdc *ccdc,
> + struct vpfe_ccdc_config_params_raw *ccdcparam)
> +{
> + struct vpfe_device *vpfe = to_vpfe(ccdc);
> + u8 max_gamma, max_data;
> +
> + if (!ccdcparam->alaw.enable)
> + return 0;
> +
> + max_gamma = ccdc_gamma_width_max_bit(ccdcparam->alaw.gamma_wd);
> + max_data = ccdc_data_size_max_bit(ccdcparam->data_sz);
> +
> + if (ccdcparam->alaw.gamma_wd > VPFE_CCDC_GAMMA_BITS_09_0 ||
> + ccdcparam->alaw.gamma_wd < VPFE_CCDC_GAMMA_BITS_15_6 ||
> + max_gamma > max_data) {
> + vpfe_dbg(1, vpfe, "Invalid data line select\n");
> + return -EINVAL;
> + }
> +
> + return 0;
> +}
> +
> +static void
> +vpfe_ccdc_update_raw_params(struct vpfe_ccdc *ccdc,
> + struct vpfe_ccdc_config_params_raw *raw_params)
> +{
> + struct vpfe_ccdc_config_params_raw *config_params =
> + &ccdc->ccdc_cfg.bayer.config_params;
> +
> + config_params = raw_params;
> +}
> +
> +/*
> + * vpfe_ccdc_restore_defaults()
> + * This function will write defaults to all CCDC registers
> + */
> +static void vpfe_ccdc_restore_defaults(struct vpfe_ccdc *ccdc)
> +{
> + int i;
> +
> + /* Disable CCDC */
> + vpfe_pcr_enable(ccdc, 0);
> +
> + /* set all registers to default value */
> + for (i = 4; i <= 0x94; i += 4)
> + vpfe_reg_write(ccdc, 0, i);
> +
> + vpfe_reg_write(ccdc, VPFE_NO_CULLING, VPFE_CULLING);
> + vpfe_reg_write(ccdc, VPFE_CCDC_GAMMA_BITS_11_2, VPFE_ALAW);
> +}
> +
> +static int vpfe_ccdc_close(struct vpfe_ccdc *ccdc, struct device *dev)
> +{
> + int dma_cntl, i, pcr;
> +
> + /* If the CCDC module is still busy wait for it to be done */
> + for (i = 0; i < 10; i++) {
> + usleep_range(5000, 6000);
> + pcr = vpfe_reg_read(ccdc, VPFE_PCR);
> + if (!pcr)
> + break;
> +
> + /* make sure it it is disabled */
> + vpfe_pcr_enable(ccdc, 0);
> + }
> +
> + /* Disable CCDC by resetting all register to default POR values */
> + vpfe_ccdc_restore_defaults(ccdc);
> +
> + /* if DMA_CNTL overflow bit is set. Clear it
> + * It appears to take a while for this to become quiescent ~20ms
> + */
> + for (i = 0; i < 10; i++) {
> + dma_cntl = vpfe_reg_read(ccdc, VPFE_DMA_CNTL);
> + if (!(dma_cntl & VPFE_DMA_CNTL_OVERFLOW))
> + break;
> +
> + /* Clear the overflow bit */
> + vpfe_reg_write(ccdc, dma_cntl, VPFE_DMA_CNTL);
> + usleep_range(5000, 6000);
> + }
> +
> + /* Disabled the module at the CONFIG level */
> + vpfe_config_enable(ccdc, 0);
> +
> + pm_runtime_put_sync(dev);
> +
> + return 0;
> +}
> +
> +static int vpfe_ccdc_set_params(struct vpfe_ccdc *ccdc, void __user *params)
> +{
> + struct vpfe_device *vpfe = container_of(ccdc, struct vpfe_device, ccdc);
> + struct vpfe_ccdc_config_params_raw raw_params;
> + int x;
> +
> + if (ccdc->ccdc_cfg.if_type != VPFE_RAW_BAYER)
> + return -EINVAL;
> +
> + x = copy_from_user(&raw_params, params, sizeof(raw_params));
> + if (x) {
> + vpfe_dbg(1, vpfe,
> + "vpfe_ccdc_set_params: error in copying ccdc params, %d\n",
> + x);
> + return -EFAULT;
> + }
> +
> + if (!vpfe_ccdc_validate_param(ccdc, &raw_params)) {
> + vpfe_ccdc_update_raw_params(ccdc, &raw_params);
> + return 0;
> + }
> +
> + return -EINVAL;
> +}
> +
> +/*
> + * vpfe_ccdc_config_ycbcr()
> + * This function will configure CCDC for YCbCr video capture
> + */
> +static void vpfe_ccdc_config_ycbcr(struct vpfe_ccdc *ccdc)
> +{
> + struct vpfe_device *vpfe = container_of(ccdc, struct vpfe_device, ccdc);
> + struct ccdc_params_ycbcr *params = &ccdc->ccdc_cfg.ycbcr;
> + u32 syn_mode;
> +
> + vpfe_dbg(3, vpfe, "vpfe_ccdc_config_ycbcr:\n");
> + /*
> + * first restore the CCDC registers to default values
> + * This is important since we assume default values to be set in
> + * a lot of registers that we didn't touch
> + */
> + vpfe_ccdc_restore_defaults(ccdc);
> +
> + /*
> + * configure pixel format, frame format, configure video frame
> + * format, enable output to SDRAM, enable internal timing generator
> + * and 8bit pack mode
> + */
> + syn_mode = (((params->pix_fmt & VPFE_SYN_MODE_INPMOD_MASK) <<
> + VPFE_SYN_MODE_INPMOD_SHIFT) |
> + ((params->frm_fmt & VPFE_SYN_FLDMODE_MASK) <<
> + VPFE_SYN_FLDMODE_SHIFT) | VPFE_VDHDEN_ENABLE |
> + VPFE_WEN_ENABLE | VPFE_DATA_PACK_ENABLE);
> +
> + /* setup BT.656 sync mode */
> + if (params->bt656_enable) {
> + vpfe_reg_write(ccdc, VPFE_REC656IF_BT656_EN, VPFE_REC656IF);
> +
> + /*
> + * configure the FID, VD, HD pin polarity,
> + * fld,hd pol positive, vd negative, 8-bit data
> + */
> + syn_mode |= VPFE_SYN_MODE_VD_POL_NEGATIVE;
> + if (ccdc->ccdc_cfg.if_type == VPFE_BT656_10BIT)
> + syn_mode |= VPFE_SYN_MODE_10BITS;
> + else
> + syn_mode |= VPFE_SYN_MODE_8BITS;
> + } else {
> + /* y/c external sync mode */
> + syn_mode |= (((params->fid_pol & VPFE_FID_POL_MASK) <<
> + VPFE_FID_POL_SHIFT) |
> + ((params->hd_pol & VPFE_HD_POL_MASK) <<
> + VPFE_HD_POL_SHIFT) |
> + ((params->vd_pol & VPFE_VD_POL_MASK) <<
> + VPFE_VD_POL_SHIFT));
> + }
> + vpfe_reg_write(ccdc, syn_mode, VPFE_SYNMODE);
> +
> + /* configure video window */
> + vpfe_ccdc_setwin(ccdc, &params->win,
> + params->frm_fmt, params->bytesperpixel);
> +
> + /*
> + * configure the order of y cb cr in SDRAM, and disable latch
> + * internal register on vsync
> + */
> + if (ccdc->ccdc_cfg.if_type == VPFE_BT656_10BIT)
> + vpfe_reg_write(ccdc,
> + (params->pix_order << VPFE_CCDCFG_Y8POS_SHIFT) |
> + VPFE_LATCH_ON_VSYNC_DISABLE |
> + VPFE_CCDCFG_BW656_10BIT, VPFE_CCDCFG);
> + else
> + vpfe_reg_write(ccdc,
> + (params->pix_order << VPFE_CCDCFG_Y8POS_SHIFT) |
> + VPFE_LATCH_ON_VSYNC_DISABLE, VPFE_CCDCFG);
> +
> + /*
> + * configure the horizontal line offset. This should be a
> + * on 32 byte boundary. So clear LSB 5 bits
> + */
> + vpfe_reg_write(ccdc, params->bytesperline, VPFE_HSIZE_OFF);
> +
> + /* configure the memory line offset */
> + if (params->buf_type == CCDC_BUFTYPE_FLD_INTERLEAVED)
> + /* two fields are interleaved in memory */
> + vpfe_reg_write(ccdc, VPFE_SDOFST_FIELD_INTERLEAVED,
> + VPFE_SDOFST);
> +}
> +
> +static void
> +vpfe_ccdc_config_black_clamp(struct vpfe_ccdc *ccdc,
> + struct vpfe_ccdc_black_clamp *bclamp)
> +{
> + u32 val;
> +
> + if (!bclamp->enable) {
> + /* configure DCSub */
> + val = (bclamp->dc_sub) & VPFE_BLK_DC_SUB_MASK;
> + vpfe_reg_write(ccdc, val, VPFE_DCSUB);
> + vpfe_reg_write(ccdc, VPFE_CLAMP_DEFAULT_VAL, VPFE_CLAMP);
> + return;
> + }
> + /*
> + * Configure gain, Start pixel, No of line to be avg,
> + * No of pixel/line to be avg, & Enable the Black clamping
> + */
> + val = ((bclamp->sgain & VPFE_BLK_SGAIN_MASK) |
> + ((bclamp->start_pixel & VPFE_BLK_ST_PXL_MASK) <<
> + VPFE_BLK_ST_PXL_SHIFT) |
> + ((bclamp->sample_ln & VPFE_BLK_SAMPLE_LINE_MASK) <<
> + VPFE_BLK_SAMPLE_LINE_SHIFT) |
> + ((bclamp->sample_pixel & VPFE_BLK_SAMPLE_LN_MASK) <<
> + VPFE_BLK_SAMPLE_LN_SHIFT) | VPFE_BLK_CLAMP_ENABLE);
> + vpfe_reg_write(ccdc, val, VPFE_CLAMP);
> + /* If Black clamping is enable then make dcsub 0 */
> + vpfe_reg_write(ccdc, VPFE_DCSUB_DEFAULT_VAL, VPFE_DCSUB);
> +}
> +
> +static void
> +vpfe_ccdc_config_black_compense(struct vpfe_ccdc *ccdc,
> + struct vpfe_ccdc_black_compensation *bcomp)
> +{
> + u32 val;
> +
> + val = ((bcomp->b & VPFE_BLK_COMP_MASK) |
> + ((bcomp->gb & VPFE_BLK_COMP_MASK) <<
> + VPFE_BLK_COMP_GB_COMP_SHIFT) |
> + ((bcomp->gr & VPFE_BLK_COMP_MASK) <<
> + VPFE_BLK_COMP_GR_COMP_SHIFT) |
> + ((bcomp->r & VPFE_BLK_COMP_MASK) <<
> + VPFE_BLK_COMP_R_COMP_SHIFT));
> + vpfe_reg_write(ccdc, val, VPFE_BLKCMP);
> +}
> +
> +/*
> + * vpfe_ccdc_config_raw()
> + * This function will configure CCDC for Raw capture mode
> + */
> +static void vpfe_ccdc_config_raw(struct vpfe_ccdc *ccdc)
> +{
> + struct vpfe_device *vpfe = container_of(ccdc, struct vpfe_device, ccdc);
> + struct vpfe_ccdc_config_params_raw *config_params =
> + &ccdc->ccdc_cfg.bayer.config_params;
> + struct ccdc_params_raw *params = &ccdc->ccdc_cfg.bayer;
> + unsigned int syn_mode;
> + unsigned int val;
> +
> + vpfe_dbg(3, vpfe, "vpfe_ccdc_config_raw:\n");
> +
> + /* Reset CCDC */
> + vpfe_ccdc_restore_defaults(ccdc);
> +
> + /* Disable latching function registers on VSYNC */
> + vpfe_reg_write(ccdc, VPFE_LATCH_ON_VSYNC_DISABLE, VPFE_CCDCFG);
> +
> + /*
> + * Configure the vertical sync polarity(SYN_MODE.VDPOL),
> + * horizontal sync polarity (SYN_MODE.HDPOL), frame id polarity
> + * (SYN_MODE.FLDPOL), frame format(progressive or interlace),
> + * data size(SYNMODE.DATSIZ), &pixel format (Input mode), output
> + * SDRAM, enable internal timing generator
> + */
> + syn_mode = (((params->vd_pol & VPFE_VD_POL_MASK) << VPFE_VD_POL_SHIFT) |
> + ((params->hd_pol & VPFE_HD_POL_MASK) << VPFE_HD_POL_SHIFT) |
> + ((params->fid_pol & VPFE_FID_POL_MASK) <<
> + VPFE_FID_POL_SHIFT) | ((params->frm_fmt &
> + VPFE_FRM_FMT_MASK) << VPFE_FRM_FMT_SHIFT) |
> + ((config_params->data_sz & VPFE_DATA_SZ_MASK) <<
> + VPFE_DATA_SZ_SHIFT) | ((params->pix_fmt &
> + VPFE_PIX_FMT_MASK) << VPFE_PIX_FMT_SHIFT) |
> + VPFE_WEN_ENABLE | VPFE_VDHDEN_ENABLE);
> +
> + /* Enable and configure aLaw register if needed */
> + if (config_params->alaw.enable) {
> + val = ((config_params->alaw.gamma_wd &
> + VPFE_ALAW_GAMMA_WD_MASK) | VPFE_ALAW_ENABLE);
> + vpfe_reg_write(ccdc, val, VPFE_ALAW);
> + vpfe_dbg(3, vpfe, "\nWriting 0x%x to ALAW...\n", val);
> + }
> +
> + /* Configure video window */
> + vpfe_ccdc_setwin(ccdc, &params->win, params->frm_fmt,
> + params->bytesperpixel);
> +
> + /* Configure Black Clamp */
> + vpfe_ccdc_config_black_clamp(ccdc, &config_params->blk_clamp);
> +
> + /* Configure Black level compensation */
> + vpfe_ccdc_config_black_compense(ccdc, &config_params->blk_comp);
> +
> + /* If data size is 8 bit then pack the data */
> + if ((config_params->data_sz == VPFE_CCDC_DATA_8BITS) ||
> + config_params->alaw.enable)
> + syn_mode |= VPFE_DATA_PACK_ENABLE;
> +
> + /*
> + * Configure Horizontal offset register. If pack 8 is enabled then
> + * 1 pixel will take 1 byte
> + */
> + vpfe_reg_write(ccdc, params->bytesperline, VPFE_HSIZE_OFF);
> +
> + vpfe_dbg(3, vpfe, "Writing %d (%x) to HSIZE_OFF\n",
> + params->bytesperline, params->bytesperline);
> +
> + /* Set value for SDOFST */
> + if (params->frm_fmt == CCDC_FRMFMT_INTERLACED) {
> + if (params->image_invert_enable) {
> + /* For interlace inverse mode */
> + vpfe_reg_write(ccdc, VPFE_INTERLACED_IMAGE_INVERT,
> + VPFE_SDOFST);
> + } else {
> + /* For interlace non inverse mode */
> + vpfe_reg_write(ccdc, VPFE_INTERLACED_NO_IMAGE_INVERT,
> + VPFE_SDOFST);
> + }
> + } else if (params->frm_fmt == CCDC_FRMFMT_PROGRESSIVE) {
> + vpfe_reg_write(ccdc, VPFE_PROGRESSIVE_NO_IMAGE_INVERT,
> + VPFE_SDOFST);
> + }
> +
> + vpfe_reg_write(ccdc, syn_mode, VPFE_SYNMODE);
> +
> + vpfe_reg_dump(ccdc);
> +}
> +
> +static inline int
> +vpfe_ccdc_set_buftype(struct vpfe_ccdc *ccdc,
> + enum ccdc_buftype buf_type)
> +{
> + if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER)
> + ccdc->ccdc_cfg.bayer.buf_type = buf_type;
> + else
> + ccdc->ccdc_cfg.ycbcr.buf_type = buf_type;
> +
> + return 0;
> +}
> +
> +static inline enum ccdc_buftype vpfe_ccdc_get_buftype(struct vpfe_ccdc *ccdc)
> +{
> + if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER)
> + return ccdc->ccdc_cfg.bayer.buf_type;
> +
> + return ccdc->ccdc_cfg.ycbcr.buf_type;
> +}
> +
> +static int vpfe_ccdc_set_pixel_format(struct vpfe_ccdc *ccdc, u32 pixfmt)
> +{
> + struct vpfe_device *vpfe = container_of(ccdc, struct vpfe_device, ccdc);
> +
> + vpfe_dbg(1, vpfe, "vpfe_ccdc_set_pixel_format: if_type: %d, pixfmt:%s\n",
> + ccdc->ccdc_cfg.if_type, print_fourcc(pixfmt));
> +
> + if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) {
> + ccdc->ccdc_cfg.bayer.pix_fmt = CCDC_PIXFMT_RAW;
> + /*
> + * Need to clear it in case it was left on
> + * after the last capture.
> + */
> + ccdc->ccdc_cfg.bayer.config_params.alaw.enable = 0;
> +
> + switch (pixfmt) {
> + case V4L2_PIX_FMT_SBGGR8:
> + ccdc->ccdc_cfg.bayer.config_params.alaw.enable = 1;
> + break;
> +
> + case V4L2_PIX_FMT_YUYV:
> + case V4L2_PIX_FMT_UYVY:
> + case V4L2_PIX_FMT_YUV420:
> + case V4L2_PIX_FMT_NV12:
> + case V4L2_PIX_FMT_RGB565X:
> + break;
> +
> + case V4L2_PIX_FMT_SBGGR16:
> + default:
> + return -EINVAL;
> + }
> + } else {
> + switch (pixfmt) {
> + case V4L2_PIX_FMT_YUYV:
> + ccdc->ccdc_cfg.ycbcr.pix_order = CCDC_PIXORDER_YCBYCR;
> + break;
> +
> + case V4L2_PIX_FMT_UYVY:
> + ccdc->ccdc_cfg.ycbcr.pix_order = CCDC_PIXORDER_CBYCRY;
> + break;
> +
> + default:
> + return -EINVAL;
> + }
> + }
> +
> + return 0;
> +}
> +
> +static u32 vpfe_ccdc_get_pixel_format(struct vpfe_ccdc *ccdc)
> +{
> + u32 pixfmt;
> +
> + if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) {
> + pixfmt = V4L2_PIX_FMT_YUYV;
> + } else {
> + if (ccdc->ccdc_cfg.ycbcr.pix_order == CCDC_PIXORDER_YCBYCR)
> + pixfmt = V4L2_PIX_FMT_YUYV;
> + else
> + pixfmt = V4L2_PIX_FMT_UYVY;
> + }
> +
> + return pixfmt;
> +}
> +
> +static int
> +vpfe_ccdc_set_image_window(struct vpfe_ccdc *ccdc,
> + struct v4l2_rect *win, unsigned int bpp)
> +{
> + if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER) {
> + ccdc->ccdc_cfg.bayer.win = *win;
> + ccdc->ccdc_cfg.bayer.bytesperpixel = bpp;
> + ccdc->ccdc_cfg.bayer.bytesperline = ALIGN(win->width * bpp, 32);
> + } else {
> + ccdc->ccdc_cfg.ycbcr.win = *win;
> + ccdc->ccdc_cfg.ycbcr.bytesperpixel = bpp;
> + ccdc->ccdc_cfg.ycbcr.bytesperline = ALIGN(win->width * bpp, 32);
> + }
> +
> + return 0;
> +}
> +
> +static inline void
> +vpfe_ccdc_get_image_window(struct vpfe_ccdc *ccdc,
> + struct v4l2_rect *win)
> +{
> + if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER)
> + *win = ccdc->ccdc_cfg.bayer.win;
> + else
> + *win = ccdc->ccdc_cfg.ycbcr.win;
> +}
> +
> +static inline unsigned int vpfe_ccdc_get_line_length(struct vpfe_ccdc *ccdc)
> +{
> + if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER)
> + return ccdc->ccdc_cfg.bayer.bytesperline;
> +
> + return ccdc->ccdc_cfg.ycbcr.bytesperline;
> +}
> +
> +static inline int
> +vpfe_ccdc_set_frame_format(struct vpfe_ccdc *ccdc,
> + enum ccdc_frmfmt frm_fmt)
> +{
> + if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER)
> + ccdc->ccdc_cfg.bayer.frm_fmt = frm_fmt;
> + else
> + ccdc->ccdc_cfg.ycbcr.frm_fmt = frm_fmt;
> +
> + return 0;
> +}
> +
> +static inline enum ccdc_frmfmt
> +vpfe_ccdc_get_frame_format(struct vpfe_ccdc *ccdc)
> +{
> + if (ccdc->ccdc_cfg.if_type == VPFE_RAW_BAYER)
> + return ccdc->ccdc_cfg.bayer.frm_fmt;
> +
> + return ccdc->ccdc_cfg.ycbcr.frm_fmt;
> +}
> +
> +static inline int vpfe_ccdc_getfid(struct vpfe_ccdc *ccdc)
> +{
> + return (vpfe_reg_read(ccdc, VPFE_SYNMODE) >> 15) & 1;
> +}
> +
> +static inline void vpfe_set_sdr_addr(struct vpfe_ccdc *ccdc, unsigned long addr)
> +{
> + vpfe_reg_write(ccdc, addr & 0xffffffe0, VPFE_SDR_ADDR);
> +}
> +
> +static int vpfe_ccdc_set_hw_if_params(struct vpfe_ccdc *ccdc,
> + struct vpfe_hw_if_param *params)
> +{
> + struct vpfe_device *vpfe = container_of(ccdc, struct vpfe_device, ccdc);
> +
> + ccdc->ccdc_cfg.if_type = params->if_type;
> +
> + switch (params->if_type) {
> + case VPFE_BT656:
> + case VPFE_YCBCR_SYNC_16:
> + case VPFE_YCBCR_SYNC_8:
> + case VPFE_BT656_10BIT:
> + ccdc->ccdc_cfg.ycbcr.vd_pol = params->vdpol;
> + ccdc->ccdc_cfg.ycbcr.hd_pol = params->hdpol;
> + break;
> +
> + case VPFE_RAW_BAYER:
> + ccdc->ccdc_cfg.bayer.vd_pol = params->vdpol;
> + ccdc->ccdc_cfg.bayer.hd_pol = params->hdpol;
> + if (params->bus_width == 10)
> + ccdc->ccdc_cfg.bayer.config_params.data_sz =
> + VPFE_CCDC_DATA_10BITS;
> + else
> + ccdc->ccdc_cfg.bayer.config_params.data_sz =
> + VPFE_CCDC_DATA_8BITS;
> + vpfe_dbg(1, vpfe, "params.bus_width: %d\n",
> + params->bus_width);
> + vpfe_dbg(1, vpfe, "config_params.data_sz: %d\n",
> + ccdc->ccdc_cfg.bayer.config_params.data_sz);
> + break;
> +
> + default:
> + return -EINVAL;
> + }
> +
> + return 0;
> +}
> +
> +static void vpfe_clear_intr(struct vpfe_ccdc *ccdc, int vdint)
> +{
> + unsigned int vpfe_int_status;
> +
> + vpfe_int_status = vpfe_reg_read(ccdc, VPFE_IRQ_STS);
> +
> + switch (vdint) {
> + /* VD0 interrupt */
> + case VPFE_VDINT0:
> + vpfe_int_status &= ~VPFE_VDINT0;
> + vpfe_int_status |= VPFE_VDINT0;
> + break;
> +
> + /* VD1 interrupt */
> + case VPFE_VDINT1:
> + vpfe_int_status &= ~VPFE_VDINT1;
> + vpfe_int_status |= VPFE_VDINT1;
> + break;
> +
> + /* VD2 interrupt */
> + case VPFE_VDINT2:
> + vpfe_int_status &= ~VPFE_VDINT2;
> + vpfe_int_status |= VPFE_VDINT2;
> + break;
> +
> + /* Clear all interrupts */
> + default:
> + vpfe_int_status &= ~(VPFE_VDINT0 |
> + VPFE_VDINT1 |
> + VPFE_VDINT2);
> + vpfe_int_status |= (VPFE_VDINT0 |
> + VPFE_VDINT1 |
> + VPFE_VDINT2);
> + break;
> + }
> + /* Clear specific VDINT from the status register */
> + vpfe_reg_write(ccdc, vpfe_int_status, VPFE_IRQ_STS);
> +
> + vpfe_int_status = vpfe_reg_read(ccdc, VPFE_IRQ_STS);
> +
> + /* Acknowledge that we are done with all interrupts */
> + vpfe_reg_write(ccdc, 1, VPFE_IRQ_EOI);
> +}
> +
> +static void vpfe_ccdc_config_defaults(struct vpfe_ccdc *ccdc)
> +{
> + ccdc->ccdc_cfg.if_type = VPFE_RAW_BAYER;
> +
> + ccdc->ccdc_cfg.ycbcr.pix_fmt = CCDC_PIXFMT_YCBCR_8BIT;
> + ccdc->ccdc_cfg.ycbcr.frm_fmt = CCDC_FRMFMT_INTERLACED;
> + ccdc->ccdc_cfg.ycbcr.fid_pol = VPFE_PINPOL_POSITIVE;
> + ccdc->ccdc_cfg.ycbcr.vd_pol = VPFE_PINPOL_POSITIVE;
> + ccdc->ccdc_cfg.ycbcr.hd_pol = VPFE_PINPOL_POSITIVE;
> + ccdc->ccdc_cfg.ycbcr.pix_order = CCDC_PIXORDER_CBYCRY;
> + ccdc->ccdc_cfg.ycbcr.buf_type = CCDC_BUFTYPE_FLD_INTERLEAVED;
> +
> + ccdc->ccdc_cfg.ycbcr.win.left = 0;
> + ccdc->ccdc_cfg.ycbcr.win.top = 0;
> + ccdc->ccdc_cfg.ycbcr.win.width = 720;
> + ccdc->ccdc_cfg.ycbcr.win.height = 576;
> + ccdc->ccdc_cfg.ycbcr.bt656_enable = 1;
> +
> + ccdc->ccdc_cfg.bayer.pix_fmt = CCDC_PIXFMT_RAW;
> + ccdc->ccdc_cfg.bayer.frm_fmt = CCDC_FRMFMT_PROGRESSIVE;
> + ccdc->ccdc_cfg.bayer.fid_pol = VPFE_PINPOL_POSITIVE;
> + ccdc->ccdc_cfg.bayer.vd_pol = VPFE_PINPOL_POSITIVE;
> + ccdc->ccdc_cfg.bayer.hd_pol = VPFE_PINPOL_POSITIVE;
> +
> + ccdc->ccdc_cfg.bayer.win.left = 0;
> + ccdc->ccdc_cfg.bayer.win.top = 0;
> + ccdc->ccdc_cfg.bayer.win.width = 800;
> + ccdc->ccdc_cfg.bayer.win.height = 600;
> + ccdc->ccdc_cfg.bayer.config_params.data_sz = VPFE_CCDC_DATA_8BITS;
> + ccdc->ccdc_cfg.bayer.config_params.alaw.gamma_wd =
> + VPFE_CCDC_GAMMA_BITS_09_0;
> +}
> +
> +/*
> + * vpfe_get_ccdc_image_format - Get image parameters based on CCDC settings
> + */
> +static int vpfe_get_ccdc_image_format(struct vpfe_device *vpfe,
> + struct v4l2_format *f)
> +{
> + struct v4l2_rect image_win;
> + enum ccdc_buftype buf_type;
> + enum ccdc_frmfmt frm_fmt;
> +
> + memset(f, 0, sizeof(*f));
> + f->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
> + vpfe_ccdc_get_image_window(&vpfe->ccdc, &image_win);
> + f->fmt.pix.width = image_win.width;
> + f->fmt.pix.height = image_win.height;
> + f->fmt.pix.bytesperline = vpfe_ccdc_get_line_length(&vpfe->ccdc);
> + f->fmt.pix.sizeimage = f->fmt.pix.bytesperline *
> + f->fmt.pix.height;
> + buf_type = vpfe_ccdc_get_buftype(&vpfe->ccdc);
> + f->fmt.pix.pixelformat = vpfe_ccdc_get_pixel_format(&vpfe->ccdc);
> + frm_fmt = vpfe_ccdc_get_frame_format(&vpfe->ccdc);
> +
> + if (frm_fmt == CCDC_FRMFMT_PROGRESSIVE) {
> + f->fmt.pix.field = V4L2_FIELD_NONE;
> + } else if (frm_fmt == CCDC_FRMFMT_INTERLACED) {
> + if (buf_type == CCDC_BUFTYPE_FLD_INTERLEAVED) {
> + f->fmt.pix.field = V4L2_FIELD_INTERLACED;
> + } else if (buf_type == CCDC_BUFTYPE_FLD_SEPARATED) {
> + f->fmt.pix.field = V4L2_FIELD_SEQ_TB;
> + } else {
> + vpfe_err(vpfe, "Invalid buf_type\n");
> + return -EINVAL;
> + }
> + } else {
> + vpfe_err(vpfe, "Invalid frm_fmt\n");
> + return -EINVAL;
> + }
> + return 0;
> +}
> +
> +static int vpfe_config_ccdc_image_format(struct vpfe_device *vpfe)
> +{
> + enum ccdc_frmfmt frm_fmt = CCDC_FRMFMT_INTERLACED;
> + int ret;
> +
> + vpfe_dbg(2, vpfe, "vpfe_config_ccdc_image_format\n");
> +
> + vpfe_dbg(1, vpfe, "pixelformat: %s\n",
> + print_fourcc(vpfe->fmt.fmt.pix.pixelformat));
> +
> + if (vpfe_ccdc_set_pixel_format(&vpfe->ccdc,
> + vpfe->fmt.fmt.pix.pixelformat) < 0) {
> + vpfe_err(vpfe, "couldn't set pix format in ccdc\n");
> + return -EINVAL;
> + }
> +
> + /* configure the image window */
> + vpfe_ccdc_set_image_window(&vpfe->ccdc, &vpfe->crop, vpfe->bpp);
> +
> + switch (vpfe->fmt.fmt.pix.field) {
> + case V4L2_FIELD_INTERLACED:
> + /* do nothing, since it is default */
> + ret = vpfe_ccdc_set_buftype(
> + &vpfe->ccdc,
> + CCDC_BUFTYPE_FLD_INTERLEAVED);
> + break;
> +
> + case V4L2_FIELD_NONE:
> + frm_fmt = CCDC_FRMFMT_PROGRESSIVE;
> + /* buffer type only applicable for interlaced scan */
> + break;
> +
> + case V4L2_FIELD_SEQ_TB:
> + ret = vpfe_ccdc_set_buftype(
> + &vpfe->ccdc,
> + CCDC_BUFTYPE_FLD_SEPARATED);
> + break;
> +
> + default:
> + return -EINVAL;
> + }
> +
> + if (ret)
> + return ret;
> +
> + return vpfe_ccdc_set_frame_format(&vpfe->ccdc, frm_fmt);
> +}
> +
> +/*
> + * vpfe_config_image_format()
> + * For a given standard, this functions sets up the default
> + * pix format & crop values in the vpfe device and ccdc. It first
> + * starts with defaults based values from the standard table.
> + * It then checks if sub device support g_mbus_fmt and then override the
> + * values based on that.Sets crop values to match with scan resolution
> + * starting at 0,0. It calls vpfe_config_ccdc_image_format() set the
> + * values in ccdc
> + */
> +static int vpfe_config_image_format(struct vpfe_device *vpfe,
> + v4l2_std_id std_id)
> +{
> + struct v4l2_pix_format *pix = &vpfe->fmt.fmt.pix;
> + int i, ret;
> +
> + for (i = 0; i < ARRAY_SIZE(vpfe_standards); i++) {
> + if (vpfe_standards[i].std_id & std_id) {
> + vpfe->std_info.active_pixels =
> + vpfe_standards[i].width;
> + vpfe->std_info.active_lines =
> + vpfe_standards[i].height;
> + vpfe->std_info.frame_format =
> + vpfe_standards[i].frame_format;
> + vpfe->std_index = i;
> +
> + break;
> + }
> + }
> +
> + if (i == ARRAY_SIZE(vpfe_standards)) {
> + vpfe_err(vpfe, "standard not supported\n");
> + return -EINVAL;
> + }
> +
> + vpfe->crop.top = vpfe->crop.left = 0;
> + vpfe->crop.width = vpfe->std_info.active_pixels;
> + vpfe->crop.height = vpfe->std_info.active_lines;
> + pix->width = vpfe->crop.width;
> + pix->height = vpfe->crop.height;
> + pix->pixelformat = V4L2_PIX_FMT_YUYV;
> +
> + /* first field and frame format based on standard frame format */
> + if (vpfe->std_info.frame_format)
> + pix->field = V4L2_FIELD_INTERLACED;
> + else
> + pix->field = V4L2_FIELD_NONE;
> +
> + ret = __vpfe_get_format(vpfe, &vpfe->fmt, &vpfe->bpp);
> + if (ret)
> + return ret;
> +
> + /* Update the crop window based on found values */
> + vpfe->crop.width = pix->width;
> + vpfe->crop.height = pix->height;
> +
> + return vpfe_config_ccdc_image_format(vpfe);
> +}
> +
> +static int vpfe_initialize_device(struct vpfe_device *vpfe)
> +{
> + struct vpfe_subdev_info *sdinfo;
> + int ret;
> +
> + sdinfo = &vpfe->cfg->sub_devs[0];
> + sdinfo->sd = vpfe->sd[0];
> + vpfe->current_input = 0;
> + vpfe->std_index = 0;
> + /* Configure the default format information */
> + ret = vpfe_config_image_format(vpfe,
> + vpfe_standards[vpfe->std_index].std_id);
> + if (ret)
> + return ret;
> +
> + pm_runtime_get_sync(vpfe->pdev);
> +
> + vpfe_config_enable(&vpfe->ccdc, 1);
> +
> + vpfe_ccdc_restore_defaults(&vpfe->ccdc);
> +
> + /* Clear all VPFE interrupts */
> + vpfe_clear_intr(&vpfe->ccdc, -1);
> +
> + return ret;
> +}
> +
> +/*
> + * vpfe_release : This function is based on the vb2_fop_release
> + * helper function.
> + * It has been augmented to handle module power management,
> + * by disabling/enabling h/w module fcntl clock when necessary.
> + */
> +static int vpfe_release(struct file *file)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + int ret;
> +
> + vpfe_dbg(2, vpfe, "vpfe_release\n");
> +
> + ret = _vb2_fop_release(file, NULL);
> +
> + if (v4l2_fh_is_singular_file(file)) {
> + mutex_lock(&vpfe->lock);
> + vpfe_ccdc_close(&vpfe->ccdc, vpfe->pdev);
> + v4l2_fh_release(file);
> + mutex_unlock(&vpfe->lock);
> + }
> +
> + return ret;
> +}
> +
> +/*
> + * vpfe_open : This function is based on the v4l2_fh_open helper function.
> + * It has been augmented to handle module power management,
> + * by disabling/enabling h/w module fcntl clock when necessary.
> + */
> +
> +static int vpfe_open(struct file *file)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + int ret;
> +
> + ret = v4l2_fh_open(file);
> + if (ret) {
> + vpfe_err(vpfe, "v4l2_fh_open failed\n");
> + return ret;
> + }
> +
> + if (!v4l2_fh_is_singular_file(file))
> + return 0;
> +
> + mutex_lock(&vpfe->lock);
> + if (vpfe_initialize_device(vpfe)) {
> + mutex_unlock(&vpfe->lock);
> + v4l2_fh_release(file);
> + return -ENODEV;
> + }
> + mutex_unlock(&vpfe->lock);
> +
> + return 0;
> +}
> +
> +/**
> + * vpfe_schedule_next_buffer: set next buffer address for capture
> + * @vpfe : ptr to vpfe device
> + *
> + * This function will get next buffer from the dma queue and
> + * set the buffer address in the vpfe register for capture.
> + * the buffer is marked active
> + *
> + * Assumes caller is holding vpfe->dma_queue_lock already
> + */
> +static inline void vpfe_schedule_next_buffer(struct vpfe_device *vpfe)
> +{
> + vpfe->next_frm = list_entry(vpfe->dma_queue.next,
> + struct vpfe_cap_buffer, list);
> + list_del(&vpfe->next_frm->list);
> +
> + vpfe_set_sdr_addr(&vpfe->ccdc,
> + vb2_dma_contig_plane_dma_addr(&vpfe->next_frm->vb, 0));
> +}
> +
> +static inline void vpfe_schedule_bottom_field(struct vpfe_device *vpfe)
> +{
> + unsigned long addr;
> +
> + addr = vb2_dma_contig_plane_dma_addr(&vpfe->next_frm->vb, 0) +
> + vpfe->field_off;
> +
> + vpfe_set_sdr_addr(&vpfe->ccdc, addr);
> +}
> +
> +/*
> + * vpfe_process_buffer_complete: process a completed buffer
> + * @vpfe : ptr to vpfe device
> + *
> + * This function time stamp the buffer and mark it as DONE. It also
> + * wake up any process waiting on the QUEUE and set the next buffer
> + * as current
> + */
> +static inline void vpfe_process_buffer_complete(struct vpfe_device *vpfe)
> +{
> + v4l2_get_timestamp(&vpfe->cur_frm->vb.v4l2_buf.timestamp);
> + vpfe->cur_frm->vb.v4l2_buf.field = vpfe->fmt.fmt.pix.field;
> + vpfe->cur_frm->vb.v4l2_buf.sequence = vpfe->sequence++;
> + vb2_buffer_done(&vpfe->cur_frm->vb, VB2_BUF_STATE_DONE);
> + vpfe->cur_frm = vpfe->next_frm;
> +}
> +
> +/*
> + * vpfe_isr : ISR handler for vpfe capture (VINT0)
> + * @irq: irq number
> + * @dev_id: dev_id ptr
> + *
> + * It changes status of the captured buffer, takes next buffer from the queue
> + * and sets its address in VPFE registers
> + */
> +static irqreturn_t vpfe_isr(int irq, void *dev)
> +{
> + struct vpfe_device *vpfe = (struct vpfe_device *)dev;
> + enum v4l2_field field;
> + int intr_status;
> + int fid;
> +
> + intr_status = vpfe_reg_read(&vpfe->ccdc, VPFE_IRQ_STS);
> +
> + if (intr_status & VPFE_VDINT0) {
> + field = vpfe->fmt.fmt.pix.field;
> +
> + if (field == V4L2_FIELD_NONE) {
> + /* handle progressive frame capture */
> + if (vpfe->cur_frm != vpfe->next_frm)
> + vpfe_process_buffer_complete(vpfe);
> + goto next_intr;
> + }
> +
> + /* interlaced or TB capture check which field
> + we are in hardware */
> + fid = vpfe_ccdc_getfid(&vpfe->ccdc);
> +
> + /* switch the software maintained field id */
> + vpfe->field ^= 1;
> + if (fid == vpfe->field) {
> + /* we are in-sync here,continue */
> + if (fid == 0) {
> + /*
> + * One frame is just being captured. If the
> + * next frame is available, release the
> + * current frame and move on
> + */
> + if (vpfe->cur_frm != vpfe->next_frm)
> + vpfe_process_buffer_complete(vpfe);
> + /*
> + * based on whether the two fields are stored
> + * interleave or separately in memory,
> + * reconfigure the CCDC memory address
> + */
> + if (field == V4L2_FIELD_SEQ_TB)
> + vpfe_schedule_bottom_field(vpfe);
> +
> + goto next_intr;
> + }
> + /*
> + * if one field is just being captured configure
> + * the next frame get the next frame from the empty
> + * queue if no frame is available hold on to the
> + * current buffer
> + */
> + spin_lock(&vpfe->dma_queue_lock);
> + if (!list_empty(&vpfe->dma_queue) &&
> + vpfe->cur_frm == vpfe->next_frm)
> + vpfe_schedule_next_buffer(vpfe);
> + spin_unlock(&vpfe->dma_queue_lock);
> + } else if (fid == 0) {
> + /*
> + * out of sync. Recover from any hardware out-of-sync.
> + * May loose one frame
> + */
> + vpfe->field = fid;
> + }
> + }
> +
> +next_intr:
> + if (intr_status & VPFE_VDINT1) {
> + spin_lock(&vpfe->dma_queue_lock);
> + if (vpfe->fmt.fmt.pix.field == V4L2_FIELD_NONE &&
> + !list_empty(&vpfe->dma_queue) &&
> + vpfe->cur_frm == vpfe->next_frm)
> + vpfe_schedule_next_buffer(vpfe);
> + spin_unlock(&vpfe->dma_queue_lock);
> + }
> +
> + vpfe_clear_intr(&vpfe->ccdc, intr_status);
> +
> + return IRQ_HANDLED;
> +}
> +
> +static inline void vpfe_detach_irq(struct vpfe_device *vpfe)
> +{
> + unsigned int intr = VPFE_VDINT0;
> + enum ccdc_frmfmt frame_format;
> +
> + frame_format = vpfe_ccdc_get_frame_format(&vpfe->ccdc);
> + if (frame_format == CCDC_FRMFMT_PROGRESSIVE)
> + intr |= VPFE_VDINT1;
> +
> + vpfe_reg_write(&vpfe->ccdc, intr, VPFE_IRQ_EN_CLR);
> +}
> +
> +static inline void vpfe_attach_irq(struct vpfe_device *vpfe)
> +{
> + unsigned int intr = VPFE_VDINT0;
> + enum ccdc_frmfmt frame_format;
> +
> + frame_format = vpfe_ccdc_get_frame_format(&vpfe->ccdc);
> + if (frame_format == CCDC_FRMFMT_PROGRESSIVE)
> + intr |= VPFE_VDINT1;
> +
> + vpfe_reg_write(&vpfe->ccdc, intr, VPFE_IRQ_EN_SET);
> +}
> +
> +static int vpfe_querycap(struct file *file, void *priv,
> + struct v4l2_capability *cap)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> +
> + vpfe_dbg(2, vpfe, "vpfe_querycap\n");
> +
> + strlcpy(cap->driver, VPFE_MODULE_NAME, sizeof(cap->driver));
> + strlcpy(cap->card, "TI AM437x VPFE", sizeof(cap->card));
> + snprintf(cap->bus_info, sizeof(cap->bus_info),
> + "platform:%s", vpfe->v4l2_dev.name);
> + cap->device_caps = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING |
> + V4L2_CAP_READWRITE;
> + cap->capabilities = cap->device_caps | V4L2_CAP_DEVICE_CAPS;
> +
> + return 0;
> +}
> +
> +/* get the format set at output pad of the adjacent subdev */
> +static int __vpfe_get_format(struct vpfe_device *vpfe,
> + struct v4l2_format *format, unsigned int *bpp)
> +{
> + struct v4l2_mbus_framefmt mbus_fmt;
> + struct vpfe_subdev_info *sdinfo;
> + struct v4l2_subdev_format fmt;
> + int ret;
> +
> + sdinfo = vpfe->current_subdev;
> + if (!sdinfo->sd)
> + return -EINVAL;
> +
> + fmt.which = V4L2_SUBDEV_FORMAT_ACTIVE;
> + fmt.pad = 0;
> +
> + ret = v4l2_subdev_call(sdinfo->sd, pad, get_fmt, NULL, &fmt);
> + if (ret && ret != -ENOIOCTLCMD && ret != -ENODEV)
> + return ret;
> +
> + if (!ret) {
> + v4l2_fill_pix_format(&format->fmt.pix, &fmt.format);
> + mbus_to_pix(vpfe, &fmt.format, &format->fmt.pix, bpp);
> + } else {
> + ret = v4l2_device_call_until_err(&vpfe->v4l2_dev,
> + sdinfo->grp_id,
> + video, g_mbus_fmt,
> + &mbus_fmt);
> + if (ret && ret != -ENOIOCTLCMD && ret != -ENODEV)
> + return ret;
> + v4l2_fill_pix_format(&format->fmt.pix, &mbus_fmt);
> + mbus_to_pix(vpfe, &mbus_fmt, &format->fmt.pix, bpp);
> + }
> +
> + format->type = vpfe->fmt.type;
> +
> + vpfe_dbg(1, vpfe,
> + "%s size %dx%d (%s) bytesperline= %d, size= %d, bpp= %d\n",
> + __func__, format->fmt.pix.width, format->fmt.pix.height,
> + print_fourcc(format->fmt.pix.pixelformat),
> + format->fmt.pix.bytesperline, format->fmt.pix.sizeimage, *bpp);
> +
> + return 0;
> +}
> +
> +/* set the format at output pad of the adjacent subdev */
> +static int __vpfe_set_format(struct vpfe_device *vpfe,
> + struct v4l2_format *format, unsigned int *bpp)
> +{
> + struct vpfe_subdev_info *sdinfo;
> + struct v4l2_subdev_format fmt;
> + int ret;
> +
> + vpfe_dbg(2, vpfe, "__vpfe_set_format\n");
> +
> + sdinfo = vpfe->current_subdev;
> + if (!sdinfo->sd)
> + return -EINVAL;
> +
> + fmt.which = V4L2_SUBDEV_FORMAT_ACTIVE;
> + fmt.pad = 0;
> +
> + ret = pix_to_mbus(vpfe, &format->fmt.pix, &fmt.format);
> + if (ret)
> + return ret;
> +
> + ret = v4l2_subdev_call(sdinfo->sd, pad, set_fmt, NULL, &fmt);
> + if (ret == -ENOIOCTLCMD)
> + return -EINVAL;
> +
> + format->type = vpfe->fmt.type;
> +
> + /* convert mbus_format to v4l2_format */
> + v4l2_fill_pix_format(&format->fmt.pix, &fmt.format);
> + mbus_to_pix(vpfe, &fmt.format, &format->fmt.pix, bpp);
> + vpfe_dbg(1, vpfe, "__vpfe_set_format size %dx%d (%s) bytesperline = %d, sizeimage = %d, bpp = %d\n",
> + format->fmt.pix.width, format->fmt.pix.height,
> + print_fourcc(format->fmt.pix.pixelformat),
> + format->fmt.pix.bytesperline, format->fmt.pix.sizeimage, *bpp);
> +
> + return 0;
> +}
> +
> +static int vpfe_g_fmt(struct file *file, void *priv,
> + struct v4l2_format *fmt)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> +
> + vpfe_dbg(2, vpfe, "vpfe_g_fmt\n");
> +
> + *fmt = vpfe->fmt;
> +
> + return 0;
> +}
> +
> +static int vpfe_enum_fmt(struct file *file, void *priv,
> + struct v4l2_fmtdesc *f)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + struct vpfe_subdev_info *sdinfo;
> + struct vpfe_fmt *fmt = NULL;
> + unsigned int k;
> +
> + vpfe_dbg(2, vpfe, "vpfe_enum_format index:%d\n",
> + f->index);
> +
> + sdinfo = vpfe->current_subdev;
> + if (!sdinfo->sd)
> + return -EINVAL;
> +
> + if (f->index > ARRAY_SIZE(formats))
> + return -EINVAL;
> +
> + for (k = 0; k < ARRAY_SIZE(formats); k++) {
> + if (formats[k].index == f->index) {
> + fmt = &formats[k];
> + break;
> + }
> + }
> + if (!fmt)
> + return -EINVAL;
> +
> + strncpy(f->description, fmt->name, sizeof(f->description) - 1);
> + f->pixelformat = fmt->fourcc;
> + f->type = vpfe->fmt.type;
> +
> + vpfe_dbg(1, vpfe, "vpfe_enum_format: mbus index: %d code: %x pixelformat: %s [%s]\n",
> + f->index, fmt->code, print_fourcc(fmt->fourcc), fmt->name);
> +
> + return 0;
> +}
> +
> +static int vpfe_try_fmt(struct file *file, void *priv,
> + struct v4l2_format *fmt)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + unsigned int bpp;
> +
> + vpfe_dbg(2, vpfe, "vpfe_try_fmt\n");
> +
> + return __vpfe_get_format(vpfe, fmt, &bpp);
> +}
> +
> +static int vpfe_s_fmt(struct file *file, void *priv,
> + struct v4l2_format *fmt)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + struct v4l2_format format;
> + unsigned int bpp;
> + int ret;
> +
> + vpfe_dbg(2, vpfe, "vpfe_s_fmt\n");
> +
> + /* If streaming is started, return error */
> + if (vb2_is_busy(&vpfe->buffer_queue)) {
> + vpfe_err(vpfe, "%s device busy\n", __func__);
> + return -EBUSY;
> + }
> +
> + ret = vpfe_try_fmt(file, priv, fmt);
> + if (ret)
> + return ret;
> +
> +
> + if (!cmp_v4l2_format(fmt, &format)) {
> + /* Sensor format is different from the requested format
> + * so we need to change it
> + */
> + ret = __vpfe_set_format(vpfe, fmt, &bpp);
> + if (ret)
> + return ret;
> + } else /* Just make sure all of the fields are consistent */
> + *fmt = format;
> +
> + /* First detach any IRQ if currently attached */
> + vpfe_detach_irq(vpfe);
> + vpfe->fmt = *fmt;
> + vpfe->bpp = bpp;
> +
> + /* Update the crop window based on found values */
> + vpfe->crop.width = fmt->fmt.pix.width;
> + vpfe->crop.height = fmt->fmt.pix.height;
> +
> + /* set image capture parameters in the ccdc */
> + return vpfe_config_ccdc_image_format(vpfe);
> +}
> +
> +static int vpfe_enum_size(struct file *file, void *priv,
> + struct v4l2_frmsizeenum *fsize)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + struct v4l2_subdev_frame_size_enum fse;
> + struct vpfe_subdev_info *sdinfo;
> + struct v4l2_mbus_framefmt mbus;
> + struct v4l2_pix_format pix;

Zero this pix struct so that there are no random values.
Same for fse and mbus.

> + struct vpfe_fmt *fmt;
> + int ret;
> +
> + vpfe_dbg(2, vpfe, "vpfe_enum_size\n");
> +
> + /* check for valid format */
> + fmt = find_format_by_pix(fsize->pixel_format);
> + if (!fmt) {
> + vpfe_dbg(3, vpfe, "Invalid pixel code: %x, default used instead\n",
> + fsize->pixel_format);
> + return -EINVAL;
> + }
> +
> + memset(fsize->reserved, 0x00, sizeof(fsize->reserved));
> +
> + sdinfo = vpfe->current_subdev;
> + if (!sdinfo->sd)
> + return -EINVAL;
> +
> + /* Construct pix from parameter and use default for the rest */
> + pix.pixelformat = fsize->pixel_format;
> + pix.width = 640;
> + pix.height = 480;
> + pix.colorspace = V4L2_COLORSPACE_SRGB;
> + pix.field = V4L2_FIELD_NONE;
> + ret = pix_to_mbus(vpfe, &pix, &mbus);
> + if (ret)
> + return ret;
> +
> + fse.index = fsize->index;
> + fse.pad = 0;
> + fse.code = mbus.code;
> + ret = v4l2_subdev_call(sdinfo->sd, pad, enum_frame_size, NULL, &fse);
> + if (ret)
> + return -EINVAL;
> +
> + vpfe_dbg(1, vpfe, "vpfe_enum_size: index: %d code: %x W:[%d,%d] H:[%d,%d]\n",
> + fse.index, fse.code, fse.min_width, fse.max_width,
> + fse.min_height, fse.max_height);
> +
> + fsize->type = V4L2_FRMSIZE_TYPE_DISCRETE;
> + fsize->discrete.width = fse.max_width;
> + fsize->discrete.height = fse.max_height;
> +
> + vpfe_dbg(1, vpfe, "vpfe_enum_size: index: %d pixformat: %s size: %dx%d\n",
> + fsize->index, print_fourcc(fsize->pixel_format),
> + fsize->discrete.width, fsize->discrete.height);
> +
> + return 0;
> +}
> +
> +/*
> + * vpfe_get_subdev_input_index - Get subdev index and subdev input index for a
> + * given app input index
> + */
> +static int
> +vpfe_get_subdev_input_index(struct vpfe_device *vpfe,
> + int *subdev_index,
> + int *subdev_input_index,
> + int app_input_index)
> +{
> + struct vpfe_config *cfg = vpfe->cfg;
> + struct vpfe_subdev_info *sdinfo;
> + int i, j = 0;
> +
> + for (i = 0; i < ARRAY_SIZE(vpfe->cfg->asd); i++) {
> + sdinfo = &cfg->sub_devs[i];
> + if (app_input_index < (j + 1)) {
> + *subdev_index = i;
> + *subdev_input_index = app_input_index - j;
> + return 0;
> + }
> + j++;
> + }
> + return -EINVAL;
> +}
> +
> +/*
> + * vpfe_get_app_input - Get app input index for a given subdev input index
> + * driver stores the input index of the current sub device and translate it
> + * when application request the current input
> + */
> +static int vpfe_get_app_input_index(struct vpfe_device *vpfe,
> + int *app_input_index)
> +{
> + struct vpfe_config *cfg = vpfe->cfg;
> + struct vpfe_subdev_info *sdinfo;
> + int i, j = 0;
> +
> + for (i = 0; i < ARRAY_SIZE(vpfe->cfg->asd); i++) {
> + sdinfo = &cfg->sub_devs[i];
> + if (!strcmp(sdinfo->name, vpfe->current_subdev->name)) {
> + if (vpfe->current_input >= 1)
> + return -1;
> + *app_input_index = j + vpfe->current_input;
> + return 0;
> + }
> + j++;
> + }
> + return -EINVAL;
> +}
> +
> +static int vpfe_enum_input(struct file *file, void *priv,
> + struct v4l2_input *inp)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + struct vpfe_subdev_info *sdinfo;
> + int subdev, index;
> +
> + vpfe_dbg(2, vpfe, "vpfe_enum_input\n");
> +
> + if (vpfe_get_subdev_input_index(vpfe,
> + &subdev,
> + &index,
> + inp->index) < 0) {
> + vpfe_dbg(1, vpfe,
> + "input information not found for the subdev\n");
> + return -EINVAL;
> + }
> + sdinfo = &vpfe->cfg->sub_devs[subdev];
> + *inp = sdinfo->inputs[index];
> +
> + return 0;
> +}
> +
> +static int vpfe_g_input(struct file *file, void *priv, unsigned int *index)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> +
> + vpfe_dbg(2, vpfe, "vpfe_g_input\n");
> +
> + return vpfe_get_app_input_index(vpfe, index);
> +}
> +
> +/* Assumes caller is holding vpfe_dev->lock */
> +static int vpfe_set_input(struct vpfe_device *vpfe, unsigned int index)
> +{
> + int subdev_index = 0, inp_index = 0;
> + struct vpfe_subdev_info *sdinfo;
> + struct vpfe_route *route;
> + u32 input, output;
> + int ret;
> +
> + vpfe_dbg(2, vpfe, "vpfe_set_input: index: %d\n", index);
> +
> + /* If streaming is started, return error */
> + if (vb2_is_busy(&vpfe->buffer_queue)) {
> + vpfe_err(vpfe, "%s device busy\n", __func__);
> + return -EBUSY;
> + }
> + ret = vpfe_get_subdev_input_index(vpfe,
> + &subdev_index,
> + &inp_index,
> + index);
> + if (ret < 0) {
> + vpfe_err(vpfe, "invalid input index: %d\n", index);
> + goto get_out;
> + }
> +
> + sdinfo = &vpfe->cfg->sub_devs[subdev_index];
> + sdinfo->sd = vpfe->sd[subdev_index];
> + route = &sdinfo->routes[inp_index];
> + if (route && sdinfo->can_route) {
> + input = route->input;
> + output = route->output;
> + if (sdinfo->sd) {
> + ret = v4l2_subdev_call(sdinfo->sd, video,
> + s_routing, input, output, 0);
> + if (ret) {
> + vpfe_err(vpfe, "s_routing failed\n");
> + ret = -EINVAL;
> + goto get_out;
> + }
> + }
> +
> + }
> +
> + vpfe->current_subdev = sdinfo;
> + if (sdinfo->sd)
> + vpfe->v4l2_dev.ctrl_handler = sdinfo->sd->ctrl_handler;
> + vpfe->current_input = index;
> + vpfe->std_index = 0;
> +
> + /* set the bus/interface parameter for the sub device in ccdc */
> + ret = vpfe_ccdc_set_hw_if_params(&vpfe->ccdc, &sdinfo->vpfe_param);
> + if (ret)
> + return ret;
> +
> + /* set the default image parameters in the device */
> + return vpfe_config_image_format(vpfe,
> + vpfe_standards[vpfe->std_index].std_id);
> +
> +get_out:
> + return ret;
> +}
> +
> +static int vpfe_s_input(struct file *file, void *priv, unsigned int index)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> +
> + vpfe_dbg(2, vpfe,
> + "vpfe_s_input: index: %d\n", index);
> +
> + return vpfe_set_input(vpfe, index);
> +}
> +
> +static int vpfe_querystd(struct file *file, void *priv, v4l2_std_id *std_id)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + struct vpfe_subdev_info *sdinfo;
> +
> + vpfe_dbg(2, vpfe, "vpfe_querystd\n");
> +
> + sdinfo = vpfe->current_subdev;
> + if (!(sdinfo->inputs[0].capabilities & V4L2_IN_CAP_STD))
> + return -ENODATA;
> +
> + /* Call querystd function of decoder device */
> + return v4l2_device_call_until_err(&vpfe->v4l2_dev, sdinfo->grp_id,
> + video, querystd, std_id);
> +}
> +
> +static int vpfe_s_std(struct file *file, void *priv, v4l2_std_id std_id)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + struct vpfe_subdev_info *sdinfo;
> + int ret;
> +
> + vpfe_dbg(2, vpfe, "vpfe_s_std\n");
> +
> + sdinfo = vpfe->current_subdev;
> + if (!(sdinfo->inputs[0].capabilities & V4L2_IN_CAP_STD))
> + return -ENODATA;
> +
> + /* If streaming is started, return error */
> + if (vb2_is_busy(&vpfe->buffer_queue)) {
> + vpfe_err(vpfe, "%s device busy\n", __func__);
> + ret = -EBUSY;
> + return ret;
> + }
> +
> + ret = v4l2_device_call_until_err(&vpfe->v4l2_dev, sdinfo->grp_id,
> + video, s_std, std_id);
> + if (ret < 0) {
> + vpfe_err(vpfe, "Failed to set standard\n");
> + return ret;
> + }
> + ret = vpfe_config_image_format(vpfe, std_id);
> +
> + return ret;
> +}
> +
> +static int vpfe_g_std(struct file *file, void *priv, v4l2_std_id *std_id)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + struct vpfe_subdev_info *sdinfo;
> +
> + vpfe_dbg(2, vpfe, "vpfe_g_std\n");
> +
> + sdinfo = vpfe->current_subdev;
> + if (sdinfo->inputs[0].capabilities != V4L2_IN_CAP_STD)
> + return -ENODATA;
> +
> + *std_id = vpfe_standards[vpfe->std_index].std_id;
> +
> + return 0;
> +}
> +
> +/*
> + * vpfe_calculate_offsets : This function calculates buffers offset
> + * for top and bottom field
> + */
> +static void vpfe_calculate_offsets(struct vpfe_device *vpfe)
> +{
> + struct v4l2_rect image_win;
> +
> + vpfe_dbg(2, vpfe, "vpfe_calculate_offsets\n");
> +
> + vpfe_ccdc_get_image_window(&vpfe->ccdc, &image_win);
> + vpfe->field_off = image_win.height * image_win.width;
> +}
> +
> +/*
> + * vpfe_queue_setup - Callback function for buffer setup.
> + * @vq: vb2_queue ptr
> + * @fmt: v4l2 format
> + * @nbuffers: ptr to number of buffers requested by application
> + * @nplanes:: contains number of distinct video planes needed to hold a frame
> + * @sizes[]: contains the size (in bytes) of each plane.
> + * @alloc_ctxs: ptr to allocation context
> + *
> + * This callback function is called when reqbuf() is called to adjust
> + * the buffer count and buffer size
> + */
> +static int vpfe_queue_setup(struct vb2_queue *vq,
> + const struct v4l2_format *fmt,
> + unsigned int *nbuffers, unsigned int *nplanes,
> + unsigned int sizes[], void *alloc_ctxs[])
> +{
> + struct vpfe_device *vpfe = vb2_get_drv_priv(vq);
> +
> + if (fmt && fmt->fmt.pix.sizeimage < vpfe->fmt.fmt.pix.sizeimage)
> + return -EINVAL;
> +
> + if (vq->num_buffers + *nbuffers < 3)
> + *nbuffers = 3 - vq->num_buffers;
> +
> + *nplanes = 1;
> + sizes[0] = fmt ? fmt->fmt.pix.sizeimage : vpfe->fmt.fmt.pix.sizeimage;
> + alloc_ctxs[0] = vpfe->alloc_ctx;
> +
> + vpfe_dbg(1, vpfe,
> + "nbuffers=%d, size=%u\n", *nbuffers, sizes[0]);
> +
> + /* Calculate field offset */
> + vpfe_calculate_offsets(vpfe);
> +
> + return 0;
> +}
> +
> +/*
> + * vpfe_buffer_prepare : callback function for buffer prepare
> + * @vb: ptr to vb2_buffer
> + *
> + * This is the callback function for buffer prepare when vb2_qbuf()
> + * function is called. The buffer is prepared and user space virtual address
> + * or user address is converted into physical address
> + */
> +static int vpfe_buffer_prepare(struct vb2_buffer *vb)
> +{
> + struct vpfe_device *vpfe = vb2_get_drv_priv(vb->vb2_queue);
> +
> + vb2_set_plane_payload(vb, 0, vpfe->fmt.fmt.pix.sizeimage);
> +
> + if (vb2_get_plane_payload(vb, 0) > vb2_plane_size(vb, 0))
> + return -EINVAL;
> +
> + vb->v4l2_buf.field = vpfe->fmt.fmt.pix.field;
> +
> + return 0;
> +}
> +
> +/*
> + * vpfe_buffer_queue : Callback function to add buffer to DMA queue
> + * @vb: ptr to vb2_buffer
> + */
> +static void vpfe_buffer_queue(struct vb2_buffer *vb)
> +{
> + struct vpfe_device *vpfe = vb2_get_drv_priv(vb->vb2_queue);
> + struct vpfe_cap_buffer *buf = to_vpfe_buffer(vb);
> + unsigned long flags = 0;
> +
> + /* add the buffer to the DMA queue */
> + spin_lock_irqsave(&vpfe->dma_queue_lock, flags);
> + list_add_tail(&buf->list, &vpfe->dma_queue);
> + spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags);
> +}
> +
> +/*
> + * vpfe_start_streaming : Starts the DMA engine for streaming
> + * @vb: ptr to vb2_buffer
> + * @count: number of buffers
> + */
> +static int vpfe_start_streaming(struct vb2_queue *vq, unsigned int count)
> +{
> + struct vpfe_device *vpfe = vb2_get_drv_priv(vq);
> + struct vpfe_cap_buffer *buf, *tmp;
> + struct vpfe_subdev_info *sdinfo;
> + unsigned long flags;
> + unsigned long addr;
> + int ret;
> +
> + spin_lock_irqsave(&vpfe->dma_queue_lock, flags);
> +
> + vpfe->field = 0;
> + vpfe->sequence = 0;
> +
> + sdinfo = vpfe->current_subdev;
> +
> + vpfe_attach_irq(vpfe);
> +
> + if (vpfe->ccdc.ccdc_cfg.if_type == VPFE_RAW_BAYER)
> + vpfe_ccdc_config_raw(&vpfe->ccdc);
> + else
> + vpfe_ccdc_config_ycbcr(&vpfe->ccdc);
> +
> + /* Get the next frame from the buffer queue */
> + vpfe->next_frm = list_entry(vpfe->dma_queue.next,
> + struct vpfe_cap_buffer, list);
> + vpfe->cur_frm = vpfe->next_frm;
> + /* Remove buffer from the buffer queue */
> + list_del(&vpfe->cur_frm->list);
> + spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags);
> +
> + addr = vb2_dma_contig_plane_dma_addr(&vpfe->cur_frm->vb, 0);
> +
> + vpfe_set_sdr_addr(&vpfe->ccdc, (unsigned long)(addr));
> +
> + vpfe_pcr_enable(&vpfe->ccdc, 1);
> +
> + ret = v4l2_subdev_call(sdinfo->sd, video, s_stream, 1);
> + if (ret < 0) {
> + vpfe_err(vpfe, "Error in attaching interrupt handle\n");
> + goto err;
> + }
> +
> + return 0;
> +
> +err:
> + list_for_each_entry_safe(buf, tmp, &vpfe->dma_queue, list) {
> + list_del(&buf->list);
> + vb2_buffer_done(&buf->vb, VB2_BUF_STATE_QUEUED);
> + }
> + spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags);
> +
> + return ret;
> +}
> +
> +/*
> + * vpfe_stop_streaming : Stop the DMA engine
> + * @vq: ptr to vb2_queue
> + *
> + * This callback stops the DMA engine and any remaining buffers
> + * in the DMA queue are released.
> + */
> +static void vpfe_stop_streaming(struct vb2_queue *vq)
> +{
> + struct vpfe_device *vpfe = vb2_get_drv_priv(vq);
> + struct vpfe_subdev_info *sdinfo;
> + unsigned long flags;
> + int ret;
> +
> + vpfe_pcr_enable(&vpfe->ccdc, 0);
> +
> + vpfe_detach_irq(vpfe);
> +
> + sdinfo = vpfe->current_subdev;
> + ret = v4l2_subdev_call(sdinfo->sd, video, s_stream, 0);
> + if (ret && ret != -ENOIOCTLCMD && ret != -ENODEV)
> + vpfe_dbg(1, vpfe, "stream off failed in subdev\n");
> +
> + /* release all active buffers */
> + spin_lock_irqsave(&vpfe->dma_queue_lock, flags);
> + if (vpfe->cur_frm == vpfe->next_frm) {
> + vb2_buffer_done(&vpfe->cur_frm->vb, VB2_BUF_STATE_ERROR);
> + } else {
> + if (vpfe->cur_frm != NULL)
> + vb2_buffer_done(&vpfe->cur_frm->vb,
> + VB2_BUF_STATE_ERROR);
> + if (vpfe->next_frm != NULL)
> + vb2_buffer_done(&vpfe->next_frm->vb,
> + VB2_BUF_STATE_ERROR);
> + }
> +
> + while (!list_empty(&vpfe->dma_queue)) {
> + vpfe->next_frm = list_entry(vpfe->dma_queue.next,
> + struct vpfe_cap_buffer, list);
> + list_del(&vpfe->next_frm->list);
> + vb2_buffer_done(&vpfe->next_frm->vb, VB2_BUF_STATE_ERROR);
> + }
> + spin_unlock_irqrestore(&vpfe->dma_queue_lock, flags);
> +}
> +
> +static int vpfe_cropcap(struct file *file, void *priv,
> + struct v4l2_cropcap *crop)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> +
> + vpfe_dbg(2, vpfe, "vpfe_cropcap\n");
> +
> + if (vpfe->std_index >= ARRAY_SIZE(vpfe_standards))
> + return -EINVAL;
> +
> + memset(crop, 0, sizeof(struct v4l2_cropcap));
> +
> + crop->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
> + crop->defrect.width = vpfe_standards[vpfe->std_index].width;
> + crop->bounds.width = crop->defrect.width;
> + crop->defrect.height = vpfe_standards[vpfe->std_index].height;
> + crop->bounds.height = crop->defrect.height;
> + crop->pixelaspect = vpfe_standards[vpfe->std_index].pixelaspect;
> +
> + return 0;
> +}
> +
> +static int
> +vpfe_g_selection(struct file *file, void *fh, struct v4l2_selection *s)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> +
> + switch (s->target) {
> + case V4L2_SEL_TGT_COMPOSE_DEFAULT:
> + case V4L2_SEL_TGT_COMPOSE_BOUNDS:
> + case V4L2_SEL_TGT_CROP_BOUNDS:
> + case V4L2_SEL_TGT_CROP_DEFAULT:
> + s->r.left = s->r.top = 0;
> + s->r.width = vpfe->crop.width;
> + s->r.height = vpfe->crop.height;
> + break;
> +
> + case V4L2_SEL_TGT_COMPOSE:

Why is COMPOSE supported when AFAICT only crop is implemented?

> + case V4L2_SEL_TGT_CROP:
> + s->r = vpfe->crop;
> + break;
> +
> + default:
> + return -EINVAL;
> + }
> +
> + return 0;
> +}
> +
> +static int enclosed_rectangle(struct v4l2_rect *a, struct v4l2_rect *b)
> +{
> + if (a->left < b->left || a->top < b->top)
> + return 0;
> +
> + if (a->left + a->width > b->left + b->width)
> + return 0;
> +
> + if (a->top + a->height > b->top + b->height)
> + return 0;
> +
> + return 1;
> +}
> +
> +static int
> +vpfe_s_selection(struct file *file, void *fh, struct v4l2_selection *s)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + struct v4l2_rect cr = vpfe->crop;
> + struct v4l2_rect r = s->r;
> +
> + /* If streaming is started, return error */
> + if (vb2_is_busy(&vpfe->buffer_queue)) {
> + vpfe_err(vpfe, "%s device busy\n", __func__);
> + return -EBUSY;
> + }
> +
> + if (s->type != V4L2_BUF_TYPE_VIDEO_CAPTURE ||
> + (s->target != V4L2_SEL_TGT_COMPOSE &&

Same question.

> + s->target != V4L2_SEL_TGT_CROP))
> + return -EINVAL;
> +
> + v4l_bound_align_image(&r.width, 0, cr.width, 0,
> + &r.height, 0, cr.height, 0, 0);
> +
> + r.left = clamp_t(unsigned int, r.left, 0, cr.width - r.width);
> + r.top = clamp_t(unsigned int, r.top, 0, cr.height - r.height);
> +
> + if (s->flags & V4L2_SEL_FLAG_LE && !enclosed_rectangle(&r, &s->r))
> + return -ERANGE;
> +
> + if (s->flags & V4L2_SEL_FLAG_GE && !enclosed_rectangle(&s->r, &r))
> + return -ERANGE;
> +
> + s->r = vpfe->crop = r;
> +
> + vpfe_ccdc_set_image_window(&vpfe->ccdc, &r, vpfe->bpp);
> + vpfe->fmt.fmt.pix.width = r.width;
> + vpfe->fmt.fmt.pix.height = r.height;
> + vpfe->fmt.fmt.pix.bytesperline = vpfe_ccdc_get_line_length(&vpfe->ccdc);
> + vpfe->fmt.fmt.pix.sizeimage = vpfe->fmt.fmt.pix.bytesperline *
> + vpfe->fmt.fmt.pix.height;
> +
> + vpfe_dbg(1, vpfe, "cropped (%d,%d)/%dx%d of %dx%d\n",
> + r.left, r.top, r.width, r.height, cr.width, cr.height);
> +
> + return 0;
> +}
> +
> +static long vpfe_ioctl_default(struct file *file, void *priv,
> + bool valid_prio, unsigned int cmd, void *param)
> +{
> + struct vpfe_device *vpfe = video_drvdata(file);
> + int ret;
> +
> + vpfe_dbg(2, vpfe, "vpfe_ioctl_default\n");
> +
> + if (!valid_prio) {
> + vpfe_err(vpfe, "%s device busy\n", __func__);
> + return -EBUSY;
> + }
> +
> + /* If streaming is started, return error */
> + if (vb2_is_busy(&vpfe->buffer_queue)) {
> + vpfe_err(vpfe, "%s device busy\n", __func__);
> + return -EBUSY;
> + }
> +
> + switch (cmd) {
> + case VIDIOC_AM437X_CCDC_CFG:
> + ret = vpfe_ccdc_set_params(&vpfe->ccdc, param);
> + if (ret) {
> + vpfe_dbg(2, vpfe,
> + "Error setting parameters in CCDC\n");
> + return ret;
> + }
> + ret = vpfe_get_ccdc_image_format(vpfe,
> + &vpfe->fmt);
> + if (ret < 0) {
> + vpfe_dbg(2, vpfe,
> + "Invalid image format at CCDC\n");
> + return ret;
> + }
> + break;
> +
> + default:
> + ret = -ENOTTY;
> + break;
> + }
> +
> + return ret;
> +}
> +
> +static const struct vb2_ops vpfe_video_qops = {
> + .wait_prepare = vb2_ops_wait_prepare,
> + .wait_finish = vb2_ops_wait_finish,
> + .queue_setup = vpfe_queue_setup,
> + .buf_prepare = vpfe_buffer_prepare,
> + .buf_queue = vpfe_buffer_queue,
> + .start_streaming = vpfe_start_streaming,
> + .stop_streaming = vpfe_stop_streaming,
> +};
> +
> +/* vpfe capture driver file operations */
> +static const struct v4l2_file_operations vpfe_fops = {
> + .owner = THIS_MODULE,
> + .open = vpfe_open,
> + .release = vpfe_release,
> + .read = vb2_fop_read,
> + .poll = vb2_fop_poll,
> + .unlocked_ioctl = video_ioctl2,
> + .mmap = vb2_fop_mmap,
> +
> +};
> +
> +/* vpfe capture ioctl operations */
> +static const struct v4l2_ioctl_ops vpfe_ioctl_ops = {
> + .vidioc_querycap = vpfe_querycap,
> + .vidioc_enum_fmt_vid_cap = vpfe_enum_fmt,
> + .vidioc_g_fmt_vid_cap = vpfe_g_fmt,
> + .vidioc_s_fmt_vid_cap = vpfe_s_fmt,
> + .vidioc_try_fmt_vid_cap = vpfe_try_fmt,
> +
> + .vidioc_enum_framesizes = vpfe_enum_size,
> +
> + .vidioc_enum_input = vpfe_enum_input,
> + .vidioc_g_input = vpfe_g_input,
> + .vidioc_s_input = vpfe_s_input,
> +
> + .vidioc_querystd = vpfe_querystd,
> + .vidioc_s_std = vpfe_s_std,
> + .vidioc_g_std = vpfe_g_std,
> +
> + .vidioc_reqbufs = vb2_ioctl_reqbufs,
> + .vidioc_create_bufs = vb2_ioctl_create_bufs,
> + .vidioc_prepare_buf = vb2_ioctl_prepare_buf,
> + .vidioc_querybuf = vb2_ioctl_querybuf,
> + .vidioc_qbuf = vb2_ioctl_qbuf,
> + .vidioc_dqbuf = vb2_ioctl_dqbuf,
> + .vidioc_expbuf = vb2_ioctl_expbuf,
> + .vidioc_streamon = vb2_ioctl_streamon,
> + .vidioc_streamoff = vb2_ioctl_streamoff,
> +
> + .vidioc_log_status = v4l2_ctrl_log_status,
> + .vidioc_subscribe_event = v4l2_ctrl_subscribe_event,
> + .vidioc_unsubscribe_event = v4l2_event_unsubscribe,
> +
> + .vidioc_cropcap = vpfe_cropcap,
> + .vidioc_g_selection = vpfe_g_selection,
> + .vidioc_s_selection = vpfe_s_selection,
> +
> + .vidioc_default = vpfe_ioctl_default,
> +};
> +
> +static int
> +vpfe_async_bound(struct v4l2_async_notifier *notifier,
> + struct v4l2_subdev *subdev,
> + struct v4l2_async_subdev *asd)
> +{
> + struct vpfe_device *vpfe = container_of(notifier->v4l2_dev,
> + struct vpfe_device, v4l2_dev);
> + struct v4l2_subdev_mbus_code_enum mbus_code;
> + struct vpfe_subdev_info *sdinfo;
> + bool found = false;
> + int i, j;
> +
> + vpfe_dbg(1, vpfe, "vpfe_async_bound\n");
> +
> + for (i = 0; i < ARRAY_SIZE(vpfe->cfg->asd); i++) {
> + sdinfo = &vpfe->cfg->sub_devs[i];
> +
> + if (!strcmp(sdinfo->name, subdev->name)) {
> + vpfe->sd[i] = subdev;
> + vpfe_info(vpfe,
> + "v4l2 sub device %s registered\n",
> + subdev->name);
> + vpfe->sd[i]->grp_id =
> + sdinfo->grp_id;
> + /* update tvnorms from the sub devices */
> + for (j = 0; j < 1; j++)
> + vpfe->video_dev->tvnorms |=
> + sdinfo->inputs[j].std;
> +
> + found = true;
> + break;
> + }
> + }
> +
> + if (!found) {
> + vpfe_info(vpfe, "sub device (%s) not matched\n", subdev->name);
> + return -EINVAL;
> + }
> +
> + for (j = 0; ; ++j) {
> + int ret;
> +
> + memset(&mbus_code, 0, sizeof(mbus_code));
> + mbus_code.index = j;
> + ret = v4l2_subdev_call(subdev, pad, enum_mbus_code,
> + NULL, &mbus_code);
> + if (ret)
> + break;
> +
> + find_format_by_code(mbus_code.code, true);
> + }
> +
> + return 0;
> +}
> +
> +static int vpfe_probe_complete(struct vpfe_device *vpfe)
> +{
> + struct video_device *vdev;
> + struct vb2_queue *q;
> + int err;
> +
> + spin_lock_init(&vpfe->dma_queue_lock);
> + mutex_init(&vpfe->lock);
> +
> + vpfe->fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
> +
> + /* set first sub device as current one */
> + vpfe->current_subdev = &vpfe->cfg->sub_devs[0];
> + vpfe->v4l2_dev.ctrl_handler = vpfe->sd[0]->ctrl_handler;
> +
> + err = vpfe_set_input(vpfe, 0);
> + if (err)
> + goto probe_out;
> +
> + /* Initialize videobuf2 queue as per the buffer type */
> + vpfe->alloc_ctx = vb2_dma_contig_init_ctx(vpfe->pdev);
> + if (IS_ERR(vpfe->alloc_ctx)) {
> + vpfe_err(vpfe, "Failed to get the context\n");
> + err = PTR_ERR(vpfe->alloc_ctx);
> + goto probe_out;
> + }
> +
> + q = &vpfe->buffer_queue;
> + q->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
> + q->io_modes = VB2_MMAP | VB2_DMABUF | VB2_READ;
> + q->drv_priv = vpfe;
> + q->ops = &vpfe_video_qops;
> + q->mem_ops = &vb2_dma_contig_memops;
> + q->buf_struct_size = sizeof(struct vpfe_cap_buffer);
> + q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
> + q->lock = &vpfe->lock;
> + q->min_buffers_needed = 1;
> +
> + err = vb2_queue_init(q);
> + if (err) {
> + vpfe_err(vpfe, "vb2_queue_init() failed\n");
> + vb2_dma_contig_cleanup_ctx(vpfe->alloc_ctx);
> + goto probe_out;
> + }
> +
> + INIT_LIST_HEAD(&vpfe->dma_queue);
> +
> + vdev = vpfe->video_dev;
> + strlcpy(vdev->name, VPFE_MODULE_NAME, sizeof(vdev->name));
> + vdev->release = video_device_release;
> + vdev->fops = &vpfe_fops;
> + vdev->ioctl_ops = &vpfe_ioctl_ops;
> + vdev->v4l2_dev = &vpfe->v4l2_dev;
> + vdev->vfl_dir = VFL_DIR_RX;
> + vdev->queue = q;
> + vdev->lock = &vpfe->lock;
> + video_set_drvdata(vdev, vpfe);
> + err = video_register_device(vpfe->video_dev, VFL_TYPE_GRABBER, -1);
> + if (err) {
> + vpfe_err(vpfe,
> + "Unable to register video device.\n");
> + goto probe_out;
> + }
> +
> + return 0;
> +
> +probe_out:
> + v4l2_device_unregister(&vpfe->v4l2_dev);
> + return err;
> +}
> +
> +static int vpfe_async_complete(struct v4l2_async_notifier *notifier)
> +{
> + struct vpfe_device *vpfe = container_of(notifier->v4l2_dev,
> + struct vpfe_device, v4l2_dev);
> +
> + return vpfe_probe_complete(vpfe);
> +}
> +
> +static struct vpfe_config *
> +vpfe_get_pdata(struct platform_device *pdev)
> +{
> + struct device_node *endpoint = NULL, *rem = NULL;
> + struct v4l2_of_endpoint bus_cfg;
> + struct vpfe_subdev_info *sdinfo;
> + struct vpfe_config *pdata;
> + unsigned int flags;
> + unsigned int i;
> + int err;
> +
> + dev_dbg(&pdev->dev, "vpfe_get_pdata\n");
> +
> + if (!IS_ENABLED(CONFIG_OF) || !pdev->dev.of_node)
> + return pdev->dev.platform_data;
> +
> + pdata = devm_kzalloc(&pdev->dev, sizeof(*pdata), GFP_KERNEL);
> + if (!pdata)
> + return NULL;
> +
> + for (i = 0; ; i++) {
> + endpoint = of_graph_get_next_endpoint(pdev->dev.of_node,
> + endpoint);
> + if (!endpoint)
> + break;
> +
> + sdinfo = &pdata->sub_devs[i];
> + sdinfo->grp_id = 0;
> +
> + /* we only support camera */
> + sdinfo->inputs[0].index = i;
> + strcpy(sdinfo->inputs[0].name, "Camera");
> + sdinfo->inputs[0].type = V4L2_INPUT_TYPE_CAMERA;
> + sdinfo->inputs[0].std = V4L2_STD_ALL;
> + sdinfo->inputs[0].capabilities = V4L2_IN_CAP_STD;
> +
> + sdinfo->can_route = 0;
> + sdinfo->routes = NULL;
> +
> + of_property_read_u32(endpoint, "ti,am437x-vpfe-interface",
> + &sdinfo->vpfe_param.if_type);
> + if (sdinfo->vpfe_param.if_type < 0 ||
> + sdinfo->vpfe_param.if_type > 4) {
> + sdinfo->vpfe_param.if_type = VPFE_RAW_BAYER;
> + }
> +
> + err = v4l2_of_parse_endpoint(endpoint, &bus_cfg);
> + if (err) {
> + dev_err(&pdev->dev, "Could not parse the endpoint\n");
> + goto done;
> + }
> +
> + sdinfo->vpfe_param.bus_width = bus_cfg.bus.parallel.bus_width;
> +
> + if (sdinfo->vpfe_param.bus_width < 8 ||
> + sdinfo->vpfe_param.bus_width > 16) {
> + dev_err(&pdev->dev, "Invalid bus width.\n");
> + goto done;
> + }
> +
> + flags = bus_cfg.bus.parallel.flags;
> +
> + if (flags & V4L2_MBUS_HSYNC_ACTIVE_HIGH)
> + sdinfo->vpfe_param.hdpol = 1;
> +
> + if (flags & V4L2_MBUS_VSYNC_ACTIVE_HIGH)
> + sdinfo->vpfe_param.vdpol = 1;
> +
> + rem = of_graph_get_remote_port_parent(endpoint);
> + if (!rem) {
> + dev_err(&pdev->dev, "Remote device at %s not found\n",
> + endpoint->full_name);
> + goto done;
> + }
> +
> + strncpy(sdinfo->name, rem->name, sizeof(sdinfo->name));
> +
> + pdata->asd[i] = devm_kzalloc(&pdev->dev,
> + sizeof(struct v4l2_async_subdev),
> + GFP_KERNEL);
> + pdata->asd[i]->match_type = V4L2_ASYNC_MATCH_OF;
> + pdata->asd[i]->match.of.node = rem;
> + of_node_put(endpoint);
> + of_node_put(rem);
> + }
> +
> + of_node_put(endpoint);
> + return pdata;
> +
> +done:
> + of_node_put(endpoint);
> + of_node_put(rem);
> + return NULL;
> +}
> +
> +/*
> + * vpfe_probe : This function creates device entries by register
> + * itself to the V4L2 driver and initializes fields of each
> + * device objects
> + */
> +static int vpfe_probe(struct platform_device *pdev)
> +{
> + struct vpfe_config *vpfe_cfg = vpfe_get_pdata(pdev);
> + struct vpfe_device *vpfe;
> + struct vpfe_ccdc *ccdc;
> + struct resource *res;
> + int ret;
> +
> + if (!vpfe_cfg) {
> + dev_err(&pdev->dev, "No platform data\n");
> + return -EINVAL;
> + }
> +
> + vpfe = devm_kzalloc(&pdev->dev, sizeof(*vpfe), GFP_KERNEL);
> + if (!vpfe)
> + return -ENOMEM;
> +
> + vpfe->pdev = &pdev->dev;
> + vpfe->cfg = vpfe_cfg;
> + ccdc = &vpfe->ccdc;
> +
> + res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
> + ccdc->ccdc_cfg.base_addr = devm_ioremap_resource(&pdev->dev, res);
> + if (IS_ERR(ccdc->ccdc_cfg.base_addr))
> + return PTR_ERR(ccdc->ccdc_cfg.base_addr);
> +
> + vpfe->irq = platform_get_irq(pdev, 0);
> + if (vpfe->irq <= 0) {
> + dev_err(&pdev->dev, "No IRQ resource\n");
> + return -ENODEV;
> + }
> +
> + ret = devm_request_irq(vpfe->pdev, vpfe->irq, vpfe_isr, 0,
> + "vpfe_capture0", vpfe);
> + if (ret) {
> + dev_err(&pdev->dev, "Unable to request interrupt\n");
> + return -EINVAL;
> + }
> +
> + vpfe->video_dev = video_device_alloc();
> + if (!vpfe->video_dev) {
> + dev_err(&pdev->dev, "Unable to alloc video device\n");
> + return -ENOMEM;
> + }
> +
> + ret = v4l2_device_register(&pdev->dev, &vpfe->v4l2_dev);
> + if (ret) {
> + vpfe_err(vpfe,
> + "Unable to register v4l2 device.\n");
> + goto probe_out_video_release;
> + }
> +
> + /* set the driver data in platform device */
> + platform_set_drvdata(pdev, vpfe);
> + /* Enabling module functional clock */
> + pm_runtime_enable(&pdev->dev);
> +
> + /* for now just enable it here instead of waiting for the open */
> + pm_runtime_get_sync(&pdev->dev);
> +
> + vpfe_ccdc_config_defaults(ccdc);
> +
> + pm_runtime_put_sync(&pdev->dev);
> +
> + vpfe->sd = devm_kzalloc(&pdev->dev, sizeof(struct v4l2_subdev *) *
> + ARRAY_SIZE(vpfe->cfg->asd), GFP_KERNEL);
> + if (!vpfe->sd) {
> + ret = -ENOMEM;
> + goto probe_out_v4l2_unregister;
> + }
> +
> + vpfe->notifier.subdevs = vpfe->cfg->asd;
> + vpfe->notifier.num_subdevs = ARRAY_SIZE(vpfe->cfg->asd);
> + vpfe->notifier.bound = vpfe_async_bound;
> + vpfe->notifier.complete = vpfe_async_complete;
> + ret = v4l2_async_notifier_register(&vpfe->v4l2_dev,
> + &vpfe->notifier);
> + if (ret) {
> + vpfe_err(vpfe, "Error registering async notifier\n");
> + ret = -EINVAL;
> + goto probe_out_v4l2_unregister;
> + }
> +
> + return 0;
> +
> +probe_out_v4l2_unregister:
> + v4l2_device_unregister(&vpfe->v4l2_dev);
> +probe_out_video_release:
> + if (!video_is_registered(vpfe->video_dev))
> + video_device_release(vpfe->video_dev);
> + return ret;
> +}
> +
> +/*
> + * vpfe_remove : It un-register device from V4L2 driver
> + */
> +static int vpfe_remove(struct platform_device *pdev)
> +{
> + struct vpfe_device *vpfe = platform_get_drvdata(pdev);
> +
> + vpfe_dbg(2, vpfe, "vpfe_remove\n");
> +
> + pm_runtime_disable(&pdev->dev);
> +
> + v4l2_async_notifier_unregister(&vpfe->notifier);
> + v4l2_device_unregister(&vpfe->v4l2_dev);
> + video_unregister_device(vpfe->video_dev);
> +
> + return 0;
> +}
> +
> +#ifdef CONFIG_PM_SLEEP
> +
> +static void vpfe_save_context(struct vpfe_ccdc *ccdc)
> +{
> + ccdc->ccdc_ctx[VPFE_PCR >> 2] = vpfe_reg_read(ccdc, VPFE_PCR);
> + ccdc->ccdc_ctx[VPFE_SYNMODE >> 2] = vpfe_reg_read(ccdc, VPFE_SYNMODE);
> + ccdc->ccdc_ctx[VPFE_SDOFST >> 2] = vpfe_reg_read(ccdc, VPFE_SDOFST);
> + ccdc->ccdc_ctx[VPFE_SDR_ADDR >> 2] = vpfe_reg_read(ccdc, VPFE_SDR_ADDR);
> + ccdc->ccdc_ctx[VPFE_CLAMP >> 2] = vpfe_reg_read(ccdc, VPFE_CLAMP);
> + ccdc->ccdc_ctx[VPFE_DCSUB >> 2] = vpfe_reg_read(ccdc, VPFE_DCSUB);
> + ccdc->ccdc_ctx[VPFE_COLPTN >> 2] = vpfe_reg_read(ccdc, VPFE_COLPTN);
> + ccdc->ccdc_ctx[VPFE_BLKCMP >> 2] = vpfe_reg_read(ccdc, VPFE_BLKCMP);
> + ccdc->ccdc_ctx[VPFE_VDINT >> 2] = vpfe_reg_read(ccdc, VPFE_VDINT);
> + ccdc->ccdc_ctx[VPFE_ALAW >> 2] = vpfe_reg_read(ccdc, VPFE_ALAW);
> + ccdc->ccdc_ctx[VPFE_REC656IF >> 2] = vpfe_reg_read(ccdc, VPFE_REC656IF);
> + ccdc->ccdc_ctx[VPFE_CCDCFG >> 2] = vpfe_reg_read(ccdc, VPFE_CCDCFG);
> + ccdc->ccdc_ctx[VPFE_CULLING >> 2] = vpfe_reg_read(ccdc, VPFE_CULLING);
> + ccdc->ccdc_ctx[VPFE_HD_VD_WID >> 2] = vpfe_reg_read(ccdc,
> + VPFE_HD_VD_WID);
> + ccdc->ccdc_ctx[VPFE_PIX_LINES >> 2] = vpfe_reg_read(ccdc,
> + VPFE_PIX_LINES);
> + ccdc->ccdc_ctx[VPFE_HORZ_INFO >> 2] = vpfe_reg_read(ccdc,
> + VPFE_HORZ_INFO);
> + ccdc->ccdc_ctx[VPFE_VERT_START >> 2] = vpfe_reg_read(ccdc,
> + VPFE_VERT_START);
> + ccdc->ccdc_ctx[VPFE_VERT_LINES >> 2] = vpfe_reg_read(ccdc,
> + VPFE_VERT_LINES);
> + ccdc->ccdc_ctx[VPFE_HSIZE_OFF >> 2] = vpfe_reg_read(ccdc,
> + VPFE_HSIZE_OFF);
> +}
> +
> +static int vpfe_suspend(struct device *dev)
> +{
> + struct platform_device *pdev = to_platform_device(dev);
> + struct vpfe_device *vpfe = platform_get_drvdata(pdev);
> + struct vpfe_ccdc *ccdc = &vpfe->ccdc;
> +
> + /* if streaming has not started we dont care */
> + if (!vb2_start_streaming_called(&vpfe->buffer_queue))
> + return 0;
> +
> + pm_runtime_get_sync(dev);
> + vpfe_config_enable(ccdc, 1);
> +
> + /* Save VPFE context */
> + vpfe_save_context(ccdc);
> +
> + /* Disable CCDC */
> + vpfe_pcr_enable(ccdc, 0);
> + vpfe_config_enable(ccdc, 0);
> +
> + /* Disable both master and slave clock */
> + pm_runtime_put_sync(dev);
> +
> + /* Select sleep pin state */
> + pinctrl_pm_select_sleep_state(dev);
> +
> + return 0;
> +}
> +
> +static void vpfe_restore_context(struct vpfe_ccdc *ccdc)
> +{
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_SYNMODE >> 2], VPFE_SYNMODE);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_CULLING >> 2], VPFE_CULLING);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_SDOFST >> 2], VPFE_SDOFST);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_SDR_ADDR >> 2], VPFE_SDR_ADDR);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_CLAMP >> 2], VPFE_CLAMP);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_DCSUB >> 2], VPFE_DCSUB);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_COLPTN >> 2], VPFE_COLPTN);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_BLKCMP >> 2], VPFE_BLKCMP);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_VDINT >> 2], VPFE_VDINT);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_ALAW >> 2], VPFE_ALAW);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_REC656IF >> 2], VPFE_REC656IF);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_CCDCFG >> 2], VPFE_CCDCFG);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_PCR >> 2], VPFE_PCR);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_HD_VD_WID >> 2],
> + VPFE_HD_VD_WID);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_PIX_LINES >> 2],
> + VPFE_PIX_LINES);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_HORZ_INFO >> 2],
> + VPFE_HORZ_INFO);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_VERT_START >> 2],
> + VPFE_VERT_START);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_VERT_LINES >> 2],
> + VPFE_VERT_LINES);
> + vpfe_reg_write(ccdc, ccdc->ccdc_ctx[VPFE_HSIZE_OFF >> 2],
> + VPFE_HSIZE_OFF);
> +}
> +
> +static int vpfe_resume(struct device *dev)
> +{
> + struct platform_device *pdev = to_platform_device(dev);
> + struct vpfe_device *vpfe = platform_get_drvdata(pdev);
> + struct vpfe_ccdc *ccdc = &vpfe->ccdc;
> +
> + /* if streaming has not started we dont care */
> + if (!vb2_start_streaming_called(&vpfe->buffer_queue))
> + return 0;
> +
> + /* Enable both master and slave clock */
> + pm_runtime_get_sync(dev);
> + vpfe_config_enable(ccdc, 1);
> +
> + /* Restore VPFE context */
> + vpfe_restore_context(ccdc);
> +
> + vpfe_config_enable(ccdc, 0);
> + pm_runtime_put_sync(dev);
> +
> + /* Select default pin state */
> + pinctrl_pm_select_default_state(dev);
> +
> + return 0;
> +}
> +
> +#endif
> +
> +static SIMPLE_DEV_PM_OPS(vpfe_pm_ops, vpfe_suspend, vpfe_resume);
> +
> +static const struct of_device_id vpfe_of_match[] = {
> + { .compatible = "ti,am437x-vpfe", },
> + { /* sentinel */ },
> +};
> +MODULE_DEVICE_TABLE(of, vpfe_of_match);
> +
> +static struct platform_driver vpfe_driver = {
> + .probe = vpfe_probe,
> + .remove = vpfe_remove,
> + .driver = {
> + .name = VPFE_MODULE_NAME,
> + .owner = THIS_MODULE,
> + .pm = &vpfe_pm_ops,
> + .of_match_table = of_match_ptr(vpfe_of_match),
> + },
> +};
> +
> +module_platform_driver(vpfe_driver);
> +
> +MODULE_AUTHOR("Texas Instruments");
> +MODULE_DESCRIPTION("TI AM437x VPFE driver");
> +MODULE_LICENSE("GPL");
> +MODULE_VERSION(VPFE_VERSION);

Regards,

Hans
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/