[PATCH 1/2] media: Add mem2mem deinterlacing driver.

Hans Verkuil hverkuil at xs4all.nl
Wed Jul 11 05:45:51 EDT 2012


Hi Javier!

Thanks for the patch.

On Wed 11 July 2012 10:55:03 Javier Martin wrote:
> Some video decoders such as tvp5150 provide separate
> video fields (V4L2_FIELD_SEQ_TB). This driver uses
> dmaengine to convert this format to V4L2_FIELD_INTERLACED_TB
> (weaving) or V4L2_FIELD_NONE (line doubling)

Which field is used for the line doubling? Top or bottom? Or is each field
doubled, thus doubling the framerate?

I also recommend adding SEQ_BT/INTERLACED_BT support: NTSC transmits the bottom
field first, so it is useful to have support for that.

Regards,

	Hans

> so that the
> image can be displayed or processed.
> 
> Of course there will be combing effect in the image but this
> can be accepted for some low quality applications.
> 
> Currently YUV420 and YUYV formats are supported but
> can be extended later.
> ---
>  drivers/media/video/Kconfig           |    8 +
>  drivers/media/video/Makefile          |    2 +
>  drivers/media/video/m2m-deinterlace.c | 1077 +++++++++++++++++++++++++++++++++
>  3 files changed, 1087 insertions(+)
>  create mode 100644 drivers/media/video/m2m-deinterlace.c
> 
> diff --git a/drivers/media/video/Kconfig b/drivers/media/video/Kconfig
> index 9cebf7b..c0b9233 100644
> --- a/drivers/media/video/Kconfig
> +++ b/drivers/media/video/Kconfig
> @@ -1188,6 +1188,14 @@ config VIDEO_CODA
>  	   Coda is a range of video codec IPs that supports
>  	   H.264, MPEG-4, and other video formats.
>  
> +config VIDEO_MEM2MEM_DEINTERLACE
> +	tristate "Deinterlace support"
> +	depends on VIDEO_DEV && VIDEO_V4L2 && DMA_ENGINE
> +	select VIDEOBUF2_DMA_CONTIG
> +	select V4L2_MEM2MEM_DEV
> +	help
> +	    Generic deinterlacing V4L2 driver.
> +
>  config VIDEO_SAMSUNG_S5P_G2D
>  	tristate "Samsung S5P and EXYNOS4 G2D 2d graphics accelerator driver"
>  	depends on VIDEO_DEV && VIDEO_V4L2 && PLAT_S5P
> diff --git a/drivers/media/video/Makefile b/drivers/media/video/Makefile
> index a04c307..b6a01b1 100644
> --- a/drivers/media/video/Makefile
> +++ b/drivers/media/video/Makefile
> @@ -189,6 +189,8 @@ obj-$(CONFIG_VIDEO_ATMEL_ISI)		+= atmel-isi.o
>  obj-$(CONFIG_VIDEO_MX2_EMMAPRP)		+= mx2_emmaprp.o
>  obj-$(CONFIG_VIDEO_CODA) 			+= coda.o
>  
> +obj-$(CONFIG_VIDEO_MEM2MEM_DEINTERLACE)	+= m2m-deinterlace.o
> +
>  obj-$(CONFIG_VIDEO_SAMSUNG_S5P_FIMC) 	+= s5p-fimc/
>  obj-$(CONFIG_VIDEO_SAMSUNG_S5P_JPEG)	+= s5p-jpeg/
>  obj-$(CONFIG_VIDEO_SAMSUNG_S5P_MFC)	+= s5p-mfc/
> diff --git a/drivers/media/video/m2m-deinterlace.c b/drivers/media/video/m2m-deinterlace.c
> new file mode 100644
> index 0000000..9642efd
> --- /dev/null
> +++ b/drivers/media/video/m2m-deinterlace.c
> @@ -0,0 +1,1077 @@
> +/*
> + * V4L2 deinterlacing support.
> + *
> + * Copyright (c) 2012 Vista Silicon S.L.
> + * Javier Martin <javier.martin at vista-silicon.com>
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by the
> + * Free Software Foundation; either version 2 of the
> + * License, or (at your option) any later version
> + */
> +
> +#include <linux/module.h>
> +#include <linux/slab.h>
> +#include <linux/interrupt.h>
> +#include <linux/dmaengine.h>
> +#include <linux/platform_device.h>
> +
> +#include <media/v4l2-mem2mem.h>
> +#include <media/v4l2-device.h>
> +#include <media/v4l2-ioctl.h>
> +#include <media/videobuf2-dma-contig.h>
> +
> +#define MEM2MEM_TEST_MODULE_NAME "mem2mem-deinterlace"
> +
> +MODULE_DESCRIPTION("mem2mem device which supports deinterlacing using dmaengine");
> +MODULE_AUTHOR("Javier Martin <javier.martin at vista-silicon.com");
> +MODULE_LICENSE("GPL");
> +MODULE_VERSION("0.0.1");
> +
> +static bool debug = true;
> +module_param(debug, bool, 0644);
> +
> +/* Flags that indicate a format can be used for capture/output */
> +#define MEM2MEM_CAPTURE	(1 << 0)
> +#define MEM2MEM_OUTPUT	(1 << 1)
> +
> +#define MEM2MEM_NAME		"m2m-deinterlace"
> +
> +#define dprintk(dev, fmt, arg...) \
> +	v4l2_dbg(1, debug, &dev->v4l2_dev, "%s: " fmt, __func__, ## arg)
> +
> +struct deinterlace_fmt {
> +	char	*name;
> +	u32	fourcc;
> +	enum v4l2_field field;
> +	/* Types the format can be used for */
> +	u32	types;
> +};
> +
> +static struct deinterlace_fmt formats[] = {
> +	{
> +		.name	= "YUV 4:2:0 Planar (weaving)",
> +		.fourcc	= V4L2_PIX_FMT_YUV420,
> +		.field	= V4L2_FIELD_INTERLACED_TB,
> +		.types	= MEM2MEM_CAPTURE,
> +	},
> +	{
> +		.name	= "YUV 4:2:0 Planar (line doubling)",
> +		.fourcc	= V4L2_PIX_FMT_YUV420,
> +		/* Line doubling, top field */
> +		.field	= V4L2_FIELD_NONE,
> +		.types	= MEM2MEM_CAPTURE,
> +	},
> +	{
> +		.name	= "YUYV 4:2:2 (weaving)",
> +		.fourcc	= V4L2_PIX_FMT_YUYV,
> +		.field	= V4L2_FIELD_INTERLACED_TB,
> +		.types	= MEM2MEM_CAPTURE,
> +	},
> +	{
> +		.name	= "YUYV 4:2:2 (line doubling)",
> +		.fourcc	= V4L2_PIX_FMT_YUYV,
> +		/* Line doubling, top field */
> +		.field	= V4L2_FIELD_NONE,
> +		.types	= MEM2MEM_CAPTURE,
> +	},
> +	{
> +		.name	= "YUV 4:2:0 Planar (top-bottom)",
> +		.fourcc	= V4L2_PIX_FMT_YUV420,
> +		.field	= V4L2_FIELD_SEQ_TB,
> +		.types	= MEM2MEM_OUTPUT,
> +	},
> +	{
> +		.name	= "YUYV 4:2:2 (top-bottom)",
> +		.fourcc	= V4L2_PIX_FMT_YUYV,
> +		.field	= V4L2_FIELD_SEQ_TB,
> +		.types	= MEM2MEM_OUTPUT,
> +	},
> +};
> +
> +/* Per-queue, driver-specific private data */
> +struct deinterlace_q_data {
> +	unsigned int		width;
> +	unsigned int		height;
> +	unsigned int		sizeimage;
> +	struct deinterlace_fmt	*fmt;
> +};
> +
> +enum {
> +	V4L2_M2M_SRC = 0,
> +	V4L2_M2M_DST = 1,
> +};
> +
> +enum {
> +	YUV420_DMA_Y_ODD,
> +	YUV420_DMA_Y_EVEN,
> +	YUV420_DMA_U_ODD,
> +	YUV420_DMA_U_EVEN,
> +	YUV420_DMA_V_ODD,
> +	YUV420_DMA_V_EVEN,
> +	YUV420_DMA_Y_ODD_DOUBLING,
> +	YUV420_DMA_U_ODD_DOUBLING,
> +	YUV420_DMA_V_ODD_DOUBLING,
> +	YUYV_DMA_ODD,
> +	YUYV_DMA_EVEN,
> +	YUYV_DMA_EVEN_DOUBLING,
> +};
> +
> +/* Source and destination queue data */
> +static struct deinterlace_q_data q_data[2];
> +
> +static struct deinterlace_q_data *get_q_data(enum v4l2_buf_type type)
> +{
> +	switch (type) {
> +	case V4L2_BUF_TYPE_VIDEO_OUTPUT:
> +		return &q_data[V4L2_M2M_SRC];
> +	case V4L2_BUF_TYPE_VIDEO_CAPTURE:
> +		return &q_data[V4L2_M2M_DST];
> +	default:
> +		BUG();
> +	}
> +	return NULL;
> +}
> +
> +#define NUM_FORMATS ARRAY_SIZE(formats)
> +
> +static struct deinterlace_fmt *find_format(struct v4l2_format *f)
> +{
> +	struct deinterlace_fmt *fmt;
> +	unsigned int k;
> +
> +	for (k = 0; k < NUM_FORMATS; k++) {
> +		fmt = &formats[k];
> +		if ((fmt->types == f->type) &&
> +			(fmt->fourcc == f->fmt.pix.pixelformat) &&
> +			(fmt->field == f->fmt.pix.field))
> +			break;
> +	}
> +
> +	if (k == NUM_FORMATS)
> +		return NULL;
> +
> +	return &formats[k];
> +}
> +
> +struct deinterlace_dev {
> +	struct v4l2_device	v4l2_dev;
> +	struct video_device	*vfd;
> +
> +	atomic_t		busy;
> +	struct mutex		dev_mutex;
> +	spinlock_t		irqlock;
> +
> +	struct dma_chan		*dma_chan;
> +
> +	struct v4l2_m2m_dev	*m2m_dev;
> +	struct vb2_alloc_ctx	*alloc_ctx;
> +};
> +
> +struct deinterlace_ctx {
> +	struct deinterlace_dev	*dev;
> +
> +	/* Abort requested by m2m */
> +	int			aborting;
> +	dma_cookie_t		cookie;
> +	struct v4l2_m2m_ctx	*m2m_ctx;
> +	struct dma_interleaved_template *xt;
> +};
> +
> +/*
> + * mem2mem callbacks
> + */
> +static int deinterlace_job_ready(void *priv)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +	struct deinterlace_dev *pcdev = ctx->dev;
> +
> +	if ((v4l2_m2m_num_src_bufs_ready(ctx->m2m_ctx) > 0)
> +	    && (v4l2_m2m_num_dst_bufs_ready(ctx->m2m_ctx) > 0)
> +	    && (atomic_read(&ctx->dev->busy) == 0)) {
> +		dprintk(pcdev, "Task ready\n");
> +		return 1;
> +	}
> +
> +	dprintk(pcdev, "Task not ready to run\n");
> +
> +	return 0;
> +}
> +
> +static void deinterlace_job_abort(void *priv)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +	struct deinterlace_dev *pcdev = ctx->dev;
> +
> +	ctx->aborting = 1;
> +
> +	dprintk(pcdev, "Aborting task\n");
> +
> +	v4l2_m2m_job_finish(pcdev->m2m_dev, ctx->m2m_ctx);
> +}
> +
> +static void deinterlace_lock(void *priv)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +	struct deinterlace_dev *pcdev = ctx->dev;
> +	mutex_lock(&pcdev->dev_mutex);
> +}
> +
> +static void deinterlace_unlock(void *priv)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +	struct deinterlace_dev *pcdev = ctx->dev;
> +	mutex_unlock(&pcdev->dev_mutex);
> +}
> +
> +static void dma_callback(void *data)
> +{
> +	struct deinterlace_ctx *curr_ctx = data;
> +	struct deinterlace_dev *pcdev = curr_ctx->dev;
> +	struct vb2_buffer *src_vb, *dst_vb;
> +
> +	atomic_set(&pcdev->busy, 0);
> +
> +	src_vb = v4l2_m2m_src_buf_remove(curr_ctx->m2m_ctx);
> +	dst_vb = v4l2_m2m_dst_buf_remove(curr_ctx->m2m_ctx);
> +
> +	v4l2_m2m_buf_done(src_vb, VB2_BUF_STATE_DONE);
> +	v4l2_m2m_buf_done(dst_vb, VB2_BUF_STATE_DONE);
> +
> +	v4l2_m2m_job_finish(pcdev->m2m_dev, curr_ctx->m2m_ctx);
> +
> +	dprintk(pcdev, "dma transfers completed.\n");
> +}
> +
> +static void deinterlace_issue_dma(struct deinterlace_ctx *ctx, int op,
> +				  int do_callback)
> +{
> +	struct deinterlace_q_data *s_q_data, *d_q_data;
> +	struct vb2_buffer *src_buf, *dst_buf;
> +	struct deinterlace_dev *pcdev = ctx->dev;
> +	struct dma_chan *chan = pcdev->dma_chan;
> +	struct dma_device *dmadev = chan->device;
> +	struct dma_async_tx_descriptor *tx;
> +	unsigned int s_width, s_height;
> +	unsigned int d_width, d_height;
> +	unsigned int d_size, s_size;
> +	dma_addr_t p_in, p_out;
> +	enum dma_ctrl_flags flags;
> +
> +	src_buf = v4l2_m2m_next_src_buf(ctx->m2m_ctx);
> +	dst_buf = v4l2_m2m_next_dst_buf(ctx->m2m_ctx);
> +
> +	s_q_data = get_q_data(V4L2_BUF_TYPE_VIDEO_OUTPUT);
> +	s_width	= s_q_data->width;
> +	s_height = s_q_data->height;
> +	s_size = s_width * s_height;
> +
> +	d_q_data = get_q_data(V4L2_BUF_TYPE_VIDEO_CAPTURE);
> +	d_width = d_q_data->width;
> +	d_height = d_q_data->height;
> +	d_size = d_width * d_height;
> +
> +	p_in = (dma_addr_t)vb2_dma_contig_plane_dma_addr(src_buf, 0);
> +	p_out = (dma_addr_t)vb2_dma_contig_plane_dma_addr(dst_buf, 0);
> +	if (!p_in || !p_out) {
> +		v4l2_err(&pcdev->v4l2_dev,
> +			 "Acquiring kernel pointers to buffers failed\n");
> +		return;
> +	}
> +
> +	switch (op) {
> +	case YUV420_DMA_Y_ODD:
> +		ctx->xt->numf = s_height / 2;
> +		ctx->xt->sgl[0].size = s_width;
> +		ctx->xt->sgl[0].icg = s_width;
> +		ctx->xt->src_start = p_in;
> +		ctx->xt->dst_start = p_out;
> +		break;
> +	case YUV420_DMA_Y_EVEN:
> +		ctx->xt->numf = s_height / 2;
> +		ctx->xt->sgl[0].size = s_width;
> +		ctx->xt->sgl[0].icg = s_width;
> +		ctx->xt->src_start = p_in + s_size / 2;
> +		ctx->xt->dst_start = p_out + s_width;
> +		break;
> +	case YUV420_DMA_U_ODD:
> +		ctx->xt->numf = s_height / 4;
> +		ctx->xt->sgl[0].size = s_width / 2;
> +		ctx->xt->sgl[0].icg = s_width / 2;
> +		ctx->xt->src_start = p_in + s_size;
> +		ctx->xt->dst_start = p_out + s_size;
> +		break;
> +	case YUV420_DMA_U_EVEN:
> +		ctx->xt->numf = s_height / 4;
> +		ctx->xt->sgl[0].size = s_width / 2;
> +		ctx->xt->sgl[0].icg = s_width / 2;
> +		ctx->xt->src_start = p_in + (9 * s_size) / 8;
> +		ctx->xt->dst_start = p_out + s_size + s_width / 2;
> +		break;
> +	case YUV420_DMA_V_ODD:
> +		ctx->xt->numf = s_height / 4;
> +		ctx->xt->sgl[0].size = s_width / 2;
> +		ctx->xt->sgl[0].icg = s_width / 2;
> +		ctx->xt->src_start = p_in + (5 * s_size) / 4;
> +		ctx->xt->dst_start = p_out + (5 * s_size) / 4;
> +		break;
> +	case YUV420_DMA_V_EVEN:
> +		ctx->xt->numf = s_height / 4;
> +		ctx->xt->sgl[0].size = s_width / 2;
> +		ctx->xt->sgl[0].icg = s_width / 2;
> +		ctx->xt->src_start = p_in + (11 * s_size) / 8;
> +		ctx->xt->dst_start = p_out + (5 * s_size) / 4 + s_width / 2;
> +		break;
> +	case YUV420_DMA_Y_ODD_DOUBLING:
> +		ctx->xt->numf = s_height / 2;
> +		ctx->xt->sgl[0].size = s_width;
> +		ctx->xt->sgl[0].icg = s_width;
> +		ctx->xt->src_start = p_in;
> +		ctx->xt->dst_start = p_out + s_width;
> +		break;
> +	case YUV420_DMA_U_ODD_DOUBLING:
> +		ctx->xt->numf = s_height / 4;
> +		ctx->xt->sgl[0].size = s_width / 2;
> +		ctx->xt->sgl[0].icg = s_width / 2;
> +		ctx->xt->src_start = p_in + s_size;
> +		ctx->xt->dst_start = p_out + s_size + s_width / 2;
> +		break;
> +	case YUV420_DMA_V_ODD_DOUBLING:
> +		ctx->xt->numf = s_height / 4;
> +		ctx->xt->sgl[0].size = s_width / 2;
> +		ctx->xt->sgl[0].icg = s_width / 2;
> +		ctx->xt->src_start = p_in + (5 * s_size) / 4;
> +		ctx->xt->dst_start = p_out + (5 * s_size) / 4 + s_width / 2;
> +		break;
> +	case YUYV_DMA_ODD:
> +		ctx->xt->numf = s_height / 2;
> +		ctx->xt->sgl[0].size = s_width * 2;
> +		ctx->xt->sgl[0].icg = s_width * 2;
> +		ctx->xt->src_start = p_in;
> +		ctx->xt->dst_start = p_out;
> +		break;
> +	case YUYV_DMA_EVEN:
> +		ctx->xt->numf = s_height / 2;
> +		ctx->xt->sgl[0].size = s_width * 2;
> +		ctx->xt->sgl[0].icg = s_width * 2;
> +		ctx->xt->src_start = p_in + s_size;
> +		ctx->xt->dst_start = p_out + s_width * 2;
> +		break;
> +	case YUYV_DMA_EVEN_DOUBLING:
> +	default:
> +		ctx->xt->numf = s_height / 2;
> +		ctx->xt->sgl[0].size = s_width * 2;
> +		ctx->xt->sgl[0].icg = s_width * 2;
> +		ctx->xt->src_start = p_in;
> +		ctx->xt->dst_start = p_out + s_width * 2;
> +		break;
> +	}
> +
> +	/* Common parameters for al transfers */
> +	ctx->xt->frame_size = 1;
> +	ctx->xt->dir = DMA_MEM_TO_MEM;
> +	ctx->xt->src_sgl = false;
> +	ctx->xt->dst_sgl = true;
> +	flags = DMA_CTRL_ACK | DMA_PREP_INTERRUPT |
> +		DMA_COMPL_SKIP_DEST_UNMAP | DMA_COMPL_SKIP_SRC_UNMAP;
> +
> +	tx = dmadev->device_prep_interleaved_dma(chan, ctx->xt, flags);
> +	if (tx == NULL) {
> +		v4l2_warn(&pcdev->v4l2_dev, "DMA interleaved prep error\n");
> +		return;
> +	}
> +
> +	if (do_callback) {
> +		tx->callback = dma_callback;
> +		tx->callback_param = ctx;
> +	}
> +
> +	ctx->cookie = dmaengine_submit(tx);
> +	if (dma_submit_error(ctx->cookie)) {
> +		v4l2_warn(&pcdev->v4l2_dev,
> +			  "DMA submit error %d with src=0x%x dst=0x%x len=0x%x\n",
> +			  ctx->cookie, p_in, p_out, s_size * 3/2);
> +		return;
> +	}
> +
> +	dma_async_issue_pending(chan);
> +}
> +
> +static void deinterlace_device_run(void *priv)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +	struct deinterlace_q_data *dst_q_data;
> +
> +	atomic_set(&ctx->dev->busy, 1);
> +
> +	dprintk(ctx->dev, "%s: DMA try issue.\n", __func__);
> +
> +	dst_q_data = get_q_data(V4L2_BUF_TYPE_VIDEO_CAPTURE);
> +
> +	switch (dst_q_data->fmt->fourcc) {
> +	case V4L2_PIX_FMT_YUV420:
> +		switch (dst_q_data->fmt->field) {
> +		case V4L2_FIELD_INTERLACED_TB:
> +			dprintk(ctx->dev, "%s: yuv420 interlaced tb.\n",
> +				__func__);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_Y_ODD, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_Y_EVEN, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_U_ODD, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_U_EVEN, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_V_ODD, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_V_EVEN, 1);
> +			break;
> +		case V4L2_FIELD_NONE:
> +		default:
> +			dprintk(ctx->dev, "%s: yuv420 interlaced line doubling.\n",
> +				__func__);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_Y_ODD, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_Y_ODD_DOUBLING, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_U_ODD, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_U_ODD_DOUBLING, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_V_ODD, 0);
> +			deinterlace_issue_dma(ctx, YUV420_DMA_V_ODD_DOUBLING, 1);
> +			break;
> +		}
> +		break;
> +	case V4L2_PIX_FMT_YUYV:
> +	default:
> +		switch (dst_q_data->fmt->field) {
> +		case V4L2_FIELD_INTERLACED_TB:
> +			dprintk(ctx->dev, "%s: yuyv interlaced_tb.\n",
> +				__func__);
> +			deinterlace_issue_dma(ctx, YUYV_DMA_ODD, 0);
> +			deinterlace_issue_dma(ctx, YUYV_DMA_EVEN, 1);
> +			break;
> +		case V4L2_FIELD_NONE:
> +		default:
> +			dprintk(ctx->dev, "%s: yuyv interlaced line doubling.\n",
> +				__func__);
> +			deinterlace_issue_dma(ctx, YUYV_DMA_ODD, 0);
> +			deinterlace_issue_dma(ctx, YUYV_DMA_EVEN_DOUBLING, 1);
> +			break;
> +		}
> +		break;
> +	}
> +
> +	dprintk(ctx->dev, "%s: DMA issue done.\n", __func__);
> +}
> +
> +/*
> + * video ioctls
> + */
> +static int vidioc_querycap(struct file *file, void *priv,
> +			   struct v4l2_capability *cap)
> +{
> +	strncpy(cap->driver, MEM2MEM_NAME, sizeof(cap->driver) - 1);
> +	strncpy(cap->card, MEM2MEM_NAME, sizeof(cap->card) - 1);
> +	cap->bus_info[0] = 0;
> +	cap->capabilities = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_VIDEO_OUTPUT
> +			  | V4L2_CAP_STREAMING;
> +
> +	return 0;
> +}
> +
> +static int enum_fmt(struct v4l2_fmtdesc *f, u32 type)
> +{
> +	int i, num;
> +	struct deinterlace_fmt *fmt;
> +
> +	num = 0;
> +
> +	for (i = 0; i < NUM_FORMATS; ++i) {
> +		if (formats[i].types & type) {
> +			/* index-th format of type type found ? */
> +			if (num == f->index)
> +				break;
> +			/* Correct type but haven't reached our index yet,
> +			 * just increment per-type index */
> +			++num;
> +		}
> +	}
> +
> +	if (i < NUM_FORMATS) {
> +		/* Format found */
> +		fmt = &formats[i];
> +		strncpy(f->description, fmt->name, sizeof(f->description) - 1);
> +		f->pixelformat = fmt->fourcc;
> +		return 0;
> +	}
> +
> +	/* Format not found */
> +	return -EINVAL;
> +}
> +
> +static int vidioc_enum_fmt_vid_cap(struct file *file, void *priv,
> +				   struct v4l2_fmtdesc *f)
> +{
> +	return enum_fmt(f, MEM2MEM_CAPTURE);
> +}
> +
> +static int vidioc_enum_fmt_vid_out(struct file *file, void *priv,
> +				   struct v4l2_fmtdesc *f)
> +{
> +	return enum_fmt(f, MEM2MEM_OUTPUT);
> +}
> +
> +static int vidioc_g_fmt(struct deinterlace_ctx *ctx, struct v4l2_format *f)
> +{
> +	struct vb2_queue *vq;
> +	struct deinterlace_q_data *q_data;
> +
> +	vq = v4l2_m2m_get_vq(ctx->m2m_ctx, f->type);
> +	if (!vq)
> +		return -EINVAL;
> +
> +	q_data = get_q_data(f->type);
> +
> +	f->fmt.pix.width	= q_data->width;
> +	f->fmt.pix.height	= q_data->height;
> +	f->fmt.pix.field	= q_data->fmt->field;
> +	f->fmt.pix.pixelformat	= q_data->fmt->fourcc;
> +
> +	switch (q_data->fmt->fourcc) {
> +	case V4L2_PIX_FMT_YUV420:
> +		f->fmt.pix.bytesperline = q_data->width * 3 / 2;
> +		break;
> +	case V4L2_PIX_FMT_YUYV:
> +	default:
> +		f->fmt.pix.bytesperline = q_data->width * 2;
> +	}
> +
> +	f->fmt.pix.sizeimage	= q_data->sizeimage;
> +
> +	return 0;
> +}
> +
> +static int vidioc_g_fmt_vid_out(struct file *file, void *priv,
> +				struct v4l2_format *f)
> +{
> +	return vidioc_g_fmt(priv, f);
> +}
> +
> +static int vidioc_g_fmt_vid_cap(struct file *file, void *priv,
> +				struct v4l2_format *f)
> +{
> +	return vidioc_g_fmt(priv, f);
> +}
> +
> +static int vidioc_try_fmt(struct v4l2_format *f, struct deinterlace_fmt *fmt)
> +{
> +	switch (f->fmt.pix.pixelformat) {
> +	case V4L2_PIX_FMT_YUV420:
> +		f->fmt.pix.bytesperline = f->fmt.pix.width * 3 / 2;
> +		break;
> +	case V4L2_PIX_FMT_YUYV:
> +	default:
> +		f->fmt.pix.bytesperline = f->fmt.pix.width * 2;
> +	}
> +	f->fmt.pix.sizeimage = f->fmt.pix.height * f->fmt.pix.bytesperline;
> +
> +	return 0;
> +}
> +
> +static int vidioc_try_fmt_vid_cap(struct file *file, void *priv,
> +				  struct v4l2_format *f)
> +{
> +	struct deinterlace_fmt *fmt;
> +	struct deinterlace_ctx *ctx = priv;
> +
> +	fmt = find_format(f);
> +	if (!fmt || !(fmt->types & MEM2MEM_CAPTURE)) {
> +		v4l2_err(&ctx->dev->v4l2_dev,
> +			 "Fourcc format (0x%08x) invalid.\n",
> +			 f->fmt.pix.pixelformat);
> +		return -EINVAL;
> +	}
> +	return vidioc_try_fmt(f, fmt);
> +}
> +
> +static int vidioc_try_fmt_vid_out(struct file *file, void *priv,
> +				  struct v4l2_format *f)
> +{
> +	struct deinterlace_fmt *fmt;
> +	struct deinterlace_ctx *ctx = priv;
> +
> +	fmt = find_format(f);
> +	if (!fmt || !(fmt->types & MEM2MEM_OUTPUT)) {
> +		v4l2_err(&ctx->dev->v4l2_dev,
> +			 "Fourcc format (0x%08x) invalid.\n",
> +			 f->fmt.pix.pixelformat);
> +		return -EINVAL;
> +	}
> +
> +	return vidioc_try_fmt(f, fmt);
> +}
> +
> +static int vidioc_s_fmt(struct deinterlace_ctx *ctx, struct v4l2_format *f)
> +{
> +	struct deinterlace_q_data *q_data;
> +	struct vb2_queue *vq;
> +
> +	vq = v4l2_m2m_get_vq(ctx->m2m_ctx, f->type);
> +	if (!vq)
> +		return -EINVAL;
> +
> +	q_data = get_q_data(f->type);
> +	if (!q_data)
> +		return -EINVAL;
> +
> +	if (vb2_is_busy(vq)) {
> +		v4l2_err(&ctx->dev->v4l2_dev, "%s queue busy\n", __func__);
> +		return -EBUSY;
> +	}
> +
> +	q_data->fmt		= find_format(f);
> +	if (!q_data->fmt) {
> +		v4l2_err(&ctx->dev->v4l2_dev,
> +			 "Couldn't set format type %d, wxh: %dx%d. fmt: %d, field: %d\n",
> +			f->type, f->fmt.pix.width, f->fmt.pix.height,
> +			f->fmt.pix.pixelformat, f->fmt.pix.field);
> +		return -EINVAL;
> +	}
> +
> +	q_data->width		= f->fmt.pix.width;
> +	q_data->height		= f->fmt.pix.height;
> +
> +	switch (f->fmt.pix.pixelformat) {
> +	case V4L2_PIX_FMT_YUV420:
> +		f->fmt.pix.bytesperline = f->fmt.pix.width * 3 / 2;
> +		break;
> +	case V4L2_PIX_FMT_YUYV:
> +	default:
> +		f->fmt.pix.bytesperline = f->fmt.pix.width * 2;
> +	}
> +
> +	dprintk(ctx->dev,
> +		"Setting format for type %d, wxh: %dx%d, fmt: %d, field: %d\n",
> +		f->type, q_data->width, q_data->height, q_data->fmt->fourcc,
> +		q_data->fmt->field);
> +
> +	return 0;
> +}
> +
> +static int vidioc_s_fmt_vid_cap(struct file *file, void *priv,
> +				struct v4l2_format *f)
> +{
> +	int ret;
> +
> +	ret = vidioc_try_fmt_vid_cap(file, priv, f);
> +	if (ret)
> +		return ret;
> +	return vidioc_s_fmt(priv, f);
> +}
> +
> +static int vidioc_s_fmt_vid_out(struct file *file, void *priv,
> +				struct v4l2_format *f)
> +{
> +	int ret;
> +
> +	ret = vidioc_try_fmt_vid_out(file, priv, f);
> +	if (ret)
> +		return ret;
> +
> +	return vidioc_s_fmt(priv, f);
> +}
> +
> +static int vidioc_reqbufs(struct file *file, void *priv,
> +			  struct v4l2_requestbuffers *reqbufs)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +
> +	return v4l2_m2m_reqbufs(file, ctx->m2m_ctx, reqbufs);
> +}
> +
> +static int vidioc_querybuf(struct file *file, void *priv,
> +			   struct v4l2_buffer *buf)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +
> +	return v4l2_m2m_querybuf(file, ctx->m2m_ctx, buf);
> +}
> +
> +static int vidioc_qbuf(struct file *file, void *priv, struct v4l2_buffer *buf)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +
> +	return v4l2_m2m_qbuf(file, ctx->m2m_ctx, buf);
> +}
> +
> +static int vidioc_dqbuf(struct file *file, void *priv, struct v4l2_buffer *buf)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +
> +	return v4l2_m2m_dqbuf(file, ctx->m2m_ctx, buf);
> +}
> +
> +static int vidioc_streamon(struct file *file, void *priv,
> +			   enum v4l2_buf_type type)
> +{
> +	struct deinterlace_q_data *s_q_data, *d_q_data;
> +	struct deinterlace_ctx *ctx = priv;
> +
> +	/* Check that src and dst queues have the same format */
> +	s_q_data = get_q_data(V4L2_BUF_TYPE_VIDEO_OUTPUT);
> +	d_q_data = get_q_data(V4L2_BUF_TYPE_VIDEO_OUTPUT);
> +	if (s_q_data->fmt->fourcc != d_q_data->fmt->fourcc)
> +		return -EINVAL;
> +
> +	return v4l2_m2m_streamon(file, ctx->m2m_ctx, type);
> +}
> +
> +static int vidioc_streamoff(struct file *file, void *priv,
> +			    enum v4l2_buf_type type)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +
> +	return v4l2_m2m_streamoff(file, ctx->m2m_ctx, type);
> +}
> +
> +static const struct v4l2_ioctl_ops deinterlace_ioctl_ops = {
> +	.vidioc_querycap	= vidioc_querycap,
> +
> +	.vidioc_enum_fmt_vid_cap = vidioc_enum_fmt_vid_cap,
> +	.vidioc_g_fmt_vid_cap	= vidioc_g_fmt_vid_cap,
> +	.vidioc_try_fmt_vid_cap	= vidioc_try_fmt_vid_cap,
> +	.vidioc_s_fmt_vid_cap	= vidioc_s_fmt_vid_cap,
> +
> +	.vidioc_enum_fmt_vid_out = vidioc_enum_fmt_vid_out,
> +	.vidioc_g_fmt_vid_out	= vidioc_g_fmt_vid_out,
> +	.vidioc_try_fmt_vid_out	= vidioc_try_fmt_vid_out,
> +	.vidioc_s_fmt_vid_out	= vidioc_s_fmt_vid_out,
> +
> +	.vidioc_reqbufs		= vidioc_reqbufs,
> +	.vidioc_querybuf	= vidioc_querybuf,
> +
> +	.vidioc_qbuf		= vidioc_qbuf,
> +	.vidioc_dqbuf		= vidioc_dqbuf,
> +
> +	.vidioc_streamon	= vidioc_streamon,
> +	.vidioc_streamoff	= vidioc_streamoff,
> +};
> +
> +
> +/*
> + * Queue operations
> + */
> +struct vb2_dc_conf {
> +	struct device           *dev;
> +};
> +
> +static int deinterlace_queue_setup(struct vb2_queue *vq,
> +				const struct v4l2_format *fmt,
> +				unsigned int *nbuffers, unsigned int *nplanes,
> +				unsigned int sizes[], void *alloc_ctxs[])
> +{
> +	struct deinterlace_ctx *ctx = vb2_get_drv_priv(vq);
> +	struct deinterlace_q_data *q_data;
> +	unsigned int size, count = *nbuffers;
> +
> +	q_data = get_q_data(vq->type);
> +
> +	switch (q_data->fmt->fourcc) {
> +	case V4L2_PIX_FMT_YUV420:
> +		size = q_data->width * q_data->height * 3 / 2;
> +		break;
> +	case V4L2_PIX_FMT_YUYV:
> +	default:
> +		size = q_data->width * q_data->height * 2;
> +	}
> +
> +	*nplanes = 1;
> +	*nbuffers = count;
> +	sizes[0] = size;
> +
> +	alloc_ctxs[0] = ctx->dev->alloc_ctx;
> +
> +	dprintk(ctx->dev, "get %d buffer(s) of size %d each.\n", count, size);
> +
> +	return 0;
> +}
> +
> +static int deinterlace_buf_prepare(struct vb2_buffer *vb)
> +{
> +	struct deinterlace_ctx *ctx = vb2_get_drv_priv(vb->vb2_queue);
> +	struct deinterlace_q_data *q_data;
> +
> +	dprintk(ctx->dev, "type: %d\n", vb->vb2_queue->type);
> +
> +	q_data = get_q_data(vb->vb2_queue->type);
> +
> +	if (vb2_plane_size(vb, 0) < q_data->sizeimage) {
> +		dprintk(ctx->dev, "%s data will not fit into plane (%lu < %lu)\n",
> +			__func__, vb2_plane_size(vb, 0), (long)q_data->sizeimage);
> +		return -EINVAL;
> +	}
> +
> +	vb2_set_plane_payload(vb, 0, q_data->sizeimage);
> +
> +	return 0;
> +}
> +
> +static void deinterlace_buf_queue(struct vb2_buffer *vb)
> +{
> +	struct deinterlace_ctx *ctx = vb2_get_drv_priv(vb->vb2_queue);
> +	v4l2_m2m_buf_queue(ctx->m2m_ctx, vb);
> +}
> +
> +static struct vb2_ops deinterlace_qops = {
> +	.queue_setup	 = deinterlace_queue_setup,
> +	.buf_prepare	 = deinterlace_buf_prepare,
> +	.buf_queue	 = deinterlace_buf_queue,
> +};
> +
> +static int queue_init(void *priv, struct vb2_queue *src_vq,
> +		      struct vb2_queue *dst_vq)
> +{
> +	struct deinterlace_ctx *ctx = priv;
> +	int ret;
> +
> +	memset(src_vq, 0, sizeof(*src_vq));
> +	src_vq->type = V4L2_BUF_TYPE_VIDEO_OUTPUT;
> +	src_vq->io_modes = VB2_MMAP | VB2_USERPTR;
> +	src_vq->drv_priv = ctx;
> +	src_vq->buf_struct_size = sizeof(struct v4l2_m2m_buffer);
> +	src_vq->ops = &deinterlace_qops;
> +	src_vq->mem_ops = &vb2_dma_contig_memops;
> +
> +	ret = vb2_queue_init(src_vq);
> +	if (ret)
> +		return ret;
> +
> +	memset(dst_vq, 0, sizeof(*dst_vq));
> +	dst_vq->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
> +	dst_vq->io_modes = VB2_MMAP | VB2_USERPTR;
> +	dst_vq->drv_priv = ctx;
> +	dst_vq->buf_struct_size = sizeof(struct v4l2_m2m_buffer);
> +	dst_vq->ops = &deinterlace_qops;
> +	dst_vq->mem_ops = &vb2_dma_contig_memops;
> +
> +	return vb2_queue_init(dst_vq);
> +}
> +
> +/*
> + * File operations
> + */
> +static int deinterlace_open(struct file *file)
> +{
> +	struct deinterlace_dev *pcdev = video_drvdata(file);
> +	struct deinterlace_ctx *ctx = NULL;
> +
> +	ctx = kzalloc(sizeof *ctx, GFP_KERNEL);
> +	if (!ctx)
> +		return -ENOMEM;
> +
> +	file->private_data = ctx;
> +	ctx->dev = pcdev;
> +
> +	ctx->m2m_ctx = v4l2_m2m_ctx_init(pcdev->m2m_dev, ctx, &queue_init);
> +	if (IS_ERR(ctx->m2m_ctx)) {
> +		int ret = PTR_ERR(ctx->m2m_ctx);
> +
> +		kfree(ctx);
> +		return ret;
> +	}
> +
> +	ctx->xt = kzalloc(sizeof(struct dma_async_tx_descriptor) +
> +				sizeof(struct data_chunk), GFP_KERNEL);
> +	if (!ctx->xt) {
> +		int ret = PTR_ERR(ctx->xt);
> +
> +		kfree(ctx);
> +		return ret;
> +	}
> +
> +	dprintk(pcdev, "Created instance %p, m2m_ctx: %p\n", ctx, ctx->m2m_ctx);
> +
> +	return 0;
> +}
> +
> +static int deinterlace_release(struct file *file)
> +{
> +	struct deinterlace_dev *pcdev = video_drvdata(file);
> +	struct deinterlace_ctx *ctx = file->private_data;
> +
> +	dprintk(pcdev, "Releasing instance %p\n", ctx);
> +
> +	v4l2_m2m_ctx_release(ctx->m2m_ctx);
> +	kfree(ctx->xt);
> +	kfree(ctx);
> +
> +	return 0;
> +}
> +
> +static unsigned int deinterlace_poll(struct file *file,
> +				 struct poll_table_struct *wait)
> +{
> +	struct deinterlace_ctx *ctx = file->private_data;
> +	int ret;
> +
> +	deinterlace_lock(ctx);
> +	ret = v4l2_m2m_poll(file, ctx->m2m_ctx, wait);
> +	deinterlace_unlock(ctx);
> +
> +	return ret;
> +}
> +
> +static int deinterlace_mmap(struct file *file, struct vm_area_struct *vma)
> +{
> +	struct deinterlace_ctx *ctx = file->private_data;
> +
> +	return v4l2_m2m_mmap(file, ctx->m2m_ctx, vma);
> +}
> +
> +static const struct v4l2_file_operations deinterlace_fops = {
> +	.owner		= THIS_MODULE,
> +	.open		= deinterlace_open,
> +	.release	= deinterlace_release,
> +	.poll		= deinterlace_poll,
> +	.unlocked_ioctl	= video_ioctl2,
> +	.mmap		= deinterlace_mmap,
> +};
> +
> +static struct video_device deinterlace_videodev = {
> +	.name		= MEM2MEM_NAME,
> +	.fops		= &deinterlace_fops,
> +	.ioctl_ops	= &deinterlace_ioctl_ops,
> +	.minor		= -1,
> +	.release	= video_device_release,
> +};
> +
> +static struct v4l2_m2m_ops m2m_ops = {
> +	.device_run	= deinterlace_device_run,
> +	.job_ready	= deinterlace_job_ready,
> +	.job_abort	= deinterlace_job_abort,
> +	.lock		= deinterlace_lock,
> +	.unlock		= deinterlace_unlock,
> +};
> +
> +static int deinterlace_probe(struct platform_device *pdev)
> +{
> +	struct deinterlace_dev *pcdev;
> +	struct video_device *vfd;
> +	dma_cap_mask_t mask;
> +	int ret = 0;
> +
> +	pcdev = kzalloc(sizeof *pcdev, GFP_KERNEL);
> +	if (!pcdev)
> +		return -ENOMEM;
> +
> +	spin_lock_init(&pcdev->irqlock);
> +
> +	dma_cap_zero(mask);
> +	dma_cap_set(DMA_INTERLEAVE, mask);
> +	pcdev->dma_chan = dma_request_channel(mask, NULL, pcdev);
> +	if (!pcdev->dma_chan)
> +		goto free_dev;
> +
> +	if (!dma_has_cap(DMA_INTERLEAVE, pcdev->dma_chan->device->cap_mask)) {
> +		v4l2_err(&pcdev->v4l2_dev, "DMA does not support INTERLEAVE\n");
> +		goto rel_dma;
> +	}
> +
> +	ret = v4l2_device_register(&pdev->dev, &pcdev->v4l2_dev);
> +	if (ret)
> +		goto rel_dma;
> +
> +	atomic_set(&pcdev->busy, 0);
> +	mutex_init(&pcdev->dev_mutex);
> +
> +	vfd = video_device_alloc();
> +	if (!vfd) {
> +		v4l2_err(&pcdev->v4l2_dev, "Failed to allocate video device\n");
> +		ret = -ENOMEM;
> +		goto unreg_dev;
> +	}
> +
> +	*vfd = deinterlace_videodev;
> +	vfd->lock = &pcdev->dev_mutex;
> +
> +	ret = video_register_device(vfd, VFL_TYPE_GRABBER, 0);
> +	if (ret) {
> +		v4l2_err(&pcdev->v4l2_dev, "Failed to register video device\n");
> +		goto rel_vdev;
> +	}
> +
> +	video_set_drvdata(vfd, pcdev);
> +	snprintf(vfd->name, sizeof(vfd->name), "%s", deinterlace_videodev.name);
> +	pcdev->vfd = vfd;
> +	v4l2_info(&pcdev->v4l2_dev, MEM2MEM_TEST_MODULE_NAME
> +			" Device registered as /dev/video%d\n", vfd->num);
> +
> +	platform_set_drvdata(pdev, pcdev);
> +
> +	pcdev->alloc_ctx = vb2_dma_contig_init_ctx(&pdev->dev);
> +	if (IS_ERR(pcdev->alloc_ctx)) {
> +		v4l2_err(&pcdev->v4l2_dev, "Failed to alloc vb2 context\n");
> +		ret = PTR_ERR(pcdev->alloc_ctx);
> +		goto err_ctx;
> +	}
> +
> +	pcdev->m2m_dev = v4l2_m2m_init(&m2m_ops);
> +	if (IS_ERR(pcdev->m2m_dev)) {
> +		v4l2_err(&pcdev->v4l2_dev, "Failed to init mem2mem device\n");
> +		ret = PTR_ERR(pcdev->m2m_dev);
> +		goto err_m2m;
> +	}
> +
> +	q_data[V4L2_M2M_SRC].fmt = &formats[1];
> +	q_data[V4L2_M2M_DST].fmt = &formats[0];
> +
> +	return 0;
> +
> +	v4l2_m2m_release(pcdev->m2m_dev);
> +err_m2m:
> +	video_unregister_device(pcdev->vfd);
> +err_ctx:
> +	vb2_dma_contig_cleanup_ctx(pcdev->alloc_ctx);
> +rel_vdev:
> +	video_device_release(vfd);
> +unreg_dev:
> +	v4l2_device_unregister(&pcdev->v4l2_dev);
> +rel_dma:
> +	dma_release_channel(pcdev->dma_chan);
> +free_dev:
> +	kfree(pcdev);
> +
> +	return ret;
> +}
> +
> +static int deinterlace_remove(struct platform_device *pdev)
> +{
> +	struct deinterlace_dev *pcdev =
> +		(struct deinterlace_dev *)platform_get_drvdata(pdev);
> +
> +	v4l2_info(&pcdev->v4l2_dev, "Removing " MEM2MEM_TEST_MODULE_NAME);
> +	v4l2_m2m_release(pcdev->m2m_dev);
> +	video_unregister_device(pcdev->vfd);
> +	v4l2_device_unregister(&pcdev->v4l2_dev);
> +	vb2_dma_contig_cleanup_ctx(pcdev->alloc_ctx);
> +	dma_release_channel(pcdev->dma_chan);
> +	kfree(pcdev);
> +
> +	return 0;
> +}
> +
> +static struct platform_driver deinterlace_pdrv = {
> +	.probe		= deinterlace_probe,
> +	.remove		= deinterlace_remove,
> +	.driver		= {
> +		.name	= MEM2MEM_NAME,
> +		.owner	= THIS_MODULE,
> +	},
> +};
> +
> +static void __exit deinterlace_exit(void)
> +{
> +	platform_driver_unregister(&deinterlace_pdrv);
> +}
> +
> +static int __init deinterlace_init(void)
> +{
> +	return platform_driver_register(&deinterlace_pdrv);
> +}
> +
> +module_init(deinterlace_init);
> +module_exit(deinterlace_exit);
> +
> 



More information about the linux-arm-kernel mailing list