[PATCH RESEND v9 18/18] media: platform: Add jpeg enc feature

Tomasz Figa tfiga at chromium.org
Wed Jul 8 07:16:06 EDT 2020


On Wed, Jul 8, 2020 at 9:14 AM Xia Jiang <xia.jiang at mediatek.com> wrote:
>
> On Tue, 2020-07-07 at 15:35 +0200, Tomasz Figa wrote:
> > On Tue, Jul 7, 2020 at 8:47 AM Xia Jiang <xia.jiang at mediatek.com> wrote:
> > >
> > > On Tue, 2020-06-30 at 16:53 +0000, Tomasz Figa wrote:
> > > > Hi Xia,
> > > >
> > > > On Tue, Jun 30, 2020 at 10:56:21AM +0800, Xia Jiang wrote:
> > > > > On Thu, 2020-06-11 at 18:46 +0000, Tomasz Figa wrote:
> > > > > > Hi Xia,
> > > > > >
> > > > > > On Thu, Jun 04, 2020 at 05:05:53PM +0800, Xia Jiang wrote:
> > > > [snip]
> > > > > > > +static void mtk_jpeg_enc_device_run(void *priv)
> > > > > > > +{
> > > > > > > +       struct mtk_jpeg_ctx *ctx = priv;
> > > > > > > +       struct mtk_jpeg_dev *jpeg = ctx->jpeg;
> > > > > > > +       struct vb2_v4l2_buffer *src_buf, *dst_buf;
> > > > > > > +       enum vb2_buffer_state buf_state = VB2_BUF_STATE_ERROR;
> > > > > > > +       unsigned long flags;
> > > > > > > +       struct mtk_jpeg_src_buf *jpeg_src_buf;
> > > > > > > +       struct mtk_jpeg_enc_bs enc_bs;
> > > > > > > +       int ret;
> > > > > > > +
> > > > > > > +       src_buf = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx);
> > > > > > > +       dst_buf = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx);
> > > > > > > +       jpeg_src_buf = mtk_jpeg_vb2_to_srcbuf(&src_buf->vb2_buf);
> > > > > > > +
> > > > > > > +       ret = pm_runtime_get_sync(jpeg->dev);
> > > > > > > +       if (ret < 0)
> > > > > > > +               goto enc_end;
> > > > > > > +
> > > > > > > +       spin_lock_irqsave(&jpeg->hw_lock, flags);
> > > > > > > +
> > > > > > > +       /*
> > > > > > > +        * Resetting the hardware every frame is to ensure that all the
> > > > > > > +        * registers are cleared. This is a hardware requirement.
> > > > > > > +        */
> > > > > > > +       mtk_jpeg_enc_reset(jpeg->reg_base);
> > > > > > > +
> > > > > > > +       mtk_jpeg_set_enc_dst(ctx, jpeg->reg_base, &dst_buf->vb2_buf, &enc_bs);
> > > > > > > +       mtk_jpeg_set_enc_src(ctx, jpeg->reg_base, &src_buf->vb2_buf);
> > > > > > > +       mtk_jpeg_enc_set_config(jpeg->reg_base, ctx->out_q.fmt->hw_format,
> > > > > > > +                               ctx->enable_exif, ctx->enc_quality,
> > > > > > > +                               ctx->restart_interval);
> > > > > > > +       mtk_jpeg_enc_start(jpeg->reg_base);
> > > > > >
> > > > > > Could we just move the above 5 functions into one function inside
> > > > > > mtk_jpeg_enc_hw.c that takes mtk_jpeg_dev pointer as its argument, let's
> > > > > > say mtk_jpeg_enc_hw_run() and simply program all the data to the registers
> > > > > > directly, without the extra level of abstractions?
> > > > > I can move the 5 functions into one function(mtk_jpeg_enc_hw_run()), but
> > > > > this function will be very long, because it contains computation code
> > > > > such as setting dst addr, blk_num, quality.
> > > > > In v4, you have adviced the following architecture:
> > > > > How about the following model, as used by many other drivers:
> > > > >
> > > > > mtk_jpeg_enc_set_src()
> > > > > {
> > > > >         // Set any registers related to source format and buffer
> > > > > }
> > > > >
> > > > > mtk_jpeg_enc_set_dst()
> > > > > {
> > > > >         // Set any registers related to destination format and buffer
> > > > > }
> > > > >
> > > > > mtk_jpeg_enc_set_params()
> > > > > {
> > > > >         // Set any registers related to additional encoding parameters
> > > > > }
> > > > >
> > > > > mtk_jpeg_enc_device_run(enc, ctx)
> > > > > {
> > > > >         mtk_jpeg_enc_set_src(enc, src_buf, src_fmt);
> > > > >         mtk_jpeg_enc_set_dst(enc, dst_buf, dst_fmt);
> > > > >         mtk_jpeg_enc_set_params(enc, ctx);
> > > > >         // Trigger the hardware run
> > > > > }
> > > > > I think that this architecture is more clear(mtk_jpeg_enc_set_config is
> > > > > equivalent to mtk_jpeg_enc_set_params).
> > > > > Should I keep the original architecture or move 5 functions into
> > > > > mtk_jpeg_enc_hw_run?
> > > >
> > > > Sounds good to me.
> > > >
> > > > My biggest issue with the code that it ends up introducing one more
> > > > level of abstraction, but with the approach you suggested, the arguments
> > > > just accept standard structs, which avoids that problem.
> > > >
> > > > [snip]
> > > > > > > +
> > > > > > > +       ctx->fh.ctrl_handler = &ctx->ctrl_hdl;
> > > > > > > +       ctx->colorspace = V4L2_COLORSPACE_JPEG,
> > > > > > > +       ctx->ycbcr_enc = V4L2_YCBCR_ENC_DEFAULT;
> > > > > > > +       ctx->quantization = V4L2_QUANTIZATION_DEFAULT;
> > > > > > > +       ctx->xfer_func = V4L2_XFER_FUNC_DEFAULT;
> > > > > >
> > > > > > Since we already have a v4l2_pix_format_mplane struct which has fields for
> > > > > > the above 4 values, could we just store them there?
> > > > > >
> > > > > > Also, I don't see this driver handling the colorspaces in any way, but it
> > > > > > seems to allow changing them from the userspace. This is incorrect, because
> > > > > > the userspace has no way to know that the colorspace is not handled.
> > > > > > Instead, the try_fmt implementation should always override the
> > > > > > userspace-provided colorspace configuration with the ones that the driver
> > > > > > assumes.
> > > > > >
> > > > > > > +       pix_mp->width = MTK_JPEG_MIN_WIDTH;
> > > > > > > +       pix_mp->height = MTK_JPEG_MIN_HEIGHT;
> > > > > > > +
> > > > > > > +       q->fmt = mtk_jpeg_find_format(V4L2_PIX_FMT_YUYV,
> > > > > > > +                                     MTK_JPEG_FMT_FLAG_ENC_OUTPUT);
> > > > > > > +       vidioc_try_fmt(container_of(pix_mp, struct v4l2_format,
> > > > > > > +                                   fmt.pix_mp), q->fmt);
> > > > > > > +       q->w = pix_mp->width;
> > > > > > > +       q->h = pix_mp->height;
> > > > > > > +       q->crop_rect.width = pix_mp->width;
> > > > > > > +       q->crop_rect.height = pix_mp->height;
> > > > > > > +       q->sizeimage[0] = pix_mp->plane_fmt[0].sizeimage;
> > > > > > > +       q->bytesperline[0] = pix_mp->plane_fmt[0].bytesperline;
> > > > > >
> > > > > > Actually, do we need this custom mtk_jpeg_q_data struct? Why couldn't we
> > > > > > just keep the same values inside the standard v4l2_pix_format_mplane
> > > > > > struct?
> > > > > I think that we need mtk_jpeg_q_data struct.If we delete it, how can we
> > > > > know these values(w, h, sizeimage, bytesperline, mtk_jpeg_fmt) belong to
> > > > > output or capture(output and capture's sizeimages are different, width
> > > > > and height are differnt too for jpeg dec )?We have
> > > > > s_fmt_vid_out_mplane/cap_mplane function to set these values.
> > > > >
> > > > > But we can use standard v4l2_pix_format_mplane struct replacing the w, h
> > > > > bytesperline, sizeimage in mtk_jpeg_q_data struct like this:
> > > > > struct mtk_jpeg_q_data{
> > > > >     struct mtk_jpeg_fmt *fmt;
> > > > >     struct v4l2_pix_format_mplane pix_mp;
> > > > >     struct v4l2_rect enc_crop_rect
> > > > > }
> > > > > Then delete ctx->colorspace ctx->ycbcr_enc ctx->quantization
> > > > > ctx->xfer_func, becuase v4l2_pix_format_mplane in q_data has contained
> > > > > them and assign them for out_q and cap_q separately.
> > > > >
> > > > > WDYT?
> > > >
> > > > Sounds good to me. I was considering just making it like
> > > >
> > > > struct mtk_jpeg_ctx {
> > > >       struct mtk_jpeg_fmt *src_fmt;
> > > >       struct v4l2_pix_format_mplane src_pix_mp;
> > > >       struct v4l2_rect src_crop;
> > > >
> > > >       struct mtk_jpeg_fmt *dst_fmt;
> > > >       struct v4l2_pix_format_mplane dst_pix_mp;
> > > >       struct v4l2_rect dst_crop;
> > > > };
> > > >
> > > > but I like your suggestion as well, as long as custom data structures
> > > > are not used to store standard information.
> > >
> > > Dear Tomasz,
> > >
> > > I used the structure like below:
> > >  struct mtk_jpeg_q_data{
> > >         struct mtk_jpeg_fmt *fmt;
> > >         struct v4l2_pix_format_mplane pix_mp;
> > >         struct v4l2_rect enc_crop_rect
> > > }
> > > Then delete ctx->colorspace ctx->ycbcr_enc ctx->quantization
> > > ctx->xfer_func, becuase v4l2_pix_format_mplane in q_data has contained
> > > them and assign them for out_q and cap_q separately.
> > >
> > > Then the v4l2_compliance test will fail, the fail log as below:
> > > Format ioctls:
> > >         test VIDIOC_ENUM_FMT/FRAMESIZES/FRAMEINTERVALS: OK
> > >         test VIDIOC_G/S_PARM: OK (Not Supported)
> > >         test VIDIOC_G_FBUF: OK (Not Supported)
> > >         test VIDIOC_G_FMT: OK
> > >         test VIDIOC_TRY_FMT: OK
> > >                 fail: v4l2-test-formats.cpp(836):
> > > fmt_cap.g_colorspace() != col
> > >         test VIDIOC_S_FMT: FAIL
> > >         test VIDIOC_G_SLICED_VBI_CAP: OK (Not Supported)
> > >         test Cropping: OK
> > >         test Composing: OK (Not Supported)
> > >         test Scaling: OK (Not Supported)
> > >
> > > The source code of  v4l2-test-formats.cpp as below:
> > >
> > > static int testM2MFormats(struct node *node)
> > > {
> > >         cv4l_fmt fmt_out;
> > >         cv4l_fmt fmt;
> > >         cv4l_fmt fmt_cap;
> > >         __u32 cap_type = node->g_type();
> > >         __u32 out_type = v4l_type_invert(cap_type);
> > >         __u32 col, ycbcr_enc, quant, xfer_func;
> > >
> > >         fail_on_test(node->g_fmt(fmt_out, out_type));
> > >         node->g_fmt(fmt_cap, cap_type);
> > >         fail_on_test(fmt_cap.g_colorspace() != fmt_out.g_colorspace());
> > >         fail_on_test(fmt_cap.g_ycbcr_enc() != fmt_out.g_ycbcr_enc());
> > >         fail_on_test(fmt_cap.g_quantization() != fmt_out.g_quantization());
> > >         fail_on_test(fmt_cap.g_xfer_func() != fmt_out.g_xfer_func());
> > >         col = fmt_out.g_colorspace() == V4L2_COLORSPACE_SMPTE170M ?
> > >                 V4L2_COLORSPACE_REC709 : V4L2_COLORSPACE_SMPTE170M;
> > >         ycbcr_enc = fmt_out.g_ycbcr_enc() == V4L2_YCBCR_ENC_601 ?
> > >                 V4L2_YCBCR_ENC_709 : V4L2_YCBCR_ENC_601;
> > >         quant = fmt_out.g_quantization() == V4L2_QUANTIZATION_LIM_RANGE ?
> > >                 V4L2_QUANTIZATION_FULL_RANGE : V4L2_QUANTIZATION_LIM_RANGE;
> > >         xfer_func = fmt_out.g_xfer_func() == V4L2_XFER_FUNC_SRGB ?
> > >                 V4L2_XFER_FUNC_709 : V4L2_XFER_FUNC_SRGB;
> > >         fmt_out.s_colorspace(col);
> > >         fmt_out.s_xfer_func(xfer_func);
> > >         fmt_out.s_ycbcr_enc(ycbcr_enc);
> > >         fmt_out.s_quantization(quant);
> > >         node->s_fmt(fmt_out);
> > >         fail_on_test(fmt_out.g_colorspace() != col);
> > >         fail_on_test(fmt_out.g_xfer_func() != xfer_func);
> > >         fail_on_test(fmt_out.g_ycbcr_enc() != ycbcr_enc);
> > >         fail_on_test(fmt_out.g_quantization() != quant);
> > >         node->g_fmt(fmt_cap);
> > >         fail_on_test(fmt_cap.g_colorspace() != col);  // line 836
> > >         fail_on_test(fmt_cap.g_xfer_func() != xfer_func);
> > >         fail_on_test(fmt_cap.g_ycbcr_enc() != ycbcr_enc);
> > >         fail_on_test(fmt_cap.g_quantization() != quant);
> > > }
> > >
> > > It needs that cap's colorspace equals out's colorspace when userspace
> > > just set out's colorspace and then get cap's colorspace. However, cap's
> > > colorspace which getted from driver equals V4L2_COLORSPACE_JPEG, because
> > > the code in g_fmt() like this:
> > > pix_mp->colorspace = q_data->pix_mp.colorspace;
> > > pix_mp->ycbcr_enc = q_data->pix_mp.ycbcr_enc;
> > > pix_mp->xfer_func = q_data->pix_mp.xfer_func;
> > > pix_mp->quantization = q_data->pix_mp.quantization;
> > >
> > > How should I handle this case? Should I store them(colorspace,
> > > sfer_func, ycbcr_enc, quatization) in ctx as the orinal desin? Then I
> > > can get them from g_fmt() like this:
> > > pix_mp->colorspace = ctx->colorspace;
> > > pix_mp->ycbcr_enc = ctx->ycbcr_enc;
> > > pix_mp->xfer_func = ctx->xfer_func;
> > > pix_mp->quantization = ctx->quantization;
> >
> > Why would there be any other colorspace accepted? I suppose that the
> > hardware only supports the JPEG color space, so it shouldn't accept
> > any other colorspace in TRY_FMT (and thus S_FMT) anyway.
> >
> > Still, for correctness, I would suggest propagating the colorspace
> > (and related) information from OUTPUT format to CAPTURE format in
> > S_FMT(OUTPUT).
> Dear Tomasz,
> If the driver doesn't accept any other colorspace from userspace(means
> the colorspace always equals V4L2_COLORSPACE_JPEG on my understanding),
> the v4l2_compliance will fail at line 831:
>   fail_on_test(fmt_out.g_colorspace() != col); // line 831
>
> The v4l2_compliance needs driver can accept other colorspace and capture
> colorspace equals output colorspace.
>
> I try to propagate the colorspace from OUTPUT format to CAPTURE foramt
> in S_FMT, the compliance test succeed.
>
> Should the jpeg driver can accept other colorspace?

Hans, could you advise on what is the proper behavior regarding
colorimetry support?

Best regards,
Tomasz



More information about the linux-arm-kernel mailing list