Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Todor Tomov | 3415 | 90.42% | 5 | 22.73% |
Andrey Konovalov | 265 | 7.02% | 4 | 18.18% |
Milen Mitkov | 64 | 1.69% | 1 | 4.55% |
Sowjanya Komatineni | 6 | 0.16% | 1 | 4.55% |
Dan Carpenter | 5 | 0.13% | 1 | 4.55% |
Vladimir Zapolskiy | 5 | 0.13% | 1 | 4.55% |
Mauro Carvalho Chehab | 4 | 0.11% | 1 | 4.55% |
Laurent Pinchart | 4 | 0.11% | 2 | 9.09% |
Hans Verkuil | 3 | 0.08% | 2 | 9.09% |
Tomi Valkeinen | 2 | 0.05% | 1 | 4.55% |
Ezequiel García | 2 | 0.05% | 1 | 4.55% |
Boris Brezillon | 1 | 0.03% | 1 | 4.55% |
Radoslav Tsvetkov | 1 | 0.03% | 1 | 4.55% |
Total | 3777 | 22 |
// SPDX-License-Identifier: GPL-2.0 /* * camss-video.c * * Qualcomm MSM Camera Subsystem - V4L2 device node * * Copyright (c) 2013-2015, The Linux Foundation. All rights reserved. * Copyright (C) 2015-2018 Linaro Ltd. */ #include <linux/slab.h> #include <media/media-entity.h> #include <media/v4l2-dev.h> #include <media/v4l2-device.h> #include <media/v4l2-ioctl.h> #include <media/v4l2-mc.h> #include <media/videobuf2-dma-sg.h> #include "camss-video.h" #include "camss.h" #define CAMSS_FRAME_MIN_WIDTH 1 #define CAMSS_FRAME_MAX_WIDTH 8191 #define CAMSS_FRAME_MIN_HEIGHT 1 #define CAMSS_FRAME_MAX_HEIGHT_RDI 8191 #define CAMSS_FRAME_MAX_HEIGHT_PIX 4096 /* ----------------------------------------------------------------------------- * Helper functions */ /* * video_mbus_to_pix_mp - Convert v4l2_mbus_framefmt to v4l2_pix_format_mplane * @mbus: v4l2_mbus_framefmt format (input) * @pix: v4l2_pix_format_mplane format (output) * @f: a pointer to formats array element to be used for the conversion * @alignment: bytesperline alignment value * * Fill the output pix structure with information from the input mbus format. * * Return 0 on success or a negative error code otherwise */ static int video_mbus_to_pix_mp(const struct v4l2_mbus_framefmt *mbus, struct v4l2_pix_format_mplane *pix, const struct camss_format_info *f, unsigned int alignment) { unsigned int i; u32 bytesperline; memset(pix, 0, sizeof(*pix)); v4l2_fill_pix_format_mplane(pix, mbus); pix->pixelformat = f->pixelformat; pix->num_planes = f->planes; for (i = 0; i < pix->num_planes; i++) { bytesperline = pix->width / f->hsub[i].numerator * f->hsub[i].denominator * f->bpp[i] / 8; bytesperline = ALIGN(bytesperline, alignment); pix->plane_fmt[i].bytesperline = bytesperline; pix->plane_fmt[i].sizeimage = pix->height / f->vsub[i].numerator * f->vsub[i].denominator * bytesperline; } return 0; } static struct v4l2_subdev *video_remote_subdev(struct camss_video *video, u32 *pad) { struct media_pad *remote; remote = media_pad_remote_pad_first(&video->pad); if (!remote || !is_media_entity_v4l2_subdev(remote->entity)) return NULL; if (pad) *pad = remote->index; return media_entity_to_v4l2_subdev(remote->entity); } static int video_get_subdev_format(struct camss_video *video, struct v4l2_format *format) { struct v4l2_subdev_format fmt = { .which = V4L2_SUBDEV_FORMAT_ACTIVE, }; struct v4l2_subdev *subdev; u32 pad; int ret; subdev = video_remote_subdev(video, &pad); if (subdev == NULL) return -EPIPE; fmt.pad = pad; ret = v4l2_subdev_call(subdev, pad, get_fmt, NULL, &fmt); if (ret) return ret; ret = camss_format_find_format(fmt.format.code, format->fmt.pix_mp.pixelformat, video->formats, video->nformats); if (ret < 0) return ret; format->type = video->type; return video_mbus_to_pix_mp(&fmt.format, &format->fmt.pix_mp, &video->formats[ret], video->bpl_alignment); } /* ----------------------------------------------------------------------------- * Video queue operations */ static int video_queue_setup(struct vb2_queue *q, unsigned int *num_buffers, unsigned int *num_planes, unsigned int sizes[], struct device *alloc_devs[]) { struct camss_video *video = vb2_get_drv_priv(q); const struct v4l2_pix_format_mplane *format = &video->active_fmt.fmt.pix_mp; unsigned int i; if (*num_planes) { if (*num_planes != format->num_planes) return -EINVAL; for (i = 0; i < *num_planes; i++) if (sizes[i] < format->plane_fmt[i].sizeimage) return -EINVAL; return 0; } *num_planes = format->num_planes; for (i = 0; i < *num_planes; i++) sizes[i] = format->plane_fmt[i].sizeimage; return 0; } static int video_buf_init(struct vb2_buffer *vb) { struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); struct camss_video *video = vb2_get_drv_priv(vb->vb2_queue); struct camss_buffer *buffer = container_of(vbuf, struct camss_buffer, vb); const struct v4l2_pix_format_mplane *format = &video->active_fmt.fmt.pix_mp; struct sg_table *sgt; unsigned int i; for (i = 0; i < format->num_planes; i++) { sgt = vb2_dma_sg_plane_desc(vb, i); if (!sgt) return -EFAULT; buffer->addr[i] = sg_dma_address(sgt->sgl); } if (format->pixelformat == V4L2_PIX_FMT_NV12 || format->pixelformat == V4L2_PIX_FMT_NV21 || format->pixelformat == V4L2_PIX_FMT_NV16 || format->pixelformat == V4L2_PIX_FMT_NV61) buffer->addr[1] = buffer->addr[0] + format->plane_fmt[0].bytesperline * format->height; return 0; } static int video_buf_prepare(struct vb2_buffer *vb) { struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); struct camss_video *video = vb2_get_drv_priv(vb->vb2_queue); const struct v4l2_pix_format_mplane *format = &video->active_fmt.fmt.pix_mp; unsigned int i; for (i = 0; i < format->num_planes; i++) { if (format->plane_fmt[i].sizeimage > vb2_plane_size(vb, i)) return -EINVAL; vb2_set_plane_payload(vb, i, format->plane_fmt[i].sizeimage); } vbuf->field = V4L2_FIELD_NONE; return 0; } static void video_buf_queue(struct vb2_buffer *vb) { struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); struct camss_video *video = vb2_get_drv_priv(vb->vb2_queue); struct camss_buffer *buffer = container_of(vbuf, struct camss_buffer, vb); video->ops->queue_buffer(video, buffer); } static int video_check_format(struct camss_video *video) { struct v4l2_pix_format_mplane *pix = &video->active_fmt.fmt.pix_mp; struct v4l2_format format; struct v4l2_pix_format_mplane *sd_pix = &format.fmt.pix_mp; int ret; sd_pix->pixelformat = pix->pixelformat; ret = video_get_subdev_format(video, &format); if (ret < 0) return ret; if (pix->pixelformat != sd_pix->pixelformat || pix->height != sd_pix->height || pix->width != sd_pix->width || pix->num_planes != sd_pix->num_planes || pix->field != format.fmt.pix_mp.field) return -EPIPE; return 0; } static int video_start_streaming(struct vb2_queue *q, unsigned int count) { struct camss_video *video = vb2_get_drv_priv(q); struct video_device *vdev = &video->vdev; struct media_entity *entity; struct media_pad *pad; struct v4l2_subdev *subdev; int ret; ret = video_device_pipeline_alloc_start(vdev); if (ret < 0) { dev_err(video->camss->dev, "Failed to start media pipeline: %d\n", ret); goto flush_buffers; } ret = video_check_format(video); if (ret < 0) goto error; entity = &vdev->entity; while (1) { pad = &entity->pads[0]; if (!(pad->flags & MEDIA_PAD_FL_SINK)) break; pad = media_pad_remote_pad_first(pad); if (!pad || !is_media_entity_v4l2_subdev(pad->entity)) break; entity = pad->entity; subdev = media_entity_to_v4l2_subdev(entity); ret = v4l2_subdev_call(subdev, video, s_stream, 1); if (ret < 0 && ret != -ENOIOCTLCMD) goto error; } return 0; error: video_device_pipeline_stop(vdev); flush_buffers: video->ops->flush_buffers(video, VB2_BUF_STATE_QUEUED); return ret; } static void video_stop_streaming(struct vb2_queue *q) { struct camss_video *video = vb2_get_drv_priv(q); struct video_device *vdev = &video->vdev; struct media_entity *entity; struct media_pad *pad; struct v4l2_subdev *subdev; int ret; entity = &vdev->entity; while (1) { pad = &entity->pads[0]; if (!(pad->flags & MEDIA_PAD_FL_SINK)) break; pad = media_pad_remote_pad_first(pad); if (!pad || !is_media_entity_v4l2_subdev(pad->entity)) break; entity = pad->entity; subdev = media_entity_to_v4l2_subdev(entity); ret = v4l2_subdev_call(subdev, video, s_stream, 0); if (entity->use_count > 1) { /* Don't stop if other instances of the pipeline are still running */ dev_dbg(video->camss->dev, "Video pipeline still used, don't stop streaming.\n"); return; } if (ret) { dev_err(video->camss->dev, "Video pipeline stop failed: %d\n", ret); return; } } video_device_pipeline_stop(vdev); video->ops->flush_buffers(video, VB2_BUF_STATE_ERROR); } static const struct vb2_ops msm_video_vb2_q_ops = { .queue_setup = video_queue_setup, .wait_prepare = vb2_ops_wait_prepare, .wait_finish = vb2_ops_wait_finish, .buf_init = video_buf_init, .buf_prepare = video_buf_prepare, .buf_queue = video_buf_queue, .start_streaming = video_start_streaming, .stop_streaming = video_stop_streaming, }; /* ----------------------------------------------------------------------------- * V4L2 ioctls */ static int video_querycap(struct file *file, void *fh, struct v4l2_capability *cap) { strscpy(cap->driver, "qcom-camss", sizeof(cap->driver)); strscpy(cap->card, "Qualcomm Camera Subsystem", sizeof(cap->card)); return 0; } static int video_enum_fmt(struct file *file, void *fh, struct v4l2_fmtdesc *f) { struct camss_video *video = video_drvdata(file); int i, j, k; u32 mcode = f->mbus_code; if (f->type != video->type) return -EINVAL; if (f->index >= video->nformats) return -EINVAL; /* * Find index "i" of "k"th unique pixelformat in formats array. * * If f->mbus_code passed to video_enum_fmt() is not zero, a device * with V4L2_CAP_IO_MC capability restricts enumeration to only the * pixel formats that can be produced from that media bus code. * This is implemented by skipping video->formats[] entries with * code != f->mbus_code (if f->mbus_code is not zero). * If the f->mbus_code passed to video_enum_fmt() is not supported, * -EINVAL is returned. * If f->mbus_code is zero, all the pixel formats are enumerated. */ k = -1; for (i = 0; i < video->nformats; i++) { if (mcode != 0 && video->formats[i].code != mcode) continue; for (j = 0; j < i; j++) { if (mcode != 0 && video->formats[j].code != mcode) continue; if (video->formats[i].pixelformat == video->formats[j].pixelformat) break; } if (j == i) k++; if (k == f->index) break; } if (k == -1 || k < f->index) /* * All the unique pixel formats matching the arguments * have been enumerated (k >= 0 and f->index > 0), or * no pixel formats match the non-zero f->mbus_code (k == -1). */ return -EINVAL; f->pixelformat = video->formats[i].pixelformat; return 0; } static int video_enum_framesizes(struct file *file, void *fh, struct v4l2_frmsizeenum *fsize) { struct camss_video *video = video_drvdata(file); int i; if (fsize->index) return -EINVAL; /* Only accept pixel format present in the formats[] table */ for (i = 0; i < video->nformats; i++) { if (video->formats[i].pixelformat == fsize->pixel_format) break; } if (i == video->nformats) return -EINVAL; fsize->type = V4L2_FRMSIZE_TYPE_CONTINUOUS; fsize->stepwise.min_width = CAMSS_FRAME_MIN_WIDTH; fsize->stepwise.max_width = CAMSS_FRAME_MAX_WIDTH; fsize->stepwise.min_height = CAMSS_FRAME_MIN_HEIGHT; fsize->stepwise.max_height = (video->line_based) ? CAMSS_FRAME_MAX_HEIGHT_PIX : CAMSS_FRAME_MAX_HEIGHT_RDI; fsize->stepwise.step_width = 1; fsize->stepwise.step_height = 1; return 0; } static int video_g_fmt(struct file *file, void *fh, struct v4l2_format *f) { struct camss_video *video = video_drvdata(file); *f = video->active_fmt; return 0; } static int __video_try_fmt(struct camss_video *video, struct v4l2_format *f) { struct v4l2_pix_format_mplane *pix_mp; const struct camss_format_info *fi; struct v4l2_plane_pix_format *p; u32 bytesperline[3] = { 0 }; u32 sizeimage[3] = { 0 }; u32 width, height; u32 bpl, lines; int i, j; pix_mp = &f->fmt.pix_mp; if (video->line_based) for (i = 0; i < pix_mp->num_planes && i < 3; i++) { p = &pix_mp->plane_fmt[i]; bytesperline[i] = clamp_t(u32, p->bytesperline, 1, 65528); sizeimage[i] = clamp_t(u32, p->sizeimage, bytesperline[i], bytesperline[i] * CAMSS_FRAME_MAX_HEIGHT_PIX); } for (j = 0; j < video->nformats; j++) if (pix_mp->pixelformat == video->formats[j].pixelformat) break; if (j == video->nformats) j = 0; /* default format */ fi = &video->formats[j]; width = pix_mp->width; height = pix_mp->height; memset(pix_mp, 0, sizeof(*pix_mp)); pix_mp->pixelformat = fi->pixelformat; pix_mp->width = clamp_t(u32, width, 1, CAMSS_FRAME_MAX_WIDTH); pix_mp->height = clamp_t(u32, height, 1, CAMSS_FRAME_MAX_HEIGHT_RDI); pix_mp->num_planes = fi->planes; for (i = 0; i < pix_mp->num_planes; i++) { bpl = pix_mp->width / fi->hsub[i].numerator * fi->hsub[i].denominator * fi->bpp[i] / 8; bpl = ALIGN(bpl, video->bpl_alignment); pix_mp->plane_fmt[i].bytesperline = bpl; pix_mp->plane_fmt[i].sizeimage = pix_mp->height / fi->vsub[i].numerator * fi->vsub[i].denominator * bpl; } pix_mp->field = V4L2_FIELD_NONE; pix_mp->colorspace = V4L2_COLORSPACE_SRGB; pix_mp->flags = 0; pix_mp->ycbcr_enc = V4L2_MAP_YCBCR_ENC_DEFAULT(pix_mp->colorspace); pix_mp->quantization = V4L2_MAP_QUANTIZATION_DEFAULT(true, pix_mp->colorspace, pix_mp->ycbcr_enc); pix_mp->xfer_func = V4L2_MAP_XFER_FUNC_DEFAULT(pix_mp->colorspace); if (video->line_based) for (i = 0; i < pix_mp->num_planes; i++) { p = &pix_mp->plane_fmt[i]; p->bytesperline = clamp_t(u32, p->bytesperline, 1, 65528); p->sizeimage = clamp_t(u32, p->sizeimage, p->bytesperline, p->bytesperline * CAMSS_FRAME_MAX_HEIGHT_PIX); lines = p->sizeimage / p->bytesperline; if (p->bytesperline < bytesperline[i]) p->bytesperline = ALIGN(bytesperline[i], 8); if (p->sizeimage < p->bytesperline * lines) p->sizeimage = p->bytesperline * lines; if (p->sizeimage < sizeimage[i]) p->sizeimage = sizeimage[i]; } return 0; } static int video_try_fmt(struct file *file, void *fh, struct v4l2_format *f) { struct camss_video *video = video_drvdata(file); return __video_try_fmt(video, f); } static int video_s_fmt(struct file *file, void *fh, struct v4l2_format *f) { struct camss_video *video = video_drvdata(file); int ret; if (vb2_is_busy(&video->vb2_q)) return -EBUSY; ret = __video_try_fmt(video, f); if (ret < 0) return ret; video->active_fmt = *f; return 0; } static int video_enum_input(struct file *file, void *fh, struct v4l2_input *input) { if (input->index > 0) return -EINVAL; strscpy(input->name, "camera", sizeof(input->name)); input->type = V4L2_INPUT_TYPE_CAMERA; return 0; } static int video_g_input(struct file *file, void *fh, unsigned int *input) { *input = 0; return 0; } static int video_s_input(struct file *file, void *fh, unsigned int input) { return input == 0 ? 0 : -EINVAL; } static const struct v4l2_ioctl_ops msm_vid_ioctl_ops = { .vidioc_querycap = video_querycap, .vidioc_enum_fmt_vid_cap = video_enum_fmt, .vidioc_enum_framesizes = video_enum_framesizes, .vidioc_g_fmt_vid_cap_mplane = video_g_fmt, .vidioc_s_fmt_vid_cap_mplane = video_s_fmt, .vidioc_try_fmt_vid_cap_mplane = video_try_fmt, .vidioc_reqbufs = vb2_ioctl_reqbufs, .vidioc_querybuf = vb2_ioctl_querybuf, .vidioc_qbuf = vb2_ioctl_qbuf, .vidioc_expbuf = vb2_ioctl_expbuf, .vidioc_dqbuf = vb2_ioctl_dqbuf, .vidioc_create_bufs = vb2_ioctl_create_bufs, .vidioc_prepare_buf = vb2_ioctl_prepare_buf, .vidioc_streamon = vb2_ioctl_streamon, .vidioc_streamoff = vb2_ioctl_streamoff, .vidioc_enum_input = video_enum_input, .vidioc_g_input = video_g_input, .vidioc_s_input = video_s_input, }; /* ----------------------------------------------------------------------------- * V4L2 file operations */ static int video_open(struct file *file) { struct video_device *vdev = video_devdata(file); struct camss_video *video = video_drvdata(file); struct v4l2_fh *vfh; int ret; mutex_lock(&video->lock); vfh = kzalloc(sizeof(*vfh), GFP_KERNEL); if (vfh == NULL) { ret = -ENOMEM; goto error_alloc; } v4l2_fh_init(vfh, vdev); v4l2_fh_add(vfh); file->private_data = vfh; ret = v4l2_pipeline_pm_get(&vdev->entity); if (ret < 0) { dev_err(video->camss->dev, "Failed to power up pipeline: %d\n", ret); goto error_pm_use; } mutex_unlock(&video->lock); return 0; error_pm_use: v4l2_fh_release(file); error_alloc: mutex_unlock(&video->lock); return ret; } static int video_release(struct file *file) { struct video_device *vdev = video_devdata(file); vb2_fop_release(file); v4l2_pipeline_pm_put(&vdev->entity); file->private_data = NULL; return 0; } static const struct v4l2_file_operations msm_vid_fops = { .owner = THIS_MODULE, .unlocked_ioctl = video_ioctl2, .open = video_open, .release = video_release, .poll = vb2_fop_poll, .mmap = vb2_fop_mmap, .read = vb2_fop_read, }; /* ----------------------------------------------------------------------------- * CAMSS video core */ static void msm_video_release(struct video_device *vdev) { struct camss_video *video = video_get_drvdata(vdev); media_entity_cleanup(&vdev->entity); mutex_destroy(&video->q_lock); mutex_destroy(&video->lock); if (atomic_dec_and_test(&video->camss->ref_count)) camss_delete(video->camss); } /* * msm_video_init_format - Helper function to initialize format * @video: struct camss_video * * Initialize pad format with default value. * * Return 0 on success or a negative error code otherwise */ static int msm_video_init_format(struct camss_video *video) { int ret; struct v4l2_format format = { .type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE, .fmt.pix_mp = { .width = 1920, .height = 1080, .pixelformat = video->formats[0].pixelformat, }, }; ret = __video_try_fmt(video, &format); if (ret < 0) return ret; video->active_fmt = format; return 0; } /* * msm_video_register - Register a video device node * @video: struct camss_video * @v4l2_dev: V4L2 device * @name: name to be used for the video device node * * Initialize and register a video device node to a V4L2 device. Also * initialize the vb2 queue. * * Return 0 on success or a negative error code otherwise */ int msm_video_register(struct camss_video *video, struct v4l2_device *v4l2_dev, const char *name) { struct media_pad *pad = &video->pad; struct video_device *vdev; struct vb2_queue *q; int ret; vdev = &video->vdev; mutex_init(&video->q_lock); q = &video->vb2_q; q->drv_priv = video; q->mem_ops = &vb2_dma_sg_memops; q->ops = &msm_video_vb2_q_ops; q->type = V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE; q->io_modes = VB2_DMABUF | VB2_MMAP | VB2_READ; q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; q->buf_struct_size = sizeof(struct camss_buffer); q->dev = video->camss->dev; q->lock = &video->q_lock; ret = vb2_queue_init(q); if (ret < 0) { dev_err(v4l2_dev->dev, "Failed to init vb2 queue: %d\n", ret); goto error_vb2_init; } pad->flags = MEDIA_PAD_FL_SINK; ret = media_entity_pads_init(&vdev->entity, 1, pad); if (ret < 0) { dev_err(v4l2_dev->dev, "Failed to init video entity: %d\n", ret); goto error_vb2_init; } mutex_init(&video->lock); ret = msm_video_init_format(video); if (ret < 0) { dev_err(v4l2_dev->dev, "Failed to init format: %d\n", ret); goto error_video_register; } vdev->fops = &msm_vid_fops; vdev->device_caps = V4L2_CAP_VIDEO_CAPTURE_MPLANE | V4L2_CAP_STREAMING | V4L2_CAP_READWRITE | V4L2_CAP_IO_MC; vdev->ioctl_ops = &msm_vid_ioctl_ops; vdev->release = msm_video_release; vdev->v4l2_dev = v4l2_dev; vdev->vfl_dir = VFL_DIR_RX; vdev->queue = &video->vb2_q; vdev->lock = &video->lock; strscpy(vdev->name, name, sizeof(vdev->name)); ret = video_register_device(vdev, VFL_TYPE_VIDEO, -1); if (ret < 0) { dev_err(v4l2_dev->dev, "Failed to register video device: %d\n", ret); goto error_video_register; } video_set_drvdata(vdev, video); atomic_inc(&video->camss->ref_count); return 0; error_video_register: media_entity_cleanup(&vdev->entity); mutex_destroy(&video->lock); error_vb2_init: mutex_destroy(&video->q_lock); return ret; } void msm_video_unregister(struct camss_video *video) { atomic_inc(&video->camss->ref_count); vb2_video_unregister_device(&video->vdev); atomic_dec(&video->camss->ref_count); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1