Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Jacob Chen | 1939 | 70.82% | 2 | 8.00% |
Heiko Stübner | 437 | 15.96% | 4 | 16.00% |
Helen Mae Koike Fornazier | 261 | 9.53% | 2 | 8.00% |
Dafna Hirschfeld | 76 | 2.78% | 11 | 44.00% |
Laurent Pinchart | 22 | 0.80% | 3 | 12.00% |
Hans Verkuil | 2 | 0.07% | 2 | 8.00% |
Paul Elder | 1 | 0.04% | 1 | 4.00% |
Total | 2738 | 25 |
// SPDX-License-Identifier: (GPL-2.0+ OR MIT) /* * Rockchip ISP1 Driver - Stats subdevice * * Copyright (C) 2017 Rockchip Electronics Co., Ltd. */ #include <media/v4l2-common.h> #include <media/v4l2-event.h> #include <media/v4l2-ioctl.h> #include <media/videobuf2-core.h> #include <media/videobuf2-vmalloc.h> /* for ISP statistics */ #include "rkisp1-common.h" #define RKISP1_STATS_DEV_NAME RKISP1_DRIVER_NAME "_stats" #define RKISP1_ISP_STATS_REQ_BUFS_MIN 2 #define RKISP1_ISP_STATS_REQ_BUFS_MAX 8 static int rkisp1_stats_enum_fmt_meta_cap(struct file *file, void *priv, struct v4l2_fmtdesc *f) { struct video_device *video = video_devdata(file); struct rkisp1_stats *stats = video_get_drvdata(video); if (f->index > 0 || f->type != video->queue->type) return -EINVAL; f->pixelformat = stats->vdev_fmt.fmt.meta.dataformat; return 0; } static int rkisp1_stats_g_fmt_meta_cap(struct file *file, void *priv, struct v4l2_format *f) { struct video_device *video = video_devdata(file); struct rkisp1_stats *stats = video_get_drvdata(video); struct v4l2_meta_format *meta = &f->fmt.meta; if (f->type != video->queue->type) return -EINVAL; memset(meta, 0, sizeof(*meta)); meta->dataformat = stats->vdev_fmt.fmt.meta.dataformat; meta->buffersize = stats->vdev_fmt.fmt.meta.buffersize; return 0; } static int rkisp1_stats_querycap(struct file *file, void *priv, struct v4l2_capability *cap) { struct video_device *vdev = video_devdata(file); strscpy(cap->driver, RKISP1_DRIVER_NAME, sizeof(cap->driver)); strscpy(cap->card, vdev->name, sizeof(cap->card)); strscpy(cap->bus_info, RKISP1_BUS_INFO, sizeof(cap->bus_info)); return 0; } /* ISP video device IOCTLs */ static const struct v4l2_ioctl_ops rkisp1_stats_ioctl = { .vidioc_reqbufs = vb2_ioctl_reqbufs, .vidioc_querybuf = vb2_ioctl_querybuf, .vidioc_create_bufs = vb2_ioctl_create_bufs, .vidioc_qbuf = vb2_ioctl_qbuf, .vidioc_dqbuf = vb2_ioctl_dqbuf, .vidioc_prepare_buf = vb2_ioctl_prepare_buf, .vidioc_expbuf = vb2_ioctl_expbuf, .vidioc_streamon = vb2_ioctl_streamon, .vidioc_streamoff = vb2_ioctl_streamoff, .vidioc_enum_fmt_meta_cap = rkisp1_stats_enum_fmt_meta_cap, .vidioc_g_fmt_meta_cap = rkisp1_stats_g_fmt_meta_cap, .vidioc_s_fmt_meta_cap = rkisp1_stats_g_fmt_meta_cap, .vidioc_try_fmt_meta_cap = rkisp1_stats_g_fmt_meta_cap, .vidioc_querycap = rkisp1_stats_querycap, .vidioc_subscribe_event = v4l2_ctrl_subscribe_event, .vidioc_unsubscribe_event = v4l2_event_unsubscribe, }; static const struct v4l2_file_operations rkisp1_stats_fops = { .mmap = vb2_fop_mmap, .unlocked_ioctl = video_ioctl2, .poll = vb2_fop_poll, .open = v4l2_fh_open, .release = vb2_fop_release }; static int rkisp1_stats_vb2_queue_setup(struct vb2_queue *vq, unsigned int *num_buffers, unsigned int *num_planes, unsigned int sizes[], struct device *alloc_devs[]) { *num_planes = 1; *num_buffers = clamp_t(u32, *num_buffers, RKISP1_ISP_STATS_REQ_BUFS_MIN, RKISP1_ISP_STATS_REQ_BUFS_MAX); sizes[0] = sizeof(struct rkisp1_stat_buffer); return 0; } static void rkisp1_stats_vb2_buf_queue(struct vb2_buffer *vb) { struct vb2_v4l2_buffer *vbuf = to_vb2_v4l2_buffer(vb); struct rkisp1_buffer *stats_buf = container_of(vbuf, struct rkisp1_buffer, vb); struct vb2_queue *vq = vb->vb2_queue; struct rkisp1_stats *stats_dev = vq->drv_priv; spin_lock_irq(&stats_dev->lock); list_add_tail(&stats_buf->queue, &stats_dev->stat); spin_unlock_irq(&stats_dev->lock); } static int rkisp1_stats_vb2_buf_prepare(struct vb2_buffer *vb) { if (vb2_plane_size(vb, 0) < sizeof(struct rkisp1_stat_buffer)) return -EINVAL; vb2_set_plane_payload(vb, 0, sizeof(struct rkisp1_stat_buffer)); return 0; } static void rkisp1_stats_vb2_stop_streaming(struct vb2_queue *vq) { struct rkisp1_stats *stats = vq->drv_priv; struct rkisp1_buffer *buf; unsigned int i; spin_lock_irq(&stats->lock); for (i = 0; i < RKISP1_ISP_STATS_REQ_BUFS_MAX; i++) { if (list_empty(&stats->stat)) break; buf = list_first_entry(&stats->stat, struct rkisp1_buffer, queue); list_del(&buf->queue); vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR); } spin_unlock_irq(&stats->lock); } static const struct vb2_ops rkisp1_stats_vb2_ops = { .queue_setup = rkisp1_stats_vb2_queue_setup, .buf_queue = rkisp1_stats_vb2_buf_queue, .buf_prepare = rkisp1_stats_vb2_buf_prepare, .wait_prepare = vb2_ops_wait_prepare, .wait_finish = vb2_ops_wait_finish, .stop_streaming = rkisp1_stats_vb2_stop_streaming, }; static int rkisp1_stats_init_vb2_queue(struct vb2_queue *q, struct rkisp1_stats *stats) { struct rkisp1_vdev_node *node; node = container_of(q, struct rkisp1_vdev_node, buf_queue); q->type = V4L2_BUF_TYPE_META_CAPTURE; q->io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF; q->drv_priv = stats; q->ops = &rkisp1_stats_vb2_ops; q->mem_ops = &vb2_vmalloc_memops; q->buf_struct_size = sizeof(struct rkisp1_buffer); q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; q->lock = &node->vlock; return vb2_queue_init(q); } static void rkisp1_stats_get_awb_meas_v10(struct rkisp1_stats *stats, struct rkisp1_stat_buffer *pbuf) { /* Protect against concurrent access from ISR? */ struct rkisp1_device *rkisp1 = stats->rkisp1; u32 reg_val; pbuf->meas_type |= RKISP1_CIF_ISP_STAT_AWB; reg_val = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AWB_WHITE_CNT_V10); pbuf->params.awb.awb_mean[0].cnt = RKISP1_CIF_ISP_AWB_GET_PIXEL_CNT(reg_val); reg_val = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AWB_MEAN_V10); pbuf->params.awb.awb_mean[0].mean_cr_or_r = RKISP1_CIF_ISP_AWB_GET_MEAN_CR_R(reg_val); pbuf->params.awb.awb_mean[0].mean_cb_or_b = RKISP1_CIF_ISP_AWB_GET_MEAN_CB_B(reg_val); pbuf->params.awb.awb_mean[0].mean_y_or_g = RKISP1_CIF_ISP_AWB_GET_MEAN_Y_G(reg_val); } static void rkisp1_stats_get_awb_meas_v12(struct rkisp1_stats *stats, struct rkisp1_stat_buffer *pbuf) { /* Protect against concurrent access from ISR? */ struct rkisp1_device *rkisp1 = stats->rkisp1; u32 reg_val; pbuf->meas_type |= RKISP1_CIF_ISP_STAT_AWB; reg_val = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AWB_WHITE_CNT_V12); pbuf->params.awb.awb_mean[0].cnt = RKISP1_CIF_ISP_AWB_GET_PIXEL_CNT(reg_val); reg_val = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AWB_MEAN_V12); pbuf->params.awb.awb_mean[0].mean_cr_or_r = RKISP1_CIF_ISP_AWB_GET_MEAN_CR_R(reg_val); pbuf->params.awb.awb_mean[0].mean_cb_or_b = RKISP1_CIF_ISP_AWB_GET_MEAN_CB_B(reg_val); pbuf->params.awb.awb_mean[0].mean_y_or_g = RKISP1_CIF_ISP_AWB_GET_MEAN_Y_G(reg_val); } static void rkisp1_stats_get_aec_meas_v10(struct rkisp1_stats *stats, struct rkisp1_stat_buffer *pbuf) { struct rkisp1_device *rkisp1 = stats->rkisp1; unsigned int i; pbuf->meas_type |= RKISP1_CIF_ISP_STAT_AUTOEXP; for (i = 0; i < RKISP1_CIF_ISP_AE_MEAN_MAX_V10; i++) pbuf->params.ae.exp_mean[i] = (u8)rkisp1_read(rkisp1, RKISP1_CIF_ISP_EXP_MEAN_00_V10 + i * 4); } static void rkisp1_stats_get_aec_meas_v12(struct rkisp1_stats *stats, struct rkisp1_stat_buffer *pbuf) { struct rkisp1_device *rkisp1 = stats->rkisp1; u32 value; int i; pbuf->meas_type |= RKISP1_CIF_ISP_STAT_AUTOEXP; for (i = 0; i < RKISP1_CIF_ISP_AE_MEAN_MAX_V12 / 4; i++) { value = rkisp1_read(rkisp1, RKISP1_CIF_ISP_EXP_MEAN_V12 + i * 4); pbuf->params.ae.exp_mean[4 * i + 0] = RKISP1_CIF_ISP_EXP_GET_MEAN_xy0_V12(value); pbuf->params.ae.exp_mean[4 * i + 1] = RKISP1_CIF_ISP_EXP_GET_MEAN_xy1_V12(value); pbuf->params.ae.exp_mean[4 * i + 2] = RKISP1_CIF_ISP_EXP_GET_MEAN_xy2_V12(value); pbuf->params.ae.exp_mean[4 * i + 3] = RKISP1_CIF_ISP_EXP_GET_MEAN_xy3_V12(value); } value = rkisp1_read(rkisp1, RKISP1_CIF_ISP_EXP_MEAN_V12 + i * 4); pbuf->params.ae.exp_mean[4 * i + 0] = RKISP1_CIF_ISP_EXP_GET_MEAN_xy0_V12(value); } static void rkisp1_stats_get_afc_meas(struct rkisp1_stats *stats, struct rkisp1_stat_buffer *pbuf) { struct rkisp1_device *rkisp1 = stats->rkisp1; struct rkisp1_cif_isp_af_stat *af; pbuf->meas_type |= RKISP1_CIF_ISP_STAT_AFM; af = &pbuf->params.af; af->window[0].sum = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AFM_SUM_A); af->window[0].lum = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AFM_LUM_A); af->window[1].sum = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AFM_SUM_B); af->window[1].lum = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AFM_LUM_B); af->window[2].sum = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AFM_SUM_C); af->window[2].lum = rkisp1_read(rkisp1, RKISP1_CIF_ISP_AFM_LUM_C); } static void rkisp1_stats_get_hst_meas_v10(struct rkisp1_stats *stats, struct rkisp1_stat_buffer *pbuf) { struct rkisp1_device *rkisp1 = stats->rkisp1; unsigned int i; pbuf->meas_type |= RKISP1_CIF_ISP_STAT_HIST; for (i = 0; i < RKISP1_CIF_ISP_HIST_BIN_N_MAX_V10; i++) { u32 reg_val = rkisp1_read(rkisp1, RKISP1_CIF_ISP_HIST_BIN_0_V10 + i * 4); pbuf->params.hist.hist_bins[i] = RKISP1_CIF_ISP_HIST_GET_BIN_V10(reg_val); } } static void rkisp1_stats_get_hst_meas_v12(struct rkisp1_stats *stats, struct rkisp1_stat_buffer *pbuf) { struct rkisp1_device *rkisp1 = stats->rkisp1; u32 value; int i; pbuf->meas_type |= RKISP1_CIF_ISP_STAT_HIST; for (i = 0; i < RKISP1_CIF_ISP_HIST_BIN_N_MAX_V12 / 2; i++) { value = rkisp1_read(rkisp1, RKISP1_CIF_ISP_HIST_BIN_V12 + i * 4); pbuf->params.hist.hist_bins[2 * i] = RKISP1_CIF_ISP_HIST_GET_BIN0_V12(value); pbuf->params.hist.hist_bins[2 * i + 1] = RKISP1_CIF_ISP_HIST_GET_BIN1_V12(value); } } static void rkisp1_stats_get_bls_meas(struct rkisp1_stats *stats, struct rkisp1_stat_buffer *pbuf) { struct rkisp1_device *rkisp1 = stats->rkisp1; const struct rkisp1_mbus_info *in_fmt = rkisp1->isp.sink_fmt; struct rkisp1_cif_isp_bls_meas_val *bls_val; bls_val = &pbuf->params.ae.bls_val; if (in_fmt->bayer_pat == RKISP1_RAW_BGGR) { bls_val->meas_b = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_A_MEASURED); bls_val->meas_gb = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_B_MEASURED); bls_val->meas_gr = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_C_MEASURED); bls_val->meas_r = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_D_MEASURED); } else if (in_fmt->bayer_pat == RKISP1_RAW_GBRG) { bls_val->meas_gb = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_A_MEASURED); bls_val->meas_b = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_B_MEASURED); bls_val->meas_r = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_C_MEASURED); bls_val->meas_gr = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_D_MEASURED); } else if (in_fmt->bayer_pat == RKISP1_RAW_GRBG) { bls_val->meas_gr = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_A_MEASURED); bls_val->meas_r = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_B_MEASURED); bls_val->meas_b = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_C_MEASURED); bls_val->meas_gb = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_D_MEASURED); } else if (in_fmt->bayer_pat == RKISP1_RAW_RGGB) { bls_val->meas_r = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_A_MEASURED); bls_val->meas_gr = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_B_MEASURED); bls_val->meas_gb = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_C_MEASURED); bls_val->meas_b = rkisp1_read(rkisp1, RKISP1_CIF_ISP_BLS_D_MEASURED); } } static const struct rkisp1_stats_ops rkisp1_v10_stats_ops = { .get_awb_meas = rkisp1_stats_get_awb_meas_v10, .get_aec_meas = rkisp1_stats_get_aec_meas_v10, .get_hst_meas = rkisp1_stats_get_hst_meas_v10, }; static struct rkisp1_stats_ops rkisp1_v12_stats_ops = { .get_awb_meas = rkisp1_stats_get_awb_meas_v12, .get_aec_meas = rkisp1_stats_get_aec_meas_v12, .get_hst_meas = rkisp1_stats_get_hst_meas_v12, }; static void rkisp1_stats_send_measurement(struct rkisp1_stats *stats, u32 isp_ris) { struct rkisp1_stat_buffer *cur_stat_buf; struct rkisp1_buffer *cur_buf = NULL; unsigned int frame_sequence = stats->rkisp1->isp.frame_sequence; u64 timestamp = ktime_get_ns(); /* get one empty buffer */ if (!list_empty(&stats->stat)) { cur_buf = list_first_entry(&stats->stat, struct rkisp1_buffer, queue); list_del(&cur_buf->queue); } if (!cur_buf) return; cur_stat_buf = (struct rkisp1_stat_buffer *) vb2_plane_vaddr(&cur_buf->vb.vb2_buf, 0); if (isp_ris & RKISP1_CIF_ISP_AWB_DONE) stats->ops->get_awb_meas(stats, cur_stat_buf); if (isp_ris & RKISP1_CIF_ISP_AFM_FIN) rkisp1_stats_get_afc_meas(stats, cur_stat_buf); if (isp_ris & RKISP1_CIF_ISP_EXP_END) { stats->ops->get_aec_meas(stats, cur_stat_buf); rkisp1_stats_get_bls_meas(stats, cur_stat_buf); } if (isp_ris & RKISP1_CIF_ISP_HIST_MEASURE_RDY) stats->ops->get_hst_meas(stats, cur_stat_buf); vb2_set_plane_payload(&cur_buf->vb.vb2_buf, 0, sizeof(struct rkisp1_stat_buffer)); cur_buf->vb.sequence = frame_sequence; cur_buf->vb.vb2_buf.timestamp = timestamp; vb2_buffer_done(&cur_buf->vb.vb2_buf, VB2_BUF_STATE_DONE); } void rkisp1_stats_isr(struct rkisp1_stats *stats, u32 isp_ris) { struct rkisp1_device *rkisp1 = stats->rkisp1; unsigned int isp_mis_tmp = 0; spin_lock(&stats->lock); rkisp1_write(rkisp1, RKISP1_CIF_ISP_ICR, RKISP1_STATS_MEAS_MASK); isp_mis_tmp = rkisp1_read(rkisp1, RKISP1_CIF_ISP_MIS); if (isp_mis_tmp & RKISP1_STATS_MEAS_MASK) rkisp1->debug.stats_error++; if (isp_ris & RKISP1_STATS_MEAS_MASK) rkisp1_stats_send_measurement(stats, isp_ris); spin_unlock(&stats->lock); } static void rkisp1_init_stats(struct rkisp1_stats *stats) { stats->vdev_fmt.fmt.meta.dataformat = V4L2_META_FMT_RK_ISP1_STAT_3A; stats->vdev_fmt.fmt.meta.buffersize = sizeof(struct rkisp1_stat_buffer); if (stats->rkisp1->info->isp_ver == RKISP1_V12) stats->ops = &rkisp1_v12_stats_ops; else stats->ops = &rkisp1_v10_stats_ops; } int rkisp1_stats_register(struct rkisp1_device *rkisp1) { struct rkisp1_stats *stats = &rkisp1->stats; struct rkisp1_vdev_node *node = &stats->vnode; struct video_device *vdev = &node->vdev; int ret; stats->rkisp1 = rkisp1; mutex_init(&node->vlock); INIT_LIST_HEAD(&stats->stat); spin_lock_init(&stats->lock); strscpy(vdev->name, RKISP1_STATS_DEV_NAME, sizeof(vdev->name)); video_set_drvdata(vdev, stats); vdev->ioctl_ops = &rkisp1_stats_ioctl; vdev->fops = &rkisp1_stats_fops; vdev->release = video_device_release_empty; vdev->lock = &node->vlock; vdev->v4l2_dev = &rkisp1->v4l2_dev; vdev->queue = &node->buf_queue; vdev->device_caps = V4L2_CAP_META_CAPTURE | V4L2_CAP_STREAMING; vdev->vfl_dir = VFL_DIR_RX; rkisp1_stats_init_vb2_queue(vdev->queue, stats); rkisp1_init_stats(stats); video_set_drvdata(vdev, stats); node->pad.flags = MEDIA_PAD_FL_SINK; ret = media_entity_pads_init(&vdev->entity, 1, &node->pad); if (ret) goto error; ret = video_register_device(vdev, VFL_TYPE_VIDEO, -1); if (ret) { dev_err(&vdev->dev, "failed to register %s, ret=%d\n", vdev->name, ret); goto error; } return 0; error: media_entity_cleanup(&vdev->entity); mutex_destroy(&node->vlock); stats->rkisp1 = NULL; return ret; } void rkisp1_stats_unregister(struct rkisp1_device *rkisp1) { struct rkisp1_stats *stats = &rkisp1->stats; struct rkisp1_vdev_node *node = &stats->vnode; struct video_device *vdev = &node->vdev; if (!stats->rkisp1) return; vb2_video_unregister_device(vdev); media_entity_cleanup(&vdev->entity); mutex_destroy(&node->vlock); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1