Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Alan Cox | 4360 | 88.17% | 2 | 8.70% |
Michael Hunold | 301 | 6.09% | 8 | 34.78% |
Hans Verkuil | 254 | 5.14% | 6 | 26.09% |
Joe Perches | 15 | 0.30% | 1 | 4.35% |
Andrew Morton | 5 | 0.10% | 1 | 4.35% |
Paul Gortmaker | 3 | 0.06% | 1 | 4.35% |
Johannes Stezenbach | 3 | 0.06% | 1 | 4.35% |
Mauro Carvalho Chehab | 2 | 0.04% | 1 | 4.35% |
Adrian Bunk | 1 | 0.02% | 1 | 4.35% |
Thomas Gleixner | 1 | 0.02% | 1 | 4.35% |
Total | 4945 | 23 |
// SPDX-License-Identifier: GPL-2.0-only #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include <linux/kernel.h> #include <linux/export.h> #include <media/drv-intf/saa7146_vv.h> static void calculate_output_format_register(struct saa7146_dev* saa, u32 palette, u32* clip_format) { /* clear out the necessary bits */ *clip_format &= 0x0000ffff; /* set these bits new */ *clip_format |= (( ((palette&0xf00)>>8) << 30) | ((palette&0x00f) << 24) | (((palette&0x0f0)>>4) << 16)); } static void calculate_hps_source_and_sync(struct saa7146_dev *dev, int source, int sync, u32* hps_ctrl) { *hps_ctrl &= ~(MASK_30 | MASK_31 | MASK_28); *hps_ctrl |= (source << 30) | (sync << 28); } static void calculate_hxo_and_hyo(struct saa7146_vv *vv, u32* hps_h_scale, u32* hps_ctrl) { int hyo = 0, hxo = 0; hyo = vv->standard->v_offset; hxo = vv->standard->h_offset; *hps_h_scale &= ~(MASK_B0 | 0xf00); *hps_h_scale |= (hxo << 0); *hps_ctrl &= ~(MASK_W0 | MASK_B2); *hps_ctrl |= (hyo << 12); } /* helper functions for the calculation of the horizontal- and vertical scaling registers, clip-format-register etc ... these functions take pointers to the (most-likely read-out original-values) and manipulate them according to the requested changes. */ /* hps_coeff used for CXY and CXUV; scale 1/1 -> scale 1/64 */ static struct { u16 hps_coeff; u16 weight_sum; } hps_h_coeff_tab [] = { {0x00, 2}, {0x02, 4}, {0x00, 4}, {0x06, 8}, {0x02, 8}, {0x08, 8}, {0x00, 8}, {0x1E, 16}, {0x0E, 8}, {0x26, 8}, {0x06, 8}, {0x42, 8}, {0x02, 8}, {0x80, 8}, {0x00, 8}, {0xFE, 16}, {0xFE, 8}, {0x7E, 8}, {0x7E, 8}, {0x3E, 8}, {0x3E, 8}, {0x1E, 8}, {0x1E, 8}, {0x0E, 8}, {0x0E, 8}, {0x06, 8}, {0x06, 8}, {0x02, 8}, {0x02, 8}, {0x00, 8}, {0x00, 8}, {0xFE, 16}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0xFE, 8}, {0x7E, 8}, {0x7E, 8}, {0x3E, 8}, {0x3E, 8}, {0x1E, 8}, {0x1E, 8}, {0x0E, 8}, {0x0E, 8}, {0x06, 8}, {0x06, 8}, {0x02, 8}, {0x02, 8}, {0x00, 8}, {0x00, 8}, {0xFE, 16} }; /* table of attenuation values for horizontal scaling */ static u8 h_attenuation[] = { 1, 2, 4, 8, 2, 4, 8, 16, 0}; /* calculate horizontal scale registers */ static int calculate_h_scale_registers(struct saa7146_dev *dev, int in_x, int out_x, int flip_lr, u32* hps_ctrl, u32* hps_v_gain, u32* hps_h_prescale, u32* hps_h_scale) { /* horizontal prescaler */ u32 dcgx = 0, xpsc = 0, xacm = 0, cxy = 0, cxuv = 0; /* horizontal scaler */ u32 xim = 0, xp = 0, xsci =0; /* vertical scale & gain */ u32 pfuv = 0; /* helper variables */ u32 h_atten = 0, i = 0; if ( 0 == out_x ) { return -EINVAL; } /* mask out vanity-bit */ *hps_ctrl &= ~MASK_29; /* calculate prescale-(xspc)-value: [n .. 1/2) : 1 [1/2 .. 1/3) : 2 [1/3 .. 1/4) : 3 ... */ if (in_x > out_x) { xpsc = in_x / out_x; } else { /* zooming */ xpsc = 1; } /* if flip_lr-bit is set, number of pixels after horizontal prescaling must be < 384 */ if ( 0 != flip_lr ) { /* set vanity bit */ *hps_ctrl |= MASK_29; while (in_x / xpsc >= 384 ) xpsc++; } /* if zooming is wanted, number of pixels after horizontal prescaling must be < 768 */ else { while ( in_x / xpsc >= 768 ) xpsc++; } /* maximum prescale is 64 (p.69) */ if ( xpsc > 64 ) xpsc = 64; /* keep xacm clear*/ xacm = 0; /* set horizontal filter parameters (CXY = CXUV) */ cxy = hps_h_coeff_tab[( (xpsc - 1) < 63 ? (xpsc - 1) : 63 )].hps_coeff; cxuv = cxy; /* calculate and set horizontal fine scale (xsci) */ /* bypass the horizontal scaler ? */ if ( (in_x == out_x) && ( 1 == xpsc ) ) xsci = 0x400; else xsci = ( (1024 * in_x) / (out_x * xpsc) ) + xpsc; /* set start phase for horizontal fine scale (xp) to 0 */ xp = 0; /* set xim, if we bypass the horizontal scaler */ if ( 0x400 == xsci ) xim = 1; else xim = 0; /* if the prescaler is bypassed, enable horizontal accumulation mode (xacm) and clear dcgx */ if( 1 == xpsc ) { xacm = 1; dcgx = 0; } else { xacm = 0; /* get best match in the table of attenuations for horizontal scaling */ h_atten = hps_h_coeff_tab[( (xpsc - 1) < 63 ? (xpsc - 1) : 63 )].weight_sum; for (i = 0; h_attenuation[i] != 0; i++) { if (h_attenuation[i] >= h_atten) break; } dcgx = i; } /* the horizontal scaling increment controls the UV filter to reduce the bandwidth to improve the display quality, so set it ... */ if ( xsci == 0x400) pfuv = 0x00; else if ( xsci < 0x600) pfuv = 0x01; else if ( xsci < 0x680) pfuv = 0x11; else if ( xsci < 0x700) pfuv = 0x22; else pfuv = 0x33; *hps_v_gain &= MASK_W0|MASK_B2; *hps_v_gain |= (pfuv << 24); *hps_h_scale &= ~(MASK_W1 | 0xf000); *hps_h_scale |= (xim << 31) | (xp << 24) | (xsci << 12); *hps_h_prescale |= (dcgx << 27) | ((xpsc-1) << 18) | (xacm << 17) | (cxy << 8) | (cxuv << 0); return 0; } static struct { u16 hps_coeff; u16 weight_sum; } hps_v_coeff_tab [] = { {0x0100, 2}, {0x0102, 4}, {0x0300, 4}, {0x0106, 8}, {0x0502, 8}, {0x0708, 8}, {0x0F00, 8}, {0x011E, 16}, {0x110E, 16}, {0x1926, 16}, {0x3906, 16}, {0x3D42, 16}, {0x7D02, 16}, {0x7F80, 16}, {0xFF00, 16}, {0x01FE, 32}, {0x01FE, 32}, {0x817E, 32}, {0x817E, 32}, {0xC13E, 32}, {0xC13E, 32}, {0xE11E, 32}, {0xE11E, 32}, {0xF10E, 32}, {0xF10E, 32}, {0xF906, 32}, {0xF906, 32}, {0xFD02, 32}, {0xFD02, 32}, {0xFF00, 32}, {0xFF00, 32}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x01FE, 64}, {0x817E, 64}, {0x817E, 64}, {0xC13E, 64}, {0xC13E, 64}, {0xE11E, 64}, {0xE11E, 64}, {0xF10E, 64}, {0xF10E, 64}, {0xF906, 64}, {0xF906, 64}, {0xFD02, 64}, {0xFD02, 64}, {0xFF00, 64}, {0xFF00, 64}, {0x01FE, 128} }; /* table of attenuation values for vertical scaling */ static u16 v_attenuation[] = { 2, 4, 8, 16, 32, 64, 128, 256, 0}; /* calculate vertical scale registers */ static int calculate_v_scale_registers(struct saa7146_dev *dev, enum v4l2_field field, int in_y, int out_y, u32* hps_v_scale, u32* hps_v_gain) { int lpi = 0; /* vertical scaling */ u32 yacm = 0, ysci = 0, yacl = 0, ypo = 0, ype = 0; /* vertical scale & gain */ u32 dcgy = 0, cya_cyb = 0; /* helper variables */ u32 v_atten = 0, i = 0; /* error, if vertical zooming */ if ( in_y < out_y ) { return -EINVAL; } /* linear phase interpolation may be used if scaling is between 1 and 1/2 (both fields used) or scaling is between 1/2 and 1/4 (if only one field is used) */ if (V4L2_FIELD_HAS_BOTH(field)) { if( 2*out_y >= in_y) { lpi = 1; } } else if (field == V4L2_FIELD_TOP || field == V4L2_FIELD_ALTERNATE || field == V4L2_FIELD_BOTTOM) { if( 4*out_y >= in_y ) { lpi = 1; } out_y *= 2; } if( 0 != lpi ) { yacm = 0; yacl = 0; cya_cyb = 0x00ff; /* calculate scaling increment */ if ( in_y > out_y ) ysci = ((1024 * in_y) / (out_y + 1)) - 1024; else ysci = 0; dcgy = 0; /* calculate ype and ypo */ ype = ysci / 16; ypo = ype + (ysci / 64); } else { yacm = 1; /* calculate scaling increment */ ysci = (((10 * 1024 * (in_y - out_y - 1)) / in_y) + 9) / 10; /* calculate ype and ypo */ ypo = ype = ((ysci + 15) / 16); /* the sequence length interval (yacl) has to be set according to the prescale value, e.g. [n .. 1/2) : 0 [1/2 .. 1/3) : 1 [1/3 .. 1/4) : 2 ... */ if ( ysci < 512) { yacl = 0; } else { yacl = ( ysci / (1024 - ysci) ); } /* get filter coefficients for cya, cyb from table hps_v_coeff_tab */ cya_cyb = hps_v_coeff_tab[ (yacl < 63 ? yacl : 63 ) ].hps_coeff; /* get best match in the table of attenuations for vertical scaling */ v_atten = hps_v_coeff_tab[ (yacl < 63 ? yacl : 63 ) ].weight_sum; for (i = 0; v_attenuation[i] != 0; i++) { if (v_attenuation[i] >= v_atten) break; } dcgy = i; } /* ypo and ype swapped in spec ? */ *hps_v_scale |= (yacm << 31) | (ysci << 21) | (yacl << 15) | (ypo << 8 ) | (ype << 1); *hps_v_gain &= ~(MASK_W0|MASK_B2); *hps_v_gain |= (dcgy << 16) | (cya_cyb << 0); return 0; } /* simple bubble-sort algorithm with duplicate elimination */ static void saa7146_set_window(struct saa7146_dev *dev, int width, int height, enum v4l2_field field) { struct saa7146_vv *vv = dev->vv_data; int source = vv->current_hps_source; int sync = vv->current_hps_sync; u32 hps_v_scale = 0, hps_v_gain = 0, hps_ctrl = 0, hps_h_prescale = 0, hps_h_scale = 0; /* set vertical scale */ hps_v_scale = 0; /* all bits get set by the function-call */ hps_v_gain = 0; /* fixme: saa7146_read(dev, HPS_V_GAIN);*/ calculate_v_scale_registers(dev, field, vv->standard->v_field*2, height, &hps_v_scale, &hps_v_gain); /* set horizontal scale */ hps_ctrl = 0; hps_h_prescale = 0; /* all bits get set in the function */ hps_h_scale = 0; calculate_h_scale_registers(dev, vv->standard->h_pixels, width, vv->hflip, &hps_ctrl, &hps_v_gain, &hps_h_prescale, &hps_h_scale); /* set hyo and hxo */ calculate_hxo_and_hyo(vv, &hps_h_scale, &hps_ctrl); calculate_hps_source_and_sync(dev, source, sync, &hps_ctrl); /* write out new register contents */ saa7146_write(dev, HPS_V_SCALE, hps_v_scale); saa7146_write(dev, HPS_V_GAIN, hps_v_gain); saa7146_write(dev, HPS_CTRL, hps_ctrl); saa7146_write(dev, HPS_H_PRESCALE,hps_h_prescale); saa7146_write(dev, HPS_H_SCALE, hps_h_scale); /* upload shadow-ram registers */ saa7146_write(dev, MC2, (MASK_05 | MASK_06 | MASK_21 | MASK_22) ); } static void saa7146_set_output_format(struct saa7146_dev *dev, unsigned long palette) { u32 clip_format = saa7146_read(dev, CLIP_FORMAT_CTRL); /* call helper function */ calculate_output_format_register(dev,palette,&clip_format); /* update the hps registers */ saa7146_write(dev, CLIP_FORMAT_CTRL, clip_format); saa7146_write(dev, MC2, (MASK_05 | MASK_21)); } /* select input-source */ void saa7146_set_hps_source_and_sync(struct saa7146_dev *dev, int source, int sync) { struct saa7146_vv *vv = dev->vv_data; u32 hps_ctrl = 0; /* read old state */ hps_ctrl = saa7146_read(dev, HPS_CTRL); hps_ctrl &= ~( MASK_31 | MASK_30 | MASK_28 ); hps_ctrl |= (source << 30) | (sync << 28); /* write back & upload register */ saa7146_write(dev, HPS_CTRL, hps_ctrl); saa7146_write(dev, MC2, (MASK_05 | MASK_21)); vv->current_hps_source = source; vv->current_hps_sync = sync; } EXPORT_SYMBOL_GPL(saa7146_set_hps_source_and_sync); void saa7146_write_out_dma(struct saa7146_dev* dev, int which, struct saa7146_video_dma* vdma) { int where = 0; if( which < 1 || which > 3) { return; } /* calculate starting address */ where = (which-1)*0x18; saa7146_write(dev, where, vdma->base_odd); saa7146_write(dev, where+0x04, vdma->base_even); saa7146_write(dev, where+0x08, vdma->prot_addr); saa7146_write(dev, where+0x0c, vdma->pitch); saa7146_write(dev, where+0x10, vdma->base_page); saa7146_write(dev, where+0x14, vdma->num_line_byte); /* upload */ saa7146_write(dev, MC2, (MASK_02<<(which-1))|(MASK_18<<(which-1))); /* printk("vdma%d.base_even: 0x%08x\n", which,vdma->base_even); printk("vdma%d.base_odd: 0x%08x\n", which,vdma->base_odd); printk("vdma%d.prot_addr: 0x%08x\n", which,vdma->prot_addr); printk("vdma%d.base_page: 0x%08x\n", which,vdma->base_page); printk("vdma%d.pitch: 0x%08x\n", which,vdma->pitch); printk("vdma%d.num_line_byte: 0x%08x\n", which,vdma->num_line_byte); */ } static int calculate_video_dma_grab_packed(struct saa7146_dev* dev, struct saa7146_buf *buf) { struct saa7146_vv *vv = dev->vv_data; struct v4l2_pix_format *pix = &vv->video_fmt; struct saa7146_video_dma vdma1; struct saa7146_format *sfmt = saa7146_format_by_fourcc(dev, pix->pixelformat); int width = pix->width; int height = pix->height; int bytesperline = pix->bytesperline; enum v4l2_field field = pix->field; int depth = sfmt->depth; DEB_CAP("[size=%dx%d,fields=%s]\n", width, height, v4l2_field_names[field]); if( bytesperline != 0) { vdma1.pitch = bytesperline*2; } else { vdma1.pitch = (width*depth*2)/8; } vdma1.num_line_byte = ((vv->standard->v_field<<16) + vv->standard->h_pixels); vdma1.base_page = buf->pt[0].dma | ME1 | sfmt->swap; if( 0 != vv->vflip ) { vdma1.prot_addr = buf->pt[0].offset; vdma1.base_even = buf->pt[0].offset+(vdma1.pitch/2)*height; vdma1.base_odd = vdma1.base_even - (vdma1.pitch/2); } else { vdma1.base_even = buf->pt[0].offset; vdma1.base_odd = vdma1.base_even + (vdma1.pitch/2); vdma1.prot_addr = buf->pt[0].offset+(vdma1.pitch/2)*height; } if (V4L2_FIELD_HAS_BOTH(field)) { } else if (field == V4L2_FIELD_ALTERNATE) { /* fixme */ if ( vv->last_field == V4L2_FIELD_TOP ) { vdma1.base_odd = vdma1.prot_addr; vdma1.pitch /= 2; } else if ( vv->last_field == V4L2_FIELD_BOTTOM ) { vdma1.base_odd = vdma1.base_even; vdma1.base_even = vdma1.prot_addr; vdma1.pitch /= 2; } } else if (field == V4L2_FIELD_TOP) { vdma1.base_odd = vdma1.prot_addr; vdma1.pitch /= 2; } else if (field == V4L2_FIELD_BOTTOM) { vdma1.base_odd = vdma1.base_even; vdma1.base_even = vdma1.prot_addr; vdma1.pitch /= 2; } if( 0 != vv->vflip ) { vdma1.pitch *= -1; } saa7146_write_out_dma(dev, 1, &vdma1); return 0; } static int calc_planar_422(struct saa7146_vv *vv, struct saa7146_buf *buf, struct saa7146_video_dma *vdma2, struct saa7146_video_dma *vdma3) { struct v4l2_pix_format *pix = &vv->video_fmt; int height = pix->height; int width = pix->width; vdma2->pitch = width; vdma3->pitch = width; /* fixme: look at bytesperline! */ if( 0 != vv->vflip ) { vdma2->prot_addr = buf->pt[1].offset; vdma2->base_even = ((vdma2->pitch/2)*height)+buf->pt[1].offset; vdma2->base_odd = vdma2->base_even - (vdma2->pitch/2); vdma3->prot_addr = buf->pt[2].offset; vdma3->base_even = ((vdma3->pitch/2)*height)+buf->pt[2].offset; vdma3->base_odd = vdma3->base_even - (vdma3->pitch/2); } else { vdma3->base_even = buf->pt[2].offset; vdma3->base_odd = vdma3->base_even + (vdma3->pitch/2); vdma3->prot_addr = (vdma3->pitch/2)*height+buf->pt[2].offset; vdma2->base_even = buf->pt[1].offset; vdma2->base_odd = vdma2->base_even + (vdma2->pitch/2); vdma2->prot_addr = (vdma2->pitch/2)*height+buf->pt[1].offset; } return 0; } static int calc_planar_420(struct saa7146_vv *vv, struct saa7146_buf *buf, struct saa7146_video_dma *vdma2, struct saa7146_video_dma *vdma3) { struct v4l2_pix_format *pix = &vv->video_fmt; int height = pix->height; int width = pix->width; vdma2->pitch = width/2; vdma3->pitch = width/2; if( 0 != vv->vflip ) { vdma2->prot_addr = buf->pt[2].offset; vdma2->base_even = ((vdma2->pitch/2)*height)+buf->pt[2].offset; vdma2->base_odd = vdma2->base_even - (vdma2->pitch/2); vdma3->prot_addr = buf->pt[1].offset; vdma3->base_even = ((vdma3->pitch/2)*height)+buf->pt[1].offset; vdma3->base_odd = vdma3->base_even - (vdma3->pitch/2); } else { vdma3->base_even = buf->pt[2].offset; vdma3->base_odd = vdma3->base_even + (vdma3->pitch); vdma3->prot_addr = (vdma3->pitch/2)*height+buf->pt[2].offset; vdma2->base_even = buf->pt[1].offset; vdma2->base_odd = vdma2->base_even + (vdma2->pitch); vdma2->prot_addr = (vdma2->pitch/2)*height+buf->pt[1].offset; } return 0; } static int calculate_video_dma_grab_planar(struct saa7146_dev* dev, struct saa7146_buf *buf) { struct saa7146_vv *vv = dev->vv_data; struct v4l2_pix_format *pix = &vv->video_fmt; struct saa7146_video_dma vdma1; struct saa7146_video_dma vdma2; struct saa7146_video_dma vdma3; struct saa7146_format *sfmt = saa7146_format_by_fourcc(dev, pix->pixelformat); int width = pix->width; int height = pix->height; enum v4l2_field field = pix->field; if (WARN_ON(!buf->pt[0].dma) || WARN_ON(!buf->pt[1].dma) || WARN_ON(!buf->pt[2].dma)) return -1; DEB_CAP("[size=%dx%d,fields=%s]\n", width, height, v4l2_field_names[field]); /* fixme: look at bytesperline! */ /* fixme: what happens for user space buffers here?. The offsets are most likely wrong, this version here only works for page-aligned buffers, modifications to the pagetable-functions are necessary...*/ vdma1.pitch = width*2; vdma1.num_line_byte = ((vv->standard->v_field<<16) + vv->standard->h_pixels); vdma1.base_page = buf->pt[0].dma | ME1; if( 0 != vv->vflip ) { vdma1.prot_addr = buf->pt[0].offset; vdma1.base_even = ((vdma1.pitch/2)*height)+buf->pt[0].offset; vdma1.base_odd = vdma1.base_even - (vdma1.pitch/2); } else { vdma1.base_even = buf->pt[0].offset; vdma1.base_odd = vdma1.base_even + (vdma1.pitch/2); vdma1.prot_addr = (vdma1.pitch/2)*height+buf->pt[0].offset; } vdma2.num_line_byte = 0; /* unused */ vdma2.base_page = buf->pt[1].dma | ME1; vdma3.num_line_byte = 0; /* unused */ vdma3.base_page = buf->pt[2].dma | ME1; switch( sfmt->depth ) { case 12: { calc_planar_420(vv,buf,&vdma2,&vdma3); break; } case 16: { calc_planar_422(vv,buf,&vdma2,&vdma3); break; } default: { return -1; } } if (V4L2_FIELD_HAS_BOTH(field)) { } else if (field == V4L2_FIELD_ALTERNATE) { /* fixme */ vdma1.base_odd = vdma1.prot_addr; vdma1.pitch /= 2; vdma2.base_odd = vdma2.prot_addr; vdma2.pitch /= 2; vdma3.base_odd = vdma3.prot_addr; vdma3.pitch /= 2; } else if (field == V4L2_FIELD_TOP) { vdma1.base_odd = vdma1.prot_addr; vdma1.pitch /= 2; vdma2.base_odd = vdma2.prot_addr; vdma2.pitch /= 2; vdma3.base_odd = vdma3.prot_addr; vdma3.pitch /= 2; } else if (field == V4L2_FIELD_BOTTOM) { vdma1.base_odd = vdma1.base_even; vdma1.base_even = vdma1.prot_addr; vdma1.pitch /= 2; vdma2.base_odd = vdma2.base_even; vdma2.base_even = vdma2.prot_addr; vdma2.pitch /= 2; vdma3.base_odd = vdma3.base_even; vdma3.base_even = vdma3.prot_addr; vdma3.pitch /= 2; } if( 0 != vv->vflip ) { vdma1.pitch *= -1; vdma2.pitch *= -1; vdma3.pitch *= -1; } saa7146_write_out_dma(dev, 1, &vdma1); if( (sfmt->flags & FORMAT_BYTE_SWAP) != 0 ) { saa7146_write_out_dma(dev, 3, &vdma2); saa7146_write_out_dma(dev, 2, &vdma3); } else { saa7146_write_out_dma(dev, 2, &vdma2); saa7146_write_out_dma(dev, 3, &vdma3); } return 0; } static void program_capture_engine(struct saa7146_dev *dev, int planar) { struct saa7146_vv *vv = dev->vv_data; int count = 0; unsigned long e_wait = vv->current_hps_sync == SAA7146_HPS_SYNC_PORT_A ? CMD_E_FID_A : CMD_E_FID_B; unsigned long o_wait = vv->current_hps_sync == SAA7146_HPS_SYNC_PORT_A ? CMD_O_FID_A : CMD_O_FID_B; /* wait for o_fid_a/b / e_fid_a/b toggle only if rps register 0 is not set*/ WRITE_RPS0(CMD_PAUSE | CMD_OAN | CMD_SIG0 | o_wait); WRITE_RPS0(CMD_PAUSE | CMD_OAN | CMD_SIG0 | e_wait); /* set rps register 0 */ WRITE_RPS0(CMD_WR_REG | (1 << 8) | (MC2/4)); WRITE_RPS0(MASK_27 | MASK_11); /* turn on video-dma1 */ WRITE_RPS0(CMD_WR_REG_MASK | (MC1/4)); WRITE_RPS0(MASK_06 | MASK_22); /* => mask */ WRITE_RPS0(MASK_06 | MASK_22); /* => values */ if( 0 != planar ) { /* turn on video-dma2 */ WRITE_RPS0(CMD_WR_REG_MASK | (MC1/4)); WRITE_RPS0(MASK_05 | MASK_21); /* => mask */ WRITE_RPS0(MASK_05 | MASK_21); /* => values */ /* turn on video-dma3 */ WRITE_RPS0(CMD_WR_REG_MASK | (MC1/4)); WRITE_RPS0(MASK_04 | MASK_20); /* => mask */ WRITE_RPS0(MASK_04 | MASK_20); /* => values */ } /* wait for o_fid_a/b / e_fid_a/b toggle */ if ( vv->last_field == V4L2_FIELD_INTERLACED ) { WRITE_RPS0(CMD_PAUSE | o_wait); WRITE_RPS0(CMD_PAUSE | e_wait); } else if ( vv->last_field == V4L2_FIELD_TOP ) { WRITE_RPS0(CMD_PAUSE | (vv->current_hps_sync == SAA7146_HPS_SYNC_PORT_A ? MASK_10 : MASK_09)); WRITE_RPS0(CMD_PAUSE | o_wait); } else if ( vv->last_field == V4L2_FIELD_BOTTOM ) { WRITE_RPS0(CMD_PAUSE | (vv->current_hps_sync == SAA7146_HPS_SYNC_PORT_A ? MASK_10 : MASK_09)); WRITE_RPS0(CMD_PAUSE | e_wait); } /* turn off video-dma1 */ WRITE_RPS0(CMD_WR_REG_MASK | (MC1/4)); WRITE_RPS0(MASK_22 | MASK_06); /* => mask */ WRITE_RPS0(MASK_22); /* => values */ if( 0 != planar ) { /* turn off video-dma2 */ WRITE_RPS0(CMD_WR_REG_MASK | (MC1/4)); WRITE_RPS0(MASK_05 | MASK_21); /* => mask */ WRITE_RPS0(MASK_21); /* => values */ /* turn off video-dma3 */ WRITE_RPS0(CMD_WR_REG_MASK | (MC1/4)); WRITE_RPS0(MASK_04 | MASK_20); /* => mask */ WRITE_RPS0(MASK_20); /* => values */ } /* generate interrupt */ WRITE_RPS0(CMD_INTERRUPT); /* stop */ WRITE_RPS0(CMD_STOP); } /* disable clipping */ static void saa7146_disable_clipping(struct saa7146_dev *dev) { u32 clip_format = saa7146_read(dev, CLIP_FORMAT_CTRL); /* mask out relevant bits (=lower word)*/ clip_format &= MASK_W1; /* upload clipping-registers*/ saa7146_write(dev, CLIP_FORMAT_CTRL, clip_format); saa7146_write(dev, MC2, (MASK_05 | MASK_21)); /* disable video dma2 */ saa7146_write(dev, MC1, MASK_21); } void saa7146_set_capture(struct saa7146_dev *dev, struct saa7146_buf *buf, struct saa7146_buf *next) { struct saa7146_vv *vv = dev->vv_data; struct v4l2_pix_format *pix = &vv->video_fmt; struct saa7146_format *sfmt = saa7146_format_by_fourcc(dev, pix->pixelformat); u32 vdma1_prot_addr; DEB_CAP("buf:%p, next:%p\n", buf, next); vdma1_prot_addr = saa7146_read(dev, PROT_ADDR1); if( 0 == vdma1_prot_addr ) { /* clear out beginning of streaming bit (rps register 0)*/ DEB_CAP("forcing sync to new frame\n"); saa7146_write(dev, MC2, MASK_27 ); } saa7146_set_window(dev, pix->width, pix->height, pix->field); saa7146_set_output_format(dev, sfmt->trans); saa7146_disable_clipping(dev); if ( vv->last_field == V4L2_FIELD_INTERLACED ) { } else if ( vv->last_field == V4L2_FIELD_TOP ) { vv->last_field = V4L2_FIELD_BOTTOM; } else if ( vv->last_field == V4L2_FIELD_BOTTOM ) { vv->last_field = V4L2_FIELD_TOP; } if( 0 != IS_PLANAR(sfmt->trans)) { calculate_video_dma_grab_planar(dev, buf); program_capture_engine(dev,1); } else { calculate_video_dma_grab_packed(dev, buf); program_capture_engine(dev,0); } /* printk("vdma%d.base_even: 0x%08x\n", 1,saa7146_read(dev,BASE_EVEN1)); printk("vdma%d.base_odd: 0x%08x\n", 1,saa7146_read(dev,BASE_ODD1)); printk("vdma%d.prot_addr: 0x%08x\n", 1,saa7146_read(dev,PROT_ADDR1)); printk("vdma%d.base_page: 0x%08x\n", 1,saa7146_read(dev,BASE_PAGE1)); printk("vdma%d.pitch: 0x%08x\n", 1,saa7146_read(dev,PITCH1)); printk("vdma%d.num_line_byte: 0x%08x\n", 1,saa7146_read(dev,NUM_LINE_BYTE1)); printk("vdma%d => vptr : 0x%08x\n", 1,saa7146_read(dev,PCI_VDP1)); */ /* write the address of the rps-program */ saa7146_write(dev, RPS_ADDR0, dev->d_rps0.dma_handle); /* turn on rps */ saa7146_write(dev, MC1, (MASK_12 | MASK_28)); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1