Lines Matching refs:stride
37 static void vc1_v_overlap_c(uint8_t *src, int stride) in vc1_v_overlap_c() argument
44 a = src[-2 * stride]; in vc1_v_overlap_c()
45 b = src[-stride]; in vc1_v_overlap_c()
47 d = src[stride]; in vc1_v_overlap_c()
51 src[-2 * stride] = a - d1; in vc1_v_overlap_c()
52 src[-stride] = av_clip_uint8(b - d2); in vc1_v_overlap_c()
54 src[stride] = d + d1; in vc1_v_overlap_c()
61 static void vc1_h_overlap_c(uint8_t *src, int stride) in vc1_h_overlap_c() argument
79 src += stride; in vc1_h_overlap_c()
147 static av_always_inline int vc1_filter_line(uint8_t *src, int stride, int pq) in vc1_filter_line() argument
149 int a0 = (2 * (src[-2 * stride] - src[1 * stride]) - in vc1_filter_line()
150 5 * (src[-1 * stride] - src[0 * stride]) + 4) >> 3; in vc1_filter_line()
155 int a1 = FFABS((2 * (src[-4 * stride] - src[-1 * stride]) - in vc1_filter_line()
156 5 * (src[-3 * stride] - src[-2 * stride]) + 4) >> 3); in vc1_filter_line()
157 int a2 = FFABS((2 * (src[ 0 * stride] - src[ 3 * stride]) - in vc1_filter_line()
158 5 * (src[ 1 * stride] - src[ 2 * stride]) + 4) >> 3); in vc1_filter_line()
160 int clip = src[-1 * stride] - src[0 * stride]; in vc1_filter_line()
177 src[-1 * stride] = av_clip_uint8(src[-1 * stride] - d); in vc1_filter_line()
178 src[ 0 * stride] = av_clip_uint8(src[ 0 * stride] + d); in vc1_filter_line()
196 static inline void vc1_loop_filter(uint8_t *src, int step, int stride, in vc1_loop_filter() argument
203 filt3 = vc1_filter_line(src + 2 * step, stride, pq); in vc1_loop_filter()
205 vc1_filter_line(src + 0 * step, stride, pq); in vc1_loop_filter()
206 vc1_filter_line(src + 1 * step, stride, pq); in vc1_loop_filter()
207 vc1_filter_line(src + 3 * step, stride, pq); in vc1_loop_filter()
213 static void vc1_v_loop_filter4_c(uint8_t *src, int stride, int pq) in vc1_v_loop_filter4_c() argument
215 vc1_loop_filter(src, 1, stride, 4, pq); in vc1_v_loop_filter4_c()
218 static void vc1_h_loop_filter4_c(uint8_t *src, int stride, int pq) in vc1_h_loop_filter4_c() argument
220 vc1_loop_filter(src, stride, 1, 4, pq); in vc1_h_loop_filter4_c()
223 static void vc1_v_loop_filter8_c(uint8_t *src, int stride, int pq) in vc1_v_loop_filter8_c() argument
225 vc1_loop_filter(src, 1, stride, 8, pq); in vc1_v_loop_filter8_c()
228 static void vc1_h_loop_filter8_c(uint8_t *src, int stride, int pq) in vc1_h_loop_filter8_c() argument
230 vc1_loop_filter(src, stride, 1, 8, pq); in vc1_h_loop_filter8_c()
233 static void vc1_v_loop_filter16_c(uint8_t *src, int stride, int pq) in vc1_v_loop_filter16_c() argument
235 vc1_loop_filter(src, 1, stride, 16, pq); in vc1_v_loop_filter16_c()
238 static void vc1_h_loop_filter16_c(uint8_t *src, int stride, int pq) in vc1_h_loop_filter16_c() argument
240 vc1_loop_filter(src, stride, 1, 16, pq); in vc1_h_loop_filter16_c()
244 static void vc1_inv_trans_8x8_dc_c(uint8_t *dest, ptrdiff_t stride, int16_t *block) in vc1_inv_trans_8x8_dc_c() argument
261 dest += stride; in vc1_inv_trans_8x8_dc_c()
335 static void vc1_inv_trans_8x4_dc_c(uint8_t *dest, ptrdiff_t stride, int16_t *block) in vc1_inv_trans_8x4_dc_c() argument
352 dest += stride; in vc1_inv_trans_8x4_dc_c()
356 static void vc1_inv_trans_8x4_c(uint8_t *dest, ptrdiff_t stride, int16_t *block) in vc1_inv_trans_8x4_c() argument
401 dest[0 * stride] = av_clip_uint8(dest[0 * stride] + ((t1 + t3) >> 7)); in vc1_inv_trans_8x4_c()
402 dest[1 * stride] = av_clip_uint8(dest[1 * stride] + ((t2 - t4) >> 7)); in vc1_inv_trans_8x4_c()
403 dest[2 * stride] = av_clip_uint8(dest[2 * stride] + ((t2 + t4) >> 7)); in vc1_inv_trans_8x4_c()
404 dest[3 * stride] = av_clip_uint8(dest[3 * stride] + ((t1 - t3) >> 7)); in vc1_inv_trans_8x4_c()
412 static void vc1_inv_trans_4x8_dc_c(uint8_t *dest, ptrdiff_t stride, int16_t *block) in vc1_inv_trans_4x8_dc_c() argument
425 dest += stride; in vc1_inv_trans_4x8_dc_c()
429 static void vc1_inv_trans_4x8_c(uint8_t *dest, ptrdiff_t stride, int16_t *block) in vc1_inv_trans_4x8_c() argument
470 dest[0 * stride] = av_clip_uint8(dest[0 * stride] + ((t5 + t1) >> 7)); in vc1_inv_trans_4x8_c()
471 dest[1 * stride] = av_clip_uint8(dest[1 * stride] + ((t6 + t2) >> 7)); in vc1_inv_trans_4x8_c()
472 dest[2 * stride] = av_clip_uint8(dest[2 * stride] + ((t7 + t3) >> 7)); in vc1_inv_trans_4x8_c()
473 dest[3 * stride] = av_clip_uint8(dest[3 * stride] + ((t8 + t4) >> 7)); in vc1_inv_trans_4x8_c()
474 dest[4 * stride] = av_clip_uint8(dest[4 * stride] + ((t8 - t4 + 1) >> 7)); in vc1_inv_trans_4x8_c()
475 dest[5 * stride] = av_clip_uint8(dest[5 * stride] + ((t7 - t3 + 1) >> 7)); in vc1_inv_trans_4x8_c()
476 dest[6 * stride] = av_clip_uint8(dest[6 * stride] + ((t6 - t2 + 1) >> 7)); in vc1_inv_trans_4x8_c()
477 dest[7 * stride] = av_clip_uint8(dest[7 * stride] + ((t5 - t1 + 1) >> 7)); in vc1_inv_trans_4x8_c()
485 static void vc1_inv_trans_4x4_dc_c(uint8_t *dest, ptrdiff_t stride, int16_t *block) in vc1_inv_trans_4x4_dc_c() argument
498 dest += stride; in vc1_inv_trans_4x4_dc_c()
502 static void vc1_inv_trans_4x4_c(uint8_t *dest, ptrdiff_t stride, int16_t *block) in vc1_inv_trans_4x4_c() argument
532 dest[0 * stride] = av_clip_uint8(dest[0 * stride] + ((t1 + t3) >> 7)); in vc1_inv_trans_4x4_c()
533 dest[1 * stride] = av_clip_uint8(dest[1 * stride] + ((t2 - t4) >> 7)); in vc1_inv_trans_4x4_c()
534 dest[2 * stride] = av_clip_uint8(dest[2 * stride] + ((t2 + t4) >> 7)); in vc1_inv_trans_4x4_c()
535 dest[3 * stride] = av_clip_uint8(dest[3 * stride] + ((t1 - t3) >> 7)); in vc1_inv_trans_4x4_c()
547 int stride, \
554 return -4 * src[-stride] + 53 * src[0] + \
555 18 * src[stride] - 3 * src[stride * 2]; \
557 return -1 * src[-stride] + 9 * src[0] + \
558 9 * src[stride] - 1 * src[stride * 2]; \
560 return -3 * src[-stride] + 18 * src[0] + \
561 53 * src[stride] - 4 * src[stride * 2]; \
570 static av_always_inline int vc1_mspel_filter(const uint8_t *src, int stride, in VC1_MSPEL_FILTER_16B()
577 return (-4 * src[-stride] + 53 * src[0] + in VC1_MSPEL_FILTER_16B()
578 18 * src[stride] - 3 * src[stride * 2] + 32 - r) >> 6; in VC1_MSPEL_FILTER_16B()
580 return (-1 * src[-stride] + 9 * src[0] + in VC1_MSPEL_FILTER_16B()
581 9 * src[stride] - 1 * src[stride * 2] + 8 - r) >> 4; in VC1_MSPEL_FILTER_16B()
583 return (-3 * src[-stride] + 18 * src[0] + in VC1_MSPEL_FILTER_16B()
584 53 * src[stride] - 4 * src[stride * 2] + 32 - r) >> 6; in VC1_MSPEL_FILTER_16B()
593 ptrdiff_t stride, \
613 tptr[i] = (vc1_mspel_ver_filter_16bits(src + i, stride, vmode) + r) >> shift; \
614 src += stride; \
623 dst += stride; \
633 OP(dst[i], vc1_mspel_filter(src + i, stride, vmode, r)); \
634 src += stride; \
635 dst += stride; \
645 dst += stride; \
646 src += stride; \
651 ptrdiff_t stride, \
671 tptr[i] = (vc1_mspel_ver_filter_16bits(src + i, stride, vmode) + r) >> shift; \
672 src += stride; \
681 dst += stride; \
691 OP(dst[i], vc1_mspel_filter(src + i, stride, vmode, r)); \
692 src += stride; \
693 dst += stride; \
703 dst += stride; \
704 src += stride; \
741 ptrdiff_t stride, int rnd) \ in VC1_MSPEL_MC()
743 put_vc1_mspel_mc(dst, src, stride, a, b, rnd); \ in VC1_MSPEL_MC()
747 ptrdiff_t stride, int rnd) \
749 avg_vc1_mspel_mc(dst, src, stride, a, b, rnd); \
753 ptrdiff_t stride, int rnd) \
755 put_vc1_mspel_mc_16(dst, src, stride, a, b, rnd); \
759 ptrdiff_t stride, int rnd) \
761 avg_vc1_mspel_mc_16(dst, src, stride, a, b, rnd); \
785 C * src[stride + a] + D * src[stride + a + 1] + 32 - 4) >> 6)
788 ptrdiff_t stride, int h, int x, int y)
807 dst += stride;
808 src += stride;
813 ptrdiff_t stride, int h, int x, int y) in put_no_rnd_vc1_chroma_mc4_c() argument
828 dst += stride; in put_no_rnd_vc1_chroma_mc4_c()
829 src += stride; in put_no_rnd_vc1_chroma_mc4_c()
836 ptrdiff_t stride, int h, int x, int y) in avg_no_rnd_vc1_chroma_mc8_c() argument
855 dst += stride; in avg_no_rnd_vc1_chroma_mc8_c()
856 src += stride; in avg_no_rnd_vc1_chroma_mc8_c()
862 ptrdiff_t stride, int h, int x, int y) in avg_no_rnd_vc1_chroma_mc4_c() argument
877 dst += stride; in avg_no_rnd_vc1_chroma_mc4_c()
878 src += stride; in avg_no_rnd_vc1_chroma_mc4_c()