rnd_avg32 942 libavcodec/dsputil.c OP(*((uint32_t*)&dst[i*dst_stride ]), rnd_avg32(a, b));\ rnd_avg32 945 libavcodec/dsputil.c OP(*((uint32_t*)&dst[i*dst_stride+4]), rnd_avg32(a, b));\ rnd_avg32 956 libavcodec/dsputil.c OP(*((uint32_t*)&dst[i*dst_stride ]), rnd_avg32(a, b));\ rnd_avg32 967 libavcodec/dsputil.c OP(*((uint16_t*)&dst[i*dst_stride ]), rnd_avg32(a, b));\ rnd_avg32 1257 libavcodec/dsputil.c #define op_avg(a, b) a = rnd_avg32(a, b) rnd_avg32 49 libavcodec/sh4/dsputil_align.c #define avg(d,s) d = rnd_avg32(s,d) rnd_avg32 29 libavcodec/sh4/qpel.c OP(LP(dst ),rnd_avg32(LP(src1 ),LP(src2 )) ); \ rnd_avg32 39 libavcodec/sh4/qpel.c OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LP(src2 )) ); \ rnd_avg32 62 libavcodec/sh4/qpel.c OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LP(src2 )) ); \ rnd_avg32 63 libavcodec/sh4/qpel.c OP(LP(dst+4),rnd_avg32(AV_RN32(src1+4),LP(src2+4)) ); \ rnd_avg32 64 libavcodec/sh4/qpel.c OP(LP(dst+8),rnd_avg32(AV_RN32(src1+8),LP(src2+8)) ); \ rnd_avg32 65 libavcodec/sh4/qpel.c OP(LP(dst+12),rnd_avg32(AV_RN32(src1+12),LP(src2+12)) ); \ rnd_avg32 86 libavcodec/sh4/qpel.c OP(LP(dst ),rnd_avg32(AV_RN32(src1 ),LP(src2 )) ); \ rnd_avg32 87 libavcodec/sh4/qpel.c OP(LP(dst+4),rnd_avg32(AV_RN32(src1+4),LP(src2+4)) ); \ rnd_avg32 108 libavcodec/sh4/qpel.c OP(LP(dst ),rnd_avg32(LP(src1 ),LP(src2 )) ); \ rnd_avg32 109 libavcodec/sh4/qpel.c OP(LP(dst+4),rnd_avg32(LP(src1+4),LP(src2+4)) ); \ rnd_avg32 132 libavcodec/sh4/qpel.c OP(LP(dst ),rnd_avg32(LP(src1 ),LP(src2 )) ); \ rnd_avg32 133 libavcodec/sh4/qpel.c OP(LP(dst+4),rnd_avg32(LP(src1+4),LP(src2+4)) ); \ rnd_avg32 134 libavcodec/sh4/qpel.c OP(LP(dst+8),rnd_avg32(LP(src1+8),LP(src2+8)) ); \ rnd_avg32 135 libavcodec/sh4/qpel.c OP(LP(dst+12),rnd_avg32(LP(src1+12),LP(src2+12)) ); \ rnd_avg32 315 libavcodec/sh4/qpel.c #define op_avg(a, b) a = rnd_avg32(a,b)