27 v16u8 in10, in11, in12, in13, in14, in15, in16, in17;
28 v16u8 in20, in21, in22, in23, in24, in25, in26, in27;
29 v8i16 out0, out1, out2, out3, out4, out5, out6, out7;
32 LD_UB8(src2,
stride, in20, in21, in22, in23, in24, in25, in26, in27);
33 ILVR_B4_SH(in10, in20, in11, in21, in12, in22, in13, in23,
34 out0, out1, out2, out3);
35 ILVR_B4_SH(in14, in24, in15, in25, in16, in26, in17, in27,
36 out4, out5, out6, out7);
37 HSUB_UB4_SH(out0, out1, out2, out3, out0, out1, out2, out3);
38 HSUB_UB4_SH(out4, out5, out6, out7, out4, out5, out6, out7);
39 ST_SH8(out0, out1, out2, out3, out4, out5, out6, out7,
block, 8);
43 int16_t *dst,
int32_t dst_stride,
51 dst_ptr = (uint8_t *) dst;
53 for (cnt = (
height >> 2); cnt--;) {
55 src += (4 * src_stride);
61 dst_ptr += (4 * 2 * dst_stride);
66 uint8_t *dst,
int32_t dst_stride,
70 const uint8_t *src_tmp;
72 v16u8
src0,
src1, src2, src3, src4, src5, src6, src7;
74 for (cnt = (
width >> 4); cnt--;) {
78 for (loop_cnt = (
height >> 3); loop_cnt--;) {
79 LD_UB8(src_tmp, src_stride,
80 src0,
src1, src2, src3, src4, src5, src6, src7);
81 src_tmp += (8 * src_stride);
85 dst_tmp += (8 * dst_stride);
94 uint8_t *dst,
int32_t dst_stride,
98 v16u8
src0,
src1, src2, src3, src4, src5, src6, src7;
101 for (cnt = (
height / 12); cnt--;) {
103 src0,
src1, src2, src3, src4, src5, src6, src7);
104 src += (8 * src_stride);
107 dst += (8 * dst_stride);
110 src += (4 * src_stride);
112 dst += (4 * dst_stride);
114 }
else if (0 ==
height % 8) {
116 }
else if (0 ==
height % 4) {
117 for (cnt = (
height >> 2); cnt--;) {
119 src += (4 * src_stride);
122 dst += (4 * dst_stride);
140 const uint8_t *src2, ptrdiff_t
stride)