[FFmpeg-devel] [PATCH 1/3] vp9/x86: 8x8 iadst SIMD (ssse3/avx) variants.

Ronald S. Bultje rsbultje at gmail.com
Thu Jan 23 02:55:08 CET 2014


Cycle measurements for intra itxfm_8x8_add on ped1080p.webm:
idct_idct:   133 -> 135 cycles (noise measurement)
idct_iadst:  900 -> 241 cycles
iadst_idct:  864 -> 215 cycles
iadst_iadst: 973 -> 310 cycles
---
 libavcodec/x86/vp9dsp_init.c |  10 ++-
 libavcodec/x86/vp9itxfm.asm  | 157 +++++++++++++++++++++++++++++++------------
 2 files changed, 121 insertions(+), 46 deletions(-)

diff --git a/libavcodec/x86/vp9dsp_init.c b/libavcodec/x86/vp9dsp_init.c
index ab3396e..9b2cfe1 100644
--- a/libavcodec/x86/vp9dsp_init.c
+++ b/libavcodec/x86/vp9dsp_init.c
@@ -167,8 +167,8 @@ itxfm_func(idct,  iadst, size, opt); \
 itxfm_func(iadst, iadst, size, opt)
 
 itxfm_func(idct, idct,  4, ssse3);
-itxfm_func(idct, idct,  8, ssse3);
-itxfm_func(idct, idct,  8, avx);
+itxfm_funcs(8, ssse3);
+itxfm_funcs(8, avx);
 itxfm_funcs(16, ssse3);
 itxfm_funcs(16, avx);
 itxfm_func(idct, idct, 32, ssse3);
@@ -249,6 +249,9 @@ av_cold void ff_vp9dsp_init_x86(VP9DSPContext *dsp)
         dsp->itxfm_add[TX_4X4][DCT_DCT] = ff_vp9_idct_idct_4x4_add_ssse3;
         if (ARCH_X86_64) {
             dsp->itxfm_add[TX_8X8][DCT_DCT] = ff_vp9_idct_idct_8x8_add_ssse3;
+            dsp->itxfm_add[TX_8X8][ADST_DCT]  = ff_vp9_idct_iadst_8x8_add_ssse3;
+            dsp->itxfm_add[TX_8X8][DCT_ADST]  = ff_vp9_iadst_idct_8x8_add_ssse3;
+            dsp->itxfm_add[TX_8X8][ADST_ADST] = ff_vp9_iadst_iadst_8x8_add_ssse3;
             dsp->itxfm_add[TX_16X16][DCT_DCT]   = ff_vp9_idct_idct_16x16_add_ssse3;
             dsp->itxfm_add[TX_16X16][ADST_DCT]  = ff_vp9_idct_iadst_16x16_add_ssse3;
             dsp->itxfm_add[TX_16X16][DCT_ADST]  = ff_vp9_iadst_idct_16x16_add_ssse3;
@@ -265,6 +268,9 @@ av_cold void ff_vp9dsp_init_x86(VP9DSPContext *dsp)
     if (EXTERNAL_AVX(cpu_flags)) {
         if (ARCH_X86_64) {
             dsp->itxfm_add[TX_8X8][DCT_DCT] = ff_vp9_idct_idct_8x8_add_avx;
+            dsp->itxfm_add[TX_8X8][ADST_DCT]  = ff_vp9_idct_iadst_8x8_add_avx;
+            dsp->itxfm_add[TX_8X8][DCT_ADST]  = ff_vp9_iadst_idct_8x8_add_avx;
+            dsp->itxfm_add[TX_8X8][ADST_ADST] = ff_vp9_iadst_iadst_8x8_add_avx;
             dsp->itxfm_add[TX_16X16][DCT_DCT] = ff_vp9_idct_idct_16x16_add_avx;
             dsp->itxfm_add[TX_16X16][ADST_DCT]  = ff_vp9_idct_iadst_16x16_add_avx;
             dsp->itxfm_add[TX_16X16][DCT_ADST]  = ff_vp9_iadst_idct_16x16_add_avx;
diff --git a/libavcodec/x86/vp9itxfm.asm b/libavcodec/x86/vp9itxfm.asm
index dfc877e..f814a61 100644
--- a/libavcodec/x86/vp9itxfm.asm
+++ b/libavcodec/x86/vp9itxfm.asm
@@ -95,6 +95,30 @@ SECTION .text
 %endif
 %endmacro
 
+%macro VP9_UNPACK_MULSUB_2D_4X 6 ; dst1 [src1], dst2 [src2], dst3, dst4, mul1, mul2
+    punpckhwd          m%4, m%2, m%1
+    punpcklwd          m%2, m%1
+    pmaddwd            m%3, m%4, [pw_m%5_%6]
+    pmaddwd            m%4, [pw_%6_%5]
+    pmaddwd            m%1, m%2, [pw_m%5_%6]
+    pmaddwd            m%2, [pw_%6_%5]
+%endmacro
+
+%macro VP9_RND_SH_SUMSUB_BA 6 ; dst1 [src1], dst2 [src2], src3, src4, tmp, round
+    SUMSUB_BA            d, %1, %2, %5
+    SUMSUB_BA            d, %3, %4, %5
+    paddd              m%1, %6
+    paddd              m%2, %6
+    paddd              m%3, %6
+    paddd              m%4, %6
+    psrad              m%1, 14
+    psrad              m%2, 14
+    psrad              m%3, 14
+    psrad              m%4, 14
+    packssdw           m%1, m%3
+    packssdw           m%2, m%4
+%endmacro
+
 %macro VP9_STORE_2X 5-6 dstq ; reg1, reg2, tmp1, tmp2, zero, dst
     movh               m%3, [%6]
     movh               m%4, [%6+strideq]
@@ -108,6 +132,18 @@ SECTION .text
     movh      [%6+strideq], m%4
 %endmacro
 
+%macro ZERO_BLOCK 4 ; mem, stride, nnzcpl, zero_reg
+%assign %%y 0
+%rep %3
+%assign %%x 0
+%rep %3*2/mmsize
+    mova      [%1+%%y+%%x], %4
+%assign %%x (%%x+mmsize)
+%endrep
+%assign %%y (%%y+%2)
+%endrep
+%endmacro
+
 ;-------------------------------------------------------------------------------------------
 ; void vp9_idct_idct_4x4_add_<opt>(uint8_t *dst, ptrdiff_t stride, int16_t *block, int eob);
 ;-------------------------------------------------------------------------------------------
@@ -372,15 +408,9 @@ cglobal vp9_idct_idct_8x8_add, 4,4,13, dst, stride, block, eob
     VP9_IDCT8_1D
     TRANSPOSE8x8W  0, 1, 2, 3, 8, 9, 10, 11, 4
     VP9_IDCT8_1D
+
     pxor                m4, m4  ; used for the block reset, and VP9_STORE_2X
-    mova      [blockq+  0], m4
-    mova      [blockq+ 16], m4
-    mova      [blockq+ 32], m4
-    mova      [blockq+ 48], m4
-    mova      [blockq+ 64], m4
-    mova      [blockq+ 80], m4
-    mova      [blockq+ 96], m4
-    mova      [blockq+112], m4
+    ZERO_BLOCK      blockq, 16, 8, m4
     VP9_IDCT8_WRITEOUT
     RET
 %endmacro
@@ -389,6 +419,81 @@ VP9_IDCT_IDCT_8x8_ADD_XMM ssse3
 VP9_IDCT_IDCT_8x8_ADD_XMM avx
 
 ;---------------------------------------------------------------------------------------------
+; void vp9_iadst_iadst_8x8_add_<opt>(uint8_t *dst, ptrdiff_t stride, int16_t *block, int eob);
+;---------------------------------------------------------------------------------------------
+
+%macro VP9_IADST8_1D 0 ; input/output=m0/1/2/3/8/9/10/11
+    VP9_UNPACK_MULSUB_2D_4X 11,  0,  4,  5, 16305,  1606    ; m11/4=t1[d], m0/5=t0[d]
+    VP9_UNPACK_MULSUB_2D_4X  3,  8,  6, 13, 10394, 12665    ; m3/6=t5[d], m8/13=t4[d]
+    VP9_RND_SH_SUMSUB_BA     8,  0, 13,  5, 14, m7          ; m8=t0[w], m0=t4[w]
+    VP9_RND_SH_SUMSUB_BA     3, 11,  6,  4, 14, m7          ; m3=t1[w], m11=t5[w]
+
+    VP9_UNPACK_MULSUB_2D_4X  9,  2,  4,  5, 14449,  7723    ; m9/4=t3[d], m2/5=t2[d]
+    VP9_UNPACK_MULSUB_2D_4X  1, 10,  6, 13,  4756, 15679    ; m1/6=t7[d], m10/13=t6[d]
+    VP9_RND_SH_SUMSUB_BA    10,  2, 13,  5, 14, m7          ; m10=t2[w], m2=t6[w]
+    VP9_RND_SH_SUMSUB_BA     1,  9,  6,  4, 14, m7          ; m1=t3[w], m9=t7[w]
+
+    ; m8=t0, m3=t1, m10=t2, m1=t3, m0=t4, m11=t5, m2=t6, m9=t7
+
+    VP9_UNPACK_MULSUB_2D_4X  0, 11,  4,  5, 15137,  6270    ; m0/4=t5[d], m11/5=t4[d]
+    VP9_UNPACK_MULSUB_2D_4X  9,  2,  6, 13,  6270, 15137    ; m9/6=t6[d], m2/13=t7[d]
+    VP9_RND_SH_SUMSUB_BA     9, 11,  6,  5, 14, m7
+    psignw                  m9, [pw_m1]                     ; m9=out1[w], m11=t6[w]
+    VP9_RND_SH_SUMSUB_BA     2,  0, 13,  4, 14, m7          ; m2=out6[w], m0=t7[w]
+
+    SUMSUB_BA                w, 10,  8, 14                  ; m10=out0[w], m8=t2[w]
+    SUMSUB_BA                w,  1,  3, 14
+    psignw                  m1, [pw_m1]                     ; m1=out7[w], m3=t3[w]
+
+    ; m10=out0, m9=out1, m8=t2, m3=t3, m11=t6, m0=t7, m2=out6, m1=out7
+
+    SUMSUB_BA                w,  3,  8,  4
+    SUMSUB_BA                w,  0, 11,  5
+    pmulhrsw                m3, m12
+    pmulhrsw               m11, m12
+    pmulhrsw                m8, m12                         ; out4
+    pmulhrsw                m0, m12                         ; out2
+    psignw                  m3, [pw_m1]                     ; out3
+    psignw                 m11, [pw_m1]                     ; out5
+
+    ; m10=out0, m9=out1, m0=out2, m3=out3, m8=out4, m11=out5, m2=out6, m1=out7
+
+    SWAP                     0, 10, 2
+    SWAP                    11,  1, 9
+%endmacro
+
+%macro IADST8_FN 5
+INIT_XMM %5
+cglobal vp9_%1_%3_8x8_add, 3, 3, 15, dst, stride, block, eob
+    mova                m0, [blockq+  0]    ; IN(0)
+    mova                m1, [blockq+ 16]    ; IN(1)
+    mova                m2, [blockq+ 32]    ; IN(2)
+    mova                m3, [blockq+ 48]    ; IN(3)
+    mova                m8, [blockq+ 64]    ; IN(4)
+    mova                m9, [blockq+ 80]    ; IN(5)
+    mova               m10, [blockq+ 96]    ; IN(6)
+    mova               m11, [blockq+112]    ; IN(7)
+
+    mova               m12, [pw_11585x2]    ; often used
+    mova                m7, [pd_8192]       ; rounding
+    VP9_%2_1D
+    TRANSPOSE8x8W  0, 1, 2, 3, 8, 9, 10, 11, 4
+    VP9_%4_1D
+
+    pxor                m4, m4  ; used for the block reset, and VP9_STORE_2X
+    ZERO_BLOCK      blockq, 16, 8, m4
+    VP9_IDCT8_WRITEOUT
+    RET
+%endmacro
+
+IADST8_FN idct,  IDCT8,  iadst, IADST8, ssse3
+IADST8_FN idct,  IDCT8,  iadst, IADST8, avx
+IADST8_FN iadst, IADST8, idct,  IDCT8,  ssse3
+IADST8_FN iadst, IADST8, idct,  IDCT8,  avx
+IADST8_FN iadst, IADST8, iadst, IADST8, ssse3
+IADST8_FN iadst, IADST8, iadst, IADST8, avx
+
+;---------------------------------------------------------------------------------------------
 ; void vp9_idct_idct_16x16_add_<opt>(uint8_t *dst, ptrdiff_t stride, int16_t *block, int eob);
 ;---------------------------------------------------------------------------------------------
 
@@ -637,18 +742,6 @@ VP9_IDCT_IDCT_8x8_ADD_XMM avx
 %endif ; %2 == 1/2
 %endmacro
 
-%macro ZERO_BLOCK 4 ; mem, stride, nnzcpl, zero_reg
-%assign %%y 0
-%rep %3
-%assign %%x 0
-%rep %3*2/mmsize
-    mova      [%1+%%y+%%x], %4
-%assign %%x (%%x+mmsize)
-%endrep
-%assign %%y (%%y+%2)
-%endrep
-%endmacro
-
 %macro VP9_STORE_2XFULL 6-7 strideq; dc, tmp1, tmp2, tmp3, tmp4, zero, stride
     mova               m%3, [dstq]
     mova               m%5, [dstq+%7]
@@ -745,30 +838,6 @@ VP9_IDCT_IDCT_16x16_ADD_XMM avx
 ; void vp9_iadst_iadst_16x16_add_<opt>(uint8_t *dst, ptrdiff_t stride, int16_t *block, int eob);
 ;---------------------------------------------------------------------------------------------
 
-%macro VP9_UNPACK_MULSUB_2D_4X 6 ; dst1 [src1], dst2 [src2], dst3, dst4, mul1, mul2
-    punpckhwd          m%4, m%2, m%1
-    punpcklwd          m%2, m%1
-    pmaddwd            m%3, m%4, [pw_m%5_%6]
-    pmaddwd            m%4, [pw_%6_%5]
-    pmaddwd            m%1, m%2, [pw_m%5_%6]
-    pmaddwd            m%2, [pw_%6_%5]
-%endmacro
-
-%macro VP9_RND_SH_SUMSUB_BA 6 ; dst1 [src1], dst2 [src2], src3, src4, tmp, round
-    SUMSUB_BA            d, %1, %2, %5
-    SUMSUB_BA            d, %3, %4, %5
-    paddd              m%1, %6
-    paddd              m%2, %6
-    paddd              m%3, %6
-    paddd              m%4, %6
-    psrad              m%1, 14
-    psrad              m%2, 14
-    psrad              m%3, 14
-    psrad              m%4, 14
-    packssdw           m%1, m%3
-    packssdw           m%2, m%4
-%endmacro
-
 %macro VP9_IADST16_1D 2 ; src, pass
 %assign %%str 16*%2
     mova                m0, [%1+ 0*32]  ; in0
-- 
1.8.4



More information about the ffmpeg-devel mailing list