[FFmpeg-cvslog] ARM: h264dsp_neon cosmetics

Mans Rullgard git at videolan.org
Sat Dec 3 03:13:09 CET 2011


ffmpeg | branch: master | Mans Rullgard <mans at mansr.com> | Fri Dec  2 16:45:00 2011 +0000| [59807fee6dfca909d966da3d2203d6c3cbf356c8] | committer: Mans Rullgard

ARM: h264dsp_neon cosmetics

- Replace 'ip' with 'r12'.
- Use correct size designators for vld1/vst1.
- Whitespace fixes.

Signed-off-by: Mans Rullgard <mans at mansr.com>

> http://git.videolan.org/gitweb.cgi/ffmpeg.git/?a=commit;h=59807fee6dfca909d966da3d2203d6c3cbf356c8
---

 libavcodec/arm/h264dsp_neon.S |  626 ++++++++++++++++++++--------------------
 1 files changed, 313 insertions(+), 313 deletions(-)

diff --git a/libavcodec/arm/h264dsp_neon.S b/libavcodec/arm/h264dsp_neon.S
index 1e97908..e23f7b1 100644
--- a/libavcodec/arm/h264dsp_neon.S
+++ b/libavcodec/arm/h264dsp_neon.S
@@ -22,13 +22,13 @@
 #include "neon.S"
 
 /* chroma_mc8(uint8_t *dst, uint8_t *src, int stride, int h, int x, int y) */
-        .macro  h264_chroma_mc8 type
+.macro  h264_chroma_mc8 type
 function ff_\type\()_h264_chroma_mc8_neon, export=1
         push            {r4-r7, lr}
         ldrd            r4,  [sp, #20]
-.ifc \type,avg
+  .ifc \type,avg
         mov             lr,  r0
-.endif
+  .endif
         pld             [r1]
         pld             [r1, r2]
 
@@ -36,7 +36,7 @@ A       muls            r7,  r4,  r5
 T       mul             r7,  r4,  r5
 T       cmp             r7,  #0
         rsb             r6,  r7,  r5,  lsl #3
-        rsb             ip,  r7,  r4,  lsl #3
+        rsb             r12, r7,  r4,  lsl #3
         sub             r4,  r7,  r4,  lsl #3
         sub             r4,  r4,  r5,  lsl #3
         add             r4,  r4,  #64
@@ -47,10 +47,10 @@ T       cmp             r7,  #0
 
         vdup.8          d0,  r4
         lsl             r4,  r2,  #1
-        vdup.8          d1,  ip
-        vld1.64         {d4, d5}, [r1], r4
+        vdup.8          d1,  r12
+        vld1.8          {d4, d5}, [r1], r4
         vdup.8          d2,  r6
-        vld1.64         {d6, d7}, [r5], r4
+        vld1.8          {d6, d7}, [r5], r4
         vdup.8          d3,  r7
 
         vext.8          d5,  d4,  d5,  #1
@@ -59,7 +59,7 @@ T       cmp             r7,  #0
 1:      pld             [r5]
         vmull.u8        q8,  d4,  d0
         vmlal.u8        q8,  d5,  d1
-        vld1.64         {d4, d5}, [r1], r4
+        vld1.8          {d4, d5}, [r1], r4
         vmlal.u8        q8,  d6,  d2
         vext.8          d5,  d4,  d5,  #1
         vmlal.u8        q8,  d7,  d3
@@ -69,57 +69,57 @@ T       cmp             r7,  #0
         vmlal.u8        q9,  d4,  d2
         vmlal.u8        q9,  d5,  d3
         vrshrn.u16      d16, q8,  #6
-        vld1.64         {d6, d7}, [r5], r4
+        vld1.8          {d6, d7}, [r5], r4
         pld             [r1]
         vrshrn.u16      d17, q9,  #6
-.ifc \type,avg
-        vld1.64         {d20}, [lr,:64], r2
-        vld1.64         {d21}, [lr,:64], r2
+  .ifc \type,avg
+        vld1.8          {d20}, [lr,:64], r2
+        vld1.8          {d21}, [lr,:64], r2
         vrhadd.u8       q8,  q8,  q10
-.endif
+  .endif
         vext.8          d7,  d6,  d7,  #1
-        vst1.64         {d16}, [r0,:64], r2
-        vst1.64         {d17}, [r0,:64], r2
+        vst1.8          {d16}, [r0,:64], r2
+        vst1.8          {d17}, [r0,:64], r2
         bgt             1b
 
         pop             {r4-r7, pc}
 
 2:      tst             r6,  r6
-        add             ip,  ip,  r6
+        add             r12, r12, r6
         vdup.8          d0,  r4
-        vdup.8          d1,  ip
+        vdup.8          d1,  r12
 
         beq             4f
 
         add             r5,  r1,  r2
         lsl             r4,  r2,  #1
-        vld1.64         {d4}, [r1], r4
-        vld1.64         {d6}, [r5], r4
+        vld1.8          {d4}, [r1], r4
+        vld1.8          {d6}, [r5], r4
 
 3:      pld             [r5]
         vmull.u8        q8,  d4,  d0
         vmlal.u8        q8,  d6,  d1
-        vld1.64         {d4}, [r1], r4
+        vld1.8          {d4}, [r1], r4
         vmull.u8        q9,  d6,  d0
         vmlal.u8        q9,  d4,  d1
-        vld1.64         {d6}, [r5], r4
+        vld1.8          {d6}, [r5], r4
         vrshrn.u16      d16, q8,  #6
         vrshrn.u16      d17, q9,  #6
-.ifc \type,avg
-        vld1.64         {d20}, [lr,:64], r2
-        vld1.64         {d21}, [lr,:64], r2
+  .ifc \type,avg
+        vld1.8          {d20}, [lr,:64], r2
+        vld1.8          {d21}, [lr,:64], r2
         vrhadd.u8       q8,  q8,  q10
-.endif
+  .endif
         subs            r3,  r3,  #2
         pld             [r1]
-        vst1.64         {d16}, [r0,:64], r2
-        vst1.64         {d17}, [r0,:64], r2
+        vst1.8          {d16}, [r0,:64], r2
+        vst1.8          {d17}, [r0,:64], r2
         bgt             3b
 
         pop             {r4-r7, pc}
 
-4:      vld1.64         {d4, d5}, [r1], r2
-        vld1.64         {d6, d7}, [r1], r2
+4:      vld1.8          {d4, d5}, [r1], r2
+        vld1.8          {d6, d7}, [r1], r2
         vext.8          d5,  d4,  d5,  #1
         vext.8          d7,  d6,  d7,  #1
 
@@ -127,36 +127,36 @@ T       cmp             r7,  #0
         subs            r3,  r3,  #2
         vmull.u8        q8,  d4,  d0
         vmlal.u8        q8,  d5,  d1
-        vld1.64         {d4, d5}, [r1], r2
+        vld1.8          {d4, d5}, [r1], r2
         vmull.u8        q9,  d6,  d0
         vmlal.u8        q9,  d7,  d1
         pld             [r1]
         vext.8          d5,  d4,  d5,  #1
         vrshrn.u16      d16, q8,  #6
         vrshrn.u16      d17, q9,  #6
-.ifc \type,avg
-        vld1.64         {d20}, [lr,:64], r2
-        vld1.64         {d21}, [lr,:64], r2
+  .ifc \type,avg
+        vld1.8          {d20}, [lr,:64], r2
+        vld1.8          {d21}, [lr,:64], r2
         vrhadd.u8       q8,  q8,  q10
-.endif
-        vld1.64         {d6, d7}, [r1], r2
+  .endif
+        vld1.8          {d6, d7}, [r1], r2
         vext.8          d7,  d6,  d7,  #1
-        vst1.64         {d16}, [r0,:64], r2
-        vst1.64         {d17}, [r0,:64], r2
+        vst1.8          {d16}, [r0,:64], r2
+        vst1.8          {d17}, [r0,:64], r2
         bgt             5b
 
         pop             {r4-r7, pc}
 endfunc
-        .endm
+.endm
 
 /* chroma_mc4(uint8_t *dst, uint8_t *src, int stride, int h, int x, int y) */
-        .macro  h264_chroma_mc4 type
+.macro  h264_chroma_mc4 type
 function ff_\type\()_h264_chroma_mc4_neon, export=1
         push            {r4-r7, lr}
         ldrd            r4,  [sp, #20]
-.ifc \type,avg
+  .ifc \type,avg
         mov             lr,  r0
-.endif
+  .endif
         pld             [r1]
         pld             [r1, r2]
 
@@ -164,7 +164,7 @@ A       muls            r7,  r4,  r5
 T       mul             r7,  r4,  r5
 T       cmp             r7,  #0
         rsb             r6,  r7,  r5,  lsl #3
-        rsb             ip,  r7,  r4,  lsl #3
+        rsb             r12, r7,  r4,  lsl #3
         sub             r4,  r7,  r4,  lsl #3
         sub             r4,  r4,  r5,  lsl #3
         add             r4,  r4,  #64
@@ -175,10 +175,10 @@ T       cmp             r7,  #0
 
         vdup.8          d0,  r4
         lsl             r4,  r2,  #1
-        vdup.8          d1,  ip
-        vld1.64         {d4},     [r1], r4
+        vdup.8          d1,  r12
+        vld1.8          {d4},     [r1], r4
         vdup.8          d2,  r6
-        vld1.64         {d6},     [r5], r4
+        vld1.8          {d6},     [r5], r4
         vdup.8          d3,  r7
 
         vext.8          d5,  d4,  d5,  #1
@@ -192,22 +192,22 @@ T       cmp             r7,  #0
 1:      pld             [r5]
         vmull.u8        q8,  d4,  d0
         vmlal.u8        q8,  d6,  d2
-        vld1.64         {d4},     [r1], r4
+        vld1.8          {d4},     [r1], r4
         vext.8          d5,  d4,  d5,  #1
         vtrn.32         d4,  d5
         vmull.u8        q9,  d6,  d0
         vmlal.u8        q9,  d4,  d2
-        vld1.64         {d6},     [r5], r4
+        vld1.8          {d6},     [r5], r4
         vadd.i16        d16, d16, d17
         vadd.i16        d17, d18, d19
         vrshrn.u16      d16, q8,  #6
         subs            r3,  r3,  #2
         pld             [r1]
-.ifc \type,avg
+  .ifc \type,avg
         vld1.32         {d20[0]}, [lr,:32], r2
         vld1.32         {d20[1]}, [lr,:32], r2
         vrhadd.u8       d16, d16, d20
-.endif
+  .endif
         vext.8          d7,  d6,  d7,  #1
         vtrn.32         d6,  d7
         vst1.32         {d16[0]}, [r0,:32], r2
@@ -217,9 +217,9 @@ T       cmp             r7,  #0
         pop             {r4-r7, pc}
 
 2:      tst             r6,  r6
-        add             ip,  ip,  r6
+        add             r12, r12, r6
         vdup.8          d0,  r4
-        vdup.8          d1,  ip
+        vdup.8          d1,  r12
         vtrn.32         d0,  d1
 
         beq             4f
@@ -238,11 +238,11 @@ T       cmp             r7,  #0
         vadd.i16        d16, d16, d17
         vadd.i16        d17, d18, d19
         vrshrn.u16      d16, q8,  #6
-.ifc \type,avg
+  .ifc \type,avg
         vld1.32         {d20[0]}, [lr,:32], r2
         vld1.32         {d20[1]}, [lr,:32], r2
         vrhadd.u8       d16, d16, d20
-.endif
+  .endif
         subs            r3,  r3,  #2
         pld             [r1]
         vst1.32         {d16[0]}, [r0,:32], r2
@@ -251,8 +251,8 @@ T       cmp             r7,  #0
 
         pop             {r4-r7, pc}
 
-4:      vld1.64         {d4},     [r1], r2
-        vld1.64         {d6},     [r1], r2
+4:      vld1.8          {d4},     [r1], r2
+        vld1.8          {d6},     [r1], r2
         vext.8          d5,  d4,  d5,  #1
         vext.8          d7,  d6,  d7,  #1
         vtrn.32         d4,  d5
@@ -261,19 +261,19 @@ T       cmp             r7,  #0
 5:      vmull.u8        q8,  d4,  d0
         vmull.u8        q9,  d6,  d0
         subs            r3,  r3,  #2
-        vld1.64         {d4},     [r1], r2
+        vld1.8          {d4},     [r1], r2
         vext.8          d5,  d4,  d5,  #1
         vtrn.32         d4,  d5
         vadd.i16        d16, d16, d17
         vadd.i16        d17, d18, d19
         pld             [r1]
         vrshrn.u16      d16, q8,  #6
-.ifc \type,avg
+  .ifc \type,avg
         vld1.32         {d20[0]}, [lr,:32], r2
         vld1.32         {d20[1]}, [lr,:32], r2
         vrhadd.u8       d16, d16, d20
-.endif
-        vld1.64         {d6},     [r1], r2
+  .endif
+        vld1.8          {d6},     [r1], r2
         vext.8          d7,  d6,  d7,  #1
         vtrn.32         d6,  d7
         pld             [r1]
@@ -283,9 +283,9 @@ T       cmp             r7,  #0
 
         pop             {r4-r7, pc}
 endfunc
-        .endm
+.endm
 
-        .macro  h264_chroma_mc2 type
+.macro  h264_chroma_mc2 type
 function ff_\type\()_h264_chroma_mc2_neon, export=1
         push            {r4-r6, lr}
         ldr             r4,  [sp, #16]
@@ -315,29 +315,29 @@ function ff_\type\()_h264_chroma_mc2_neon, export=1
         vtrn.16         q2,  q3
         vmull.u8        q8,  d4,  d0
         vmlal.u8        q8,  d5,  d1
-.ifc \type,avg
+  .ifc \type,avg
         vld1.16         {d18[0]}, [r0,:16], r2
         vld1.16         {d18[1]}, [r0,:16]
         sub             r0,  r0,  r2
-.endif
+  .endif
         vtrn.32         d16, d17
         vadd.i16        d16, d16, d17
         vrshrn.u16      d16, q8,  #6
-.ifc \type,avg
+  .ifc \type,avg
         vrhadd.u8       d16, d16, d18
-.endif
+  .endif
         vst1.16         {d16[0]}, [r0,:16], r2
         vst1.16         {d16[1]}, [r0,:16], r2
         subs            r3,  r3,  #2
         bgt             1b
         pop             {r4-r6, pc}
 2:
-.ifc \type,put
+  .ifc \type,put
         ldrh_post       r5,  r1,  r2
         strh_post       r5,  r0,  r2
         ldrh_post       r6,  r1,  r2
         strh_post       r6,  r0,  r2
-.else
+  .else
         vld1.16         {d16[0]}, [r1], r2
         vld1.16         {d16[1]}, [r1], r2
         vld1.16         {d18[0]}, [r0,:16], r2
@@ -346,7 +346,7 @@ function ff_\type\()_h264_chroma_mc2_neon, export=1
         vrhadd.u8       d16, d16, d18
         vst1.16         {d16[0]}, [r0,:16], r2
         vst1.16         {d16[1]}, [r0,:16], r2
-.endif
+  .endif
         subs            r3,  r3,  #2
         bgt             2b
         pop             {r4-r6, pc}
@@ -362,22 +362,22 @@ endfunc
 
         /* H.264 loop filter */
 
-        .macro h264_loop_filter_start
-        ldr             ip,  [sp]
+.macro  h264_loop_filter_start
+        ldr             r12, [sp]
         tst             r2,  r2
-        ldr             ip,  [ip]
+        ldr             r12, [r12]
         it              ne
         tstne           r3,  r3
-        vmov.32         d24[0], ip
-        and             ip,  ip,  ip, lsl #16
+        vmov.32         d24[0], r12
+        and             r12, r12, r12, lsl #16
         it              eq
         bxeq            lr
-        ands            ip,  ip,  ip, lsl #8
+        ands            r12, r12, r12, lsl #8
         it              lt
         bxlt            lr
-        .endm
+.endm
 
-        .macro h264_loop_filter_luma
+.macro  h264_loop_filter_luma
         vdup.8          q11, r2         @ alpha
         vmovl.u8        q12, d24
         vabd.u8         q6,  q8,  q0    @ abs(p0 - q0)
@@ -443,29 +443,29 @@ endfunc
         vqmovun.s16     d17, q6
         vqmovun.s16     d0,  q11
         vqmovun.s16     d1,  q12
-        .endm
+.endm
 
 function ff_h264_v_loop_filter_luma_neon, export=1
         h264_loop_filter_start
 
-        vld1.64         {d0, d1},  [r0,:128], r1
-        vld1.64         {d2, d3},  [r0,:128], r1
-        vld1.64         {d4, d5},  [r0,:128], r1
+        vld1.8          {d0, d1},  [r0,:128], r1
+        vld1.8          {d2, d3},  [r0,:128], r1
+        vld1.8          {d4, d5},  [r0,:128], r1
         sub             r0,  r0,  r1, lsl #2
         sub             r0,  r0,  r1, lsl #1
-        vld1.64         {d20,d21}, [r0,:128], r1
-        vld1.64         {d18,d19}, [r0,:128], r1
-        vld1.64         {d16,d17}, [r0,:128], r1
+        vld1.8          {d20,d21}, [r0,:128], r1
+        vld1.8          {d18,d19}, [r0,:128], r1
+        vld1.8          {d16,d17}, [r0,:128], r1
 
         vpush           {d8-d15}
 
         h264_loop_filter_luma
 
         sub             r0,  r0,  r1, lsl #1
-        vst1.64         {d8, d9},  [r0,:128], r1
-        vst1.64         {d16,d17}, [r0,:128], r1
-        vst1.64         {d0, d1},  [r0,:128], r1
-        vst1.64         {d10,d11}, [r0,:128]
+        vst1.8          {d8, d9},  [r0,:128], r1
+        vst1.8          {d16,d17}, [r0,:128], r1
+        vst1.8          {d0, d1},  [r0,:128], r1
+        vst1.8          {d10,d11}, [r0,:128]
 
         vpop            {d8-d15}
         bx              lr
@@ -475,22 +475,22 @@ function ff_h264_h_loop_filter_luma_neon, export=1
         h264_loop_filter_start
 
         sub             r0,  r0,  #4
-        vld1.64         {d6},  [r0], r1
-        vld1.64         {d20}, [r0], r1
-        vld1.64         {d18}, [r0], r1
-        vld1.64         {d16}, [r0], r1
-        vld1.64         {d0},  [r0], r1
-        vld1.64         {d2},  [r0], r1
-        vld1.64         {d4},  [r0], r1
-        vld1.64         {d26}, [r0], r1
-        vld1.64         {d7},  [r0], r1
-        vld1.64         {d21}, [r0], r1
-        vld1.64         {d19}, [r0], r1
-        vld1.64         {d17}, [r0], r1
-        vld1.64         {d1},  [r0], r1
-        vld1.64         {d3},  [r0], r1
-        vld1.64         {d5},  [r0], r1
-        vld1.64         {d27}, [r0], r1
+        vld1.8          {d6},  [r0], r1
+        vld1.8          {d20}, [r0], r1
+        vld1.8          {d18}, [r0], r1
+        vld1.8          {d16}, [r0], r1
+        vld1.8          {d0},  [r0], r1
+        vld1.8          {d2},  [r0], r1
+        vld1.8          {d4},  [r0], r1
+        vld1.8          {d26}, [r0], r1
+        vld1.8          {d7},  [r0], r1
+        vld1.8          {d21}, [r0], r1
+        vld1.8          {d19}, [r0], r1
+        vld1.8          {d17}, [r0], r1
+        vld1.8          {d1},  [r0], r1
+        vld1.8          {d3},  [r0], r1
+        vld1.8          {d5},  [r0], r1
+        vld1.8          {d27}, [r0], r1
 
         transpose_8x8   q3, q10, q9, q8, q0, q1, q2, q13
 
@@ -523,7 +523,7 @@ function ff_h264_h_loop_filter_luma_neon, export=1
         bx              lr
 endfunc
 
-        .macro h264_loop_filter_chroma
+.macro  h264_loop_filter_chroma
         vdup.8          d22, r2         @ alpha
         vmovl.u8        q12, d24
         vabd.u8         d26, d16, d0    @ abs(p0 - q0)
@@ -552,22 +552,22 @@ endfunc
         vsubw.s8        q11, q11, d4
         vqmovun.s16     d16, q14
         vqmovun.s16     d0,  q11
-        .endm
+.endm
 
 function ff_h264_v_loop_filter_chroma_neon, export=1
         h264_loop_filter_start
 
         sub             r0,  r0,  r1, lsl #1
-        vld1.64         {d18}, [r0,:64], r1
-        vld1.64         {d16}, [r0,:64], r1
-        vld1.64         {d0},  [r0,:64], r1
-        vld1.64         {d2},  [r0,:64]
+        vld1.8          {d18}, [r0,:64], r1
+        vld1.8          {d16}, [r0,:64], r1
+        vld1.8          {d0},  [r0,:64], r1
+        vld1.8          {d2},  [r0,:64]
 
         h264_loop_filter_chroma
 
         sub             r0,  r0,  r1, lsl #1
-        vst1.64         {d16}, [r0,:64], r1
-        vst1.64         {d0},  [r0,:64], r1
+        vst1.8          {d16}, [r0,:64], r1
+        vst1.8          {d0},  [r0,:64], r1
 
         bx              lr
 endfunc
@@ -612,20 +612,20 @@ endfunc
 
         /* H.264 qpel MC */
 
-        .macro  lowpass_const r
+.macro  lowpass_const   r
         movw            \r,  #5
         movt            \r,  #20
         vmov.32         d6[0], \r
-        .endm
+.endm
 
-        .macro  lowpass_8 r0, r1, r2, r3, d0, d1, narrow=1
-.if \narrow
+.macro  lowpass_8       r0,  r1,  r2,  r3,  d0,  d1,  narrow=1
+  .if \narrow
         t0 .req q0
         t1 .req q8
-.else
+  .else
         t0 .req \d0
         t1 .req \d1
-.endif
+  .endif
         vext.8          d2,  \r0, \r1, #2
         vext.8          d3,  \r0, \r1, #3
         vaddl.u8        q1,  d2,  d3
@@ -646,20 +646,20 @@ endfunc
         vaddl.u8        t1,  \r2, d31
         vmla.i16        t1,  q9,  d6[1]
         vmls.i16        t1,  q10, d6[0]
-.if \narrow
+  .if \narrow
         vqrshrun.s16    \d0, t0,  #5
         vqrshrun.s16    \d1, t1,  #5
-.endif
+  .endif
         .unreq  t0
         .unreq  t1
-        .endm
+.endm
 
-        .macro  lowpass_8_1 r0, r1, d0, narrow=1
-.if \narrow
+.macro  lowpass_8_1     r0,  r1,  d0,  narrow=1
+  .if \narrow
         t0 .req q0
-.else
+  .else
         t0 .req \d0
-.endif
+  .endif
         vext.8          d2,  \r0, \r1, #2
         vext.8          d3,  \r0, \r1, #3
         vaddl.u8        q1,  d2,  d3
@@ -670,13 +670,13 @@ endfunc
         vaddl.u8        t0,  \r0, d30
         vmla.i16        t0,  q1,  d6[1]
         vmls.i16        t0,  q2,  d6[0]
-.if \narrow
+  .if \narrow
         vqrshrun.s16    \d0, t0,  #5
-.endif
+  .endif
         .unreq  t0
-        .endm
+.endm
 
-        .macro  lowpass_8.16 r0, r1, l0, h0, l1, h1, d
+.macro  lowpass_8.16    r0,  r1,  l0,  h0,  l1,  h1,  d
         vext.16         q1,  \r0, \r1, #2
         vext.16         q0,  \r0, \r1, #3
         vaddl.s16       q9,  d2,  d0
@@ -711,59 +711,59 @@ endfunc
         vrshrn.s32      d19, q1,  #10
 
         vqmovun.s16     \d,  q9
-        .endm
+.endm
 
 function put_h264_qpel16_h_lowpass_neon_packed
         mov             r4,  lr
-        mov             ip,  #16
+        mov             r12, #16
         mov             r3,  #8
         bl              put_h264_qpel8_h_lowpass_neon
         sub             r1,  r1,  r2, lsl #4
         add             r1,  r1,  #8
-        mov             ip,  #16
+        mov             r12, #16
         mov             lr,  r4
         b               put_h264_qpel8_h_lowpass_neon
 endfunc
 
-        .macro h264_qpel_h_lowpass type
+.macro  h264_qpel_h_lowpass type
 function \type\()_h264_qpel16_h_lowpass_neon
         push            {lr}
-        mov             ip,  #16
+        mov             r12, #16
         bl              \type\()_h264_qpel8_h_lowpass_neon
         sub             r0,  r0,  r3, lsl #4
         sub             r1,  r1,  r2, lsl #4
         add             r0,  r0,  #8
         add             r1,  r1,  #8
-        mov             ip,  #16
+        mov             r12, #16
         pop             {lr}
 endfunc
 
 function \type\()_h264_qpel8_h_lowpass_neon
-1:      vld1.64         {d0, d1},  [r1], r2
-        vld1.64         {d16,d17}, [r1], r2
-        subs            ip,  ip,  #2
+1:      vld1.8          {d0, d1},  [r1], r2
+        vld1.8          {d16,d17}, [r1], r2
+        subs            r12, r12, #2
         lowpass_8       d0,  d1,  d16, d17, d0,  d16
-.ifc \type,avg
+  .ifc \type,avg
         vld1.8          {d2},     [r0,:64], r3
         vrhadd.u8       d0,  d0,  d2
         vld1.8          {d3},     [r0,:64]
         vrhadd.u8       d16, d16, d3
         sub             r0,  r0,  r3
-.endif
-        vst1.64         {d0},     [r0,:64], r3
-        vst1.64         {d16},    [r0,:64], r3
+  .endif
+        vst1.8          {d0},     [r0,:64], r3
+        vst1.8          {d16},    [r0,:64], r3
         bne             1b
         bx              lr
 endfunc
-        .endm
+.endm
 
         h264_qpel_h_lowpass put
         h264_qpel_h_lowpass avg
 
-        .macro h264_qpel_h_lowpass_l2 type
+.macro  h264_qpel_h_lowpass_l2 type
 function \type\()_h264_qpel16_h_lowpass_l2_neon
         push            {lr}
-        mov             ip,  #16
+        mov             r12, #16
         bl              \type\()_h264_qpel8_h_lowpass_l2_neon
         sub             r0,  r0,  r2, lsl #4
         sub             r1,  r1,  r2, lsl #4
@@ -771,31 +771,31 @@ function \type\()_h264_qpel16_h_lowpass_l2_neon
         add             r0,  r0,  #8
         add             r1,  r1,  #8
         add             r3,  r3,  #8
-        mov             ip,  #16
+        mov             r12, #16
         pop             {lr}
 endfunc
 
 function \type\()_h264_qpel8_h_lowpass_l2_neon
-1:      vld1.64         {d0, d1},  [r1], r2
-        vld1.64         {d16,d17}, [r1], r2
-        vld1.64         {d28},     [r3], r2
-        vld1.64         {d29},     [r3], r2
-        subs            ip,  ip,  #2
+1:      vld1.8          {d0, d1},  [r1], r2
+        vld1.8          {d16,d17}, [r1], r2
+        vld1.8          {d28},     [r3], r2
+        vld1.8          {d29},     [r3], r2
+        subs            r12, r12, #2
         lowpass_8       d0,  d1,  d16, d17, d0,  d1
         vrhadd.u8       q0,  q0,  q14
-.ifc \type,avg
+  .ifc \type,avg
         vld1.8          {d2},      [r0,:64], r2
         vrhadd.u8       d0,  d0,  d2
         vld1.8          {d3},      [r0,:64]
         vrhadd.u8       d1,  d1,  d3
         sub             r0,  r0,  r2
-.endif
-        vst1.64         {d0},      [r0,:64], r2
-        vst1.64         {d1},      [r0,:64], r2
+  .endif
+        vst1.8          {d0},      [r0,:64], r2
+        vst1.8          {d1},      [r0,:64], r2
         bne             1b
         bx              lr
 endfunc
-        .endm
+.endm
 
         h264_qpel_h_lowpass_l2 put
         h264_qpel_h_lowpass_l2 avg
@@ -815,7 +815,7 @@ function put_h264_qpel16_v_lowpass_neon_packed
         b               put_h264_qpel8_v_lowpass_neon
 endfunc
 
-        .macro h264_qpel_v_lowpass type
+.macro  h264_qpel_v_lowpass type
 function \type\()_h264_qpel16_v_lowpass_neon
         mov             r4,  lr
         bl              \type\()_h264_qpel8_v_lowpass_neon
@@ -832,19 +832,19 @@ function \type\()_h264_qpel16_v_lowpass_neon
 endfunc
 
 function \type\()_h264_qpel8_v_lowpass_neon
-        vld1.64         {d8},  [r1], r3
-        vld1.64         {d10}, [r1], r3
-        vld1.64         {d12}, [r1], r3
-        vld1.64         {d14}, [r1], r3
-        vld1.64         {d22}, [r1], r3
-        vld1.64         {d24}, [r1], r3
-        vld1.64         {d26}, [r1], r3
-        vld1.64         {d28}, [r1], r3
-        vld1.64         {d9},  [r1], r3
-        vld1.64         {d11}, [r1], r3
-        vld1.64         {d13}, [r1], r3
-        vld1.64         {d15}, [r1], r3
-        vld1.64         {d23}, [r1]
+        vld1.8          {d8},  [r1], r3
+        vld1.8          {d10}, [r1], r3
+        vld1.8          {d12}, [r1], r3
+        vld1.8          {d14}, [r1], r3
+        vld1.8          {d22}, [r1], r3
+        vld1.8          {d24}, [r1], r3
+        vld1.8          {d26}, [r1], r3
+        vld1.8          {d28}, [r1], r3
+        vld1.8          {d9},  [r1], r3
+        vld1.8          {d11}, [r1], r3
+        vld1.8          {d13}, [r1], r3
+        vld1.8          {d15}, [r1], r3
+        vld1.8          {d23}, [r1]
 
         transpose_8x8   q4,  q5,  q6,  q7,  q11, q12, q13, q14
         lowpass_8       d8,  d9,  d10, d11, d8,  d10
@@ -853,7 +853,7 @@ function \type\()_h264_qpel8_v_lowpass_neon
         lowpass_8       d26, d27, d28, d29, d26, d28
         transpose_8x8   d8,  d10, d12, d14, d22, d24, d26, d28
 
-.ifc \type,avg
+  .ifc \type,avg
         vld1.8          {d9},  [r0,:64], r2
         vrhadd.u8       d8,  d8,  d9
         vld1.8          {d11}, [r0,:64], r2
@@ -871,34 +871,34 @@ function \type\()_h264_qpel8_v_lowpass_neon
         vld1.8          {d29}, [r0,:64], r2
         vrhadd.u8       d28, d28, d29
         sub             r0,  r0,  r2,  lsl #3
-.endif
+  .endif
 
-        vst1.64         {d8},  [r0,:64], r2
-        vst1.64         {d10}, [r0,:64], r2
-        vst1.64         {d12}, [r0,:64], r2
-        vst1.64         {d14}, [r0,:64], r2
-        vst1.64         {d22}, [r0,:64], r2
-        vst1.64         {d24}, [r0,:64], r2
-        vst1.64         {d26}, [r0,:64], r2
-        vst1.64         {d28}, [r0,:64], r2
+        vst1.8          {d8},  [r0,:64], r2
+        vst1.8          {d10}, [r0,:64], r2
+        vst1.8          {d12}, [r0,:64], r2
+        vst1.8          {d14}, [r0,:64], r2
+        vst1.8          {d22}, [r0,:64], r2
+        vst1.8          {d24}, [r0,:64], r2
+        vst1.8          {d26}, [r0,:64], r2
+        vst1.8          {d28}, [r0,:64], r2
 
         bx              lr
 endfunc
-        .endm
+.endm
 
         h264_qpel_v_lowpass put
         h264_qpel_v_lowpass avg
 
-        .macro h264_qpel_v_lowpass_l2 type
+.macro  h264_qpel_v_lowpass_l2 type
 function \type\()_h264_qpel16_v_lowpass_l2_neon
         mov             r4,  lr
         bl              \type\()_h264_qpel8_v_lowpass_l2_neon
         sub             r1,  r1,  r3, lsl #2
         bl              \type\()_h264_qpel8_v_lowpass_l2_neon
         sub             r0,  r0,  r3, lsl #4
-        sub             ip,  ip,  r2, lsl #4
+        sub             r12, r12, r2, lsl #4
         add             r0,  r0,  #8
-        add             ip,  ip,  #8
+        add             r12, r12, #8
         sub             r1,  r1,  r3, lsl #4
         sub             r1,  r1,  r3, lsl #2
         add             r1,  r1,  #8
@@ -908,19 +908,19 @@ function \type\()_h264_qpel16_v_lowpass_l2_neon
 endfunc
 
 function \type\()_h264_qpel8_v_lowpass_l2_neon
-        vld1.64         {d8},  [r1], r3
-        vld1.64         {d10}, [r1], r3
-        vld1.64         {d12}, [r1], r3
-        vld1.64         {d14}, [r1], r3
-        vld1.64         {d22}, [r1], r3
-        vld1.64         {d24}, [r1], r3
-        vld1.64         {d26}, [r1], r3
-        vld1.64         {d28}, [r1], r3
-        vld1.64         {d9},  [r1], r3
-        vld1.64         {d11}, [r1], r3
-        vld1.64         {d13}, [r1], r3
-        vld1.64         {d15}, [r1], r3
-        vld1.64         {d23}, [r1]
+        vld1.8          {d8},  [r1], r3
+        vld1.8          {d10}, [r1], r3
+        vld1.8          {d12}, [r1], r3
+        vld1.8          {d14}, [r1], r3
+        vld1.8          {d22}, [r1], r3
+        vld1.8          {d24}, [r1], r3
+        vld1.8          {d26}, [r1], r3
+        vld1.8          {d28}, [r1], r3
+        vld1.8          {d9},  [r1], r3
+        vld1.8          {d11}, [r1], r3
+        vld1.8          {d13}, [r1], r3
+        vld1.8          {d15}, [r1], r3
+        vld1.8          {d23}, [r1]
 
         transpose_8x8   q4,  q5,  q6,  q7,  q11, q12, q13, q14
         lowpass_8       d8,  d9,  d10, d11, d8,  d9
@@ -929,20 +929,20 @@ function \type\()_h264_qpel8_v_lowpass_l2_neon
         lowpass_8       d26, d27, d28, d29, d26, d27
         transpose_8x8   d8,  d9,  d12, d13, d22, d23, d26, d27
 
-        vld1.64         {d0},  [ip], r2
-        vld1.64         {d1},  [ip], r2
-        vld1.64         {d2},  [ip], r2
-        vld1.64         {d3},  [ip], r2
-        vld1.64         {d4},  [ip], r2
+        vld1.8          {d0},  [r12], r2
+        vld1.8          {d1},  [r12], r2
+        vld1.8          {d2},  [r12], r2
+        vld1.8          {d3},  [r12], r2
+        vld1.8          {d4},  [r12], r2
         vrhadd.u8       q0,  q0,  q4
-        vld1.64         {d5},  [ip], r2
+        vld1.8          {d5},  [r12], r2
         vrhadd.u8       q1,  q1,  q6
-        vld1.64         {d10}, [ip], r2
+        vld1.8          {d10}, [r12], r2
         vrhadd.u8       q2,  q2,  q11
-        vld1.64         {d11}, [ip], r2
+        vld1.8          {d11}, [r12], r2
         vrhadd.u8       q5,  q5,  q13
 
-.ifc \type,avg
+  .ifc \type,avg
         vld1.8          {d16}, [r0,:64], r3
         vrhadd.u8       d0,  d0,  d16
         vld1.8          {d17}, [r0,:64], r3
@@ -960,51 +960,51 @@ function \type\()_h264_qpel8_v_lowpass_l2_neon
         vld1.8          {d17}, [r0,:64], r3
         vrhadd.u8       d11, d11, d17
         sub             r0,  r0,  r3,  lsl #3
-.endif
+  .endif
 
-        vst1.64         {d0},  [r0,:64], r3
-        vst1.64         {d1},  [r0,:64], r3
-        vst1.64         {d2},  [r0,:64], r3
-        vst1.64         {d3},  [r0,:64], r3
-        vst1.64         {d4},  [r0,:64], r3
-        vst1.64         {d5},  [r0,:64], r3
-        vst1.64         {d10}, [r0,:64], r3
-        vst1.64         {d11}, [r0,:64], r3
+        vst1.8          {d0},  [r0,:64], r3
+        vst1.8          {d1},  [r0,:64], r3
+        vst1.8          {d2},  [r0,:64], r3
+        vst1.8          {d3},  [r0,:64], r3
+        vst1.8          {d4},  [r0,:64], r3
+        vst1.8          {d5},  [r0,:64], r3
+        vst1.8          {d10}, [r0,:64], r3
+        vst1.8          {d11}, [r0,:64], r3
 
         bx              lr
 endfunc
-        .endm
+.endm
 
         h264_qpel_v_lowpass_l2 put
         h264_qpel_v_lowpass_l2 avg
 
 function put_h264_qpel8_hv_lowpass_neon_top
-        lowpass_const   ip
-        mov             ip,  #12
-1:      vld1.64         {d0, d1},  [r1], r3
-        vld1.64         {d16,d17}, [r1], r3
-        subs            ip,  ip,  #2
+        lowpass_const   r12
+        mov             r12, #12
+1:      vld1.8          {d0, d1},  [r1], r3
+        vld1.8          {d16,d17}, [r1], r3
+        subs            r12, r12, #2
         lowpass_8       d0,  d1,  d16, d17, q11, q12, narrow=0
-        vst1.64         {d22-d25}, [r4,:128]!
+        vst1.8          {d22-d25}, [r4,:128]!
         bne             1b
 
-        vld1.64         {d0, d1},  [r1]
+        vld1.8          {d0, d1},  [r1]
         lowpass_8_1     d0,  d1,  q12, narrow=0
 
-        mov             ip,  #-16
-        add             r4,  r4,  ip
-        vld1.64         {d30,d31}, [r4,:128], ip
-        vld1.64         {d20,d21}, [r4,:128], ip
-        vld1.64         {d18,d19}, [r4,:128], ip
-        vld1.64         {d16,d17}, [r4,:128], ip
-        vld1.64         {d14,d15}, [r4,:128], ip
-        vld1.64         {d12,d13}, [r4,:128], ip
-        vld1.64         {d10,d11}, [r4,:128], ip
-        vld1.64         {d8, d9},  [r4,:128], ip
-        vld1.64         {d6, d7},  [r4,:128], ip
-        vld1.64         {d4, d5},  [r4,:128], ip
-        vld1.64         {d2, d3},  [r4,:128], ip
-        vld1.64         {d0, d1},  [r4,:128]
+        mov             r12, #-16
+        add             r4,  r4,  r12
+        vld1.8          {d30,d31}, [r4,:128], r12
+        vld1.8          {d20,d21}, [r4,:128], r12
+        vld1.8          {d18,d19}, [r4,:128], r12
+        vld1.8          {d16,d17}, [r4,:128], r12
+        vld1.8          {d14,d15}, [r4,:128], r12
+        vld1.8          {d12,d13}, [r4,:128], r12
+        vld1.8          {d10,d11}, [r4,:128], r12
+        vld1.8          {d8, d9},  [r4,:128], r12
+        vld1.8          {d6, d7},  [r4,:128], r12
+        vld1.8          {d4, d5},  [r4,:128], r12
+        vld1.8          {d2, d3},  [r4,:128], r12
+        vld1.8          {d0, d1},  [r4,:128]
 
         swap4           d1,  d3,  d5,  d7,  d8,  d10, d12, d14
         transpose16_4x4 q0,  q1,  q2,  q3,  q4,  q5,  q6,  q7
@@ -1012,31 +1012,31 @@ function put_h264_qpel8_hv_lowpass_neon_top
         swap4           d17, d19, d21, d31, d24, d26, d28, d22
         transpose16_4x4 q8,  q9,  q10, q15, q12, q13, q14, q11
 
-        vst1.64         {d30,d31}, [r4,:128]!
-        vst1.64         {d6, d7},  [r4,:128]!
-        vst1.64         {d20,d21}, [r4,:128]!
-        vst1.64         {d4, d5},  [r4,:128]!
-        vst1.64         {d18,d19}, [r4,:128]!
-        vst1.64         {d2, d3},  [r4,:128]!
-        vst1.64         {d16,d17}, [r4,:128]!
-        vst1.64         {d0, d1},  [r4,:128]
+        vst1.8          {d30,d31}, [r4,:128]!
+        vst1.8          {d6, d7},  [r4,:128]!
+        vst1.8          {d20,d21}, [r4,:128]!
+        vst1.8          {d4, d5},  [r4,:128]!
+        vst1.8          {d18,d19}, [r4,:128]!
+        vst1.8          {d2, d3},  [r4,:128]!
+        vst1.8          {d16,d17}, [r4,:128]!
+        vst1.8          {d0, d1},  [r4,:128]
 
         lowpass_8.16    q4,  q12, d8,  d9,  d24, d25, d8
         lowpass_8.16    q5,  q13, d10, d11, d26, d27, d9
         lowpass_8.16    q6,  q14, d12, d13, d28, d29, d10
         lowpass_8.16    q7,  q11, d14, d15, d22, d23, d11
 
-        vld1.64         {d16,d17}, [r4,:128], ip
-        vld1.64         {d30,d31}, [r4,:128], ip
+        vld1.8          {d16,d17}, [r4,:128], r12
+        vld1.8          {d30,d31}, [r4,:128], r12
         lowpass_8.16    q8,  q15, d16, d17, d30, d31, d12
-        vld1.64         {d16,d17}, [r4,:128], ip
-        vld1.64         {d30,d31}, [r4,:128], ip
+        vld1.8          {d16,d17}, [r4,:128], r12
+        vld1.8          {d30,d31}, [r4,:128], r12
         lowpass_8.16    q8,  q15, d16, d17, d30, d31, d13
-        vld1.64         {d16,d17}, [r4,:128], ip
-        vld1.64         {d30,d31}, [r4,:128], ip
+        vld1.8          {d16,d17}, [r4,:128], r12
+        vld1.8          {d30,d31}, [r4,:128], r12
         lowpass_8.16    q8,  q15, d16, d17, d30, d31, d14
-        vld1.64         {d16,d17}, [r4,:128], ip
-        vld1.64         {d30,d31}, [r4,:128]
+        vld1.8          {d16,d17}, [r4,:128], r12
+        vld1.8          {d30,d31}, [r4,:128]
         lowpass_8.16    q8,  q15, d16, d17, d30, d31, d15
 
         transpose_8x8   d12, d13, d14, d15, d8,  d9,  d10, d11
@@ -1044,11 +1044,11 @@ function put_h264_qpel8_hv_lowpass_neon_top
         bx              lr
 endfunc
 
-        .macro h264_qpel8_hv_lowpass type
+.macro  h264_qpel8_hv_lowpass type
 function \type\()_h264_qpel8_hv_lowpass_neon
         mov             r10, lr
         bl              put_h264_qpel8_hv_lowpass_neon_top
-.ifc \type,avg
+  .ifc \type,avg
         vld1.8          {d0},      [r0,:64], r2
         vrhadd.u8       d12, d12, d0
         vld1.8          {d1},      [r0,:64], r2
@@ -1066,39 +1066,39 @@ function \type\()_h264_qpel8_hv_lowpass_neon
         vld1.8          {d7},      [r0,:64], r2
         vrhadd.u8       d11, d11, d7
         sub             r0,  r0,  r2,  lsl #3
-.endif
+  .endif
 
-        vst1.64         {d12},     [r0,:64], r2
-        vst1.64         {d13},     [r0,:64], r2
-        vst1.64         {d14},     [r0,:64], r2
-        vst1.64         {d15},     [r0,:64], r2
-        vst1.64         {d8},      [r0,:64], r2
-        vst1.64         {d9},      [r0,:64], r2
-        vst1.64         {d10},     [r0,:64], r2
-        vst1.64         {d11},     [r0,:64], r2
+        vst1.8          {d12},     [r0,:64], r2
+        vst1.8          {d13},     [r0,:64], r2
+        vst1.8          {d14},     [r0,:64], r2
+        vst1.8          {d15},     [r0,:64], r2
+        vst1.8          {d8},      [r0,:64], r2
+        vst1.8          {d9},      [r0,:64], r2
+        vst1.8          {d10},     [r0,:64], r2
+        vst1.8          {d11},     [r0,:64], r2
 
         mov             lr,  r10
         bx              lr
 endfunc
-        .endm
+.endm
 
         h264_qpel8_hv_lowpass put
         h264_qpel8_hv_lowpass avg
 
-        .macro h264_qpel8_hv_lowpass_l2 type
+.macro  h264_qpel8_hv_lowpass_l2 type
 function \type\()_h264_qpel8_hv_lowpass_l2_neon
         mov             r10, lr
         bl              put_h264_qpel8_hv_lowpass_neon_top
 
-        vld1.64         {d0, d1},  [r2,:128]!
-        vld1.64         {d2, d3},  [r2,:128]!
+        vld1.8          {d0, d1},  [r2,:128]!
+        vld1.8          {d2, d3},  [r2,:128]!
         vrhadd.u8       q0,  q0,  q6
-        vld1.64         {d4, d5},  [r2,:128]!
+        vld1.8          {d4, d5},  [r2,:128]!
         vrhadd.u8       q1,  q1,  q7
-        vld1.64         {d6, d7},  [r2,:128]!
+        vld1.8          {d6, d7},  [r2,:128]!
         vrhadd.u8       q2,  q2,  q4
         vrhadd.u8       q3,  q3,  q5
-.ifc \type,avg
+  .ifc \type,avg
         vld1.8          {d16},     [r0,:64], r3
         vrhadd.u8       d0,  d0,  d16
         vld1.8          {d17},     [r0,:64], r3
@@ -1116,25 +1116,25 @@ function \type\()_h264_qpel8_hv_lowpass_l2_neon
         vld1.8          {d23},     [r0,:64], r3
         vrhadd.u8       d7,  d7,  d23
         sub             r0,  r0,  r3,  lsl #3
-.endif
-        vst1.64         {d0},      [r0,:64], r3
-        vst1.64         {d1},      [r0,:64], r3
-        vst1.64         {d2},      [r0,:64], r3
-        vst1.64         {d3},      [r0,:64], r3
-        vst1.64         {d4},      [r0,:64], r3
-        vst1.64         {d5},      [r0,:64], r3
-        vst1.64         {d6},      [r0,:64], r3
-        vst1.64         {d7},      [r0,:64], r3
+  .endif
+        vst1.8          {d0},      [r0,:64], r3
+        vst1.8          {d1},      [r0,:64], r3
+        vst1.8          {d2},      [r0,:64], r3
+        vst1.8          {d3},      [r0,:64], r3
+        vst1.8          {d4},      [r0,:64], r3
+        vst1.8          {d5},      [r0,:64], r3
+        vst1.8          {d6},      [r0,:64], r3
+        vst1.8          {d7},      [r0,:64], r3
 
         mov             lr,  r10
         bx              lr
 endfunc
-        .endm
+.endm
 
         h264_qpel8_hv_lowpass_l2 put
         h264_qpel8_hv_lowpass_l2 avg
 
-        .macro h264_qpel16_hv type
+.macro  h264_qpel16_hv  type
 function \type\()_h264_qpel16_hv_lowpass_neon
         mov             r9,  lr
         bl              \type\()_h264_qpel8_hv_lowpass_neon
@@ -1167,17 +1167,17 @@ function \type\()_h264_qpel16_hv_lowpass_l2_neon
         mov             lr,  r9
         b               \type\()_h264_qpel8_hv_lowpass_l2_neon
 endfunc
-        .endm
+.endm
 
         h264_qpel16_hv put
         h264_qpel16_hv avg
 
-        .macro h264_qpel8 type
+.macro  h264_qpel8      type
 function ff_\type\()_h264_qpel8_mc10_neon, export=1
         lowpass_const   r3
         mov             r3,  r1
         sub             r1,  r1,  #2
-        mov             ip,  #8
+        mov             r12, #8
         b               \type\()_h264_qpel8_h_lowpass_l2_neon
 endfunc
 
@@ -1185,7 +1185,7 @@ function ff_\type\()_h264_qpel8_mc20_neon, export=1
         lowpass_const   r3
         sub             r1,  r1,  #2
         mov             r3,  r2
-        mov             ip,  #8
+        mov             r12, #8
         b               \type\()_h264_qpel8_h_lowpass_neon
 endfunc
 
@@ -1193,13 +1193,13 @@ function ff_\type\()_h264_qpel8_mc30_neon, export=1
         lowpass_const   r3
         add             r3,  r1,  #1
         sub             r1,  r1,  #2
-        mov             ip,  #8
+        mov             r12, #8
         b               \type\()_h264_qpel8_h_lowpass_l2_neon
 endfunc
 
 function ff_\type\()_h264_qpel8_mc01_neon, export=1
         push            {lr}
-        mov             ip,  r1
+        mov             r12, r1
 \type\()_h264_qpel8_mc01:
         lowpass_const   r3
         mov             r3,  r2
@@ -1222,12 +1222,12 @@ T       mov             sp,  r0
         mov             r0,  sp
         sub             r1,  r1,  #2
         mov             r3,  #8
-        mov             ip,  #8
+        mov             r12, #8
         vpush           {d8-d15}
         bl              put_h264_qpel8_h_lowpass_neon
         ldrd            r0,  [r11], #8
         mov             r3,  r2
-        add             ip,  sp,  #64
+        add             r12, sp,  #64
         sub             r1,  r1,  r2, lsl #1
         mov             r2,  #8
         bl              \type\()_h264_qpel8_v_lowpass_l2_neon
@@ -1248,7 +1248,7 @@ T       mov             sp,  r0
         sub             r1,  r1,  #2
         mov             r3,  #8
         mov             r0,  sp
-        mov             ip,  #8
+        mov             r12, #8
         vpush           {d8-d15}
         bl              put_h264_qpel8_h_lowpass_neon
         mov             r4,  r0
@@ -1333,7 +1333,7 @@ endfunc
 
 function ff_\type\()_h264_qpel8_mc03_neon, export=1
         push            {lr}
-        add             ip,  r1,  r2
+        add             r12, r1,  r2
         b               \type\()_h264_qpel8_mc01
 endfunc
 
@@ -1356,12 +1356,12 @@ function ff_\type\()_h264_qpel8_mc33_neon, export=1
         sub             r1,  r1,  #1
         b               \type\()_h264_qpel8_mc11
 endfunc
-        .endm
+.endm
 
         h264_qpel8 put
         h264_qpel8 avg
 
-        .macro h264_qpel16 type
+.macro  h264_qpel16     type
 function ff_\type\()_h264_qpel16_mc10_neon, export=1
         lowpass_const   r3
         mov             r3,  r1
@@ -1385,7 +1385,7 @@ endfunc
 
 function ff_\type\()_h264_qpel16_mc01_neon, export=1
         push            {r4, lr}
-        mov             ip,  r1
+        mov             r12, r1
 \type\()_h264_qpel16_mc01:
         lowpass_const   r3
         mov             r3,  r2
@@ -1412,7 +1412,7 @@ T       mov             sp,  r0
         bl              put_h264_qpel16_h_lowpass_neon
         ldrd            r0,  [r11], #8
         mov             r3,  r2
-        add             ip,  sp,  #64
+        add             r12, sp,  #64
         sub             r1,  r1,  r2, lsl #1
         mov             r2,  #16
         bl              \type\()_h264_qpel16_v_lowpass_l2_neon
@@ -1515,7 +1515,7 @@ endfunc
 
 function ff_\type\()_h264_qpel16_mc03_neon, export=1
         push            {r4, lr}
-        add             ip,  r1,  r2
+        add             r12, r1,  r2
         b               \type\()_h264_qpel16_mc01
 endfunc
 
@@ -1538,14 +1538,14 @@ function ff_\type\()_h264_qpel16_mc33_neon, export=1
         sub             r1,  r1,  #1
         b               \type\()_h264_qpel16_mc11
 endfunc
-        .endm
+.endm
 
         h264_qpel16 put
         h264_qpel16 avg
 
 @ Biweighted prediction
 
-        .macro  biweight_16 macs, macd
+.macro  biweight_16     macs, macd
         vdup.8          d0,  r4
         vdup.8          d1,  r5
         vmov            q2,  q8
@@ -1583,9 +1583,9 @@ endfunc
         vst1.8          {d24-d25},[r6,:128], r2
         bne             1b
         pop             {r4-r6, pc}
-        .endm
+.endm
 
-        .macro  biweight_8 macs, macd
+.macro  biweight_8      macs, macd
         vdup.8          d0,  r4
         vdup.8          d1,  r5
         vmov            q1,  q8
@@ -1613,9 +1613,9 @@ endfunc
         vst1.8          {d4},[r6,:64], r2
         bne             1b
         pop             {r4-r6, pc}
-        .endm
+.endm
 
-        .macro  biweight_4 macs, macd
+.macro  biweight_4      macs, macd
         vdup.8          d0,  r4
         vdup.8          d1,  r5
         vmov            q1,  q8
@@ -1655,9 +1655,9 @@ endfunc
         vst1.32         {d2[0]},[r6,:32], r2
         vst1.32         {d2[1]},[r6,:32], r2
         pop             {r4-r6, pc}
-        .endm
+.endm
 
-        .macro  biweight_func w
+.macro  biweight_func   w
 function ff_biweight_h264_pixels_\w\()_neon, export=1
         push            {r4-r6, lr}
         ldr             r12, [sp, #16]
@@ -1687,7 +1687,7 @@ function ff_biweight_h264_pixels_\w\()_neon, export=1
 40:     rsb             r5,  r5,  #0
         biweight_\w     vmlsl.u8, vmlal.u8
 endfunc
-        .endm
+.endm
 
         biweight_func   16
         biweight_func   8
@@ -1695,7 +1695,7 @@ endfunc
 
 @ Weighted prediction
 
-        .macro  weight_16 add
+.macro  weight_16       add
         vdup.8          d0,  r12
 1:      subs            r2,  r2,  #2
         vld1.8          {d20-d21},[r0,:128], r1
@@ -1722,9 +1722,9 @@ endfunc
         vst1.8          {d24-d25},[r4,:128], r1
         bne             1b
         pop             {r4, pc}
-        .endm
+.endm
 
-        .macro  weight_8 add
+.macro  weight_8        add
         vdup.8          d0,  r12
 1:      subs            r2,  r2,  #2
         vld1.8          {d4},[r0,:64], r1
@@ -1743,9 +1743,9 @@ endfunc
         vst1.8          {d4},[r4,:64], r1
         bne             1b
         pop             {r4, pc}
-        .endm
+.endm
 
-        .macro  weight_4 add
+.macro  weight_4        add
         vdup.8          d0,  r12
         vmov            q1,  q8
         vmov            q10, q8
@@ -1779,9 +1779,9 @@ endfunc
         vst1.32         {d2[0]},[r4,:32], r1
         vst1.32         {d2[1]},[r4,:32], r1
         pop             {r4, pc}
-        .endm
+.endm
 
-        .macro  weight_func w
+.macro  weight_func     w
 function ff_weight_h264_pixels_\w\()_neon, export=1
         push            {r4, lr}
         ldr             r12, [sp, #8]
@@ -1806,7 +1806,7 @@ function ff_weight_h264_pixels_\w\()_neon, export=1
 10:     rsb             r12, r12, #0
         weight_\w       vsub.s16
 endfunc
-        .endm
+.endm
 
         weight_func     16
         weight_func     8



More information about the ffmpeg-cvslog mailing list