[Ffmpeg-cvslog] r7840 - trunk/libavcodec/armv4l/simple_idct_armv6.S

mru subversion
Mon Feb 5 22:18:11 CET 2007


Author: mru
Date: Mon Feb  5 22:18:11 2007
New Revision: 7840

Modified:
   trunk/libavcodec/armv4l/simple_idct_armv6.S

Log:
oops, revert accidental checkin


Modified: trunk/libavcodec/armv4l/simple_idct_armv6.S
==============================================================================
--- trunk/libavcodec/armv4l/simple_idct_armv6.S	(original)
+++ trunk/libavcodec/armv4l/simple_idct_armv6.S	Mon Feb  5 22:18:11 2007
@@ -47,19 +47,6 @@
 w46:    .long W46
 w57:    .long W57
 
-        .macro idct_row_start shift
-        ldr    ip, [pc, #(w42-.-8)]  /* ip = W4 | (W2 << 16) */
-        ldr    lr, [pc, #(w46-.-8)]  /* lr = W4 | (W6 << 16) */
-        ldr    v7, [pc, #(w57-.-8)]  /* v7 = W5 | (W7 << 16) */
-        mov    a2, #(1<<(\shift-1))
-        smlad  v1, a3, ip, a2
-        smlsd  v4, a3, ip, a2
-        ldr    ip, [pc, #(w13-.-8)]  /* ip = W1 | (W3 << 16) */
-        smlad  v2, a3, lr, a2
-        smlsd  v3, a3, lr, a2
-        smusdx fp, a4, v7            /* fp = B3 = W7*row[1] - W5*row[3] */
-        smuad  v5, a4, ip            /* v5 = B0 = W1*row[1] + W3*row[3] */
-        .endm
 /*
   Compute partial IDCT of single row.
   shift = left-shift amount
@@ -71,6 +58,17 @@
   Output in registers v1--v8
 */
         .macro idct_row shift
+        ldr    lr, [pc, #(w46-.-8)]  /* lr = W4 | (W6 << 16) */
+        mov    a2, #(1<<(\shift-1))
+        smlad  v1, a3, ip, a2
+        smlsd  v4, a3, ip, a2
+        ldr    ip, [pc, #(w13-.-8)]  /* ip = W1 | (W3 << 16) */
+        ldr    v7, [pc, #(w57-.-8)]  /* v7 = W5 | (W7 << 16) */
+        smlad  v2, a3, lr, a2
+        smlsd  v3, a3, lr, a2
+
+        smuad  v5, a4, ip            /* v5 = B0 = W1*row[1] + W3*row[3] */
+        smusdx fp, a4, v7            /* fp = B3 = W7*row[1] - W5*row[3] */
         ldr    lr, [a1, #12]         /* lr = row[7,5] */
         pkhtb  a3, ip, v7, asr #16   /* a4 = W7 | (W3 << 16) */
         pkhbt  a2, ip, v7, lsl #16   /* a2 = W1 | (W5 << 16) */
@@ -101,6 +99,16 @@
   Output in registers v1--v8
 */
         .macro idct_row4 shift
+        ldr    lr, [pc, #(w46-.-8)]  /* lr = W4 | (W6 << 16) */
+        ldr    v7, [pc, #(w57-.-8)]  /* v7 = W5 | (W7 << 16) */
+        mov    a2, #(1<<(\shift-1))
+        smlad  v1, a3, ip, a2
+        smlsd  v4, a3, ip, a2
+        ldr    ip, [pc, #(w13-.-8)]  /* ip = W1 | (W3 << 16) */
+        smlad  v2, a3, lr, a2
+        smlsd  v3, a3, lr, a2
+        smusdx fp, a4, v7            /* fp = B3 = W7*row[1] - W5*row[3] */
+        smuad  v5, a4, ip            /* v5 = B0 = W1*row[1] + W3*row[3] */
         pkhtb  a3, ip, v7, asr #16   /* a4 = W7 | (W3 << 16) */
         pkhbt  a2, ip, v7, lsl #16   /* a2 = W1 | (W5 << 16) */
         smusdx v6, a3, a4            /* v6 = -B1 = W7*row[3] - W3*row[1] */
@@ -185,100 +193,28 @@
         .align
         .func idct_row_armv6
 idct_row_armv6:
-        ldr    fp, [a1, #12]         /* fp = row[7,5] */
-        ldr    v7, [a1, #4]          /* v7 = row[6,4] */
+        str    lr, [sp, #-4]!
+
+        ldr    lr, [a1, #12]         /* lr = row[7,5] */
+        ldr    ip, [a1, #4]          /* ip = row[6,4] */
         ldr    a4, [a1, #8]          /* a4 = row[3,1] */
         ldr    a3, [a1]              /* a3 = row[2,0] */
-        mov    ip, #(1<<(ROW_SHIFT-1))
-        orrs   v5, fp, v7
-        cmpeq  v5, a4
-        cmpeq  v5, a3, lsr #16
+        orrs   lr, lr, ip
+        cmpeq  lr, a4
+        cmpeq  lr, a3, lsr #16
         beq    1f
-        cmp    v5, #0
-        stmfd  sp!, {a2, lr}
-        ldr    v5, [pc, #(w42-.-8)]  /* v5 = W4 | (W2 << 16) */
-        ldr    v6, [pc, #(w46-.-8)]  /* v6 = W4 | (W6 << 16) */
-        ldr    v7, [pc, #(w57-.-8)]  /* v7 = W5 | (W7 << 16) */
-
-        smlad  v1, a3, v5, ip
-        smlsd  v4, a3, v5, ip
-        ldr    a2, [pc, #(w13-.-8)]  /* a2 = W1 | (W3 << 16) */
-        smlad  v2, a3, v6, ip
-        smlsd  v3, a3, v6, ip
-        smusdx lr, a4, v7            /* lr = B3 = W7*row[1] - W5*row[3] */
-        smuad  v5, a4, a2            /* v5 = B0 = W1*row[1] + W3*row[3] */
-
-        pkhtb  a3, a2, v7, asr #16   /* a3 = W7 | (W3 << 16) */
-        pkhbt  ip, a2, v7, lsl #16   /* ip = W1 | (W5 << 16) */
-        smusdx v6, a3, a4            /* v6 = -B1 = W7*row[3] - W3*row[1] */
-        smusdx a4, a4, ip            /* v7 = B2 = W5*row[1] - W1*row[3] */
-        beq    3f
-
-        smlad  v5, fp, v7, v5        /* B0 += W5*row[5] + W7*row[7] */
-        smlad  v7, fp, a3, a4        /* B2 += W7*row[5] + W3*row[7] */
-        ldr    a4, [pc, #(w42n-.-8)] /* a4 = -W4 | (-W2 << 16) */
-        ldr    a3, [a1, #4]          /* a3 = row[6,4] */
-        smlsdx lr, fp, a2, lr        /* B3 += W3*row[5] - W1*row[7] */
-        ldr    a2, [pc, #(w46-.-8)]  /* a2 = W4 | (W6 << 16) */
-        smlad  v6, fp, ip, v6        /* B1 -= W1*row[5] + W5*row[7] */
-
-        smlad  v2, a3, a4, v2        /* A1 += -W4*row[4] - W2*row[6] */
-        smlsd  v3, a3, a4, v3        /* A2 += -W4*row[4] + W2*row[6] */
-        smlad  v1, a3, a2, v1        /* A0 += W4*row[4] + W6*row[6] */
-        smlsd  v4, a3, a2, v4        /* A3 += W4*row[4] - W6*row[6] */
-
-        ldr    a2, [sp], #4
-        add    a4, v1, v5            /* a4 = A0 + B0 */
-        sub    a3, v1, v5            /* a3 = A0 - B0 */
-        mov    v1, a4, asr #ROW_SHIFT
-        mov    v5, a3, asr #ROW_SHIFT
-
-        sub    a4, v2, v6            /* a4 = A1 + B1 */
-        add    a3, v2, v6            /* a3 = A1 - B1 */
-        mov    v2, a4, asr #ROW_SHIFT
-        mov    v6, a3, asr #ROW_SHIFT
-
-        add    a4, v3, v7            /* a4 = A2 + B2 */
-        sub    a3, v3, v7            /* a3 = A2 - B2 */
-        mov    v3, a4, asr #ROW_SHIFT
-        mov    v7, a3, asr #ROW_SHIFT
-
-        add    a4, v4, lr            /* a4 = A3 + B3 */
-        sub    a3, v4, lr            /* a3 = A3 - B3 */
-        mov    v4, a4, asr #ROW_SHIFT
-        mov    fp, a3, asr #ROW_SHIFT
+        str    a2, [sp, #-4]!
+        ldr    ip, [pc, #(w42-.-8)]  /* ip = W4 | (W2 << 16) */
+        cmp    lr, #0
+        beq    2f
 
-        strh   v1, [a2]
-        strh   v2, [a2, #(16*2)]
-        strh   v3, [a2, #(16*4)]
-        strh   v4, [a2, #(16*6)]
-        strh   fp, [a2, #(16*1)]
-        strh   v7, [a2, #(16*3)]
-        strh   v6, [a2, #(16*5)]
-        strh   v5, [a2, #(16*7)]
+        idct_row   ROW_SHIFT
+        b      3f
 
-        ldr    pc, [sp], #4
+2:      idct_row4  ROW_SHIFT
 
 3:      ldr    a2, [sp], #4
-        add    v7, v1, v5            /* v7 = A0 + B0 */
-        sub    a3, v1, v5            /* a3 = A0 - B0 */
-        mov    v1, v7, asr #ROW_SHIFT
-        mov    v5, a3, asr #ROW_SHIFT
-
-        sub    v7, v2, v6            /* v7 = A1 + B1 */
-        add    a3, v2, v6            /* a3 = A1 - B1 */
-        mov    v2, v7, asr #ROW_SHIFT
-        mov    v6, a3, asr #ROW_SHIFT
-
-        add    v7, v3, a4            /* v7 = A2 + B2 */
-        sub    a3, v3, a4            /* a3 = A2 - B2 */
-        mov    v3, v7, asr #ROW_SHIFT
-        mov    v7, a3, asr #ROW_SHIFT
-
-        add    a4, v4, lr            /* xx = A3 + B3 */
-        sub    a3, v4, lr            /* a3 = A3 - B3 */
-        mov    v4, a4, asr #ROW_SHIFT
-        mov    fp, a3, asr #ROW_SHIFT
+        idct_finish_shift ROW_SHIFT
 
         strh   v1, [a2]
         strh   v2, [a2, #(16*2)]
@@ -300,7 +236,7 @@
         strh   a3, [a2, #(16*3)]
         strh   a3, [a2, #(16*5)]
         strh   a3, [a2, #(16*7)]
-        mov    pc, lr
+        ldr    pc, [sp], #4
         .endfunc
 
 /*
@@ -314,8 +250,8 @@
         stmfd  sp!, {a2, lr}
 
         ldr    a3, [a1]              /* a3 = row[2,0] */
+        ldr    ip, [pc, #(w42-.-8)]  /* ip = W4 | (W2 << 16) */
         ldr    a4, [a1, #8]          /* a4 = row[3,1] */
-        idct_row_start COL_SHIFT
         idct_row COL_SHIFT
         ldr    a2, [sp], #4
         idct_finish_shift COL_SHIFT
@@ -344,8 +280,8 @@
         stmfd  sp!, {a2, a3, lr}
 
         ldr    a3, [a1]              /* a3 = row[2,0] */
+        ldr    ip, [pc, #(w42-.-8)]  /* ip = W4 | (W2 << 16) */
         ldr    a4, [a1, #8]          /* a4 = row[3,1] */
-        idct_row_start COL_SHIFT
         idct_row COL_SHIFT
         ldmfd  sp!, {a2, a3}
         idct_finish_shift_sat COL_SHIFT
@@ -376,8 +312,8 @@
         stmfd  sp!, {a2, a3, lr}
 
         ldr    a3, [a1]              /* a3 = row[2,0] */
+        ldr    ip, [pc, #(w42-.-8)]  /* ip = W4 | (W2 << 16) */
         ldr    a4, [a1, #8]          /* a4 = row[3,1] */
-        idct_row_start COL_SHIFT
         idct_row COL_SHIFT
         ldmfd  sp!, {a2, a3}
         idct_finish




More information about the ffmpeg-cvslog mailing list