Loading media/libstagefright/codecs/avc/enc/src/sad_halfpel_inline.h +2 −2 Original line number Diff line number Diff line Loading @@ -81,7 +81,7 @@ extern "C" "rsbs %1, %1, %2, asr #1\n\t" "rsbmi %1, %1, #0\n\t" "add %0, %0, %1" : "=r"(sad), "=r"(tmp) : "+r"(sad), "+r"(tmp) : "r"(tmp2) ); Loading @@ -94,7 +94,7 @@ extern "C" "rsbs %1, %2, %1, asr #2\n\t" "rsbmi %1, %1, #0\n\t" "add %0, %0, %1" : "=r"(sad), "=r"(tmp) : "+r"(sad), "+r"(tmp) : "r"(tmp2) ); Loading media/libstagefright/codecs/avc/enc/src/sad_inline.h +7 −7 Original line number Diff line number Diff line Loading @@ -347,8 +347,8 @@ SadMBOffset1: "rsbs %1, %1, %2\n\t" "rsbmi %1, %1, #0\n\t" "add %0, %0, %1" : "=r"(sad) : "r"(tmp), "r"(tmp2) : "+r"(sad), "+r"(tmp) : "r"(tmp2) ); return sad; } Loading @@ -366,7 +366,7 @@ SadMBOffset1: "RSB %1, %1, %1, lsl #8\n\t" "ADD %0, %0, %1, asr #7\n\t" "EOR %0, %0, %1, asr #7" : "=r"(src1), "=&r"(x7) : "+r"(src1), "=&r"(x7) : "r"(src2), "r"(mask) ); Loading @@ -385,7 +385,7 @@ SadMBOffset1: "RSB %1, %1, %1, lsl #8\n\t" "SUB %0, %0, %1, asr #7\n\t" "EOR %0, %0, %1, asr #7" : "=r"(src1), "=&r"(x7) : "+r"(src1), "=&r"(x7) : "r"(src2), "r"(mask) ); Loading @@ -399,7 +399,7 @@ SadMBOffset1: "SBC %0, %0, %3\n\t" \ "BIC %3, %4, %3\n\t" \ "ADD %2, %2, %3, lsr #8" \ : "=&r" (x5), "=&r" (x10), "=&r" (x4), "=&r" (x11) \ : "+r" (x5), "+r" (x10), "+r" (x4), "+r" (x11) \ : "r" (x6) \ ); Loading Loading @@ -467,10 +467,10 @@ LOOP_SAD0: /****** process 8 pixels ******/ x11 = *((int32*)(ref + 4)); __asm__ volatile("LDR %0, [%1], %2": "=&r"(x10), "=r"(ref): "r"(lx)); __asm__ volatile("LDR %0, [%1], %2": "=&r"(x10), "+r"(ref): "r"(lx)); //x10 = *((int32*)ref); ref+=lx; x14 = *((int32*)(blk + 4)); __asm__ volatile("LDR %0, [%1], #16": "=&r"(x12), "=r"(blk)); __asm__ volatile("LDR %0, [%1], #16": "=&r"(x12), "+r"(blk)); /* process x11 & x14 */ x11 = sad_4pixel(x11, x14, x9); Loading media/libstagefright/codecs/avc/enc/src/sad_mb_offset.h +8 −8 Original line number Diff line number Diff line Loading @@ -239,7 +239,7 @@ LOOP_SAD2: #elif (NUMBER==1) LOOP_SAD1: #endif __asm__ volatile("BIC %0, %0, #3": "=r"(ref)); __asm__ volatile("BIC %0, %0, #3": "+r"(ref)); /****** process 8 pixels ******/ x11 = *((int32*)(ref + 12)); x12 = *((int32*)(ref + 16)); Loading @@ -252,7 +252,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #24\n\t" "MVN %1, %1, lsr #8\n\t" "BIC %1, %1, %2, lsl #24" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #elif (SHIFT==16) Loading @@ -261,7 +261,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #16\n\t" "MVN %1, %1, lsr #16\n\t" "BIC %1, %1, %2, lsl #16" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #elif (SHIFT==24) Loading @@ -270,7 +270,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #8\n\t" "MVN %1, %1, lsr #24\n\t" "BIC %1, %1, %2, lsl #8" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #endif Loading @@ -297,7 +297,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #24\n\t" "MVN %1, %1, lsr #8\n\t" "BIC %1, %1, %2, lsl #24" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #elif (SHIFT==16) Loading @@ -306,7 +306,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #16\n\t" "MVN %1, %1, lsr #16\n\t" "BIC %1, %1, %2, lsl #16" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #elif (SHIFT==24) Loading @@ -315,11 +315,11 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #8\n\t" "MVN %1, %1, lsr #24\n\t" "BIC %1, %1, %2, lsl #8" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #endif __asm__ volatile("LDR %0, [%1], #16": "=&r"(x12), "=r"(blk)); __asm__ volatile("LDR %0, [%1], #16": "=&r"(x12), "+r"(blk)); /* process x11 & x14 */ x11 = sad_4pixelN(x11, x14, x9); Loading Loading
media/libstagefright/codecs/avc/enc/src/sad_halfpel_inline.h +2 −2 Original line number Diff line number Diff line Loading @@ -81,7 +81,7 @@ extern "C" "rsbs %1, %1, %2, asr #1\n\t" "rsbmi %1, %1, #0\n\t" "add %0, %0, %1" : "=r"(sad), "=r"(tmp) : "+r"(sad), "+r"(tmp) : "r"(tmp2) ); Loading @@ -94,7 +94,7 @@ extern "C" "rsbs %1, %2, %1, asr #2\n\t" "rsbmi %1, %1, #0\n\t" "add %0, %0, %1" : "=r"(sad), "=r"(tmp) : "+r"(sad), "+r"(tmp) : "r"(tmp2) ); Loading
media/libstagefright/codecs/avc/enc/src/sad_inline.h +7 −7 Original line number Diff line number Diff line Loading @@ -347,8 +347,8 @@ SadMBOffset1: "rsbs %1, %1, %2\n\t" "rsbmi %1, %1, #0\n\t" "add %0, %0, %1" : "=r"(sad) : "r"(tmp), "r"(tmp2) : "+r"(sad), "+r"(tmp) : "r"(tmp2) ); return sad; } Loading @@ -366,7 +366,7 @@ SadMBOffset1: "RSB %1, %1, %1, lsl #8\n\t" "ADD %0, %0, %1, asr #7\n\t" "EOR %0, %0, %1, asr #7" : "=r"(src1), "=&r"(x7) : "+r"(src1), "=&r"(x7) : "r"(src2), "r"(mask) ); Loading @@ -385,7 +385,7 @@ SadMBOffset1: "RSB %1, %1, %1, lsl #8\n\t" "SUB %0, %0, %1, asr #7\n\t" "EOR %0, %0, %1, asr #7" : "=r"(src1), "=&r"(x7) : "+r"(src1), "=&r"(x7) : "r"(src2), "r"(mask) ); Loading @@ -399,7 +399,7 @@ SadMBOffset1: "SBC %0, %0, %3\n\t" \ "BIC %3, %4, %3\n\t" \ "ADD %2, %2, %3, lsr #8" \ : "=&r" (x5), "=&r" (x10), "=&r" (x4), "=&r" (x11) \ : "+r" (x5), "+r" (x10), "+r" (x4), "+r" (x11) \ : "r" (x6) \ ); Loading Loading @@ -467,10 +467,10 @@ LOOP_SAD0: /****** process 8 pixels ******/ x11 = *((int32*)(ref + 4)); __asm__ volatile("LDR %0, [%1], %2": "=&r"(x10), "=r"(ref): "r"(lx)); __asm__ volatile("LDR %0, [%1], %2": "=&r"(x10), "+r"(ref): "r"(lx)); //x10 = *((int32*)ref); ref+=lx; x14 = *((int32*)(blk + 4)); __asm__ volatile("LDR %0, [%1], #16": "=&r"(x12), "=r"(blk)); __asm__ volatile("LDR %0, [%1], #16": "=&r"(x12), "+r"(blk)); /* process x11 & x14 */ x11 = sad_4pixel(x11, x14, x9); Loading
media/libstagefright/codecs/avc/enc/src/sad_mb_offset.h +8 −8 Original line number Diff line number Diff line Loading @@ -239,7 +239,7 @@ LOOP_SAD2: #elif (NUMBER==1) LOOP_SAD1: #endif __asm__ volatile("BIC %0, %0, #3": "=r"(ref)); __asm__ volatile("BIC %0, %0, #3": "+r"(ref)); /****** process 8 pixels ******/ x11 = *((int32*)(ref + 12)); x12 = *((int32*)(ref + 16)); Loading @@ -252,7 +252,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #24\n\t" "MVN %1, %1, lsr #8\n\t" "BIC %1, %1, %2, lsl #24" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #elif (SHIFT==16) Loading @@ -261,7 +261,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #16\n\t" "MVN %1, %1, lsr #16\n\t" "BIC %1, %1, %2, lsl #16" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #elif (SHIFT==24) Loading @@ -270,7 +270,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #8\n\t" "MVN %1, %1, lsr #24\n\t" "BIC %1, %1, %2, lsl #8" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #endif Loading @@ -297,7 +297,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #24\n\t" "MVN %1, %1, lsr #8\n\t" "BIC %1, %1, %2, lsl #24" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #elif (SHIFT==16) Loading @@ -306,7 +306,7 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #16\n\t" "MVN %1, %1, lsr #16\n\t" "BIC %1, %1, %2, lsl #16" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #elif (SHIFT==24) Loading @@ -315,11 +315,11 @@ LOOP_SAD1: "BIC %0, %0, %1, lsl #8\n\t" "MVN %1, %1, lsr #24\n\t" "BIC %1, %1, %2, lsl #8" : "=&r"(x10), "=&r"(x11) : "+r"(x10), "+r"(x11) : "r"(x12) ); #endif __asm__ volatile("LDR %0, [%1], #16": "=&r"(x12), "=r"(blk)); __asm__ volatile("LDR %0, [%1], #16": "=&r"(x12), "+r"(blk)); /* process x11 & x14 */ x11 = sad_4pixelN(x11, x14, x9); Loading