Loading media/libstagefright/codecs/amrwbenc/inc/basic_op.h +2 −2 Original line number Diff line number Diff line Loading @@ -51,11 +51,11 @@ #define vo_shr_r(var1, var2) ((var1+((Word16)(1L<<(var2-1))))>>var2) #define vo_sub(a,b) (a - b) #define vo_L_deposit_h(a) ((Word32)((a) << 16)) #define vo_round(a) ((a + 0x00008000) >> 16) #define vo_round(a) ((((a) >> 15) + 1) >> 1) #define vo_extract_l(a) ((Word16)(a)) #define vo_L_add(a,b) (a + b) #define vo_L_sub(a,b) (a - b) #define vo_mult_r(a,b) ((( a * b ) + 0x4000 ) >> 15 ) #define vo_mult_r(a,b) (((( a * b ) >> 14) + 1 ) >> 1 ) #define vo_negate(a) (-a) #define vo_L_shr_r(L_var1, var2) ((L_var1+((Word32)(1L<<(var2-1))))>>var2) Loading media/libstagefright/codecs/amrwbenc/src/c4t64fx.c +22 −22 Original line number Diff line number Diff line Loading @@ -267,13 +267,13 @@ void ACELP_4t64_fx( for (i = 0; i < L_SUBFR/4; i++) { s = (k_cn* (*p0++))+(k_dn * (*p1++)); s = L_add((k_cn* (*p0++)), (k_dn * (*p1++))); *p2++ = s >> 7; s = (k_cn* (*p0++))+(k_dn * (*p1++)); s = L_add((k_cn* (*p0++)), (k_dn * (*p1++))); *p2++ = s >> 7; s = (k_cn* (*p0++))+(k_dn * (*p1++)); s = L_add((k_cn* (*p0++)), (k_dn * (*p1++))); *p2++ = s >> 7; s = (k_cn* (*p0++))+(k_dn * (*p1++)); s = L_add((k_cn* (*p0++)), (k_dn * (*p1++))); *p2++ = s >> 7; } Loading Loading @@ -342,7 +342,7 @@ void ACELP_4t64_fx( { *h++ = 0; *h_inv++ = 0; L_tmp += (H[i] * H[i]) << 1; L_tmp = L_add(L_tmp, (H[i] * H[i]) << 1); } /* scale h[] down (/2) when energy of h[] is high with many pulses used */ val = extract_h(L_tmp); Loading Loading @@ -386,16 +386,16 @@ void ACELP_4t64_fx( cor = 0x00008000L; /* for rounding */ for (i = 0; i < NB_POS; i++) { cor += vo_L_mult((*ptr_h1), (*ptr_h1)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h1))); ptr_h1++; *p3-- = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h1)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h1))); ptr_h1++; *p2-- = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h1)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h1))); ptr_h1++; *p1-- = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h1)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h1))); ptr_h1++; *p0-- = extract_h(cor); } Loading Loading @@ -425,19 +425,19 @@ void ACELP_4t64_fx( for (i = k + 1; i < NB_POS; i++) { cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p3 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p2 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p1 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p0 = extract_h(cor); Loading @@ -447,15 +447,15 @@ void ACELP_4t64_fx( p1 -= (NB_POS + 1); p0 -= (NB_POS + 1); } cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p3 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p2 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p1 = extract_h(cor); Loading @@ -482,19 +482,19 @@ void ACELP_4t64_fx( for (i = k + 1; i < NB_POS; i++) { cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p3 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p2 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p1 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p0 = extract_h(cor); Loading @@ -504,7 +504,7 @@ void ACELP_4t64_fx( p1 -= (NB_POS + 1); p0 -= (NB_POS + 1); } cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p3 = extract_h(cor); Loading Loading @@ -698,7 +698,7 @@ void ACELP_4t64_fx( } /* memorise the best codevector */ ps = vo_mult(ps, ps); s = vo_L_msu(vo_L_mult(alpk, ps), psk, alp); s = L_sub(vo_L_mult(alpk, ps), vo_L_mult(psk, alp)); if (s > 0) { psk = ps; Loading Loading
media/libstagefright/codecs/amrwbenc/inc/basic_op.h +2 −2 Original line number Diff line number Diff line Loading @@ -51,11 +51,11 @@ #define vo_shr_r(var1, var2) ((var1+((Word16)(1L<<(var2-1))))>>var2) #define vo_sub(a,b) (a - b) #define vo_L_deposit_h(a) ((Word32)((a) << 16)) #define vo_round(a) ((a + 0x00008000) >> 16) #define vo_round(a) ((((a) >> 15) + 1) >> 1) #define vo_extract_l(a) ((Word16)(a)) #define vo_L_add(a,b) (a + b) #define vo_L_sub(a,b) (a - b) #define vo_mult_r(a,b) ((( a * b ) + 0x4000 ) >> 15 ) #define vo_mult_r(a,b) (((( a * b ) >> 14) + 1 ) >> 1 ) #define vo_negate(a) (-a) #define vo_L_shr_r(L_var1, var2) ((L_var1+((Word32)(1L<<(var2-1))))>>var2) Loading
media/libstagefright/codecs/amrwbenc/src/c4t64fx.c +22 −22 Original line number Diff line number Diff line Loading @@ -267,13 +267,13 @@ void ACELP_4t64_fx( for (i = 0; i < L_SUBFR/4; i++) { s = (k_cn* (*p0++))+(k_dn * (*p1++)); s = L_add((k_cn* (*p0++)), (k_dn * (*p1++))); *p2++ = s >> 7; s = (k_cn* (*p0++))+(k_dn * (*p1++)); s = L_add((k_cn* (*p0++)), (k_dn * (*p1++))); *p2++ = s >> 7; s = (k_cn* (*p0++))+(k_dn * (*p1++)); s = L_add((k_cn* (*p0++)), (k_dn * (*p1++))); *p2++ = s >> 7; s = (k_cn* (*p0++))+(k_dn * (*p1++)); s = L_add((k_cn* (*p0++)), (k_dn * (*p1++))); *p2++ = s >> 7; } Loading Loading @@ -342,7 +342,7 @@ void ACELP_4t64_fx( { *h++ = 0; *h_inv++ = 0; L_tmp += (H[i] * H[i]) << 1; L_tmp = L_add(L_tmp, (H[i] * H[i]) << 1); } /* scale h[] down (/2) when energy of h[] is high with many pulses used */ val = extract_h(L_tmp); Loading Loading @@ -386,16 +386,16 @@ void ACELP_4t64_fx( cor = 0x00008000L; /* for rounding */ for (i = 0; i < NB_POS; i++) { cor += vo_L_mult((*ptr_h1), (*ptr_h1)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h1))); ptr_h1++; *p3-- = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h1)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h1))); ptr_h1++; *p2-- = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h1)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h1))); ptr_h1++; *p1-- = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h1)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h1))); ptr_h1++; *p0-- = extract_h(cor); } Loading Loading @@ -425,19 +425,19 @@ void ACELP_4t64_fx( for (i = k + 1; i < NB_POS; i++) { cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p3 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p2 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p1 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p0 = extract_h(cor); Loading @@ -447,15 +447,15 @@ void ACELP_4t64_fx( p1 -= (NB_POS + 1); p0 -= (NB_POS + 1); } cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p3 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p2 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p1 = extract_h(cor); Loading @@ -482,19 +482,19 @@ void ACELP_4t64_fx( for (i = k + 1; i < NB_POS; i++) { cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p3 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p2 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p1 = extract_h(cor); cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p0 = extract_h(cor); Loading @@ -504,7 +504,7 @@ void ACELP_4t64_fx( p1 -= (NB_POS + 1); p0 -= (NB_POS + 1); } cor += vo_L_mult((*ptr_h1), (*ptr_h2)); cor = L_add(cor, vo_L_mult((*ptr_h1), (*ptr_h2))); ptr_h1++; ptr_h2++; *p3 = extract_h(cor); Loading Loading @@ -698,7 +698,7 @@ void ACELP_4t64_fx( } /* memorise the best codevector */ ps = vo_mult(ps, ps); s = vo_L_msu(vo_L_mult(alpk, ps), psk, alp); s = L_sub(vo_L_mult(alpk, ps), vo_L_mult(psk, alp)); if (s > 0) { psk = ps; Loading