|
|
|
@ -105,28 +105,28 @@ static av_always_inline int get_cabac_bypass_sign_x86(CABACContext *c, int val) |
|
|
|
|
{ |
|
|
|
|
x86_reg tmp; |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"movl %a5(%2), %k1 \n\t" |
|
|
|
|
"movl %a3(%2), %%eax \n\t" |
|
|
|
|
"shl $17, %k1 \n\t" |
|
|
|
|
"add %%eax, %%eax \n\t" |
|
|
|
|
"sub %k1, %%eax \n\t" |
|
|
|
|
"cltd \n\t" |
|
|
|
|
"and %%edx, %k1 \n\t" |
|
|
|
|
"add %k1, %%eax \n\t" |
|
|
|
|
"xor %%edx, %%ecx \n\t" |
|
|
|
|
"sub %%edx, %%ecx \n\t" |
|
|
|
|
"test %%ax, %%ax \n\t" |
|
|
|
|
" jnz 1f \n\t" |
|
|
|
|
"mov %a4(%2), %1 \n\t" |
|
|
|
|
"subl $0xFFFF, %%eax \n\t" |
|
|
|
|
"movzwl (%1), %%edx \n\t" |
|
|
|
|
"bswap %%edx \n\t" |
|
|
|
|
"shrl $15, %%edx \n\t" |
|
|
|
|
"add $2, %1 \n\t" |
|
|
|
|
"addl %%edx, %%eax \n\t" |
|
|
|
|
"mov %1, %a4(%2) \n\t" |
|
|
|
|
"1: \n\t" |
|
|
|
|
"movl %%eax, %a3(%2) \n\t" |
|
|
|
|
"movl %a5(%2), %k1 \n\t" |
|
|
|
|
"movl %a3(%2), %%eax \n\t" |
|
|
|
|
"shl $17, %k1 \n\t" |
|
|
|
|
"add %%eax, %%eax \n\t" |
|
|
|
|
"sub %k1, %%eax \n\t" |
|
|
|
|
"cltd \n\t" |
|
|
|
|
"and %%edx, %k1 \n\t" |
|
|
|
|
"add %k1, %%eax \n\t" |
|
|
|
|
"xor %%edx, %%ecx \n\t" |
|
|
|
|
"sub %%edx, %%ecx \n\t" |
|
|
|
|
"test %%ax, %%ax \n\t" |
|
|
|
|
"jnz 1f \n\t" |
|
|
|
|
"mov %a4(%2), %1 \n\t" |
|
|
|
|
"subl $0xFFFF, %%eax \n\t" |
|
|
|
|
"movzwl (%1), %%edx \n\t" |
|
|
|
|
"bswap %%edx \n\t" |
|
|
|
|
"shrl $15, %%edx \n\t" |
|
|
|
|
"add $2, %1 \n\t" |
|
|
|
|
"addl %%edx, %%eax \n\t" |
|
|
|
|
"mov %1, %a4(%2) \n\t" |
|
|
|
|
"1: \n\t" |
|
|
|
|
"movl %%eax, %a3(%2) \n\t" |
|
|
|
|
|
|
|
|
|
: "+c"(val), "=&r"(tmp) |
|
|
|
|
: "r"(c), |
|
|
|
|