/external/neven/Embedded/common/src/b_BitFeatureEm/ |
Functions.h | 43 #define bbf_BIT_SUM_8( vA ) ( bbf_bit8TblG[ vA ] ) 46 #define bbf_BIT_SUM_16( vA ) ( bbf_bit8TblG[ vA & 0x00FF ] + bbf_bit8TblG[ ( vA >> 8 ) & 0x00FF ] ) 49 #define bbf_BIT_SUM_32( vA ) ( bbf_bit8TblG[ vA & 0x00FF ] + bbf_bit8TblG[ ( vA >> 8 ) & 0x00FF ] + bbf_bit8TblG[ ( vA >> 16 ) & 0x00FF ] + bbf_bit8TblG[ ( vA >> 24 ) & 0x00FF ] [all...] |
/art/runtime/interpreter/mterp/mips/ |
op_shl_long_2addr.S | 2 * Long integer shift, 2addr version. vA is 64-bit value/result, vB is 5 /* shl-long/2addr vA, vB */ 10 LOAD64(a0, a1, t2) # a0/a1 <- vA/vA+1 23 SET_VREG64_GOTO(v0, v1, rOBJ, t0) # vA/vA+1 <- v0/v1 27 SET_VREG64_GOTO(zero, v0, rOBJ, t0) # vA/vA+1 <- rlo/rhi
|
op_shr_long_2addr.S | 2 * Long integer shift, 2addr version. vA is 64-bit value/result, vB is 5 /* shr-long/2addr vA, vB */ 10 LOAD64(a0, a1, t0) # a0/a1 <- vA/vA+1 22 SET_VREG64_GOTO(v0, v1, t2, t0) # vA/vA+1 <- v0/v1 27 SET_VREG64_GOTO(v1, a3, t2, t0) # vA/vA+1 <- rlo/rhi
|
op_ushr_long_2addr.S | 2 * Long integer shift, 2addr version. vA is 64-bit value/result, vB is 5 /* ushr-long/2addr vA, vB */ 10 LOAD64(a0, a1, t0) # a0/a1 <- vA/vA+1 23 SET_VREG64_GOTO(v0, v1, t3, t0) # vA/vA+1 <- v0/v1 27 SET_VREG64_GOTO(v1, zero, t3, t0) # vA/vA+1 <- rlo/rhi
|
funopWider.S | 7 /* unop vA, vB */ 14 SET_VREG64_F_GOTO(fv0, fv0f, rOBJ, t0) # vA/vA+1 <- fv0
|
binopWide2addr.S | 14 /* binop/2addr vA, vB */ 20 LOAD64($arg0, $arg1, t0) # a0/a1 <- vA/vA+1 30 SET_VREG64_GOTO($result0, $result1, rOBJ, t0) # vA/vA+1 <- $result0/$result1
|
unopWide.S | 9 /* unop vA, vB */ 13 LOAD64(a0, a1, a3) # a0/a1 <- vA 18 SET_VREG64_GOTO($result0, $result1, rOBJ, t0) # vA/vA+1 <- a0/a1
|
bincmp.S | 7 /* if-cmp vA, vB, +CCCC */ 11 GET_VREG(a0, a0) # a0 <- vA 13 b${condition} a0, a3, MterpCommonTakenBranchNoFlags # compare (vA, vB)
|
binop2addr.S | 14 /* binop/2addr vA, vB */ 17 GET_VREG(a0, rOBJ) # a0 <- vA 28 SET_VREG_GOTO($result, rOBJ, t0) # vA <- $result
|
fbinopWide2addr.S | 10 /* binop/2addr vA, vB */ 21 SET_VREG64_F_GOTO(fv0, fv0f, rOBJ, t0) # vA/vA+1 <- fv0
|
op_long_to_double.S | 4 /* unop vA, vB */ 20 SET_VREG64_F_GOTO(fv0, fv0f, rOBJ, t0) # vA/vA+1 <- result
|
unopWider.S | 8 /* unop vA, vB */ 16 SET_VREG64_GOTO($result0, $result1, rOBJ, t0) # vA/vA+1 <- a0/a1
|
/art/runtime/interpreter/mterp/x86_64/ |
bindiv2addr.S | 5 /* div/rem/2addr vA, vB */ 10 GET_WIDE_VREG %rax, rINSTq # eax <- vA 13 GET_VREG %eax, rINSTq # eax <- vA 24 SET_WIDE_VREG $result, rINSTq # vA <- result 26 SET_VREG $result, rINSTq # vA <- result
|
op_mul_int_2addr.S | 1 /* mul vA, vB */ 5 GET_VREG %eax, %rcx # eax <- vA
|
op_mul_long_2addr.S | 1 /* mul vA, vB */ 5 GET_WIDE_VREG %rax, %rcx # rax <- vA
|
zcmp.S | 9 cmpl $$0, VREG_ADDRESS(rINSTq) # compare (vA, 0)
|
/art/runtime/interpreter/mterp/x86/ |
zcmp.S | 9 cmpl $$0, VREG_ADDRESS(rINST) # compare (vA, 0)
|
/art/runtime/interpreter/mterp/arm/ |
bincmp.S | 7 /* if-cmp vA, vB, +CCCC */ 11 GET_VREG r0, r0 @ r0<- vA 13 cmp r0, r3 @ compare (vA, vB)
|
fbinop2addr.S | 8 /* binop/2addr vA, vB */ 12 VREG_INDEX_TO_ADDR r9, r9 @ r9<- &vA 15 flds s0, [r9] @ s0<- vA
|
fbinopWide2addr.S | 9 /* binop/2addr vA, vB */ 15 VREG_INDEX_TO_ADDR r9, r9 @ r9<- &vA 17 fldd d0, [r9] @ d0<- vA
|
funop.S | 7 /* unop vA, vB */ 15 VREG_INDEX_TO_ADDR r9, r9 @ r9<- &vA 16 fsts s1, [r9] @ vA<- s1
|
funopNarrower.S | 7 /* unop vA, vB */ 15 VREG_INDEX_TO_ADDR r9, r9 @ r9<- &vA 16 fsts s0, [r9] @ vA<- s0
|
funopWider.S | 7 /* unop vA, vB */ 16 VREG_INDEX_TO_ADDR r9, r9 @ r9<- &vA 17 fstd d0, [r9] @ vA<- d0
|
/art/runtime/interpreter/mterp/mips64/ |
binop2addr.S | 17 /* binop/2addr vA, vB */ 20 GET_VREG a0, a2 # a0 <- vA 29 SET_VREG $result, a2 # vA <- $result
|
binopWide2addr.S | 17 /* binop/2addr vA, vB */ 20 GET_VREG_WIDE a0, a2 # a0 <- vA 29 SET_VREG_WIDE $result, a2 # vA <- $result
|