/external/valgrind/main/VEX/priv/ |
host_generic_simd64.c | 72 return toUShort(0xFFFF & (hi32 >> 16)); 76 return toUShort(0xFFFF & hi32); 80 return toUShort(0xFFFF & (lo32 >> 16)); 84 return toUShort(0xFFFF & lo32); 257 return toUShort(xx==yy ? 0xFFFF : 0); 272 return toUShort(xx>yy ? 0xFFFF : 0); 287 return toUShort(xx==0 ? 0 : 0xFFFF); 344 return toUShort(v << n); 349 return toUShort((((UShort)v) >> n)); 354 return toUShort(((Short)v) >> n) [all...] |
guest_x86_helpers.c | 549 UShort, toUShort ); 550 case X86G_CC_OP_UMULW: ACTIONS_UMUL( 16, UShort, toUShort, 556 Short, toUShort ); 557 case X86G_CC_OP_SMULW: ACTIONS_SMUL( 16, Short, toUShort, [all...] |
guest_amd64_helpers.c | 640 UShort, toUShort ); 641 case AMD64G_CC_OP_UMULW: ACTIONS_UMUL( 16, UShort, toUShort, 649 Short, toUShort ); 650 case AMD64G_CC_OP_SMULW: ACTIONS_SMUL( 16, Short, toUShort, [all...] |
host_generic_simd128.c | 73 return toUShort((xx > yy) ? xx : yy); 78 return toUShort((xx < yy) ? xx : yy);
|
host_ppc_isel.c | 487 PPCRH_Imm(True,toUShort(n)) )); 495 PPCRH_Imm(True,toUShort(n)) )); 511 PPCRH_Imm(True,toUShort(16)) )); 628 PPCRH_Imm(True/*signed*/, toUShort(bias)))); 632 PPCRH_Imm(False/*unsigned*/, toUShort(nElems-1)))); 637 PPCRH_Imm(False/*unsigned*/, toUShort(shift)))); 641 PPCRH_Imm(True/*signed*/, toUShort(descr->base)))); [all...] |
host_mips_isel.c | 193 toUShort(n)))); 201 MIPSRH_Imm(True, toUShort(n)))); [all...] |
ir_opt.c | [all...] |
host_x86_defs.c | [all...] |
guest_generic_bb_to_IR.c | 423 = toUShort(toUInt( vge->len[vge->n_used-1] + dres.len )); [all...] |
guest_x86_toIR.c | [all...] |
guest_amd64_toIR.c | [all...] |
/external/valgrind/main/VEX/pub/ |
libvex_basictypes.h | 102 static inline UShort toUShort ( Int x ) {
|