Home | History | Annotate | Download | only in memcheck

Lines Matching refs:Ity_I64

198          Ity_I32 or Ity_I64 only. */
363 case Ity_I64:
366 case Ity_F64: return Ity_I64;
382 case Ity_I64: return IRExpr_Const(IRConst_U64(0));
489 return assignNew('V', mce, Ity_I64, binop(Iop_And64, a1, a2));
521 return assignNew('V', mce, Ity_I64, binop(Iop_Or64, a1, a2));
528 tmp1 = assignNew('V', mce, Ity_I64, unop(Iop_128to64, a1));
529 tmp2 = assignNew('V', mce, Ity_I64, unop(Iop_128HIto64, a1));
530 tmp3 = assignNew('V', mce, Ity_I64, unop(Iop_128to64, a2));
531 tmp4 = assignNew('V', mce, Ity_I64, unop(Iop_128HIto64, a2));
532 tmp5 = assignNew('V', mce, Ity_I64, binop(Iop_Or64, tmp1, tmp3));
533 tmp6 = assignNew('V', mce, Ity_I64, binop(Iop_Or64, tmp2, tmp4));
549 case Ity_I64: return mkUifU64(mce, a1, a2);
577 return assignNew('V', mce, Ity_I64, unop(Iop_Left64, a1));
614 return assignNew('V', mce, Ity_I64, binop(Iop_Or64, data, vbits));
670 'V', mce, Ity_I64,
672 assignNew('V', mce, Ity_I64, unop(Iop_Not64, data)),
707 if (src_ty == Ity_I64 && dst_ty == Ity_I64)
708 return assignNew('V', mce, Ity_I64, unop(Iop_CmpwNEZ64, vbits));
710 if (src_ty == Ity_I32 && dst_ty == Ity_I64) {
712 return assignNew('V', mce, Ity_I64, binop(Iop_32HLto64, tmp, tmp));
730 case Ity_I64:
736 IRAtom* tmp2 = assignNew('V', mce, Ity_I64, unop(Iop_128HIto64, vbits));
737 IRAtom* tmp3 = assignNew('V', mce, Ity_I64, unop(Iop_128to64, vbits));
738 IRAtom* tmp4 = assignNew('V', mce, Ity_I64, binop(Iop_Or64, tmp2, tmp3));
758 case Ity_I64:
759 return assignNew('V', mce, Ity_I64, unop(Iop_1Sto64, tmp1));
761 tmp1 = assignNew('V', mce, Ity_I64, unop(Iop_1Sto64, tmp1));
765 tmp1 = assignNew('V', mce, Ity_I64, unop(Iop_1Sto64, tmp1));
843 case Ity_I64:
940 IRType ty = m64 ? Ity_I64 : Ity_I32;
1073 if (mce->hWordTy == Ity_I64) {
1074 origin = assignNew( 'B', mce, Ity_I64, unop(Iop_32Uto64, origin) );
1370 if (t1 == Ity_I64 && t2 == Ity_I64 && finalVty == Ity_I64) {
1372 at = mkUifU(mce, Ity_I64, va1, va2);
1373 at = mkPCastTo(mce, Ity_I64, at);
1378 if (t1 == Ity_I64 && t2 == Ity_I64 && finalVty == Ity_I32) {
1380 at = mkUifU(mce, Ity_I64, va1, va2);
1422 if (t1 == Ity_I32 && t2 == Ity_I64 && t3 == Ity_I64
1423 && finalVty == Ity_I64) {
1428 at = mkPCastTo(mce, Ity_I64, va1);
1430 at = mkUifU(mce, Ity_I64, at, va2);
1431 at = mkUifU(mce, Ity_I64, at, va3);
1433 at = mkPCastTo(mce, Ity_I64, at);
1438 if (t1 == Ity_I32 && t2 == Ity_I64 && t3 == Ity_I64
1441 at = mkPCastTo(mce, Ity_I64, va1);
1442 at = mkUifU(mce, Ity_I64, at, va2);
1443 at = mkUifU(mce, Ity_I64, at, va3);
1521 if (t1 == Ity_I32 && t2 == Ity_I64 && t3 == Ity_I64 && t4 == Ity_I64
1522 && finalVty == Ity_I64) {
1527 at = mkPCastTo(mce, Ity_I64, va1);
1529 at = mkUifU(mce, Ity_I64, at, va2);
1530 at = mkUifU(mce, Ity_I64, at, va3);
1531 at = mkUifU(mce, Ity_I64, at, va4);
1533 at = mkPCastTo(mce, Ity_I64, at);
1591 if (typeOfIRExpr(mce->sb->tyenv, exprvec[i]) != Ity_I64)
1595 mergeTy = mergeTy64 ? Ity_I64 : Ity_I32;
1651 case Ity_I64:
1786 return assignNew('V', mce, Ity_I64, unop(Iop_CmpNEZ32x2, at));
1791 return assignNew('V', mce, Ity_I64, unop(Iop_CmpNEZ16x4, at));
1796 return assignNew('V', mce, Ity_I64, unop(Iop_CmpNEZ8x8, at));
1919 at = assignNew('V', mce, Ity_I64, unop(Iop_V128to64, at));
1920 at = mkPCastTo(mce, Ity_I64, at);
1930 at = assignNew('V', mce, Ity_I64, unop(Iop_V128to64, vatomX));
1931 at = mkPCastTo(mce, Ity_I64, at);
1945 at = assignNew('V', mce, Ity_I64, mkPCast32x2(mce, at));
1954 at = assignNew('V', mce, Ity_I64, mkPCast32x2(mce, vatomX));
2084 at1 = assignNew('V', mce, Ity_I64, pcast(mce, vatom1));
2085 at2 = assignNew('V', mce, Ity_I64, pcast(mce, vatom2));
2086 at3 = assignNew('V', mce, Ity_I64, binop(vanilla_narrow, at1, at2));
2103 at1 = assignNew('V', mce, Ity_I64, unop(narrow_op, vatom1));
2124 at2 = assignNew('V', mce, Ity_I64, unop(vanilla_narrow, at1));
2226 at = mkPCastTo(mce, Ity_I64, at);
2284 return mkLazy4(mce, Ity_I64, vatom1, vatom2, vatom3, vatom4);
2338 return mkLazy3(mce, Ity_I64, vatom1, vatom2, vatom3);
2354 return assignNew('V', mce, Ity_I64, triop(op, vatom1, vatom2, atom3));
2359 return assignNew('V', mce, Ity_I64, triop(op, vatom1, atom2, vatom3));
2425 return assignNew('V', mce, Ity_I64, binop(op, vatom1, atom2));
2534 return assignNew('V', mce, Ity_I64, binop(Iop_PwMax32Ux2, mkPCast32x2(mce, vatom1),
2541 return assignNew('V', mce, Ity_I64, binop(Iop_PwMax16Ux4, mkPCast16x4(mce, vatom1),
2548 return assignNew('V', mce, Ity_I64, binop(Iop_PwMax8Ux8, mkPCast8x8(mce, vatom1),
2554 assignNew('V', mce, Ity_I64, binop(Iop_PwAdd32x2, mkPCast32x2(mce, vatom1),
2559 assignNew('V', mce, Ity_I64, binop(op, mkPCast16x4(mce, vatom1),
2564 assignNew('V', mce, Ity_I64, binop(op, mkPCast8x8(mce, vatom1),
2572 assignNew('V', mce, Ity_I64, binop(op, vatom1, atom2)),
2581 assignNew('V', mce, Ity_I64, binop(op, vatom1, atom2)),
2590 assignNew('V', mce, Ity_I64, binop(op, vatom1, atom2)),
2609 return assignNew('V', mce, Ity_I64, binop(op, vatom1, vatom2));
2627 assignNew('V', mce, Ity_I64, binop(op, vatom1, atom2)),
2950 return assignNew('V', mce, Ity_I64, binop(op, vatom1, atom2));
3013 return mkLazy2(mce, Ity_I64, vatom1, vatom2);
3031 return mkLazy2(mce, Ity_I64, vatom1, vatom2);
3053 return mkLazy2(mce, Ity_I64, vatom1, vatom2);
3078 return mkLazy2(mce, Ity_I64, vatom1, vatom2);
3087 return assignNew('V', mce, Ity_I64, binop(op, vatom1, vatom2));
3093 IRAtom* vHi64 = mkPCastTo(mce, Ity_I64, vLo64);
3101 return assignNew('V', mce, Ity_I64, binop(Iop_32HLto64, vHi32, vLo32));
3129 return mkLazy2(mce, Ity_I64, vatom1, vatom2);
3156 return expensiveAddSub(mce,True,Ity_I64,
3162 return expensiveAddSub(mce,False,Ity_I64,
3183 return expensiveCmpEQorNE(mce,Ity_I64, vatom1,vatom2, atom1,atom2 );
3217 return scalarShift( mce, Ity_I64, op, vatom1,vatom2, atom1,atom2 );
3233 and_or_ty = Ity_I64; improve = mkImproveAND64; goto do_And_Or;
3249 and_or_ty = Ity_I64; improve = mkImproveOR64; goto do_And_Or;
3345 return assignNew('V', mce, Ity_I64, unop(Iop_128HIto64, vatom));
3347 return assignNew('V', mce, Ity_I64, unop(Iop_128to64, vatom));
3371 return mkPCastTo(mce, Ity_I64, vatom);
3401 return assignNew('V', mce, Ity_I64, unop(op, vatom));
3491 return mkPCastTo(mce, Ity_I64, vatom);
3520 return mkPCastTo(mce, Ity_I64,
3521 assignNew('V', mce, Ity_I64, unop(op, mkPCast32x2(mce, vatom))));
3526 assignNew('V', mce, Ity_I64, unop(op, mkPCast16x4(mce, vatom))));
3531 assignNew('V', mce, Ity_I64, unop(op, mkPCast8x8(mce, vatom))));
3581 case Ity_I64: helper = &MC_(helperc_LOADV64le);
3598 case Ity_I64: helper = &MC_(helperc_LOADV64be);
3622 tl_assert( tyAddr == Ity_I32 || tyAddr == Ity_I64 );
3653 case Ity_I64:
3657 v64lo = expr2vbits_Load_WRK(mce, end, Ity_I64, addr, bias);
3658 v64hi = expr2vbits_Load_WRK(mce, end, Ity_I64, addr, bias+8);
3660 v64hi = expr2vbits_Load_WRK(mce, end, Ity_I64, addr, bias);
3661 v64lo = expr2vbits_Load_WRK(mce, end, Ity_I64, addr, bias+8);
3794 if (tyH == Ity_I64) {
3838 tl_assert( tyAddr == Ity_I32 || tyAddr == Ity_I64 );
3867 case Ity_I64: c = IRConst_U64 (V_BITS64_DEFINED); break;
3885 case Ity_I64: helper = &MC_(helperc_STOREV64le);
3902 case Ity_I64: helper = &MC_(helperc_STOREV64be);
3940 vdataLo64 = assignNew('V', mce, Ity_I64, unop(Iop_V128to64, vdata));
3948 vdataHi64 = assignNew('V', mce, Ity_I64, unop(Iop_V128HIto64, vdata));
3975 if (ty == Ity_I64) {
4010 case 8: return Ity_I64;
4098 tl_assert(tyAddr == Ity_I32 || tyAddr == Ity_I64);
4445 case Ity_I64: elemSzB = 8; opCasCmpEQ = Iop_CasCmpEQ64; break;
4547 case Ity_I64:
4697 tl_assert(resTy == Ity_I64 || resTy == Ity_I32
4707 tl_assert(dataTy == Ity_I64 || dataTy == Ity_I32
5377 if (mce->hWordTy == Ity_I64) {
5409 if (mce->hWordTy == Ity_I64)
5410 dataB = assignNew( 'B', mce, Ity_I64, unop(Iop_32Uto64, dataB));
5443 if (eTy == Ity_I64)
5453 if (dstTy == Ity_I64)
5454 return assignNew( 'B', mce, Ity_I64, unop(Iop_32Uto64, e) );
5523 tl_assert(mce->hWordTy == Ity_I32 || mce->hWordTy == Ity_I64);
5833 tl_assert(resTy == Ity_I64 || resTy == Ity_I32