Home | History | Annotate | Download | only in X86

Lines Matching full:movsd

542 // Move Instructions. Register-to-register movss/movsd is not used for FR32/64
544 // movss/movsd is not modeled as an INSERT_SUBREG because INSERT_SUBREG requires
546 // don't use movss/movsd for copies.
607 defm MOVSD : sse12_move<FR64, X86Movsd, v2f64, f64mem, "movsd",
615 defm MOVSD : sse12_move_rm<FR64, f64mem, loadf64, "movsd",
755 // MOVSD to the lower bits.
779 // Shuffle with MOVSD
3194 // With SSE 4.1, blendi is preferred to movsd, so match that too.
3239 // extracted scalar math op with insert via movsd
3246 // vector math op with insert via movsd
3252 // With SSE 4.1, blendi is preferred to movsd, so match those too.
3269 // extracted scalar math op with insert via movsd
3283 // vector math op with insert via movsd
7159 // FIXME: Prefer a movss or movsd over a blendps when optimizing for size or
7205 // FIXME: Prefer a movss or movsd over a blendps when optimizing for size or