1 ; RUN: opt -instcombine -S < %s | FileCheck %s 2 3 %overflow.result = type {i8, i1} 4 %ov.result.32 = type { i32, i1 } 5 6 7 declare %overflow.result @llvm.uadd.with.overflow.i8(i8, i8) nounwind readnone 8 declare %overflow.result @llvm.umul.with.overflow.i8(i8, i8) nounwind readnone 9 declare %ov.result.32 @llvm.sadd.with.overflow.i32(i32, i32) nounwind readnone 10 declare %ov.result.32 @llvm.uadd.with.overflow.i32(i32, i32) nounwind readnone 11 declare %ov.result.32 @llvm.ssub.with.overflow.i32(i32, i32) nounwind readnone 12 declare %ov.result.32 @llvm.usub.with.overflow.i32(i32, i32) nounwind readnone 13 declare %ov.result.32 @llvm.smul.with.overflow.i32(i32, i32) nounwind readnone 14 declare %ov.result.32 @llvm.umul.with.overflow.i32(i32, i32) nounwind readnone 15 declare double @llvm.powi.f64(double, i32) nounwind readonly 16 declare i32 @llvm.cttz.i32(i32, i1) nounwind readnone 17 declare i32 @llvm.ctlz.i32(i32, i1) nounwind readnone 18 declare i32 @llvm.ctpop.i32(i32) nounwind readnone 19 declare i8 @llvm.ctlz.i8(i8, i1) nounwind readnone 20 declare double @llvm.cos.f64(double %Val) nounwind readonly 21 declare double @llvm.sin.f64(double %Val) nounwind readonly 22 declare double @llvm.floor.f64(double %Val) nounwind readonly 23 declare double @llvm.ceil.f64(double %Val) nounwind readonly 24 declare double @llvm.trunc.f64(double %Val) nounwind readonly 25 declare double @llvm.rint.f64(double %Val) nounwind readonly 26 declare double @llvm.nearbyint.f64(double %Val) nounwind readonly 27 28 define i8 @uaddtest1(i8 %A, i8 %B) { 29 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %A, i8 %B) 30 %y = extractvalue %overflow.result %x, 0 31 ret i8 %y 32 ; CHECK-LABEL: @uaddtest1( 33 ; CHECK-NEXT: %y = add i8 %A, %B 34 ; CHECK-NEXT: ret i8 %y 35 } 36 37 define i8 @uaddtest2(i8 %A, i8 %B, i1* %overflowPtr) { 38 %and.A = and i8 %A, 127 39 %and.B = and i8 %B, 127 40 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %and.A, i8 %and.B) 41 %y = extractvalue %overflow.result %x, 0 42 %z = extractvalue %overflow.result %x, 1 43 store i1 %z, i1* %overflowPtr 44 ret i8 %y 45 ; CHECK-LABEL: @uaddtest2( 46 ; CHECK-NEXT: %and.A = and i8 %A, 127 47 ; CHECK-NEXT: %and.B = and i8 %B, 127 48 ; CHECK-NEXT: %x = add nuw i8 %and.A, %and.B 49 ; CHECK-NEXT: store i1 false, i1* %overflowPtr 50 ; CHECK-NEXT: ret i8 %x 51 } 52 53 define i8 @uaddtest3(i8 %A, i8 %B, i1* %overflowPtr) { 54 %or.A = or i8 %A, -128 55 %or.B = or i8 %B, -128 56 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %or.A, i8 %or.B) 57 %y = extractvalue %overflow.result %x, 0 58 %z = extractvalue %overflow.result %x, 1 59 store i1 %z, i1* %overflowPtr 60 ret i8 %y 61 ; CHECK-LABEL: @uaddtest3( 62 ; CHECK-NEXT: %or.A = or i8 %A, -128 63 ; CHECK-NEXT: %or.B = or i8 %B, -128 64 ; CHECK-NEXT: %x = add i8 %or.A, %or.B 65 ; CHECK-NEXT: store i1 true, i1* %overflowPtr 66 ; CHECK-NEXT: ret i8 %x 67 } 68 69 define i8 @uaddtest4(i8 %A, i1* %overflowPtr) { 70 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 undef, i8 %A) 71 %y = extractvalue %overflow.result %x, 0 72 %z = extractvalue %overflow.result %x, 1 73 store i1 %z, i1* %overflowPtr 74 ret i8 %y 75 ; CHECK-LABEL: @uaddtest4( 76 ; CHECK-NEXT: ret i8 undef 77 } 78 79 define i8 @uaddtest5(i8 %A, i1* %overflowPtr) { 80 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 0, i8 %A) 81 %y = extractvalue %overflow.result %x, 0 82 %z = extractvalue %overflow.result %x, 1 83 store i1 %z, i1* %overflowPtr 84 ret i8 %y 85 ; CHECK-LABEL: @uaddtest5( 86 ; CHECK: ret i8 %A 87 } 88 89 define i1 @uaddtest6(i8 %A, i8 %B) { 90 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %A, i8 -4) 91 %z = extractvalue %overflow.result %x, 1 92 ret i1 %z 93 ; CHECK-LABEL: @uaddtest6( 94 ; CHECK-NEXT: %z = icmp ugt i8 %A, 3 95 ; CHECK-NEXT: ret i1 %z 96 } 97 98 define i8 @uaddtest7(i8 %A, i8 %B) { 99 %x = call %overflow.result @llvm.uadd.with.overflow.i8(i8 %A, i8 %B) 100 %z = extractvalue %overflow.result %x, 0 101 ret i8 %z 102 ; CHECK-LABEL: @uaddtest7( 103 ; CHECK-NEXT: %z = add i8 %A, %B 104 ; CHECK-NEXT: ret i8 %z 105 } 106 107 ; PR20194 108 define %ov.result.32 @saddtest_nsw(i8 %a, i8 %b) { 109 %A = sext i8 %a to i32 110 %B = sext i8 %b to i32 111 %x = call %ov.result.32 @llvm.sadd.with.overflow.i32(i32 %A, i32 %B) 112 ret %ov.result.32 %x 113 ; CHECK-LABEL: @saddtest_nsw 114 ; CHECK: %x = add nsw i32 %A, %B 115 ; CHECK-NEXT: %1 = insertvalue %ov.result.32 { i32 undef, i1 false }, i32 %x, 0 116 ; CHECK-NEXT: ret %ov.result.32 %1 117 } 118 119 define %ov.result.32 @uaddtest_nuw(i32 %a, i32 %b) { 120 %A = and i32 %a, 2147483647 121 %B = and i32 %b, 2147483647 122 %x = call %ov.result.32 @llvm.uadd.with.overflow.i32(i32 %A, i32 %B) 123 ret %ov.result.32 %x 124 ; CHECK-LABEL: @uaddtest_nuw 125 ; CHECK: %x = add nuw i32 %A, %B 126 ; CHECK-NEXT: %1 = insertvalue %ov.result.32 { i32 undef, i1 false }, i32 %x, 0 127 ; CHECK-NEXT: ret %ov.result.32 %1 128 } 129 130 define %ov.result.32 @ssubtest_nsw(i8 %a, i8 %b) { 131 %A = sext i8 %a to i32 132 %B = sext i8 %b to i32 133 %x = call %ov.result.32 @llvm.ssub.with.overflow.i32(i32 %A, i32 %B) 134 ret %ov.result.32 %x 135 ; CHECK-LABEL: @ssubtest_nsw 136 ; CHECK: %x = sub nsw i32 %A, %B 137 ; CHECK-NEXT: %1 = insertvalue %ov.result.32 { i32 undef, i1 false }, i32 %x, 0 138 ; CHECK-NEXT: ret %ov.result.32 %1 139 } 140 141 define %ov.result.32 @usubtest_nuw(i32 %a, i32 %b) { 142 %A = or i32 %a, 2147483648 143 %B = and i32 %b, 2147483647 144 %x = call %ov.result.32 @llvm.usub.with.overflow.i32(i32 %A, i32 %B) 145 ret %ov.result.32 %x 146 ; CHECK-LABEL: @usubtest_nuw 147 ; CHECK: %x = sub nuw i32 %A, %B 148 ; CHECK-NEXT: %1 = insertvalue %ov.result.32 { i32 undef, i1 false }, i32 %x, 0 149 ; CHECK-NEXT: ret %ov.result.32 %1 150 } 151 152 define %ov.result.32 @smultest1_nsw(i32 %a, i32 %b) { 153 %A = and i32 %a, 4095 ; 0xfff 154 %B = and i32 %b, 524287; 0x7ffff 155 %x = call %ov.result.32 @llvm.smul.with.overflow.i32(i32 %A, i32 %B) 156 ret %ov.result.32 %x 157 ; CHECK-LABEL: @smultest1_nsw 158 ; CHECK: %x = mul nuw nsw i32 %A, %B 159 ; CHECK-NEXT: %1 = insertvalue %ov.result.32 { i32 undef, i1 false }, i32 %x, 0 160 ; CHECK-NEXT: ret %ov.result.32 %1 161 } 162 163 define %ov.result.32 @smultest2_nsw(i32 %a, i32 %b) { 164 %A = ashr i32 %a, 16 165 %B = ashr i32 %b, 16 166 %x = call %ov.result.32 @llvm.smul.with.overflow.i32(i32 %A, i32 %B) 167 ret %ov.result.32 %x 168 ; CHECK-LABEL: @smultest2_nsw 169 ; CHECK: %x = mul nsw i32 %A, %B 170 ; CHECK-NEXT: %1 = insertvalue %ov.result.32 { i32 undef, i1 false }, i32 %x, 0 171 ; CHECK-NEXT: ret %ov.result.32 %1 172 } 173 174 define %ov.result.32 @smultest3_sw(i32 %a, i32 %b) { 175 %A = ashr i32 %a, 16 176 %B = ashr i32 %b, 15 177 %x = call %ov.result.32 @llvm.smul.with.overflow.i32(i32 %A, i32 %B) 178 ret %ov.result.32 %x 179 ; CHECK-LABEL: @smultest3_sw 180 ; CHECK: %x = call %ov.result.32 @llvm.smul.with.overflow.i32(i32 %A, i32 %B) 181 ; CHECK-NEXT: ret %ov.result.32 %x 182 } 183 184 define %ov.result.32 @umultest_nuw(i32 %a, i32 %b) { 185 %A = and i32 %a, 65535 ; 0xffff 186 %B = and i32 %b, 65535 ; 0xffff 187 %x = call %ov.result.32 @llvm.umul.with.overflow.i32(i32 %A, i32 %B) 188 ret %ov.result.32 %x 189 ; CHECK-LABEL: @umultest_nuw 190 ; CHECK: %x = mul nuw i32 %A, %B 191 ; CHECK-NEXT: %1 = insertvalue %ov.result.32 { i32 undef, i1 false }, i32 %x, 0 192 ; CHECK-NEXT: ret %ov.result.32 %1 193 } 194 195 define i8 @umultest1(i8 %A, i1* %overflowPtr) { 196 %x = call %overflow.result @llvm.umul.with.overflow.i8(i8 0, i8 %A) 197 %y = extractvalue %overflow.result %x, 0 198 %z = extractvalue %overflow.result %x, 1 199 store i1 %z, i1* %overflowPtr 200 ret i8 %y 201 ; CHECK-LABEL: @umultest1( 202 ; CHECK-NEXT: store i1 false, i1* %overflowPtr 203 ; CHECK-NEXT: ret i8 0 204 } 205 206 define i8 @umultest2(i8 %A, i1* %overflowPtr) { 207 %x = call %overflow.result @llvm.umul.with.overflow.i8(i8 1, i8 %A) 208 %y = extractvalue %overflow.result %x, 0 209 %z = extractvalue %overflow.result %x, 1 210 store i1 %z, i1* %overflowPtr 211 ret i8 %y 212 ; CHECK-LABEL: @umultest2( 213 ; CHECK-NEXT: store i1 false, i1* %overflowPtr 214 ; CHECK-NEXT: ret i8 %A 215 } 216 217 define i32 @umultest3(i32 %n) nounwind { 218 %shr = lshr i32 %n, 2 219 %mul = call %ov.result.32 @llvm.umul.with.overflow.i32(i32 %shr, i32 3) 220 %ov = extractvalue %ov.result.32 %mul, 1 221 %res = extractvalue %ov.result.32 %mul, 0 222 %ret = select i1 %ov, i32 -1, i32 %res 223 ret i32 %ret 224 ; CHECK-LABEL: @umultest3( 225 ; CHECK-NEXT: shr 226 ; CHECK-NEXT: mul nuw 227 ; CHECK-NEXT: ret 228 } 229 230 define i32 @umultest4(i32 %n) nounwind { 231 %shr = lshr i32 %n, 1 232 %mul = call %ov.result.32 @llvm.umul.with.overflow.i32(i32 %shr, i32 4) 233 %ov = extractvalue %ov.result.32 %mul, 1 234 %res = extractvalue %ov.result.32 %mul, 0 235 %ret = select i1 %ov, i32 -1, i32 %res 236 ret i32 %ret 237 ; CHECK-LABEL: @umultest4( 238 ; CHECK: umul.with.overflow 239 } 240 241 define %ov.result.32 @umultest5(i32 %x, i32 %y) nounwind { 242 %or_x = or i32 %x, 2147483648 243 %or_y = or i32 %y, 2147483648 244 %mul = call %ov.result.32 @llvm.umul.with.overflow.i32(i32 %or_x, i32 %or_y) 245 ret %ov.result.32 %mul 246 ; CHECK-LABEL: @umultest5( 247 ; CHECK-NEXT: %[[or_x:.*]] = or i32 %x, -2147483648 248 ; CHECK-NEXT: %[[or_y:.*]] = or i32 %y, -2147483648 249 ; CHECK-NEXT: %[[mul:.*]] = mul i32 %[[or_x]], %[[or_y]] 250 ; CHECK-NEXT: %[[ret:.*]] = insertvalue %ov.result.32 { i32 undef, i1 true }, i32 %[[mul]], 0 251 ; CHECK-NEXT: ret %ov.result.32 %[[ret]] 252 } 253 254 define void @powi(double %V, double *%P) { 255 entry: 256 %A = tail call double @llvm.powi.f64(double %V, i32 -1) nounwind 257 store volatile double %A, double* %P 258 259 %B = tail call double @llvm.powi.f64(double %V, i32 0) nounwind 260 store volatile double %B, double* %P 261 262 %C = tail call double @llvm.powi.f64(double %V, i32 1) nounwind 263 store volatile double %C, double* %P 264 ret void 265 ; CHECK-LABEL: @powi( 266 ; CHECK: %A = fdiv double 1.0{{.*}}, %V 267 ; CHECK: store volatile double %A, 268 ; CHECK: store volatile double 1.0 269 ; CHECK: store volatile double %V 270 } 271 272 define i32 @cttz(i32 %a) { 273 entry: 274 %or = or i32 %a, 8 275 %and = and i32 %or, -8 276 %count = tail call i32 @llvm.cttz.i32(i32 %and, i1 true) nounwind readnone 277 ret i32 %count 278 ; CHECK-LABEL: @cttz( 279 ; CHECK-NEXT: entry: 280 ; CHECK-NEXT: ret i32 3 281 } 282 283 define i8 @ctlz(i8 %a) { 284 entry: 285 %or = or i8 %a, 32 286 %and = and i8 %or, 63 287 %count = tail call i8 @llvm.ctlz.i8(i8 %and, i1 true) nounwind readnone 288 ret i8 %count 289 ; CHECK-LABEL: @ctlz( 290 ; CHECK-NEXT: entry: 291 ; CHECK-NEXT: ret i8 2 292 } 293 294 define void @cmp.simplify(i32 %a, i32 %b, i1* %c) { 295 entry: 296 %lz = tail call i32 @llvm.ctlz.i32(i32 %a, i1 false) nounwind readnone 297 %lz.cmp = icmp eq i32 %lz, 32 298 store volatile i1 %lz.cmp, i1* %c 299 %tz = tail call i32 @llvm.cttz.i32(i32 %a, i1 false) nounwind readnone 300 %tz.cmp = icmp ne i32 %tz, 32 301 store volatile i1 %tz.cmp, i1* %c 302 %pop = tail call i32 @llvm.ctpop.i32(i32 %b) nounwind readnone 303 %pop.cmp = icmp eq i32 %pop, 0 304 store volatile i1 %pop.cmp, i1* %c 305 ret void 306 ; CHECK: @cmp.simplify 307 ; CHECK-NEXT: entry: 308 ; CHECK-NEXT: %lz.cmp = icmp eq i32 %a, 0 309 ; CHECK-NEXT: store volatile i1 %lz.cmp, i1* %c 310 ; CHECK-NEXT: %tz.cmp = icmp ne i32 %a, 0 311 ; CHECK-NEXT: store volatile i1 %tz.cmp, i1* %c 312 ; CHECK-NEXT: %pop.cmp = icmp eq i32 %b, 0 313 ; CHECK-NEXT: store volatile i1 %pop.cmp, i1* %c 314 } 315 316 define i32 @cttz_simplify1a(i32 %x) nounwind readnone ssp { 317 %tmp1 = tail call i32 @llvm.ctlz.i32(i32 %x, i1 false) 318 %shr3 = lshr i32 %tmp1, 5 319 ret i32 %shr3 320 321 ; CHECK-LABEL: @cttz_simplify1a( 322 ; CHECK: icmp eq i32 %x, 0 323 ; CHECK-NEXT: zext i1 324 ; CHECK-NEXT: ret i32 325 } 326 327 define i32 @cttz_simplify1b(i32 %x) nounwind readnone ssp { 328 %tmp1 = tail call i32 @llvm.ctlz.i32(i32 %x, i1 true) 329 %shr3 = lshr i32 %tmp1, 5 330 ret i32 %shr3 331 332 ; CHECK-LABEL: @cttz_simplify1b( 333 ; CHECK-NEXT: ret i32 0 334 } 335 336 define i32 @ctlz_undef(i32 %Value) nounwind { 337 %ctlz = call i32 @llvm.ctlz.i32(i32 0, i1 true) 338 ret i32 %ctlz 339 340 ; CHECK-LABEL: @ctlz_undef( 341 ; CHECK-NEXT: ret i32 undef 342 } 343 344 define i32 @cttz_undef(i32 %Value) nounwind { 345 %cttz = call i32 @llvm.cttz.i32(i32 0, i1 true) 346 ret i32 %cttz 347 348 ; CHECK-LABEL: @cttz_undef( 349 ; CHECK-NEXT: ret i32 undef 350 } 351 352 define i32 @ctlz_select(i32 %Value) nounwind { 353 %tobool = icmp ne i32 %Value, 0 354 %ctlz = call i32 @llvm.ctlz.i32(i32 %Value, i1 true) 355 %s = select i1 %tobool, i32 %ctlz, i32 32 356 ret i32 %s 357 358 ; CHECK-LABEL: @ctlz_select( 359 ; CHECK-NEXT: call i32 @llvm.ctlz.i32(i32 %Value, i1 false) 360 ; CHECK-NEXT: ret i32 361 } 362 363 define i32 @cttz_select(i32 %Value) nounwind { 364 %tobool = icmp ne i32 %Value, 0 365 %cttz = call i32 @llvm.cttz.i32(i32 %Value, i1 true) 366 %s = select i1 %tobool, i32 %cttz, i32 32 367 ret i32 %s 368 369 ; CHECK-LABEL: @cttz_select( 370 ; CHECK-NEXT: call i32 @llvm.cttz.i32(i32 %Value, i1 false) 371 ; CHECK-NEXT: ret i32 372 } 373 374 ; CHECK-LABEL: @overflow_div_add( 375 ; CHECK: ret i1 false 376 define i1 @overflow_div_add(i32 %v1, i32 %v2) nounwind { 377 entry: 378 %div = sdiv i32 %v1, 2 379 %t = call %ov.result.32 @llvm.sadd.with.overflow.i32(i32 %div, i32 1) 380 %obit = extractvalue %ov.result.32 %t, 1 381 ret i1 %obit 382 } 383 384 ; CHECK-LABEL: @overflow_div_sub( 385 ; CHECK: ret i1 false 386 define i1 @overflow_div_sub(i32 %v1, i32 %v2) nounwind { 387 entry: 388 ; Check cases where the known sign bits are larger than the word size. 389 %a = ashr i32 %v1, 18 390 %div = sdiv i32 %a, 65536 391 %t = call %ov.result.32 @llvm.ssub.with.overflow.i32(i32 %div, i32 1) 392 %obit = extractvalue %ov.result.32 %t, 1 393 ret i1 %obit 394 } 395 396 ; CHECK-LABEL: @overflow_mod_mul( 397 ; CHECK: ret i1 false 398 define i1 @overflow_mod_mul(i32 %v1, i32 %v2) nounwind { 399 entry: 400 %rem = srem i32 %v1, 1000 401 %t = call %ov.result.32 @llvm.smul.with.overflow.i32(i32 %rem, i32 %rem) 402 %obit = extractvalue %ov.result.32 %t, 1 403 ret i1 %obit 404 } 405 406 ; CHECK-LABEL: @overflow_mod_overflow_mul( 407 ; CHECK-NOT: ret i1 false 408 define i1 @overflow_mod_overflow_mul(i32 %v1, i32 %v2) nounwind { 409 entry: 410 %rem = srem i32 %v1, 65537 411 ; This may overflow because the result of the mul operands may be greater than 16bits 412 ; and the result greater than 32. 413 %t = call %ov.result.32 @llvm.smul.with.overflow.i32(i32 %rem, i32 %rem) 414 %obit = extractvalue %ov.result.32 %t, 1 415 ret i1 %obit 416 } 417 418 define %ov.result.32 @ssubtest_reorder(i8 %a) { 419 %A = sext i8 %a to i32 420 %x = call %ov.result.32 @llvm.ssub.with.overflow.i32(i32 0, i32 %A) 421 ret %ov.result.32 %x 422 ; CHECK-LABEL: @ssubtest_reorder 423 ; CHECK: %x = sub nsw i32 0, %A 424 ; CHECK-NEXT: %1 = insertvalue %ov.result.32 { i32 undef, i1 false }, i32 %x, 0 425 ; CHECK-NEXT: ret %ov.result.32 %1 426 } 427 428 define %ov.result.32 @never_overflows_ssub_test0(i32 %a) { 429 %x = call %ov.result.32 @llvm.ssub.with.overflow.i32(i32 %a, i32 0) 430 ret %ov.result.32 %x 431 ; CHECK-LABEL: @never_overflows_ssub_test0 432 ; CHECK-NEXT: %[[x:.*]] = insertvalue %ov.result.32 { i32 undef, i1 false }, i32 %a, 0 433 ; CHECK-NEXT: ret %ov.result.32 %[[x]] 434 } 435 436 define void @cos(double *%P) { 437 entry: 438 %B = tail call double @llvm.cos.f64(double 0.0) nounwind 439 store volatile double %B, double* %P 440 441 ret void 442 ; CHECK-LABEL: @cos( 443 ; CHECK: store volatile double 1.000000e+00, double* %P 444 } 445 446 define void @sin(double *%P) { 447 entry: 448 %B = tail call double @llvm.sin.f64(double 0.0) nounwind 449 store volatile double %B, double* %P 450 451 ret void 452 ; CHECK-LABEL: @sin( 453 ; CHECK: store volatile double 0.000000e+00, double* %P 454 } 455 456 define void @floor(double *%P) { 457 entry: 458 %B = tail call double @llvm.floor.f64(double 1.5) nounwind 459 store volatile double %B, double* %P 460 %C = tail call double @llvm.floor.f64(double -1.5) nounwind 461 store volatile double %C, double* %P 462 ret void 463 ; CHECK-LABEL: @floor( 464 ; CHECK: store volatile double 1.000000e+00, double* %P, align 8 465 ; CHECK: store volatile double -2.000000e+00, double* %P, align 8 466 } 467 468 define void @ceil(double *%P) { 469 entry: 470 %B = tail call double @llvm.ceil.f64(double 1.5) nounwind 471 store volatile double %B, double* %P 472 %C = tail call double @llvm.ceil.f64(double -1.5) nounwind 473 store volatile double %C, double* %P 474 ret void 475 ; CHECK-LABEL: @ceil( 476 ; CHECK: store volatile double 2.000000e+00, double* %P, align 8 477 ; CHECK: store volatile double -1.000000e+00, double* %P, align 8 478 } 479 480 define void @trunc(double *%P) { 481 entry: 482 %B = tail call double @llvm.trunc.f64(double 1.5) nounwind 483 store volatile double %B, double* %P 484 %C = tail call double @llvm.trunc.f64(double -1.5) nounwind 485 store volatile double %C, double* %P 486 ret void 487 ; CHECK-LABEL: @trunc( 488 ; CHECK: store volatile double 1.000000e+00, double* %P, align 8 489 ; CHECK: store volatile double -1.000000e+00, double* %P, align 8 490 } 491 492 define void @rint(double *%P) { 493 entry: 494 %B = tail call double @llvm.rint.f64(double 1.5) nounwind 495 store volatile double %B, double* %P 496 %C = tail call double @llvm.rint.f64(double -1.5) nounwind 497 store volatile double %C, double* %P 498 ret void 499 ; CHECK-LABEL: @rint( 500 ; CHECK: store volatile double 2.000000e+00, double* %P, align 8 501 ; CHECK: store volatile double -2.000000e+00, double* %P, align 8 502 } 503 504 define void @nearbyint(double *%P) { 505 entry: 506 %B = tail call double @llvm.nearbyint.f64(double 1.5) nounwind 507 store volatile double %B, double* %P 508 %C = tail call double @llvm.nearbyint.f64(double -1.5) nounwind 509 store volatile double %C, double* %P 510 ret void 511 ; CHECK-LABEL: @nearbyint( 512 ; CHECK: store volatile double 2.000000e+00, double* %P, align 8 513 ; CHECK: store volatile double -2.000000e+00, double* %P, align 8 514 } 515