; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; RUN: opt -simplifycfg -simplifycfg-require-and-preserve-domtree=1 -simplifycfg-max-small-block-size=6 -S < %s | FileCheck %s ; RUN: opt -passes=simplifycfg -simplifycfg-max-small-block-size=6 -S < %s | FileCheck %s target datalayout = "e-p:64:64-p5:32:32-A5" declare void @llvm.assume(i1) declare i1 @llvm.type.test(i8*, metadata) nounwind readnone define void @test_01(i1 %c, i64* align 1 %ptr) local_unnamed_addr #0 { ; CHECK-LABEL: @test_01( ; CHECK-NEXT: br i1 [[C:%.*]], label [[TRUE2_CRITEDGE:%.*]], label [[FALSE1:%.*]] ; CHECK: false1: ; CHECK-NEXT: store volatile i64 1, i64* [[PTR:%.*]], align 4 ; CHECK-NEXT: [[PTRINT:%.*]] = ptrtoint i64* [[PTR]] to i64 ; CHECK-NEXT: [[MASKEDPTR:%.*]] = and i64 [[PTRINT]], 7 ; CHECK-NEXT: [[MASKCOND:%.*]] = icmp eq i64 [[MASKEDPTR]], 0 ; CHECK-NEXT: tail call void @llvm.assume(i1 [[MASKCOND]]) ; CHECK-NEXT: store volatile i64 0, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 3, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET:%.*]] ; CHECK: common.ret: ; CHECK-NEXT: ret void ; CHECK: true2.critedge: ; CHECK-NEXT: [[PTRINT_C:%.*]] = ptrtoint i64* [[PTR]] to i64 ; CHECK-NEXT: [[MASKEDPTR_C:%.*]] = and i64 [[PTRINT_C]], 7 ; CHECK-NEXT: [[MASKCOND_C:%.*]] = icmp eq i64 [[MASKEDPTR_C]], 0 ; CHECK-NEXT: tail call void @llvm.assume(i1 [[MASKCOND_C]]) ; CHECK-NEXT: store volatile i64 0, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 2, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET]] ; br i1 %c, label %true1, label %false1 true1: ; preds = %false1, %0 %ptrint = ptrtoint i64* %ptr to i64 %maskedptr = and i64 %ptrint, 7 %maskcond = icmp eq i64 %maskedptr, 0 tail call void @llvm.assume(i1 %maskcond) store volatile i64 0, i64* %ptr, align 8 br i1 %c, label %true2, label %false2 false1: ; preds = %0 store volatile i64 1, i64* %ptr, align 4 br label %true1 true2: ; preds = %true1 store volatile i64 2, i64* %ptr, align 8 ret void false2: ; preds = %true1 store volatile i64 3, i64* %ptr, align 8 ret void } ; Corner case: the block has max possible size for which we still do PRE. define void @test_02(i1 %c, i64* align 1 %ptr) local_unnamed_addr #0 { ; CHECK-LABEL: @test_02( ; CHECK-NEXT: br i1 [[C:%.*]], label [[TRUE2_CRITEDGE:%.*]], label [[FALSE1:%.*]] ; CHECK: false1: ; CHECK-NEXT: store volatile i64 1, i64* [[PTR:%.*]], align 4 ; CHECK-NEXT: [[PTRINT:%.*]] = ptrtoint i64* [[PTR]] to i64 ; CHECK-NEXT: [[MASKEDPTR:%.*]] = and i64 [[PTRINT]], 7 ; CHECK-NEXT: [[MASKCOND:%.*]] = icmp eq i64 [[MASKEDPTR]], 0 ; CHECK-NEXT: tail call void @llvm.assume(i1 [[MASKCOND]]) ; CHECK-NEXT: store volatile i64 0, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 3, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET:%.*]] ; CHECK: common.ret: ; CHECK-NEXT: ret void ; CHECK: true2.critedge: ; CHECK-NEXT: [[PTRINT_C:%.*]] = ptrtoint i64* [[PTR]] to i64 ; CHECK-NEXT: [[MASKEDPTR_C:%.*]] = and i64 [[PTRINT_C]], 7 ; CHECK-NEXT: [[MASKCOND_C:%.*]] = icmp eq i64 [[MASKEDPTR_C]], 0 ; CHECK-NEXT: tail call void @llvm.assume(i1 [[MASKCOND_C]]) ; CHECK-NEXT: store volatile i64 0, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 2, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET]] ; br i1 %c, label %true1, label %false1 true1: ; preds = %false1, %0 %ptrint = ptrtoint i64* %ptr to i64 %maskedptr = and i64 %ptrint, 7 %maskcond = icmp eq i64 %maskedptr, 0 tail call void @llvm.assume(i1 %maskcond) store volatile i64 0, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 br i1 %c, label %true2, label %false2 false1: ; preds = %0 store volatile i64 1, i64* %ptr, align 4 br label %true1 true2: ; preds = %true1 store volatile i64 2, i64* %ptr, align 8 ret void false2: ; preds = %true1 store volatile i64 3, i64* %ptr, align 8 ret void } ; This block is too huge for PRE. define void @test_03(i1 %c, i64* align 1 %ptr) local_unnamed_addr #0 { ; CHECK-LABEL: @test_03( ; CHECK-NEXT: br i1 [[C:%.*]], label [[TRUE1:%.*]], label [[FALSE1:%.*]] ; CHECK: true1: ; CHECK-NEXT: [[PTRINT:%.*]] = ptrtoint i64* [[PTR:%.*]] to i64 ; CHECK-NEXT: [[MASKEDPTR:%.*]] = and i64 [[PTRINT]], 7 ; CHECK-NEXT: [[MASKCOND:%.*]] = icmp eq i64 [[MASKEDPTR]], 0 ; CHECK-NEXT: tail call void @llvm.assume(i1 [[MASKCOND]]) ; CHECK-NEXT: store volatile i64 0, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: br i1 [[C]], label [[TRUE2:%.*]], label [[FALSE2:%.*]] ; CHECK: false1: ; CHECK-NEXT: store volatile i64 1, i64* [[PTR]], align 4 ; CHECK-NEXT: br label [[TRUE1]] ; CHECK: common.ret: ; CHECK-NEXT: ret void ; CHECK: true2: ; CHECK-NEXT: store volatile i64 2, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET:%.*]] ; CHECK: false2: ; CHECK-NEXT: store volatile i64 3, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET]] ; br i1 %c, label %true1, label %false1 true1: ; preds = %false1, %0 %ptrint = ptrtoint i64* %ptr to i64 %maskedptr = and i64 %ptrint, 7 %maskcond = icmp eq i64 %maskedptr, 0 tail call void @llvm.assume(i1 %maskcond) store volatile i64 0, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 br i1 %c, label %true2, label %false2 false1: ; preds = %0 store volatile i64 1, i64* %ptr, align 4 br label %true1 true2: ; preds = %true1 store volatile i64 2, i64* %ptr, align 8 ret void false2: ; preds = %true1 store volatile i64 3, i64* %ptr, align 8 ret void } ; Try the max block size for PRE again but with the bitcast/type test/assume ; sequence used for whole program devirt. define void @test_04(i1 %c, i64* align 1 %ptr, [3 x i8*]* %vtable) local_unnamed_addr #0 { ; CHECK-LABEL: @test_04( ; CHECK-NEXT: br i1 [[C:%.*]], label [[TRUE2_CRITEDGE:%.*]], label [[FALSE1:%.*]] ; CHECK: false1: ; CHECK-NEXT: store volatile i64 1, i64* [[PTR:%.*]], align 4 ; CHECK-NEXT: [[VTABLEI8:%.*]] = bitcast [3 x i8*]* [[VTABLE:%.*]] to i8* ; CHECK-NEXT: [[P:%.*]] = call i1 @llvm.type.test(i8* [[VTABLEI8]], metadata !"foo") ; CHECK-NEXT: tail call void @llvm.assume(i1 [[P]]) ; CHECK-NEXT: store volatile i64 0, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 3, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET:%.*]] ; CHECK: common.ret: ; CHECK-NEXT: ret void ; CHECK: true2.critedge: ; CHECK-NEXT: [[VTABLEI8_C:%.*]] = bitcast [3 x i8*]* [[VTABLE]] to i8* ; CHECK-NEXT: [[P_C:%.*]] = call i1 @llvm.type.test(i8* [[VTABLEI8_C]], metadata !"foo") ; CHECK-NEXT: tail call void @llvm.assume(i1 [[P_C]]) ; CHECK-NEXT: store volatile i64 0, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 2, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET]] ; br i1 %c, label %true1, label %false1 true1: ; preds = %false1, %0 %vtablei8 = bitcast [3 x i8*]* %vtable to i8* %p = call i1 @llvm.type.test(i8* %vtablei8, metadata !"foo") tail call void @llvm.assume(i1 %p) store volatile i64 0, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 br i1 %c, label %true2, label %false2 false1: ; preds = %0 store volatile i64 1, i64* %ptr, align 4 br label %true1 true2: ; preds = %true1 store volatile i64 2, i64* %ptr, align 8 ret void false2: ; preds = %true1 store volatile i64 3, i64* %ptr, align 8 ret void } ; The load, icmp and assume should not count towards the limit, they are ; ephemeral. define void @test_non_speculatable(i1 %c, i64* align 1 %ptr, i8* %ptr2) local_unnamed_addr #0 { ; CHECK-LABEL: @test_non_speculatable( ; CHECK-NEXT: br i1 [[C:%.*]], label [[TRUE2_CRITEDGE:%.*]], label [[FALSE1:%.*]] ; CHECK: false1: ; CHECK-NEXT: store volatile i64 1, i64* [[PTR:%.*]], align 4 ; CHECK-NEXT: [[V:%.*]] = load i8, i8* [[PTR2:%.*]], align 1 ; CHECK-NEXT: [[C2:%.*]] = icmp eq i8 [[V]], 42 ; CHECK-NEXT: call void @llvm.assume(i1 [[C2]]) ; CHECK-NEXT: store volatile i64 0, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 3, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET:%.*]] ; CHECK: common.ret: ; CHECK-NEXT: ret void ; CHECK: true2.critedge: ; CHECK-NEXT: [[V_C:%.*]] = load i8, i8* [[PTR2]], align 1 ; CHECK-NEXT: [[C2_C:%.*]] = icmp eq i8 [[V_C]], 42 ; CHECK-NEXT: call void @llvm.assume(i1 [[C2_C]]) ; CHECK-NEXT: store volatile i64 0, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 -1, i64* [[PTR]], align 8 ; CHECK-NEXT: store volatile i64 2, i64* [[PTR]], align 8 ; CHECK-NEXT: br label [[COMMON_RET]] ; br i1 %c, label %true1, label %false1 true1: ; preds = %false1, %0 %v = load i8, i8* %ptr2 %c2 = icmp eq i8 %v, 42 call void @llvm.assume(i1 %c2) store volatile i64 0, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 store volatile i64 -1, i64* %ptr, align 8 br i1 %c, label %true2, label %false2 false1: ; preds = %0 store volatile i64 1, i64* %ptr, align 4 br label %true1 true2: ; preds = %true1 store volatile i64 2, i64* %ptr, align 8 ret void false2: ; preds = %true1 store volatile i64 3, i64* %ptr, align 8 ret void }