1
+ ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py UTC_ARGS: --version 5
1
2
; RUN: opt < %s -force-vector-width=2 -force-vector-interleave=2 -passes=loop-vectorize -S | FileCheck %s
2
3
3
4
target datalayout = "e-m:e-i64:64-i128:128-n32:64-S128"
4
5
5
- ; CHECK-LABEL: @dead_instructions_01
6
6
;
7
7
; This test ensures that we don't generate trivially dead instructions prior to
8
8
; instruction simplification. We don't need to generate instructions
9
9
; corresponding to the original induction variable update or branch condition,
10
10
; since we rewrite the loop structure.
11
11
;
12
- ; CHECK: vector.body:
13
- ; CHECK: %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
14
- ; CHECK: %[[I0:.+]] = add i64 %index, 0
15
- ; CHECK: %[[I2:.+]] = add i64 %index, 2
16
- ; CHECK: getelementptr inbounds i64, ptr %a, i64 %[[I0]]
17
- ; CHECK: getelementptr inbounds i64, ptr %a, i64 %[[I2]]
18
- ; CHECK-NOT: add nuw nsw i64 %[[I0]], 1
19
- ; CHECK-NOT: add nuw nsw i64 %[[I2]], 1
20
- ; CHECK-NOT: icmp slt i64 {{.*}}, %n
21
- ; CHECK: %index.next = add nuw i64 %index, 4
22
- ; CHECK: %[[CMP:.+]] = icmp eq i64 %index.next, %n.vec
23
- ; CHECK: br i1 %[[CMP]], label %middle.block, label %vector.body
24
- ;
25
12
define i64 @dead_instructions_01 (ptr %a , i64 %n ) {
13
+ ; CHECK-LABEL: define i64 @dead_instructions_01(
14
+ ; CHECK-SAME: ptr [[A:%.*]], i64 [[N:%.*]]) {
15
+ ; CHECK-NEXT: [[ENTRY:.*]]:
16
+ ; CHECK-NEXT: [[SMAX:%.*]] = call i64 @llvm.smax.i64(i64 [[N]], i64 1)
17
+ ; CHECK-NEXT: [[MIN_ITERS_CHECK:%.*]] = icmp ult i64 [[SMAX]], 4
18
+ ; CHECK-NEXT: br i1 [[MIN_ITERS_CHECK]], label %[[SCALAR_PH:.*]], label %[[VECTOR_PH:.*]]
19
+ ; CHECK: [[VECTOR_PH]]:
20
+ ; CHECK-NEXT: [[N_MOD_VF:%.*]] = urem i64 [[SMAX]], 4
21
+ ; CHECK-NEXT: [[N_VEC:%.*]] = sub i64 [[SMAX]], [[N_MOD_VF]]
22
+ ; CHECK-NEXT: br label %[[VECTOR_BODY:.*]]
23
+ ; CHECK: [[VECTOR_BODY]]:
24
+ ; CHECK-NEXT: [[INDEX:%.*]] = phi i64 [ 0, %[[VECTOR_PH]] ], [ [[INDEX_NEXT:%.*]], %[[VECTOR_BODY]] ]
25
+ ; CHECK-NEXT: [[VEC_PHI:%.*]] = phi <2 x i64> [ zeroinitializer, %[[VECTOR_PH]] ], [ [[TMP6:%.*]], %[[VECTOR_BODY]] ]
26
+ ; CHECK-NEXT: [[VEC_PHI1:%.*]] = phi <2 x i64> [ zeroinitializer, %[[VECTOR_PH]] ], [ [[TMP7:%.*]], %[[VECTOR_BODY]] ]
27
+ ; CHECK-NEXT: [[TMP0:%.*]] = add i64 [[INDEX]], 0
28
+ ; CHECK-NEXT: [[TMP1:%.*]] = add i64 [[INDEX]], 2
29
+ ; CHECK-NEXT: [[TMP2:%.*]] = getelementptr inbounds i64, ptr [[A]], i64 [[TMP0]]
30
+ ; CHECK-NEXT: [[TMP3:%.*]] = getelementptr inbounds i64, ptr [[A]], i64 [[TMP1]]
31
+ ; CHECK-NEXT: [[TMP4:%.*]] = getelementptr inbounds i64, ptr [[TMP2]], i32 0
32
+ ; CHECK-NEXT: [[TMP5:%.*]] = getelementptr inbounds i64, ptr [[TMP2]], i32 2
33
+ ; CHECK-NEXT: [[WIDE_LOAD:%.*]] = load <2 x i64>, ptr [[TMP4]], align 8
34
+ ; CHECK-NEXT: [[WIDE_LOAD2:%.*]] = load <2 x i64>, ptr [[TMP5]], align 8
35
+ ; CHECK-NEXT: [[TMP6]] = add <2 x i64> [[WIDE_LOAD]], [[VEC_PHI]]
36
+ ; CHECK-NEXT: [[TMP7]] = add <2 x i64> [[WIDE_LOAD2]], [[VEC_PHI1]]
37
+ ; CHECK-NEXT: [[INDEX_NEXT]] = add nuw i64 [[INDEX]], 4
38
+ ; CHECK-NEXT: [[TMP8:%.*]] = icmp eq i64 [[INDEX_NEXT]], [[N_VEC]]
39
+ ; CHECK-NEXT: br i1 [[TMP8]], label %[[MIDDLE_BLOCK:.*]], label %[[VECTOR_BODY]], !llvm.loop [[LOOP0:![0-9]+]]
40
+ ; CHECK: [[MIDDLE_BLOCK]]:
41
+ ; CHECK-NEXT: [[BIN_RDX:%.*]] = add <2 x i64> [[TMP7]], [[TMP6]]
42
+ ; CHECK-NEXT: [[TMP9:%.*]] = call i64 @llvm.vector.reduce.add.v2i64(<2 x i64> [[BIN_RDX]])
43
+ ; CHECK-NEXT: [[CMP_N:%.*]] = icmp eq i64 [[SMAX]], [[N_VEC]]
44
+ ; CHECK-NEXT: br i1 [[CMP_N]], label %[[FOR_END:.*]], label %[[SCALAR_PH]]
45
+ ; CHECK: [[SCALAR_PH]]:
46
+ ; CHECK-NEXT: [[BC_RESUME_VAL:%.*]] = phi i64 [ [[N_VEC]], %[[MIDDLE_BLOCK]] ], [ 0, %[[ENTRY]] ]
47
+ ; CHECK-NEXT: [[BC_MERGE_RDX:%.*]] = phi i64 [ 0, %[[ENTRY]] ], [ [[TMP9]], %[[MIDDLE_BLOCK]] ]
48
+ ; CHECK-NEXT: br label %[[FOR_BODY:.*]]
49
+ ; CHECK: [[FOR_BODY]]:
50
+ ; CHECK-NEXT: [[I:%.*]] = phi i64 [ [[I_NEXT:%.*]], %[[FOR_BODY]] ], [ [[BC_RESUME_VAL]], %[[SCALAR_PH]] ]
51
+ ; CHECK-NEXT: [[R:%.*]] = phi i64 [ [[TMP2:%.*]], %[[FOR_BODY]] ], [ [[BC_MERGE_RDX]], %[[SCALAR_PH]] ]
52
+ ; CHECK-NEXT: [[TMP0:%.*]] = getelementptr inbounds i64, ptr [[A]], i64 [[I]]
53
+ ; CHECK-NEXT: [[TMP1:%.*]] = load i64, ptr [[TMP0]], align 8
54
+ ; CHECK-NEXT: [[TMP2]] = add i64 [[TMP1]], [[R]]
55
+ ; CHECK-NEXT: [[I_NEXT]] = add nuw nsw i64 [[I]], 1
56
+ ; CHECK-NEXT: [[COND:%.*]] = icmp slt i64 [[I_NEXT]], [[N]]
57
+ ; CHECK-NEXT: br i1 [[COND]], label %[[FOR_BODY]], label %[[FOR_END]], !llvm.loop [[LOOP3:![0-9]+]]
58
+ ; CHECK: [[FOR_END]]:
59
+ ; CHECK-NEXT: [[TMP3:%.*]] = phi i64 [ [[TMP2]], %[[FOR_BODY]] ], [ [[TMP9]], %[[MIDDLE_BLOCK]] ]
60
+ ; CHECK-NEXT: ret i64 [[TMP3]]
61
+ ;
26
62
entry:
27
63
br label %for.body
28
64
@@ -42,16 +78,44 @@ for.end:
42
78
}
43
79
44
80
45
- ; CHECK-LABEL: @pr47390
46
81
;
47
82
; This test ensures that the primary induction is not considered dead when
48
83
; acting as the 'add' of another induction, and otherwise feeding only its own
49
84
; 'add' (recognized earlier as 'dead'), when the tail of the loop is folded by
50
85
; masking. Such masking uses the primary induction.
51
86
;
52
- ; CHECK: vector.body:
53
- ;
54
87
define void @pr47390 (ptr %a ) {
88
+ ; CHECK-LABEL: define void @pr47390(
89
+ ; CHECK-SAME: ptr [[A:%.*]]) {
90
+ ; CHECK-NEXT: [[ENTRY:.*]]:
91
+ ; CHECK-NEXT: br i1 false, label %[[SCALAR_PH:.*]], label %[[VECTOR_PH:.*]]
92
+ ; CHECK: [[VECTOR_PH]]:
93
+ ; CHECK-NEXT: br label %[[VECTOR_BODY:.*]]
94
+ ; CHECK: [[VECTOR_BODY]]:
95
+ ; CHECK-NEXT: [[INDEX:%.*]] = phi i32 [ 0, %[[VECTOR_PH]] ], [ [[INDEX_NEXT:%.*]], %[[VECTOR_BODY]] ]
96
+ ; CHECK-NEXT: [[INDEX_NEXT]] = add i32 [[INDEX]], 4
97
+ ; CHECK-NEXT: [[TMP0:%.*]] = icmp eq i32 [[INDEX_NEXT]], 8
98
+ ; CHECK-NEXT: br i1 [[TMP0]], label %[[MIDDLE_BLOCK:.*]], label %[[VECTOR_BODY]], !llvm.loop [[LOOP4:![0-9]+]]
99
+ ; CHECK: [[MIDDLE_BLOCK]]:
100
+ ; CHECK-NEXT: br i1 true, label %[[EXIT:.*]], label %[[SCALAR_PH]]
101
+ ; CHECK: [[SCALAR_PH]]:
102
+ ; CHECK-NEXT: [[BC_RESUME_VAL:%.*]] = phi i32 [ 8, %[[MIDDLE_BLOCK]] ], [ 0, %[[ENTRY]] ]
103
+ ; CHECK-NEXT: [[BC_RESUME_VAL1:%.*]] = phi i32 [ 7, %[[MIDDLE_BLOCK]] ], [ -1, %[[ENTRY]] ]
104
+ ; CHECK-NEXT: [[BC_RESUME_VAL2:%.*]] = phi i32 [ 9, %[[MIDDLE_BLOCK]] ], [ 1, %[[ENTRY]] ]
105
+ ; CHECK-NEXT: br label %[[LOOP:.*]]
106
+ ; CHECK: [[EXIT]]:
107
+ ; CHECK-NEXT: ret void
108
+ ; CHECK: [[LOOP]]:
109
+ ; CHECK-NEXT: [[PRIMARY:%.*]] = phi i32 [ [[BC_RESUME_VAL]], %[[SCALAR_PH]] ], [ [[PRIMARY_ADD:%.*]], %[[LOOP]] ]
110
+ ; CHECK-NEXT: [[USE_PRIMARY:%.*]] = phi i32 [ [[BC_RESUME_VAL1]], %[[SCALAR_PH]] ], [ [[PRIMARY]], %[[LOOP]] ]
111
+ ; CHECK-NEXT: [[SECONDARY:%.*]] = phi i32 [ [[BC_RESUME_VAL2]], %[[SCALAR_PH]] ], [ [[SECONDARY_ADD:%.*]], %[[LOOP]] ]
112
+ ; CHECK-NEXT: [[PRIMARY_ADD]] = add i32 [[PRIMARY]], 1
113
+ ; CHECK-NEXT: [[SECONDARY_ADD]] = add i32 [[SECONDARY]], 1
114
+ ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i32, ptr [[A]], i32 [[SECONDARY]]
115
+ ; CHECK-NEXT: [[LOAD:%.*]] = load i32, ptr [[GEP]], align 8
116
+ ; CHECK-NEXT: [[CMP:%.*]] = icmp eq i32 [[SECONDARY]], 5
117
+ ; CHECK-NEXT: br i1 [[CMP]], label %[[EXIT]], label %[[LOOP]], !llvm.loop [[LOOP5:![0-9]+]]
118
+ ;
55
119
entry:
56
120
br label %loop
57
121
@@ -69,3 +133,92 @@ loop:
69
133
%cmp = icmp eq i32 %secondary , 5
70
134
br i1 %cmp , label %exit , label %loop
71
135
}
136
+
137
+ ; Test with a dead load and dead vector poiner.
138
+ define void @dead_load_and_vector_pointer (ptr %a , ptr %b ) {
139
+ ; CHECK-LABEL: define void @dead_load_and_vector_pointer(
140
+ ; CHECK-SAME: ptr [[A:%.*]], ptr [[B:%.*]]) {
141
+ ; CHECK-NEXT: [[ENTRY:.*]]:
142
+ ; CHECK-NEXT: br i1 false, label %[[SCALAR_PH:.*]], label %[[VECTOR_MEMCHECK:.*]]
143
+ ; CHECK: [[VECTOR_MEMCHECK]]:
144
+ ; CHECK-NEXT: [[SCEVGEP:%.*]] = getelementptr i8, ptr [[A]], i64 516
145
+ ; CHECK-NEXT: [[SCEVGEP1:%.*]] = getelementptr i8, ptr [[B]], i64 516
146
+ ; CHECK-NEXT: [[BOUND0:%.*]] = icmp ult ptr [[A]], [[SCEVGEP1]]
147
+ ; CHECK-NEXT: [[BOUND1:%.*]] = icmp ult ptr [[B]], [[SCEVGEP]]
148
+ ; CHECK-NEXT: [[FOUND_CONFLICT:%.*]] = and i1 [[BOUND0]], [[BOUND1]]
149
+ ; CHECK-NEXT: br i1 [[FOUND_CONFLICT]], label %[[SCALAR_PH]], label %[[VECTOR_PH:.*]]
150
+ ; CHECK: [[VECTOR_PH]]:
151
+ ; CHECK-NEXT: br label %[[VECTOR_BODY:.*]]
152
+ ; CHECK: [[VECTOR_BODY]]:
153
+ ; CHECK-NEXT: [[INDEX:%.*]] = phi i32 [ 0, %[[VECTOR_PH]] ], [ [[INDEX_NEXT:%.*]], %[[VECTOR_BODY]] ]
154
+ ; CHECK-NEXT: [[TMP0:%.*]] = add i32 [[INDEX]], 0
155
+ ; CHECK-NEXT: [[TMP1:%.*]] = add i32 [[INDEX]], 2
156
+ ; CHECK-NEXT: [[TMP2:%.*]] = getelementptr inbounds i32, ptr [[A]], i32 [[TMP0]]
157
+ ; CHECK-NEXT: [[TMP3:%.*]] = getelementptr inbounds i32, ptr [[A]], i32 [[TMP1]]
158
+ ; CHECK-NEXT: [[TMP4:%.*]] = getelementptr inbounds i32, ptr [[TMP2]], i32 0
159
+ ; CHECK-NEXT: [[TMP5:%.*]] = getelementptr inbounds i32, ptr [[TMP2]], i32 2
160
+ ; CHECK-NEXT: [[WIDE_LOAD:%.*]] = load <2 x i32>, ptr [[TMP4]], align 8, !alias.scope [[META6:![0-9]+]], !noalias [[META9:![0-9]+]]
161
+ ; CHECK-NEXT: [[WIDE_LOAD2:%.*]] = load <2 x i32>, ptr [[TMP5]], align 8, !alias.scope [[META6]], !noalias [[META9]]
162
+ ; CHECK-NEXT: [[TMP6:%.*]] = add <2 x i32> [[WIDE_LOAD]], <i32 1, i32 1>
163
+ ; CHECK-NEXT: [[TMP7:%.*]] = add <2 x i32> [[WIDE_LOAD2]], <i32 1, i32 1>
164
+ ; CHECK-NEXT: store <2 x i32> [[TMP6]], ptr [[TMP4]], align 4, !alias.scope [[META6]], !noalias [[META9]]
165
+ ; CHECK-NEXT: store <2 x i32> [[TMP7]], ptr [[TMP5]], align 4, !alias.scope [[META6]], !noalias [[META9]]
166
+ ; CHECK-NEXT: [[TMP8:%.*]] = getelementptr inbounds i32, ptr [[B]], i32 [[TMP0]]
167
+ ; CHECK-NEXT: [[TMP9:%.*]] = getelementptr inbounds i32, ptr [[B]], i32 [[TMP1]]
168
+ ; CHECK-NEXT: [[TMP10:%.*]] = getelementptr inbounds i32, ptr [[TMP8]], i32 0
169
+ ; CHECK-NEXT: [[TMP11:%.*]] = getelementptr inbounds i32, ptr [[TMP8]], i32 2
170
+ ; CHECK-NEXT: [[INDEX_NEXT]] = add nuw i32 [[INDEX]], 4
171
+ ; CHECK-NEXT: [[TMP12:%.*]] = icmp eq i32 [[INDEX_NEXT]], 128
172
+ ; CHECK-NEXT: br i1 [[TMP12]], label %[[MIDDLE_BLOCK:.*]], label %[[VECTOR_BODY]], !llvm.loop [[LOOP11:![0-9]+]]
173
+ ; CHECK: [[MIDDLE_BLOCK]]:
174
+ ; CHECK-NEXT: br i1 false, label %[[EXIT:.*]], label %[[SCALAR_PH]]
175
+ ; CHECK: [[SCALAR_PH]]:
176
+ ; CHECK-NEXT: [[BC_RESUME_VAL:%.*]] = phi i32 [ 128, %[[MIDDLE_BLOCK]] ], [ 0, %[[ENTRY]] ], [ 0, %[[VECTOR_MEMCHECK]] ]
177
+ ; CHECK-NEXT: br label %[[LOOP:.*]]
178
+ ; CHECK: [[EXIT]]:
179
+ ; CHECK-NEXT: ret void
180
+ ; CHECK: [[LOOP]]:
181
+ ; CHECK-NEXT: [[PRIMARY:%.*]] = phi i32 [ [[BC_RESUME_VAL]], %[[SCALAR_PH]] ], [ [[PRIMARY_ADD:%.*]], %[[LOOP]] ]
182
+ ; CHECK-NEXT: [[PRIMARY_ADD]] = add i32 [[PRIMARY]], 1
183
+ ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds i32, ptr [[A]], i32 [[PRIMARY]]
184
+ ; CHECK-NEXT: [[LOAD:%.*]] = load i32, ptr [[GEP]], align 8
185
+ ; CHECK-NEXT: [[ADD:%.*]] = add i32 [[LOAD]], 1
186
+ ; CHECK-NEXT: store i32 [[ADD]], ptr [[GEP]], align 4
187
+ ; CHECK-NEXT: [[GEP_B:%.*]] = getelementptr inbounds i32, ptr [[B]], i32 [[PRIMARY]]
188
+ ; CHECK-NEXT: [[LOAD2:%.*]] = load i32, ptr [[GEP_B]], align 4
189
+ ; CHECK-NEXT: [[CMP:%.*]] = icmp eq i32 [[PRIMARY]], 128
190
+ ; CHECK-NEXT: br i1 [[CMP]], label %[[EXIT]], label %[[LOOP]], !llvm.loop [[LOOP12:![0-9]+]]
191
+ ;
192
+ entry:
193
+ br label %loop
194
+
195
+ exit:
196
+ ret void
197
+
198
+ loop:
199
+ %primary = phi i32 [ 0 , %entry ], [ %primary_add , %loop ]
200
+ %primary_add = add i32 %primary , 1
201
+ %gep = getelementptr inbounds i32 , ptr %a , i32 %primary
202
+ %load = load i32 , ptr %gep , align 8
203
+ %add = add i32 %load , 1
204
+ store i32 %add , ptr %gep
205
+ %gep.b = getelementptr inbounds i32 , ptr %b , i32 %primary
206
+ %load2 = load i32 , ptr %gep.b
207
+ %cmp = icmp eq i32 %primary , 128
208
+ br i1 %cmp , label %exit , label %loop
209
+ }
210
+ ;.
211
+ ; CHECK: [[LOOP0]] = distinct !{[[LOOP0]], [[META1:![0-9]+]], [[META2:![0-9]+]]}
212
+ ; CHECK: [[META1]] = !{!"llvm.loop.isvectorized", i32 1}
213
+ ; CHECK: [[META2]] = !{!"llvm.loop.unroll.runtime.disable"}
214
+ ; CHECK: [[LOOP3]] = distinct !{[[LOOP3]], [[META2]], [[META1]]}
215
+ ; CHECK: [[LOOP4]] = distinct !{[[LOOP4]], [[META1]], [[META2]]}
216
+ ; CHECK: [[LOOP5]] = distinct !{[[LOOP5]], [[META2]], [[META1]]}
217
+ ; CHECK: [[META6]] = !{[[META7:![0-9]+]]}
218
+ ; CHECK: [[META7]] = distinct !{[[META7]], [[META8:![0-9]+]]}
219
+ ; CHECK: [[META8]] = distinct !{[[META8]], !"LVerDomain"}
220
+ ; CHECK: [[META9]] = !{[[META10:![0-9]+]]}
221
+ ; CHECK: [[META10]] = distinct !{[[META10]], [[META8]]}
222
+ ; CHECK: [[LOOP11]] = distinct !{[[LOOP11]], [[META1]], [[META2]]}
223
+ ; CHECK: [[LOOP12]] = distinct !{[[LOOP12]], [[META1]]}
224
+ ;.
0 commit comments