@@ -121,13 +121,11 @@ lookupFoldTableImpl(ArrayRef<X86FoldTableEntry> Table, unsigned RegOp) {
121
121
return nullptr ;
122
122
}
123
123
124
- const X86FoldTableEntry *
125
- llvm::lookupTwoAddrFoldTable (unsigned RegOp) {
124
+ const X86FoldTableEntry *llvm::lookupTwoAddrFoldTable (unsigned RegOp) {
126
125
return lookupFoldTableImpl (Table2Addr, RegOp);
127
126
}
128
127
129
- const X86FoldTableEntry *
130
- llvm::lookupFoldTable (unsigned RegOp, unsigned OpNum) {
128
+ const X86FoldTableEntry *llvm::lookupFoldTable (unsigned RegOp, unsigned OpNum) {
131
129
ArrayRef<X86FoldTableEntry> FoldTable;
132
130
if (OpNum == 0 )
133
131
FoldTable = ArrayRef (Table0);
@@ -181,19 +179,19 @@ struct X86MemUnfoldTable {
181
179
// Broadcast tables.
182
180
for (const X86FoldTableEntry &Entry : BroadcastTable1)
183
181
// Index 1, folded broadcast
184
- addTableEntry (Entry, TB_INDEX_1 | TB_FOLDED_LOAD | TB_FOLDED_BCAST );
182
+ addTableEntry (Entry, TB_INDEX_1 | TB_FOLDED_LOAD);
185
183
186
184
for (const X86FoldTableEntry &Entry : BroadcastTable2)
187
185
// Index 2, folded broadcast
188
- addTableEntry (Entry, TB_INDEX_2 | TB_FOLDED_LOAD | TB_FOLDED_BCAST );
186
+ addTableEntry (Entry, TB_INDEX_2 | TB_FOLDED_LOAD);
189
187
190
188
for (const X86FoldTableEntry &Entry : BroadcastTable3)
191
189
// Index 3, folded broadcast
192
- addTableEntry (Entry, TB_INDEX_3 | TB_FOLDED_LOAD | TB_FOLDED_BCAST );
190
+ addTableEntry (Entry, TB_INDEX_3 | TB_FOLDED_LOAD);
193
191
194
192
for (const X86FoldTableEntry &Entry : BroadcastTable4)
195
193
// Index 4, folded broadcast
196
- addTableEntry (Entry, TB_INDEX_4 | TB_FOLDED_LOAD | TB_FOLDED_BCAST );
194
+ addTableEntry (Entry, TB_INDEX_4 | TB_FOLDED_LOAD);
197
195
198
196
// Sort the memory->reg unfold table.
199
197
array_pod_sort (Table.begin (), Table.end ());
@@ -203,18 +201,16 @@ struct X86MemUnfoldTable {
203
201
" Memory unfolding table is not unique!" );
204
202
}
205
203
206
- void addTableEntry (const X86FoldTableEntry &Entry,
207
- uint16_t ExtraFlags) {
204
+ void addTableEntry (const X86FoldTableEntry &Entry, uint16_t ExtraFlags) {
208
205
// NOTE: This swaps the KeyOp and DstOp in the table so we can sort it.
209
206
if ((Entry.Flags & TB_NO_REVERSE) == 0 )
210
207
Table.push_back ({Entry.DstOp , Entry.KeyOp ,
211
- static_cast <uint16_t >(Entry.Flags | ExtraFlags) });
208
+ static_cast <uint16_t >(Entry.Flags | ExtraFlags)});
212
209
}
213
210
};
214
- }
211
+ } // namespace
215
212
216
- const X86FoldTableEntry *
217
- llvm::lookupUnfoldTable (unsigned MemOp) {
213
+ const X86FoldTableEntry *llvm::lookupUnfoldTable (unsigned MemOp) {
218
214
static X86MemUnfoldTable MemUnfoldTable;
219
215
auto &Table = MemUnfoldTable.Table ;
220
216
auto I = llvm::lower_bound (Table, MemOp);
@@ -238,8 +234,8 @@ struct X86BroadcastFoldTable {
238
234
unsigned BcstOp = Reg2Bcst.DstOp ;
239
235
if (const X86FoldTableEntry *Reg2Mem = lookupFoldTable (RegOp, 2 )) {
240
236
unsigned MemOp = Reg2Mem->DstOp ;
241
- uint16_t Flags = Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_2 |
242
- TB_FOLDED_LOAD | TB_FOLDED_BCAST ;
237
+ uint16_t Flags =
238
+ Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_2 | TB_FOLDED_LOAD ;
243
239
Table.push_back ({MemOp, BcstOp, Flags});
244
240
}
245
241
}
@@ -248,8 +244,8 @@ struct X86BroadcastFoldTable {
248
244
unsigned BcstOp = Reg2Bcst.DstOp ;
249
245
if (const X86FoldTableEntry *Reg2Mem = lookupFoldTable (RegOp, 2 )) {
250
246
unsigned MemOp = Reg2Mem->DstOp ;
251
- uint16_t Flags = Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_2 |
252
- TB_FOLDED_LOAD | TB_FOLDED_BCAST ;
247
+ uint16_t Flags =
248
+ Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_2 | TB_FOLDED_LOAD ;
253
249
Table.push_back ({MemOp, BcstOp, Flags});
254
250
}
255
251
}
@@ -259,8 +255,8 @@ struct X86BroadcastFoldTable {
259
255
unsigned BcstOp = Reg2Bcst.DstOp ;
260
256
if (const X86FoldTableEntry *Reg2Mem = lookupFoldTable (RegOp, 3 )) {
261
257
unsigned MemOp = Reg2Mem->DstOp ;
262
- uint16_t Flags = Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_3 |
263
- TB_FOLDED_LOAD | TB_FOLDED_BCAST ;
258
+ uint16_t Flags =
259
+ Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_3 | TB_FOLDED_LOAD ;
264
260
Table.push_back ({MemOp, BcstOp, Flags});
265
261
}
266
262
}
@@ -269,8 +265,8 @@ struct X86BroadcastFoldTable {
269
265
unsigned BcstOp = Reg2Bcst.DstOp ;
270
266
if (const X86FoldTableEntry *Reg2Mem = lookupFoldTable (RegOp, 3 )) {
271
267
unsigned MemOp = Reg2Mem->DstOp ;
272
- uint16_t Flags = Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_3 |
273
- TB_FOLDED_LOAD | TB_FOLDED_BCAST ;
268
+ uint16_t Flags =
269
+ Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_3 | TB_FOLDED_LOAD ;
274
270
Table.push_back ({MemOp, BcstOp, Flags});
275
271
}
276
272
}
@@ -280,8 +276,8 @@ struct X86BroadcastFoldTable {
280
276
unsigned BcstOp = Reg2Bcst.DstOp ;
281
277
if (const X86FoldTableEntry *Reg2Mem = lookupFoldTable (RegOp, 4 )) {
282
278
unsigned MemOp = Reg2Mem->DstOp ;
283
- uint16_t Flags = Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_4 |
284
- TB_FOLDED_LOAD | TB_FOLDED_BCAST ;
279
+ uint16_t Flags =
280
+ Reg2Mem-> Flags | Reg2Bcst. Flags | TB_INDEX_4 | TB_FOLDED_LOAD ;
285
281
Table.push_back ({MemOp, BcstOp, Flags});
286
282
}
287
283
}
0 commit comments