LLVM 23.0.0git
SelectionDAG.cpp
Go to the documentation of this file.
1//===- SelectionDAG.cpp - Implement the SelectionDAG data structures ------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This implements the SelectionDAG class.
10//
11//===----------------------------------------------------------------------===//
12
14#include "SDNodeDbgValue.h"
15#include "llvm/ADT/APFloat.h"
16#include "llvm/ADT/APInt.h"
17#include "llvm/ADT/APSInt.h"
18#include "llvm/ADT/ArrayRef.h"
19#include "llvm/ADT/BitVector.h"
20#include "llvm/ADT/DenseSet.h"
21#include "llvm/ADT/FoldingSet.h"
22#include "llvm/ADT/STLExtras.h"
25#include "llvm/ADT/Twine.h"
51#include "llvm/IR/Constant.h"
52#include "llvm/IR/Constants.h"
53#include "llvm/IR/DataLayout.h"
55#include "llvm/IR/DebugLoc.h"
57#include "llvm/IR/Function.h"
58#include "llvm/IR/GlobalValue.h"
59#include "llvm/IR/Metadata.h"
60#include "llvm/IR/Type.h"
64#include "llvm/Support/Debug.h"
73#include <algorithm>
74#include <cassert>
75#include <cstdint>
76#include <cstdlib>
77#include <limits>
78#include <optional>
79#include <string>
80#include <utility>
81#include <vector>
82
83using namespace llvm;
84using namespace llvm::SDPatternMatch;
85
86/// makeVTList - Return an instance of the SDVTList struct initialized with the
87/// specified members.
88static SDVTList makeVTList(const EVT *VTs, unsigned NumVTs) {
89 SDVTList Res = {VTs, NumVTs};
90 return Res;
91}
92
93// Default null implementations of the callbacks.
97
98void SelectionDAG::DAGNodeDeletedListener::anchor() {}
99void SelectionDAG::DAGNodeInsertedListener::anchor() {}
100
101#define DEBUG_TYPE "selectiondag"
102
103static cl::opt<bool> EnableMemCpyDAGOpt("enable-memcpy-dag-opt",
104 cl::Hidden, cl::init(true),
105 cl::desc("Gang up loads and stores generated by inlining of memcpy"));
106
107static cl::opt<int> MaxLdStGlue("ldstmemcpy-glue-max",
108 cl::desc("Number limit for gluing ld/st of memcpy."),
109 cl::Hidden, cl::init(0));
110
112 MaxSteps("has-predecessor-max-steps", cl::Hidden, cl::init(8192),
113 cl::desc("DAG combiner limit number of steps when searching DAG "
114 "for predecessor nodes"));
115
117 LLVM_DEBUG(dbgs() << Msg; V.getNode()->dump(G););
118}
119
121
122//===----------------------------------------------------------------------===//
123// ConstantFPSDNode Class
124//===----------------------------------------------------------------------===//
125
126/// isExactlyValue - We don't rely on operator== working on double values, as
127/// it returns true for things that are clearly not equal, like -0.0 and 0.0.
128/// As such, this method can be used to do an exact bit-for-bit comparison of
129/// two floating point values.
131 return getValueAPF().bitwiseIsEqual(V);
132}
133
135 const APFloat& Val) {
136 assert(VT.isFloatingPoint() && "Can only convert between FP types");
137
138 // convert modifies in place, so make a copy.
139 APFloat Val2 = APFloat(Val);
140 bool losesInfo;
142 &losesInfo);
143 return !losesInfo;
144}
145
146//===----------------------------------------------------------------------===//
147// ISD Namespace
148//===----------------------------------------------------------------------===//
149
150bool ISD::isConstantSplatVector(const SDNode *N, APInt &SplatVal) {
151 if (N->getOpcode() == ISD::SPLAT_VECTOR) {
152 if (auto OptAPInt = N->getOperand(0)->bitcastToAPInt()) {
153 unsigned EltSize =
154 N->getValueType(0).getVectorElementType().getSizeInBits();
155 SplatVal = OptAPInt->trunc(EltSize);
156 return true;
157 }
158 }
159
160 auto *BV = dyn_cast<BuildVectorSDNode>(N);
161 if (!BV)
162 return false;
163
164 APInt SplatUndef;
165 unsigned SplatBitSize;
166 bool HasUndefs;
167 unsigned EltSize = N->getValueType(0).getVectorElementType().getSizeInBits();
168 // Endianness does not matter here. We are checking for a splat given the
169 // element size of the vector, and if we find such a splat for little endian
170 // layout, then that should be valid also for big endian (as the full vector
171 // size is known to be a multiple of the element size).
172 const bool IsBigEndian = false;
173 return BV->isConstantSplat(SplatVal, SplatUndef, SplatBitSize, HasUndefs,
174 EltSize, IsBigEndian) &&
175 EltSize == SplatBitSize;
176}
177
178// FIXME: AllOnes and AllZeros duplicate a lot of code. Could these be
179// specializations of the more general isConstantSplatVector()?
180
181bool ISD::isConstantSplatVectorAllOnes(const SDNode *N, bool BuildVectorOnly) {
182 // Look through a bit convert.
183 while (N->getOpcode() == ISD::BITCAST)
184 N = N->getOperand(0).getNode();
185
186 if (!BuildVectorOnly && N->getOpcode() == ISD::SPLAT_VECTOR) {
187 APInt SplatVal;
188 return isConstantSplatVector(N, SplatVal) && SplatVal.isAllOnes();
189 }
190
191 if (N->getOpcode() != ISD::BUILD_VECTOR) return false;
192
193 unsigned i = 0, e = N->getNumOperands();
194
195 // Skip over all of the undef values.
196 while (i != e && N->getOperand(i).isUndef())
197 ++i;
198
199 // Do not accept an all-undef vector.
200 if (i == e) return false;
201
202 // Do not accept build_vectors that aren't all constants or which have non-~0
203 // elements. We have to be a bit careful here, as the type of the constant
204 // may not be the same as the type of the vector elements due to type
205 // legalization (the elements are promoted to a legal type for the target and
206 // a vector of a type may be legal when the base element type is not).
207 // We only want to check enough bits to cover the vector elements, because
208 // we care if the resultant vector is all ones, not whether the individual
209 // constants are.
210 SDValue NotZero = N->getOperand(i);
211 if (auto OptAPInt = NotZero->bitcastToAPInt()) {
212 unsigned EltSize = N->getValueType(0).getScalarSizeInBits();
213 if (OptAPInt->countr_one() < EltSize)
214 return false;
215 } else
216 return false;
217
218 // Okay, we have at least one ~0 value, check to see if the rest match or are
219 // undefs. Even with the above element type twiddling, this should be OK, as
220 // the same type legalization should have applied to all the elements.
221 for (++i; i != e; ++i)
222 if (N->getOperand(i) != NotZero && !N->getOperand(i).isUndef())
223 return false;
224 return true;
225}
226
227bool ISD::isConstantSplatVectorAllZeros(const SDNode *N, bool BuildVectorOnly) {
228 // Look through a bit convert.
229 while (N->getOpcode() == ISD::BITCAST)
230 N = N->getOperand(0).getNode();
231
232 if (!BuildVectorOnly && N->getOpcode() == ISD::SPLAT_VECTOR) {
233 APInt SplatVal;
234 return isConstantSplatVector(N, SplatVal) && SplatVal.isZero();
235 }
236
237 if (N->getOpcode() != ISD::BUILD_VECTOR) return false;
238
239 bool IsAllUndef = true;
240 for (const SDValue &Op : N->op_values()) {
241 if (Op.isUndef())
242 continue;
243 IsAllUndef = false;
244 // Do not accept build_vectors that aren't all constants or which have non-0
245 // elements. We have to be a bit careful here, as the type of the constant
246 // may not be the same as the type of the vector elements due to type
247 // legalization (the elements are promoted to a legal type for the target
248 // and a vector of a type may be legal when the base element type is not).
249 // We only want to check enough bits to cover the vector elements, because
250 // we care if the resultant vector is all zeros, not whether the individual
251 // constants are.
252 if (auto OptAPInt = Op->bitcastToAPInt()) {
253 unsigned EltSize = N->getValueType(0).getScalarSizeInBits();
254 if (OptAPInt->countr_zero() < EltSize)
255 return false;
256 } else
257 return false;
258 }
259
260 // Do not accept an all-undef vector.
261 if (IsAllUndef)
262 return false;
263 return true;
264}
265
267 return isConstantSplatVectorAllOnes(N, /*BuildVectorOnly*/ true);
268}
269
271 return isConstantSplatVectorAllZeros(N, /*BuildVectorOnly*/ true);
272}
273
275 if (N->getOpcode() != ISD::BUILD_VECTOR)
276 return false;
277
278 for (const SDValue &Op : N->op_values()) {
279 if (Op.isUndef())
280 continue;
282 return false;
283 }
284 return true;
285}
286
288 if (N->getOpcode() != ISD::BUILD_VECTOR)
289 return false;
290
291 for (const SDValue &Op : N->op_values()) {
292 if (Op.isUndef())
293 continue;
295 return false;
296 }
297 return true;
298}
299
300bool ISD::isVectorShrinkable(const SDNode *N, unsigned NewEltSize,
301 bool Signed) {
302 assert(N->getValueType(0).isVector() && "Expected a vector!");
303
304 unsigned EltSize = N->getValueType(0).getScalarSizeInBits();
305 if (EltSize <= NewEltSize)
306 return false;
307
308 if (N->getOpcode() == ISD::ZERO_EXTEND) {
309 return (N->getOperand(0).getValueType().getScalarSizeInBits() <=
310 NewEltSize) &&
311 !Signed;
312 }
313 if (N->getOpcode() == ISD::SIGN_EXTEND) {
314 return (N->getOperand(0).getValueType().getScalarSizeInBits() <=
315 NewEltSize) &&
316 Signed;
317 }
318 if (N->getOpcode() != ISD::BUILD_VECTOR)
319 return false;
320
321 for (const SDValue &Op : N->op_values()) {
322 if (Op.isUndef())
323 continue;
325 return false;
326
327 APInt C = Op->getAsAPIntVal().trunc(EltSize);
328 if (Signed && C.trunc(NewEltSize).sext(EltSize) != C)
329 return false;
330 if (!Signed && C.trunc(NewEltSize).zext(EltSize) != C)
331 return false;
332 }
333
334 return true;
335}
336
338 // Return false if the node has no operands.
339 // This is "logically inconsistent" with the definition of "all" but
340 // is probably the desired behavior.
341 if (N->getNumOperands() == 0)
342 return false;
343 return all_of(N->op_values(), [](SDValue Op) { return Op.isUndef(); });
344}
345
347 return N->getOpcode() == ISD::FREEZE && N->getOperand(0).isUndef();
348}
349
350template <typename ConstNodeType>
352 std::function<bool(ConstNodeType *)> Match,
353 bool AllowUndefs, bool AllowTruncation) {
354 // FIXME: Add support for scalar UNDEF cases?
355 if (auto *C = dyn_cast<ConstNodeType>(Op))
356 return Match(C);
357
358 // FIXME: Add support for vector UNDEF cases?
359 if (ISD::BUILD_VECTOR != Op.getOpcode() &&
360 ISD::SPLAT_VECTOR != Op.getOpcode())
361 return false;
362
363 EVT SVT = Op.getValueType().getScalarType();
364 for (unsigned i = 0, e = Op.getNumOperands(); i != e; ++i) {
365 if (AllowUndefs && Op.getOperand(i).isUndef()) {
366 if (!Match(nullptr))
367 return false;
368 continue;
369 }
370
371 auto *Cst = dyn_cast<ConstNodeType>(Op.getOperand(i));
372 if (!Cst || (!AllowTruncation && Cst->getValueType(0) != SVT) ||
373 !Match(Cst))
374 return false;
375 }
376 return true;
377}
378// Build used template types.
380 SDValue, std::function<bool(ConstantSDNode *)>, bool, bool);
382 SDValue, std::function<bool(ConstantFPSDNode *)>, bool, bool);
383
385 SDValue LHS, SDValue RHS,
386 std::function<bool(ConstantSDNode *, ConstantSDNode *)> Match,
387 bool AllowUndefs, bool AllowTypeMismatch) {
388 if (!AllowTypeMismatch && LHS.getValueType() != RHS.getValueType())
389 return false;
390
391 // TODO: Add support for scalar UNDEF cases?
392 if (auto *LHSCst = dyn_cast<ConstantSDNode>(LHS))
393 if (auto *RHSCst = dyn_cast<ConstantSDNode>(RHS))
394 return Match(LHSCst, RHSCst);
395
396 // TODO: Add support for vector UNDEF cases?
397 if (LHS.getOpcode() != RHS.getOpcode() ||
398 (LHS.getOpcode() != ISD::BUILD_VECTOR &&
399 LHS.getOpcode() != ISD::SPLAT_VECTOR))
400 return false;
401
402 EVT SVT = LHS.getValueType().getScalarType();
403 for (unsigned i = 0, e = LHS.getNumOperands(); i != e; ++i) {
404 SDValue LHSOp = LHS.getOperand(i);
405 SDValue RHSOp = RHS.getOperand(i);
406 bool LHSUndef = AllowUndefs && LHSOp.isUndef();
407 bool RHSUndef = AllowUndefs && RHSOp.isUndef();
408 auto *LHSCst = dyn_cast<ConstantSDNode>(LHSOp);
409 auto *RHSCst = dyn_cast<ConstantSDNode>(RHSOp);
410 if ((!LHSCst && !LHSUndef) || (!RHSCst && !RHSUndef))
411 return false;
412 if (!AllowTypeMismatch && (LHSOp.getValueType() != SVT ||
413 LHSOp.getValueType() != RHSOp.getValueType()))
414 return false;
415 if (!Match(LHSCst, RHSCst))
416 return false;
417 }
418 return true;
419}
420
422 switch (MinMaxOpc) {
423 default:
424 llvm_unreachable("unrecognized opcode");
425 case ISD::UMIN:
426 return ISD::UMAX;
427 case ISD::UMAX:
428 return ISD::UMIN;
429 case ISD::SMIN:
430 return ISD::SMAX;
431 case ISD::SMAX:
432 return ISD::SMIN;
433 }
434}
435
437 switch (MinMaxOpc) {
438 default:
439 llvm_unreachable("unrecognized min/max opcode");
440 case ISD::SMIN:
441 return ISD::UMIN;
442 case ISD::SMAX:
443 return ISD::UMAX;
444 case ISD::UMIN:
445 return ISD::SMIN;
446 case ISD::UMAX:
447 return ISD::SMAX;
448 }
449}
450
452 switch (VecReduceOpcode) {
453 default:
454 llvm_unreachable("Expected VECREDUCE opcode");
457 case ISD::VP_REDUCE_FADD:
458 case ISD::VP_REDUCE_SEQ_FADD:
459 return ISD::FADD;
462 case ISD::VP_REDUCE_FMUL:
463 case ISD::VP_REDUCE_SEQ_FMUL:
464 return ISD::FMUL;
466 case ISD::VP_REDUCE_ADD:
467 return ISD::ADD;
469 case ISD::VP_REDUCE_MUL:
470 return ISD::MUL;
472 case ISD::VP_REDUCE_AND:
473 return ISD::AND;
475 case ISD::VP_REDUCE_OR:
476 return ISD::OR;
478 case ISD::VP_REDUCE_XOR:
479 return ISD::XOR;
481 case ISD::VP_REDUCE_SMAX:
482 return ISD::SMAX;
484 case ISD::VP_REDUCE_SMIN:
485 return ISD::SMIN;
487 case ISD::VP_REDUCE_UMAX:
488 return ISD::UMAX;
490 case ISD::VP_REDUCE_UMIN:
491 return ISD::UMIN;
493 case ISD::VP_REDUCE_FMAX:
494 return ISD::FMAXNUM;
496 case ISD::VP_REDUCE_FMIN:
497 return ISD::FMINNUM;
499 case ISD::VP_REDUCE_FMAXIMUM:
500 return ISD::FMAXIMUM;
502 case ISD::VP_REDUCE_FMINIMUM:
503 return ISD::FMINIMUM;
504 }
505}
506
507bool ISD::isVPOpcode(unsigned Opcode) {
508 switch (Opcode) {
509 default:
510 return false;
511#define BEGIN_REGISTER_VP_SDNODE(VPSD, ...) \
512 case ISD::VPSD: \
513 return true;
514#include "llvm/IR/VPIntrinsics.def"
515 }
516}
517
518bool ISD::isVPBinaryOp(unsigned Opcode) {
519 switch (Opcode) {
520 default:
521 break;
522#define BEGIN_REGISTER_VP_SDNODE(VPSD, ...) case ISD::VPSD:
523#define VP_PROPERTY_BINARYOP return true;
524#define END_REGISTER_VP_SDNODE(VPSD) break;
525#include "llvm/IR/VPIntrinsics.def"
526 }
527 return false;
528}
529
530bool ISD::isVPReduction(unsigned Opcode) {
531 switch (Opcode) {
532 default:
533 return false;
534 case ISD::VP_REDUCE_ADD:
535 case ISD::VP_REDUCE_MUL:
536 case ISD::VP_REDUCE_AND:
537 case ISD::VP_REDUCE_OR:
538 case ISD::VP_REDUCE_XOR:
539 case ISD::VP_REDUCE_SMAX:
540 case ISD::VP_REDUCE_SMIN:
541 case ISD::VP_REDUCE_UMAX:
542 case ISD::VP_REDUCE_UMIN:
543 case ISD::VP_REDUCE_FMAX:
544 case ISD::VP_REDUCE_FMIN:
545 case ISD::VP_REDUCE_FMAXIMUM:
546 case ISD::VP_REDUCE_FMINIMUM:
547 case ISD::VP_REDUCE_FADD:
548 case ISD::VP_REDUCE_FMUL:
549 case ISD::VP_REDUCE_SEQ_FADD:
550 case ISD::VP_REDUCE_SEQ_FMUL:
551 return true;
552 }
553}
554
555/// The operand position of the vector mask.
556std::optional<unsigned> ISD::getVPMaskIdx(unsigned Opcode) {
557 switch (Opcode) {
558 default:
559 return std::nullopt;
560#define BEGIN_REGISTER_VP_SDNODE(VPSD, LEGALPOS, TDNAME, MASKPOS, ...) \
561 case ISD::VPSD: \
562 return MASKPOS;
563#include "llvm/IR/VPIntrinsics.def"
564 }
565}
566
567/// The operand position of the explicit vector length parameter.
568std::optional<unsigned> ISD::getVPExplicitVectorLengthIdx(unsigned Opcode) {
569 switch (Opcode) {
570 default:
571 return std::nullopt;
572#define BEGIN_REGISTER_VP_SDNODE(VPSD, LEGALPOS, TDNAME, MASKPOS, EVLPOS) \
573 case ISD::VPSD: \
574 return EVLPOS;
575#include "llvm/IR/VPIntrinsics.def"
576 }
577}
578
579std::optional<unsigned> ISD::getBaseOpcodeForVP(unsigned VPOpcode,
580 bool hasFPExcept) {
581 // FIXME: Return strict opcodes in case of fp exceptions.
582 switch (VPOpcode) {
583 default:
584 return std::nullopt;
585#define BEGIN_REGISTER_VP_SDNODE(VPOPC, ...) case ISD::VPOPC:
586#define VP_PROPERTY_FUNCTIONAL_SDOPC(SDOPC) return ISD::SDOPC;
587#define END_REGISTER_VP_SDNODE(VPOPC) break;
588#include "llvm/IR/VPIntrinsics.def"
589 }
590 return std::nullopt;
591}
592
593std::optional<unsigned> ISD::getVPForBaseOpcode(unsigned Opcode) {
594 switch (Opcode) {
595 default:
596 return std::nullopt;
597#define BEGIN_REGISTER_VP_SDNODE(VPOPC, ...) break;
598#define VP_PROPERTY_FUNCTIONAL_SDOPC(SDOPC) case ISD::SDOPC:
599#define END_REGISTER_VP_SDNODE(VPOPC) return ISD::VPOPC;
600#include "llvm/IR/VPIntrinsics.def"
601 }
602}
603
605 switch (ExtType) {
606 case ISD::EXTLOAD:
607 return IsFP ? ISD::FP_EXTEND : ISD::ANY_EXTEND;
608 case ISD::SEXTLOAD:
609 return ISD::SIGN_EXTEND;
610 case ISD::ZEXTLOAD:
611 return ISD::ZERO_EXTEND;
612 default:
613 break;
614 }
615
616 llvm_unreachable("Invalid LoadExtType");
617}
618
620 // To perform this operation, we just need to swap the L and G bits of the
621 // operation.
622 unsigned OldL = (Operation >> 2) & 1;
623 unsigned OldG = (Operation >> 1) & 1;
624 return ISD::CondCode((Operation & ~6) | // Keep the N, U, E bits
625 (OldL << 1) | // New G bit
626 (OldG << 2)); // New L bit.
627}
628
630 unsigned Operation = Op;
631 if (isIntegerLike)
632 Operation ^= 7; // Flip L, G, E bits, but not U.
633 else
634 Operation ^= 15; // Flip all of the condition bits.
635
637 Operation &= ~8; // Don't let N and U bits get set.
638
639 return ISD::CondCode(Operation);
640}
641
645
647 bool isIntegerLike) {
648 return getSetCCInverseImpl(Op, isIntegerLike);
649}
650
651/// For an integer comparison, return 1 if the comparison is a signed operation
652/// and 2 if the result is an unsigned comparison. Return zero if the operation
653/// does not depend on the sign of the input (setne and seteq).
654static int isSignedOp(ISD::CondCode Opcode) {
655 switch (Opcode) {
656 default: llvm_unreachable("Illegal integer setcc operation!");
657 case ISD::SETEQ:
658 case ISD::SETNE: return 0;
659 case ISD::SETLT:
660 case ISD::SETLE:
661 case ISD::SETGT:
662 case ISD::SETGE: return 1;
663 case ISD::SETULT:
664 case ISD::SETULE:
665 case ISD::SETUGT:
666 case ISD::SETUGE: return 2;
667 }
668}
669
671 EVT Type) {
672 bool IsInteger = Type.isInteger();
673 if (IsInteger && (isSignedOp(Op1) | isSignedOp(Op2)) == 3)
674 // Cannot fold a signed integer setcc with an unsigned integer setcc.
675 return ISD::SETCC_INVALID;
676
677 unsigned Op = Op1 | Op2; // Combine all of the condition bits.
678
679 // If the N and U bits get set, then the resultant comparison DOES suddenly
680 // care about orderedness, and it is true when ordered.
681 if (Op > ISD::SETTRUE2)
682 Op &= ~16; // Clear the U bit if the N bit is set.
683
684 // Canonicalize illegal integer setcc's.
685 if (IsInteger && Op == ISD::SETUNE) // e.g. SETUGT | SETULT
686 Op = ISD::SETNE;
687
688 return ISD::CondCode(Op);
689}
690
692 EVT Type) {
693 bool IsInteger = Type.isInteger();
694 if (IsInteger && (isSignedOp(Op1) | isSignedOp(Op2)) == 3)
695 // Cannot fold a signed setcc with an unsigned setcc.
696 return ISD::SETCC_INVALID;
697
698 // Combine all of the condition bits.
699 ISD::CondCode Result = ISD::CondCode(Op1 & Op2);
700
701 // Canonicalize illegal integer setcc's.
702 if (IsInteger) {
703 switch (Result) {
704 default: break;
705 case ISD::SETUO : Result = ISD::SETFALSE; break; // SETUGT & SETULT
706 case ISD::SETOEQ: // SETEQ & SETU[LG]E
707 case ISD::SETUEQ: Result = ISD::SETEQ ; break; // SETUGE & SETULE
708 case ISD::SETOLT: Result = ISD::SETULT ; break; // SETULT & SETNE
709 case ISD::SETOGT: Result = ISD::SETUGT ; break; // SETUGT & SETNE
710 }
711 }
712
713 return Result;
714}
715
716//===----------------------------------------------------------------------===//
717// SDNode Profile Support
718//===----------------------------------------------------------------------===//
719
720/// AddNodeIDOpcode - Add the node opcode to the NodeID data.
721static void AddNodeIDOpcode(FoldingSetNodeID &ID, unsigned OpC) {
722 ID.AddInteger(OpC);
723}
724
725/// AddNodeIDValueTypes - Value type lists are intern'd so we can represent them
726/// solely with their pointer.
728 ID.AddPointer(VTList.VTs);
729}
730
731/// AddNodeIDOperands - Various routines for adding operands to the NodeID data.
734 for (const auto &Op : Ops) {
735 ID.AddPointer(Op.getNode());
736 ID.AddInteger(Op.getResNo());
737 }
738}
739
740/// AddNodeIDOperands - Various routines for adding operands to the NodeID data.
743 for (const auto &Op : Ops) {
744 ID.AddPointer(Op.getNode());
745 ID.AddInteger(Op.getResNo());
746 }
747}
748
749static void AddNodeIDNode(FoldingSetNodeID &ID, unsigned OpC,
750 SDVTList VTList, ArrayRef<SDValue> OpList) {
751 AddNodeIDOpcode(ID, OpC);
752 AddNodeIDValueTypes(ID, VTList);
753 AddNodeIDOperands(ID, OpList);
754}
755
756/// If this is an SDNode with special info, add this info to the NodeID data.
758 switch (N->getOpcode()) {
761 case ISD::MCSymbol:
762 llvm_unreachable("Should only be used on nodes with operands");
763 default: break; // Normal nodes don't need extra info.
765 case ISD::Constant: {
767 ID.AddPointer(C->getConstantIntValue());
768 ID.AddBoolean(C->isOpaque());
769 break;
770 }
772 case ISD::ConstantFP:
773 ID.AddPointer(cast<ConstantFPSDNode>(N)->getConstantFPValue());
774 break;
780 ID.AddPointer(GA->getGlobal());
781 ID.AddInteger(GA->getOffset());
782 ID.AddInteger(GA->getTargetFlags());
783 break;
784 }
785 case ISD::BasicBlock:
787 break;
788 case ISD::Register:
789 ID.AddInteger(cast<RegisterSDNode>(N)->getReg().id());
790 break;
792 ID.AddPointer(cast<RegisterMaskSDNode>(N)->getRegMask());
793 break;
794 case ISD::SRCVALUE:
795 ID.AddPointer(cast<SrcValueSDNode>(N)->getValue());
796 break;
797 case ISD::FrameIndex:
799 ID.AddInteger(cast<FrameIndexSDNode>(N)->getIndex());
800 break;
802 ID.AddInteger(cast<PseudoProbeSDNode>(N)->getGuid());
803 ID.AddInteger(cast<PseudoProbeSDNode>(N)->getIndex());
804 ID.AddInteger(cast<PseudoProbeSDNode>(N)->getAttributes());
805 break;
806 case ISD::JumpTable:
808 ID.AddInteger(cast<JumpTableSDNode>(N)->getIndex());
809 ID.AddInteger(cast<JumpTableSDNode>(N)->getTargetFlags());
810 break;
814 ID.AddInteger(CP->getAlign().value());
815 ID.AddInteger(CP->getOffset());
818 else
819 ID.AddPointer(CP->getConstVal());
820 ID.AddInteger(CP->getTargetFlags());
821 break;
822 }
823 case ISD::TargetIndex: {
825 ID.AddInteger(TI->getIndex());
826 ID.AddInteger(TI->getOffset());
827 ID.AddInteger(TI->getTargetFlags());
828 break;
829 }
830 case ISD::LOAD: {
831 const LoadSDNode *LD = cast<LoadSDNode>(N);
832 ID.AddInteger(LD->getMemoryVT().getRawBits());
833 ID.AddInteger(LD->getRawSubclassData());
834 ID.AddInteger(LD->getPointerInfo().getAddrSpace());
835 ID.AddInteger(LD->getMemOperand()->getFlags());
836 break;
837 }
838 case ISD::STORE: {
839 const StoreSDNode *ST = cast<StoreSDNode>(N);
840 ID.AddInteger(ST->getMemoryVT().getRawBits());
841 ID.AddInteger(ST->getRawSubclassData());
842 ID.AddInteger(ST->getPointerInfo().getAddrSpace());
843 ID.AddInteger(ST->getMemOperand()->getFlags());
844 break;
845 }
846 case ISD::VP_LOAD: {
847 const VPLoadSDNode *ELD = cast<VPLoadSDNode>(N);
848 ID.AddInteger(ELD->getMemoryVT().getRawBits());
849 ID.AddInteger(ELD->getRawSubclassData());
850 ID.AddInteger(ELD->getPointerInfo().getAddrSpace());
851 ID.AddInteger(ELD->getMemOperand()->getFlags());
852 break;
853 }
854 case ISD::VP_LOAD_FF: {
855 const auto *LD = cast<VPLoadFFSDNode>(N);
856 ID.AddInteger(LD->getMemoryVT().getRawBits());
857 ID.AddInteger(LD->getRawSubclassData());
858 ID.AddInteger(LD->getPointerInfo().getAddrSpace());
859 ID.AddInteger(LD->getMemOperand()->getFlags());
860 break;
861 }
862 case ISD::VP_STORE: {
863 const VPStoreSDNode *EST = cast<VPStoreSDNode>(N);
864 ID.AddInteger(EST->getMemoryVT().getRawBits());
865 ID.AddInteger(EST->getRawSubclassData());
866 ID.AddInteger(EST->getPointerInfo().getAddrSpace());
867 ID.AddInteger(EST->getMemOperand()->getFlags());
868 break;
869 }
870 case ISD::EXPERIMENTAL_VP_STRIDED_LOAD: {
872 ID.AddInteger(SLD->getMemoryVT().getRawBits());
873 ID.AddInteger(SLD->getRawSubclassData());
874 ID.AddInteger(SLD->getPointerInfo().getAddrSpace());
875 break;
876 }
877 case ISD::EXPERIMENTAL_VP_STRIDED_STORE: {
879 ID.AddInteger(SST->getMemoryVT().getRawBits());
880 ID.AddInteger(SST->getRawSubclassData());
881 ID.AddInteger(SST->getPointerInfo().getAddrSpace());
882 break;
883 }
884 case ISD::VP_GATHER: {
886 ID.AddInteger(EG->getMemoryVT().getRawBits());
887 ID.AddInteger(EG->getRawSubclassData());
888 ID.AddInteger(EG->getPointerInfo().getAddrSpace());
889 ID.AddInteger(EG->getMemOperand()->getFlags());
890 break;
891 }
892 case ISD::VP_SCATTER: {
894 ID.AddInteger(ES->getMemoryVT().getRawBits());
895 ID.AddInteger(ES->getRawSubclassData());
896 ID.AddInteger(ES->getPointerInfo().getAddrSpace());
897 ID.AddInteger(ES->getMemOperand()->getFlags());
898 break;
899 }
900 case ISD::MLOAD: {
902 ID.AddInteger(MLD->getMemoryVT().getRawBits());
903 ID.AddInteger(MLD->getRawSubclassData());
904 ID.AddInteger(MLD->getPointerInfo().getAddrSpace());
905 ID.AddInteger(MLD->getMemOperand()->getFlags());
906 break;
907 }
908 case ISD::MSTORE: {
910 ID.AddInteger(MST->getMemoryVT().getRawBits());
911 ID.AddInteger(MST->getRawSubclassData());
912 ID.AddInteger(MST->getPointerInfo().getAddrSpace());
913 ID.AddInteger(MST->getMemOperand()->getFlags());
914 break;
915 }
916 case ISD::MGATHER: {
918 ID.AddInteger(MG->getMemoryVT().getRawBits());
919 ID.AddInteger(MG->getRawSubclassData());
920 ID.AddInteger(MG->getPointerInfo().getAddrSpace());
921 ID.AddInteger(MG->getMemOperand()->getFlags());
922 break;
923 }
924 case ISD::MSCATTER: {
926 ID.AddInteger(MS->getMemoryVT().getRawBits());
927 ID.AddInteger(MS->getRawSubclassData());
928 ID.AddInteger(MS->getPointerInfo().getAddrSpace());
929 ID.AddInteger(MS->getMemOperand()->getFlags());
930 break;
931 }
934 case ISD::ATOMIC_SWAP:
946 case ISD::ATOMIC_LOAD:
947 case ISD::ATOMIC_STORE: {
948 const AtomicSDNode *AT = cast<AtomicSDNode>(N);
949 ID.AddInteger(AT->getMemoryVT().getRawBits());
950 ID.AddInteger(AT->getRawSubclassData());
951 ID.AddInteger(AT->getPointerInfo().getAddrSpace());
952 ID.AddInteger(AT->getMemOperand()->getFlags());
953 break;
954 }
955 case ISD::VECTOR_SHUFFLE: {
956 ArrayRef<int> Mask = cast<ShuffleVectorSDNode>(N)->getMask();
957 for (int M : Mask)
958 ID.AddInteger(M);
959 break;
960 }
961 case ISD::ADDRSPACECAST: {
963 ID.AddInteger(ASC->getSrcAddressSpace());
964 ID.AddInteger(ASC->getDestAddressSpace());
965 break;
966 }
968 case ISD::BlockAddress: {
970 ID.AddPointer(BA->getBlockAddress());
971 ID.AddInteger(BA->getOffset());
972 ID.AddInteger(BA->getTargetFlags());
973 break;
974 }
975 case ISD::AssertAlign:
976 ID.AddInteger(cast<AssertAlignSDNode>(N)->getAlign().value());
977 break;
978 case ISD::PREFETCH:
981 // Handled by MemIntrinsicSDNode check after the switch.
982 break;
984 ID.AddPointer(cast<MDNodeSDNode>(N)->getMD());
985 break;
986 } // end switch (N->getOpcode())
987
988 // MemIntrinsic nodes could also have subclass data, address spaces, and flags
989 // to check.
990 if (auto *MN = dyn_cast<MemIntrinsicSDNode>(N)) {
991 ID.AddInteger(MN->getRawSubclassData());
992 ID.AddInteger(MN->getMemoryVT().getRawBits());
993 for (const MachineMemOperand *MMO : MN->memoperands()) {
994 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
995 ID.AddInteger(MMO->getFlags());
996 }
997 }
998}
999
1000/// AddNodeIDNode - Generic routine for adding a nodes info to the NodeID
1001/// data.
1003 AddNodeIDOpcode(ID, N->getOpcode());
1004 // Add the return value info.
1005 AddNodeIDValueTypes(ID, N->getVTList());
1006 // Add the operand info.
1007 AddNodeIDOperands(ID, N->ops());
1008
1009 // Handle SDNode leafs with special info.
1011}
1012
1013//===----------------------------------------------------------------------===//
1014// SelectionDAG Class
1015//===----------------------------------------------------------------------===//
1016
1017/// doNotCSE - Return true if CSE should not be performed for this node.
1018static bool doNotCSE(SDNode *N) {
1019 if (N->getValueType(0) == MVT::Glue)
1020 return true; // Never CSE anything that produces a glue result.
1021
1022 switch (N->getOpcode()) {
1023 default: break;
1024 case ISD::HANDLENODE:
1025 case ISD::EH_LABEL:
1026 return true; // Never CSE these nodes.
1027 }
1028
1029 // Check that remaining values produced are not flags.
1030 for (unsigned i = 1, e = N->getNumValues(); i != e; ++i)
1031 if (N->getValueType(i) == MVT::Glue)
1032 return true; // Never CSE anything that produces a glue result.
1033
1034 return false;
1035}
1036
1037/// RemoveDeadNodes - This method deletes all unreachable nodes in the
1038/// SelectionDAG.
1040 // Create a dummy node (which is not added to allnodes), that adds a reference
1041 // to the root node, preventing it from being deleted.
1042 HandleSDNode Dummy(getRoot());
1043
1044 SmallVector<SDNode*, 128> DeadNodes;
1045
1046 // Add all obviously-dead nodes to the DeadNodes worklist.
1047 for (SDNode &Node : allnodes())
1048 if (Node.use_empty())
1049 DeadNodes.push_back(&Node);
1050
1051 RemoveDeadNodes(DeadNodes);
1052
1053 // If the root changed (e.g. it was a dead load, update the root).
1054 setRoot(Dummy.getValue());
1055}
1056
1057/// RemoveDeadNodes - This method deletes the unreachable nodes in the
1058/// given list, and any nodes that become unreachable as a result.
1060
1061 // Process the worklist, deleting the nodes and adding their uses to the
1062 // worklist.
1063 while (!DeadNodes.empty()) {
1064 SDNode *N = DeadNodes.pop_back_val();
1065 // Skip to next node if we've already managed to delete the node. This could
1066 // happen if replacing a node causes a node previously added to the node to
1067 // be deleted.
1068 if (N->getOpcode() == ISD::DELETED_NODE)
1069 continue;
1070
1071 for (DAGUpdateListener *DUL = UpdateListeners; DUL; DUL = DUL->Next)
1072 DUL->NodeDeleted(N, nullptr);
1073
1074 // Take the node out of the appropriate CSE map.
1075 RemoveNodeFromCSEMaps(N);
1076
1077 // Next, brutally remove the operand list. This is safe to do, as there are
1078 // no cycles in the graph.
1079 for (SDNode::op_iterator I = N->op_begin(), E = N->op_end(); I != E; ) {
1080 SDUse &Use = *I++;
1081 SDNode *Operand = Use.getNode();
1082 Use.set(SDValue());
1083
1084 // Now that we removed this operand, see if there are no uses of it left.
1085 if (Operand->use_empty())
1086 DeadNodes.push_back(Operand);
1087 }
1088
1089 DeallocateNode(N);
1090 }
1091}
1092
1094 SmallVector<SDNode*, 16> DeadNodes(1, N);
1095
1096 // Create a dummy node that adds a reference to the root node, preventing
1097 // it from being deleted. (This matters if the root is an operand of the
1098 // dead node.)
1099 HandleSDNode Dummy(getRoot());
1100
1101 RemoveDeadNodes(DeadNodes);
1102}
1103
1105 // First take this out of the appropriate CSE map.
1106 RemoveNodeFromCSEMaps(N);
1107
1108 // Finally, remove uses due to operands of this node, remove from the
1109 // AllNodes list, and delete the node.
1110 DeleteNodeNotInCSEMaps(N);
1111}
1112
1113void SelectionDAG::DeleteNodeNotInCSEMaps(SDNode *N) {
1114 assert(N->getIterator() != AllNodes.begin() &&
1115 "Cannot delete the entry node!");
1116 assert(N->use_empty() && "Cannot delete a node that is not dead!");
1117
1118 // Drop all of the operands and decrement used node's use counts.
1119 N->DropOperands();
1120
1121 DeallocateNode(N);
1122}
1123
1124void SDDbgInfo::add(SDDbgValue *V, bool isParameter) {
1125 assert(!(V->isVariadic() && isParameter));
1126 if (isParameter)
1127 ByvalParmDbgValues.push_back(V);
1128 else
1129 DbgValues.push_back(V);
1130 for (const SDNode *Node : V->getSDNodes())
1131 if (Node)
1132 DbgValMap[Node].push_back(V);
1133}
1134
1136 DbgValMapType::iterator I = DbgValMap.find(Node);
1137 if (I == DbgValMap.end())
1138 return;
1139 for (auto &Val: I->second)
1140 Val->setIsInvalidated();
1141 DbgValMap.erase(I);
1142}
1143
1144void SelectionDAG::DeallocateNode(SDNode *N) {
1145 // If we have operands, deallocate them.
1147
1148 NodeAllocator.Deallocate(AllNodes.remove(N));
1149
1150 // Set the opcode to DELETED_NODE to help catch bugs when node
1151 // memory is reallocated.
1152 // FIXME: There are places in SDag that have grown a dependency on the opcode
1153 // value in the released node.
1154 __asan_unpoison_memory_region(&N->NodeType, sizeof(N->NodeType));
1155 N->NodeType = ISD::DELETED_NODE;
1156
1157 // If any of the SDDbgValue nodes refer to this SDNode, invalidate
1158 // them and forget about that node.
1159 DbgInfo->erase(N);
1160
1161 // Invalidate extra info.
1162 SDEI.erase(N);
1163}
1164
1165#ifndef NDEBUG
1166/// VerifySDNode - Check the given SDNode. Aborts if it is invalid.
1167void SelectionDAG::verifyNode(SDNode *N) const {
1168 switch (N->getOpcode()) {
1169 default:
1170 if (N->isTargetOpcode())
1172 break;
1173 case ISD::BUILD_PAIR: {
1174 EVT VT = N->getValueType(0);
1175 assert(N->getNumValues() == 1 && "Too many results!");
1176 assert(!VT.isVector() && (VT.isInteger() || VT.isFloatingPoint()) &&
1177 "Wrong return type!");
1178 assert(N->getNumOperands() == 2 && "Wrong number of operands!");
1179 assert(N->getOperand(0).getValueType() == N->getOperand(1).getValueType() &&
1180 "Mismatched operand types!");
1181 assert(N->getOperand(0).getValueType().isInteger() == VT.isInteger() &&
1182 "Wrong operand type!");
1183 assert(VT.getSizeInBits() == 2 * N->getOperand(0).getValueSizeInBits() &&
1184 "Wrong return type size");
1185 break;
1186 }
1187 case ISD::BUILD_VECTOR: {
1188 assert(N->getNumValues() == 1 && "Too many results!");
1189 assert(N->getValueType(0).isVector() && "Wrong return type!");
1190 assert(N->getNumOperands() == N->getValueType(0).getVectorNumElements() &&
1191 "Wrong number of operands!");
1192 EVT EltVT = N->getValueType(0).getVectorElementType();
1193 for (const SDUse &Op : N->ops()) {
1194 assert((Op.getValueType() == EltVT ||
1195 (EltVT.isInteger() && Op.getValueType().isInteger() &&
1196 EltVT.bitsLE(Op.getValueType()))) &&
1197 "Wrong operand type!");
1198 assert(Op.getValueType() == N->getOperand(0).getValueType() &&
1199 "Operands must all have the same type");
1200 }
1201 break;
1202 }
1203 case ISD::SADDO:
1204 case ISD::UADDO:
1205 case ISD::SSUBO:
1206 case ISD::USUBO:
1207 assert(N->getNumValues() == 2 && "Wrong number of results!");
1208 assert(N->getVTList().NumVTs == 2 && N->getNumOperands() == 2 &&
1209 "Invalid add/sub overflow op!");
1210 assert(N->getVTList().VTs[0].isInteger() &&
1211 N->getVTList().VTs[1].isInteger() &&
1212 N->getOperand(0).getValueType() == N->getOperand(1).getValueType() &&
1213 N->getOperand(0).getValueType() == N->getVTList().VTs[0] &&
1214 "Binary operator types must match!");
1215 break;
1216 }
1217}
1218#endif // NDEBUG
1219
1220/// Insert a newly allocated node into the DAG.
1221///
1222/// Handles insertion into the all nodes list and CSE map, as well as
1223/// verification and other common operations when a new node is allocated.
1224void SelectionDAG::InsertNode(SDNode *N) {
1225 AllNodes.push_back(N);
1226#ifndef NDEBUG
1227 N->PersistentId = NextPersistentId++;
1228 verifyNode(N);
1229#endif
1230 for (DAGUpdateListener *DUL = UpdateListeners; DUL; DUL = DUL->Next)
1231 DUL->NodeInserted(N);
1232}
1233
1234/// RemoveNodeFromCSEMaps - Take the specified node out of the CSE map that
1235/// correspond to it. This is useful when we're about to delete or repurpose
1236/// the node. We don't want future request for structurally identical nodes
1237/// to return N anymore.
1238bool SelectionDAG::RemoveNodeFromCSEMaps(SDNode *N) {
1239 bool Erased = false;
1240 switch (N->getOpcode()) {
1241 case ISD::HANDLENODE: return false; // noop.
1242 case ISD::CONDCODE:
1243 assert(CondCodeNodes[cast<CondCodeSDNode>(N)->get()] &&
1244 "Cond code doesn't exist!");
1245 Erased = CondCodeNodes[cast<CondCodeSDNode>(N)->get()] != nullptr;
1246 CondCodeNodes[cast<CondCodeSDNode>(N)->get()] = nullptr;
1247 break;
1249 Erased = ExternalSymbols.erase(cast<ExternalSymbolSDNode>(N)->getSymbol());
1250 break;
1252 ExternalSymbolSDNode *ESN = cast<ExternalSymbolSDNode>(N);
1253 Erased = TargetExternalSymbols.erase(std::pair<std::string, unsigned>(
1254 ESN->getSymbol(), ESN->getTargetFlags()));
1255 break;
1256 }
1257 case ISD::MCSymbol: {
1258 auto *MCSN = cast<MCSymbolSDNode>(N);
1259 Erased = MCSymbols.erase(MCSN->getMCSymbol());
1260 break;
1261 }
1262 case ISD::VALUETYPE: {
1263 EVT VT = cast<VTSDNode>(N)->getVT();
1264 if (VT.isExtended()) {
1265 Erased = ExtendedValueTypeNodes.erase(VT);
1266 } else {
1267 Erased = ValueTypeNodes[VT.getSimpleVT().SimpleTy] != nullptr;
1268 ValueTypeNodes[VT.getSimpleVT().SimpleTy] = nullptr;
1269 }
1270 break;
1271 }
1272 default:
1273 // Remove it from the CSE Map.
1274 assert(N->getOpcode() != ISD::DELETED_NODE && "DELETED_NODE in CSEMap!");
1275 assert(N->getOpcode() != ISD::EntryToken && "EntryToken in CSEMap!");
1276 Erased = CSEMap.RemoveNode(N);
1277 break;
1278 }
1279#ifndef NDEBUG
1280 // Verify that the node was actually in one of the CSE maps, unless it has a
1281 // glue result (which cannot be CSE'd) or is one of the special cases that are
1282 // not subject to CSE.
1283 if (!Erased && N->getValueType(N->getNumValues()-1) != MVT::Glue &&
1284 !N->isMachineOpcode() && !doNotCSE(N)) {
1285 N->dump(this);
1286 dbgs() << "\n";
1287 llvm_unreachable("Node is not in map!");
1288 }
1289#endif
1290 return Erased;
1291}
1292
1293/// AddModifiedNodeToCSEMaps - The specified node has been removed from the CSE
1294/// maps and modified in place. Add it back to the CSE maps, unless an identical
1295/// node already exists, in which case transfer all its users to the existing
1296/// node. This transfer can potentially trigger recursive merging.
1297void
1298SelectionDAG::AddModifiedNodeToCSEMaps(SDNode *N) {
1299 // For node types that aren't CSE'd, just act as if no identical node
1300 // already exists.
1301 if (!doNotCSE(N)) {
1302 SDNode *Existing = CSEMap.GetOrInsertNode(N);
1303 if (Existing != N) {
1304 // If there was already an existing matching node, use ReplaceAllUsesWith
1305 // to replace the dead one with the existing one. This can cause
1306 // recursive merging of other unrelated nodes down the line.
1307 Existing->intersectFlagsWith(N->getFlags());
1308 if (auto *MemNode = dyn_cast<MemSDNode>(Existing))
1309 MemNode->refineRanges(cast<MemSDNode>(N)->memoperands());
1310 ReplaceAllUsesWith(N, Existing);
1311
1312 // N is now dead. Inform the listeners and delete it.
1313 for (DAGUpdateListener *DUL = UpdateListeners; DUL; DUL = DUL->Next)
1314 DUL->NodeDeleted(N, Existing);
1315 DeleteNodeNotInCSEMaps(N);
1316 return;
1317 }
1318 }
1319
1320 // If the node doesn't already exist, we updated it. Inform listeners.
1321 for (DAGUpdateListener *DUL = UpdateListeners; DUL; DUL = DUL->Next)
1322 DUL->NodeUpdated(N);
1323}
1324
1325/// FindModifiedNodeSlot - Find a slot for the specified node if its operands
1326/// were replaced with those specified. If this node is never memoized,
1327/// return null, otherwise return a pointer to the slot it would take. If a
1328/// node already exists with these operands, the slot will be non-null.
1329SDNode *SelectionDAG::FindModifiedNodeSlot(SDNode *N, SDValue Op,
1330 void *&InsertPos) {
1331 if (doNotCSE(N))
1332 return nullptr;
1333
1334 SDValue Ops[] = { Op };
1335 FoldingSetNodeID ID;
1336 AddNodeIDNode(ID, N->getOpcode(), N->getVTList(), Ops);
1338 SDNode *Node = FindNodeOrInsertPos(ID, SDLoc(N), InsertPos);
1339 if (Node)
1340 Node->intersectFlagsWith(N->getFlags());
1341 return Node;
1342}
1343
1344/// FindModifiedNodeSlot - Find a slot for the specified node if its operands
1345/// were replaced with those specified. If this node is never memoized,
1346/// return null, otherwise return a pointer to the slot it would take. If a
1347/// node already exists with these operands, the slot will be non-null.
1348SDNode *SelectionDAG::FindModifiedNodeSlot(SDNode *N,
1349 SDValue Op1, SDValue Op2,
1350 void *&InsertPos) {
1351 if (doNotCSE(N))
1352 return nullptr;
1353
1354 SDValue Ops[] = { Op1, Op2 };
1355 FoldingSetNodeID ID;
1356 AddNodeIDNode(ID, N->getOpcode(), N->getVTList(), Ops);
1358 SDNode *Node = FindNodeOrInsertPos(ID, SDLoc(N), InsertPos);
1359 if (Node)
1360 Node->intersectFlagsWith(N->getFlags());
1361 return Node;
1362}
1363
1364/// FindModifiedNodeSlot - Find a slot for the specified node if its operands
1365/// were replaced with those specified. If this node is never memoized,
1366/// return null, otherwise return a pointer to the slot it would take. If a
1367/// node already exists with these operands, the slot will be non-null.
1368SDNode *SelectionDAG::FindModifiedNodeSlot(SDNode *N, ArrayRef<SDValue> Ops,
1369 void *&InsertPos) {
1370 if (doNotCSE(N))
1371 return nullptr;
1372
1373 FoldingSetNodeID ID;
1374 AddNodeIDNode(ID, N->getOpcode(), N->getVTList(), Ops);
1376 SDNode *Node = FindNodeOrInsertPos(ID, SDLoc(N), InsertPos);
1377 if (Node)
1378 Node->intersectFlagsWith(N->getFlags());
1379 return Node;
1380}
1381
1383 Type *Ty = VT == MVT::iPTR ? PointerType::get(*getContext(), 0)
1384 : VT.getTypeForEVT(*getContext());
1385
1386 return getDataLayout().getABITypeAlign(Ty);
1387}
1388
1389// EntryNode could meaningfully have debug info if we can find it...
1391 : TM(tm), OptLevel(OL), EntryNode(ISD::EntryToken, 0, DebugLoc(),
1392 getVTList(MVT::Other, MVT::Glue)),
1393 Root(getEntryNode()) {
1394 InsertNode(&EntryNode);
1395 DbgInfo = new SDDbgInfo();
1396}
1397
1399 OptimizationRemarkEmitter &NewORE, Pass *PassPtr,
1400 const TargetLibraryInfo *LibraryInfo,
1401 const LibcallLoweringInfo *LibcallsInfo,
1402 UniformityInfo *NewUA, ProfileSummaryInfo *PSIin,
1404 FunctionVarLocs const *VarLocs) {
1405 MF = &NewMF;
1406 SDAGISelPass = PassPtr;
1407 ORE = &NewORE;
1410 LibInfo = LibraryInfo;
1411 Libcalls = LibcallsInfo;
1412 Context = &MF->getFunction().getContext();
1413 UA = NewUA;
1414 PSI = PSIin;
1415 BFI = BFIin;
1416 MMI = &MMIin;
1417 FnVarLocs = VarLocs;
1418}
1419
1421 assert(!UpdateListeners && "Dangling registered DAGUpdateListeners");
1422 allnodes_clear();
1423 OperandRecycler.clear(OperandAllocator);
1424 delete DbgInfo;
1425}
1426
1428 return llvm::shouldOptimizeForSize(FLI->MBB->getBasicBlock(), PSI, BFI);
1429}
1430
1431void SelectionDAG::allnodes_clear() {
1432 assert(&*AllNodes.begin() == &EntryNode);
1433 AllNodes.remove(AllNodes.begin());
1434 while (!AllNodes.empty())
1435 DeallocateNode(&AllNodes.front());
1436#ifndef NDEBUG
1437 NextPersistentId = 0;
1438#endif
1439}
1440
1441SDNode *SelectionDAG::FindNodeOrInsertPos(const FoldingSetNodeID &ID,
1442 void *&InsertPos) {
1443 SDNode *N = CSEMap.FindNodeOrInsertPos(ID, InsertPos);
1444 if (N) {
1445 switch (N->getOpcode()) {
1446 default: break;
1447 case ISD::Constant:
1448 case ISD::ConstantFP:
1449 llvm_unreachable("Querying for Constant and ConstantFP nodes requires "
1450 "debug location. Use another overload.");
1451 }
1452 }
1453 return N;
1454}
1455
1456SDNode *SelectionDAG::FindNodeOrInsertPos(const FoldingSetNodeID &ID,
1457 const SDLoc &DL, void *&InsertPos) {
1458 SDNode *N = CSEMap.FindNodeOrInsertPos(ID, InsertPos);
1459 if (N) {
1460 switch (N->getOpcode()) {
1461 case ISD::Constant:
1462 case ISD::ConstantFP:
1463 // Erase debug location from the node if the node is used at several
1464 // different places. Do not propagate one location to all uses as it
1465 // will cause a worse single stepping debugging experience.
1466 if (N->getDebugLoc() != DL.getDebugLoc())
1467 N->setDebugLoc(DebugLoc());
1468 break;
1469 default:
1470 // When the node's point of use is located earlier in the instruction
1471 // sequence than its prior point of use, update its debug info to the
1472 // earlier location.
1473 if (DL.getIROrder() && DL.getIROrder() < N->getIROrder())
1474 N->setDebugLoc(DL.getDebugLoc());
1475 break;
1476 }
1477 }
1478 return N;
1479}
1480
1482 allnodes_clear();
1483 OperandRecycler.clear(OperandAllocator);
1484 OperandAllocator.Reset();
1485 CSEMap.clear();
1486
1487 ExtendedValueTypeNodes.clear();
1488 ExternalSymbols.clear();
1489 TargetExternalSymbols.clear();
1490 MCSymbols.clear();
1491 SDEI.clear();
1492 llvm::fill(CondCodeNodes, nullptr);
1493 llvm::fill(ValueTypeNodes, nullptr);
1494
1495 EntryNode.UseList = nullptr;
1496 InsertNode(&EntryNode);
1497 Root = getEntryNode();
1498 DbgInfo->clear();
1499}
1500
1502 return VT.bitsGT(Op.getValueType())
1503 ? getNode(ISD::FP_EXTEND, DL, VT, Op)
1504 : getNode(ISD::FP_ROUND, DL, VT, Op,
1505 getIntPtrConstant(0, DL, /*isTarget=*/true));
1506}
1507
1508std::pair<SDValue, SDValue>
1510 const SDLoc &DL, EVT VT) {
1511 assert(!VT.bitsEq(Op.getValueType()) &&
1512 "Strict no-op FP extend/round not allowed.");
1513 SDValue Res =
1514 VT.bitsGT(Op.getValueType())
1515 ? getNode(ISD::STRICT_FP_EXTEND, DL, {VT, MVT::Other}, {Chain, Op})
1516 : getNode(ISD::STRICT_FP_ROUND, DL, {VT, MVT::Other},
1517 {Chain, Op, getIntPtrConstant(0, DL, /*isTarget=*/true)});
1518
1519 return std::pair<SDValue, SDValue>(Res, SDValue(Res.getNode(), 1));
1520}
1521
1523 return VT.bitsGT(Op.getValueType()) ?
1524 getNode(ISD::ANY_EXTEND, DL, VT, Op) :
1525 getNode(ISD::TRUNCATE, DL, VT, Op);
1526}
1527
1529 return VT.bitsGT(Op.getValueType()) ?
1530 getNode(ISD::SIGN_EXTEND, DL, VT, Op) :
1531 getNode(ISD::TRUNCATE, DL, VT, Op);
1532}
1533
1535 return VT.bitsGT(Op.getValueType()) ?
1536 getNode(ISD::ZERO_EXTEND, DL, VT, Op) :
1537 getNode(ISD::TRUNCATE, DL, VT, Op);
1538}
1539
1541 EVT VT) {
1542 assert(!VT.isVector());
1543 auto Type = Op.getValueType();
1544 SDValue DestOp;
1545 if (Type == VT)
1546 return Op;
1547 auto Size = Op.getValueSizeInBits();
1548 DestOp = getBitcast(EVT::getIntegerVT(*Context, Size), Op);
1549 if (DestOp.getValueType() == VT)
1550 return DestOp;
1551
1552 return getAnyExtOrTrunc(DestOp, DL, VT);
1553}
1554
1556 EVT VT) {
1557 assert(!VT.isVector());
1558 auto Type = Op.getValueType();
1559 SDValue DestOp;
1560 if (Type == VT)
1561 return Op;
1562 auto Size = Op.getValueSizeInBits();
1563 DestOp = getBitcast(MVT::getIntegerVT(Size), Op);
1564 if (DestOp.getValueType() == VT)
1565 return DestOp;
1566
1567 return getSExtOrTrunc(DestOp, DL, VT);
1568}
1569
1571 EVT VT) {
1572 assert(!VT.isVector());
1573 auto Type = Op.getValueType();
1574 SDValue DestOp;
1575 if (Type == VT)
1576 return Op;
1577 auto Size = Op.getValueSizeInBits();
1578 DestOp = getBitcast(MVT::getIntegerVT(Size), Op);
1579 if (DestOp.getValueType() == VT)
1580 return DestOp;
1581
1582 return getZExtOrTrunc(DestOp, DL, VT);
1583}
1584
1586 EVT OpVT) {
1587 if (VT.bitsLE(Op.getValueType()))
1588 return getNode(ISD::TRUNCATE, SL, VT, Op);
1589
1590 TargetLowering::BooleanContent BType = TLI->getBooleanContents(OpVT);
1591 return getNode(TLI->getExtendForContent(BType), SL, VT, Op);
1592}
1593
1595 EVT OpVT = Op.getValueType();
1596 assert(VT.isInteger() && OpVT.isInteger() &&
1597 "Cannot getZeroExtendInReg FP types");
1598 assert(VT.isVector() == OpVT.isVector() &&
1599 "getZeroExtendInReg type should be vector iff the operand "
1600 "type is vector!");
1601 assert((!VT.isVector() ||
1603 "Vector element counts must match in getZeroExtendInReg");
1604 assert(VT.getScalarType().bitsLE(OpVT.getScalarType()) && "Not extending!");
1605 if (OpVT == VT)
1606 return Op;
1607 // TODO: Use computeKnownBits instead of AssertZext.
1608 if (Op.getOpcode() == ISD::AssertZext && cast<VTSDNode>(Op.getOperand(1))
1609 ->getVT()
1610 .getScalarType()
1611 .bitsLE(VT.getScalarType()))
1612 return Op;
1614 VT.getScalarSizeInBits());
1615 return getNode(ISD::AND, DL, OpVT, Op, getConstant(Imm, DL, OpVT));
1616}
1617
1619 SDValue EVL, const SDLoc &DL,
1620 EVT VT) {
1621 EVT OpVT = Op.getValueType();
1622 assert(VT.isInteger() && OpVT.isInteger() &&
1623 "Cannot getVPZeroExtendInReg FP types");
1624 assert(VT.isVector() && OpVT.isVector() &&
1625 "getVPZeroExtendInReg type and operand type should be vector!");
1627 "Vector element counts must match in getZeroExtendInReg");
1628 assert(VT.getScalarType().bitsLE(OpVT.getScalarType()) && "Not extending!");
1629 if (OpVT == VT)
1630 return Op;
1632 VT.getScalarSizeInBits());
1633 return getNode(ISD::VP_AND, DL, OpVT, Op, getConstant(Imm, DL, OpVT), Mask,
1634 EVL);
1635}
1636
1638 // Only unsigned pointer semantics are supported right now. In the future this
1639 // might delegate to TLI to check pointer signedness.
1640 return getZExtOrTrunc(Op, DL, VT);
1641}
1642
1644 // Only unsigned pointer semantics are supported right now. In the future this
1645 // might delegate to TLI to check pointer signedness.
1646 return getZeroExtendInReg(Op, DL, VT);
1647}
1648
1650 return getNode(ISD::SUB, DL, VT, getConstant(0, DL, VT), Val);
1651}
1652
1653/// getNOT - Create a bitwise NOT operation as (XOR Val, -1).
1655 return getNode(ISD::XOR, DL, VT, Val, getAllOnesConstant(DL, VT));
1656}
1657
1659 SDValue TrueValue = getBoolConstant(true, DL, VT, VT);
1660 return getNode(ISD::XOR, DL, VT, Val, TrueValue);
1661}
1662
1664 SDValue Mask, SDValue EVL, EVT VT) {
1665 SDValue TrueValue = getBoolConstant(true, DL, VT, VT);
1666 return getNode(ISD::VP_XOR, DL, VT, Val, TrueValue, Mask, EVL);
1667}
1668
1670 SDValue Mask, SDValue EVL) {
1671 return getVPZExtOrTrunc(DL, VT, Op, Mask, EVL);
1672}
1673
1675 SDValue Mask, SDValue EVL) {
1676 if (VT.bitsGT(Op.getValueType()))
1677 return getNode(ISD::VP_ZERO_EXTEND, DL, VT, Op, Mask, EVL);
1678 if (VT.bitsLT(Op.getValueType()))
1679 return getNode(ISD::VP_TRUNCATE, DL, VT, Op, Mask, EVL);
1680 return Op;
1681}
1682
1684 EVT OpVT) {
1685 if (!V)
1686 return getConstant(0, DL, VT);
1687
1688 switch (TLI->getBooleanContents(OpVT)) {
1691 return getConstant(1, DL, VT);
1693 return getAllOnesConstant(DL, VT);
1694 }
1695 llvm_unreachable("Unexpected boolean content enum!");
1696}
1697
1699 bool isT, bool isO) {
1700 return getConstant(APInt(VT.getScalarSizeInBits(), Val, /*isSigned=*/false),
1701 DL, VT, isT, isO);
1702}
1703
1705 bool isT, bool isO) {
1706 return getConstant(*ConstantInt::get(*Context, Val), DL, VT, isT, isO);
1707}
1708
1710 EVT VT, bool isT, bool isO) {
1711 assert(VT.isInteger() && "Cannot create FP integer constant!");
1712
1713 EVT EltVT = VT.getScalarType();
1714 const ConstantInt *Elt = &Val;
1715
1716 // Vector splats are explicit within the DAG, with ConstantSDNode holding the
1717 // to-be-splatted scalar ConstantInt.
1718 if (isa<VectorType>(Elt->getType()))
1719 Elt = ConstantInt::get(*getContext(), Elt->getValue());
1720
1721 // In some cases the vector type is legal but the element type is illegal and
1722 // needs to be promoted, for example v8i8 on ARM. In this case, promote the
1723 // inserted value (the type does not need to match the vector element type).
1724 // Any extra bits introduced will be truncated away.
1725 if (VT.isVector() && TLI->getTypeAction(*getContext(), EltVT) ==
1727 EltVT = TLI->getTypeToTransformTo(*getContext(), EltVT);
1728 APInt NewVal;
1729 if (TLI->isSExtCheaperThanZExt(VT.getScalarType(), EltVT))
1730 NewVal = Elt->getValue().sextOrTrunc(EltVT.getSizeInBits());
1731 else
1732 NewVal = Elt->getValue().zextOrTrunc(EltVT.getSizeInBits());
1733 Elt = ConstantInt::get(*getContext(), NewVal);
1734 }
1735 // In other cases the element type is illegal and needs to be expanded, for
1736 // example v2i64 on MIPS32. In this case, find the nearest legal type, split
1737 // the value into n parts and use a vector type with n-times the elements.
1738 // Then bitcast to the type requested.
1739 // Legalizing constants too early makes the DAGCombiner's job harder so we
1740 // only legalize if the DAG tells us we must produce legal types.
1741 else if (NewNodesMustHaveLegalTypes && VT.isVector() &&
1742 TLI->getTypeAction(*getContext(), EltVT) ==
1744 const APInt &NewVal = Elt->getValue();
1745 EVT ViaEltVT = TLI->getTypeToTransformTo(*getContext(), EltVT);
1746 unsigned ViaEltSizeInBits = ViaEltVT.getSizeInBits();
1747
1748 // For scalable vectors, try to use a SPLAT_VECTOR_PARTS node.
1749 if (VT.isScalableVector() ||
1750 TLI->isOperationLegal(ISD::SPLAT_VECTOR, VT)) {
1751 assert(EltVT.getSizeInBits() % ViaEltSizeInBits == 0 &&
1752 "Can only handle an even split!");
1753 unsigned Parts = EltVT.getSizeInBits() / ViaEltSizeInBits;
1754
1755 SmallVector<SDValue, 2> ScalarParts;
1756 for (unsigned i = 0; i != Parts; ++i)
1757 ScalarParts.push_back(getConstant(
1758 NewVal.extractBits(ViaEltSizeInBits, i * ViaEltSizeInBits), DL,
1759 ViaEltVT, isT, isO));
1760
1761 return getNode(ISD::SPLAT_VECTOR_PARTS, DL, VT, ScalarParts);
1762 }
1763
1764 unsigned ViaVecNumElts = VT.getSizeInBits() / ViaEltSizeInBits;
1765 EVT ViaVecVT = EVT::getVectorVT(*getContext(), ViaEltVT, ViaVecNumElts);
1766
1767 // Check the temporary vector is the correct size. If this fails then
1768 // getTypeToTransformTo() probably returned a type whose size (in bits)
1769 // isn't a power-of-2 factor of the requested type size.
1770 assert(ViaVecVT.getSizeInBits() == VT.getSizeInBits());
1771
1772 SmallVector<SDValue, 2> EltParts;
1773 for (unsigned i = 0; i < ViaVecNumElts / VT.getVectorNumElements(); ++i)
1774 EltParts.push_back(getConstant(
1775 NewVal.extractBits(ViaEltSizeInBits, i * ViaEltSizeInBits), DL,
1776 ViaEltVT, isT, isO));
1777
1778 // EltParts is currently in little endian order. If we actually want
1779 // big-endian order then reverse it now.
1780 if (getDataLayout().isBigEndian())
1781 std::reverse(EltParts.begin(), EltParts.end());
1782
1783 // The elements must be reversed when the element order is different
1784 // to the endianness of the elements (because the BITCAST is itself a
1785 // vector shuffle in this situation). However, we do not need any code to
1786 // perform this reversal because getConstant() is producing a vector
1787 // splat.
1788 // This situation occurs in MIPS MSA.
1789
1791 for (unsigned i = 0, e = VT.getVectorNumElements(); i != e; ++i)
1792 llvm::append_range(Ops, EltParts);
1793
1794 SDValue V =
1795 getNode(ISD::BITCAST, DL, VT, getBuildVector(ViaVecVT, DL, Ops));
1796 return V;
1797 }
1798
1799 assert(Elt->getBitWidth() == EltVT.getSizeInBits() &&
1800 "APInt size does not match type size!");
1801 unsigned Opc = isT ? ISD::TargetConstant : ISD::Constant;
1802 SDVTList VTs = getVTList(EltVT);
1804 AddNodeIDNode(ID, Opc, VTs, {});
1805 ID.AddPointer(Elt);
1806 ID.AddBoolean(isO);
1807 void *IP = nullptr;
1808 SDNode *N = nullptr;
1809 if ((N = FindNodeOrInsertPos(ID, DL, IP)))
1810 if (!VT.isVector())
1811 return SDValue(N, 0);
1812
1813 if (!N) {
1814 N = newSDNode<ConstantSDNode>(isT, isO, Elt, VTs);
1815 CSEMap.InsertNode(N, IP);
1816 InsertNode(N);
1817 NewSDValueDbgMsg(SDValue(N, 0), "Creating constant: ", this);
1818 }
1819
1820 SDValue Result(N, 0);
1821 if (VT.isVector())
1822 Result = getSplat(VT, DL, Result);
1823 return Result;
1824}
1825
1827 bool isT, bool isO) {
1828 unsigned Size = VT.getScalarSizeInBits();
1829 return getConstant(APInt(Size, Val, /*isSigned=*/true), DL, VT, isT, isO);
1830}
1831
1833 bool IsOpaque) {
1835 IsTarget, IsOpaque);
1836}
1837
1839 bool isTarget) {
1840 return getConstant(Val, DL, TLI->getPointerTy(getDataLayout()), isTarget);
1841}
1842
1844 const SDLoc &DL) {
1845 assert(VT.isInteger() && "Shift amount is not an integer type!");
1846 EVT ShiftVT = TLI->getShiftAmountTy(VT, getDataLayout());
1847 return getConstant(Val, DL, ShiftVT);
1848}
1849
1851 const SDLoc &DL) {
1852 assert(Val.ult(VT.getScalarSizeInBits()) && "Out of range shift");
1853 return getShiftAmountConstant(Val.getZExtValue(), VT, DL);
1854}
1855
1857 bool isTarget) {
1858 return getConstant(Val, DL, TLI->getVectorIdxTy(getDataLayout()), isTarget);
1859}
1860
1862 bool isTarget) {
1863 return getConstantFP(*ConstantFP::get(*getContext(), V), DL, VT, isTarget);
1864}
1865
1867 EVT VT, bool isTarget) {
1868 assert(VT.isFloatingPoint() && "Cannot create integer FP constant!");
1869
1870 EVT EltVT = VT.getScalarType();
1871 const ConstantFP *Elt = &V;
1872
1873 // Vector splats are explicit within the DAG, with ConstantFPSDNode holding
1874 // the to-be-splatted scalar ConstantFP.
1875 if (isa<VectorType>(Elt->getType()))
1876 Elt = ConstantFP::get(*getContext(), Elt->getValue());
1877
1878 // Do the map lookup using the actual bit pattern for the floating point
1879 // value, so that we don't have problems with 0.0 comparing equal to -0.0, and
1880 // we don't have issues with SNANs.
1881 unsigned Opc = isTarget ? ISD::TargetConstantFP : ISD::ConstantFP;
1882 SDVTList VTs = getVTList(EltVT);
1884 AddNodeIDNode(ID, Opc, VTs, {});
1885 ID.AddPointer(Elt);
1886 void *IP = nullptr;
1887 SDNode *N = nullptr;
1888 if ((N = FindNodeOrInsertPos(ID, DL, IP)))
1889 if (!VT.isVector())
1890 return SDValue(N, 0);
1891
1892 if (!N) {
1893 N = newSDNode<ConstantFPSDNode>(isTarget, Elt, VTs);
1894 CSEMap.InsertNode(N, IP);
1895 InsertNode(N);
1896 }
1897
1898 SDValue Result(N, 0);
1899 if (VT.isVector())
1900 Result = getSplat(VT, DL, Result);
1901 NewSDValueDbgMsg(Result, "Creating fp constant: ", this);
1902 return Result;
1903}
1904
1906 bool isTarget) {
1907 EVT EltVT = VT.getScalarType();
1908 if (EltVT == MVT::f32)
1909 return getConstantFP(APFloat((float)Val), DL, VT, isTarget);
1910 if (EltVT == MVT::f64)
1911 return getConstantFP(APFloat(Val), DL, VT, isTarget);
1912 if (EltVT == MVT::f80 || EltVT == MVT::f128 || EltVT == MVT::ppcf128 ||
1913 EltVT == MVT::f16 || EltVT == MVT::bf16) {
1914 bool Ignored;
1915 APFloat APF = APFloat(Val);
1917 &Ignored);
1918 return getConstantFP(APF, DL, VT, isTarget);
1919 }
1920 llvm_unreachable("Unsupported type in getConstantFP");
1921}
1922
1924 EVT VT, int64_t Offset, bool isTargetGA,
1925 unsigned TargetFlags) {
1926 assert((TargetFlags == 0 || isTargetGA) &&
1927 "Cannot set target flags on target-independent globals");
1928
1929 // Truncate (with sign-extension) the offset value to the pointer size.
1931 if (BitWidth < 64)
1933
1934 unsigned Opc;
1935 if (GV->isThreadLocal())
1937 else
1939
1940 SDVTList VTs = getVTList(VT);
1942 AddNodeIDNode(ID, Opc, VTs, {});
1943 ID.AddPointer(GV);
1944 ID.AddInteger(Offset);
1945 ID.AddInteger(TargetFlags);
1946 void *IP = nullptr;
1947 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP))
1948 return SDValue(E, 0);
1949
1950 auto *N = newSDNode<GlobalAddressSDNode>(
1951 Opc, DL.getIROrder(), DL.getDebugLoc(), GV, VTs, Offset, TargetFlags);
1952 CSEMap.InsertNode(N, IP);
1953 InsertNode(N);
1954 return SDValue(N, 0);
1955}
1956
1958 SDVTList VTs = getVTList(MVT::Untyped);
1961 ID.AddPointer(GV);
1962 void *IP = nullptr;
1963 if (SDNode *E = FindNodeOrInsertPos(ID, SDLoc(), IP))
1964 return SDValue(E, 0);
1965
1966 auto *N = newSDNode<DeactivationSymbolSDNode>(GV, VTs);
1967 CSEMap.InsertNode(N, IP);
1968 InsertNode(N);
1969 return SDValue(N, 0);
1970}
1971
1972SDValue SelectionDAG::getFrameIndex(int FI, EVT VT, bool isTarget) {
1973 unsigned Opc = isTarget ? ISD::TargetFrameIndex : ISD::FrameIndex;
1974 SDVTList VTs = getVTList(VT);
1976 AddNodeIDNode(ID, Opc, VTs, {});
1977 ID.AddInteger(FI);
1978 void *IP = nullptr;
1979 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
1980 return SDValue(E, 0);
1981
1982 auto *N = newSDNode<FrameIndexSDNode>(FI, VTs, isTarget);
1983 CSEMap.InsertNode(N, IP);
1984 InsertNode(N);
1985 return SDValue(N, 0);
1986}
1987
1988SDValue SelectionDAG::getJumpTable(int JTI, EVT VT, bool isTarget,
1989 unsigned TargetFlags) {
1990 assert((TargetFlags == 0 || isTarget) &&
1991 "Cannot set target flags on target-independent jump tables");
1992 unsigned Opc = isTarget ? ISD::TargetJumpTable : ISD::JumpTable;
1993 SDVTList VTs = getVTList(VT);
1995 AddNodeIDNode(ID, Opc, VTs, {});
1996 ID.AddInteger(JTI);
1997 ID.AddInteger(TargetFlags);
1998 void *IP = nullptr;
1999 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2000 return SDValue(E, 0);
2001
2002 auto *N = newSDNode<JumpTableSDNode>(JTI, VTs, isTarget, TargetFlags);
2003 CSEMap.InsertNode(N, IP);
2004 InsertNode(N);
2005 return SDValue(N, 0);
2006}
2007
2009 const SDLoc &DL) {
2011 return getNode(ISD::JUMP_TABLE_DEBUG_INFO, DL, MVT::Other, Chain,
2012 getTargetConstant(static_cast<uint64_t>(JTI), DL, PTy, true));
2013}
2014
2016 MaybeAlign Alignment, int Offset,
2017 bool isTarget, unsigned TargetFlags) {
2018 assert((TargetFlags == 0 || isTarget) &&
2019 "Cannot set target flags on target-independent globals");
2020 if (!Alignment)
2021 Alignment = shouldOptForSize()
2022 ? getDataLayout().getABITypeAlign(C->getType())
2023 : getDataLayout().getPrefTypeAlign(C->getType());
2024 unsigned Opc = isTarget ? ISD::TargetConstantPool : ISD::ConstantPool;
2025 SDVTList VTs = getVTList(VT);
2027 AddNodeIDNode(ID, Opc, VTs, {});
2028 ID.AddInteger(Alignment->value());
2029 ID.AddInteger(Offset);
2030 ID.AddPointer(C);
2031 ID.AddInteger(TargetFlags);
2032 void *IP = nullptr;
2033 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2034 return SDValue(E, 0);
2035
2036 auto *N = newSDNode<ConstantPoolSDNode>(isTarget, C, VTs, Offset, *Alignment,
2037 TargetFlags);
2038 CSEMap.InsertNode(N, IP);
2039 InsertNode(N);
2040 SDValue V = SDValue(N, 0);
2041 NewSDValueDbgMsg(V, "Creating new constant pool: ", this);
2042 return V;
2043}
2044
2046 MaybeAlign Alignment, int Offset,
2047 bool isTarget, unsigned TargetFlags) {
2048 assert((TargetFlags == 0 || isTarget) &&
2049 "Cannot set target flags on target-independent globals");
2050 if (!Alignment)
2051 Alignment = getDataLayout().getPrefTypeAlign(C->getType());
2052 unsigned Opc = isTarget ? ISD::TargetConstantPool : ISD::ConstantPool;
2053 SDVTList VTs = getVTList(VT);
2055 AddNodeIDNode(ID, Opc, VTs, {});
2056 ID.AddInteger(Alignment->value());
2057 ID.AddInteger(Offset);
2058 C->addSelectionDAGCSEId(ID);
2059 ID.AddInteger(TargetFlags);
2060 void *IP = nullptr;
2061 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2062 return SDValue(E, 0);
2063
2064 auto *N = newSDNode<ConstantPoolSDNode>(isTarget, C, VTs, Offset, *Alignment,
2065 TargetFlags);
2066 CSEMap.InsertNode(N, IP);
2067 InsertNode(N);
2068 return SDValue(N, 0);
2069}
2070
2073 AddNodeIDNode(ID, ISD::BasicBlock, getVTList(MVT::Other), {});
2074 ID.AddPointer(MBB);
2075 void *IP = nullptr;
2076 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2077 return SDValue(E, 0);
2078
2079 auto *N = newSDNode<BasicBlockSDNode>(MBB);
2080 CSEMap.InsertNode(N, IP);
2081 InsertNode(N);
2082 return SDValue(N, 0);
2083}
2084
2086 if (VT.isSimple() && (unsigned)VT.getSimpleVT().SimpleTy >=
2087 ValueTypeNodes.size())
2088 ValueTypeNodes.resize(VT.getSimpleVT().SimpleTy+1);
2089
2090 SDNode *&N = VT.isExtended() ?
2091 ExtendedValueTypeNodes[VT] : ValueTypeNodes[VT.getSimpleVT().SimpleTy];
2092
2093 if (N) return SDValue(N, 0);
2094 N = newSDNode<VTSDNode>(VT);
2095 InsertNode(N);
2096 return SDValue(N, 0);
2097}
2098
2100 SDNode *&N = ExternalSymbols[Sym];
2101 if (N) return SDValue(N, 0);
2102 N = newSDNode<ExternalSymbolSDNode>(false, Sym, 0, getVTList(VT));
2103 InsertNode(N);
2104 return SDValue(N, 0);
2105}
2106
2107SDValue SelectionDAG::getExternalSymbol(RTLIB::LibcallImpl Libcall, EVT VT) {
2109 return getExternalSymbol(SymName.data(), VT);
2110}
2111
2113 SDNode *&N = MCSymbols[Sym];
2114 if (N)
2115 return SDValue(N, 0);
2116 N = newSDNode<MCSymbolSDNode>(Sym, getVTList(VT));
2117 InsertNode(N);
2118 return SDValue(N, 0);
2119}
2120
2122 unsigned TargetFlags) {
2123 SDNode *&N =
2124 TargetExternalSymbols[std::pair<std::string, unsigned>(Sym, TargetFlags)];
2125 if (N) return SDValue(N, 0);
2126 N = newSDNode<ExternalSymbolSDNode>(true, Sym, TargetFlags, getVTList(VT));
2127 InsertNode(N);
2128 return SDValue(N, 0);
2129}
2130
2132 EVT VT, unsigned TargetFlags) {
2134 return getTargetExternalSymbol(SymName.data(), VT, TargetFlags);
2135}
2136
2138 if ((unsigned)Cond >= CondCodeNodes.size())
2139 CondCodeNodes.resize(Cond+1);
2140
2141 if (!CondCodeNodes[Cond]) {
2142 auto *N = newSDNode<CondCodeSDNode>(Cond);
2143 CondCodeNodes[Cond] = N;
2144 InsertNode(N);
2145 }
2146
2147 return SDValue(CondCodeNodes[Cond], 0);
2148}
2149
2151 assert(MulImm.getBitWidth() == VT.getSizeInBits() &&
2152 "APInt size does not match type size!");
2153
2154 if (MulImm == 0)
2155 return getConstant(0, DL, VT);
2156
2157 const MachineFunction &MF = getMachineFunction();
2158 const Function &F = MF.getFunction();
2159 ConstantRange CR = getVScaleRange(&F, 64);
2160 if (const APInt *C = CR.getSingleElement())
2161 return getConstant(MulImm * C->getZExtValue(), DL, VT);
2162
2163 return getNode(ISD::VSCALE, DL, VT, getConstant(MulImm, DL, VT));
2164}
2165
2166/// \returns a value of type \p VT that represents the runtime value of \p
2167/// Quantity, i.e. scaled by vscale if it's scalable, or a fixed constant
2168/// otherwise. Quantity should be a FixedOrScalableQuantity, i.e. ElementCount
2169/// or TypeSize.
2170template <typename Ty>
2172 EVT VT, Ty Quantity) {
2173 if (Quantity.isScalable())
2174 return DAG.getVScale(
2175 DL, VT, APInt(VT.getSizeInBits(), Quantity.getKnownMinValue()));
2176
2177 return DAG.getConstant(Quantity.getKnownMinValue(), DL, VT);
2178}
2179
2181 ElementCount EC) {
2182 return getFixedOrScalableQuantity(*this, DL, VT, EC);
2183}
2184
2186 return getFixedOrScalableQuantity(*this, DL, VT, TS);
2187}
2188
2190 ElementCount EC) {
2191 EVT IdxVT = TLI->getVectorIdxTy(getDataLayout());
2192 EVT MaskVT = TLI->getSetCCResultType(getDataLayout(), *getContext(), DataVT);
2193 return getNode(ISD::GET_ACTIVE_LANE_MASK, DL, MaskVT,
2194 getConstant(0, DL, IdxVT), getElementCount(DL, IdxVT, EC));
2195}
2196
2198 APInt One(ResVT.getScalarSizeInBits(), 1);
2199 return getStepVector(DL, ResVT, One);
2200}
2201
2203 const APInt &StepVal) {
2204 assert(ResVT.getScalarSizeInBits() == StepVal.getBitWidth());
2205 if (ResVT.isScalableVector())
2206 return getNode(
2207 ISD::STEP_VECTOR, DL, ResVT,
2208 getTargetConstant(StepVal, DL, ResVT.getVectorElementType()));
2209
2210 SmallVector<SDValue, 16> OpsStepConstants;
2211 for (uint64_t i = 0; i < ResVT.getVectorNumElements(); i++)
2212 OpsStepConstants.push_back(
2213 getConstant(StepVal * i, DL, ResVT.getVectorElementType()));
2214 return getBuildVector(ResVT, DL, OpsStepConstants);
2215}
2216
2217/// Swaps the values of N1 and N2. Swaps all indices in the shuffle mask M that
2218/// point at N1 to point at N2 and indices that point at N2 to point at N1.
2223
2225 SDValue N2, ArrayRef<int> Mask) {
2226 assert(VT.getVectorNumElements() == Mask.size() &&
2227 "Must have the same number of vector elements as mask elements!");
2228 assert(VT == N1.getValueType() && VT == N2.getValueType() &&
2229 "Invalid VECTOR_SHUFFLE");
2230
2231 // Canonicalize shuffle undef, undef -> undef
2232 if (N1.isUndef() && N2.isUndef())
2233 return getUNDEF(VT);
2234
2235 // Validate that all indices in Mask are within the range of the elements
2236 // input to the shuffle.
2237 int NElts = Mask.size();
2238 assert(llvm::all_of(Mask,
2239 [&](int M) { return M < (NElts * 2) && M >= -1; }) &&
2240 "Index out of range");
2241
2242 // Copy the mask so we can do any needed cleanup.
2243 SmallVector<int, 8> MaskVec(Mask);
2244
2245 // Canonicalize shuffle v, v -> v, undef
2246 if (N1 == N2) {
2247 N2 = getUNDEF(VT);
2248 for (int i = 0; i != NElts; ++i)
2249 if (MaskVec[i] >= NElts) MaskVec[i] -= NElts;
2250 }
2251
2252 // Canonicalize shuffle undef, v -> v, undef. Commute the shuffle mask.
2253 if (N1.isUndef())
2254 commuteShuffle(N1, N2, MaskVec);
2255
2256 if (TLI->hasVectorBlend()) {
2257 // If shuffling a splat, try to blend the splat instead. We do this here so
2258 // that even when this arises during lowering we don't have to re-handle it.
2259 auto BlendSplat = [&](BuildVectorSDNode *BV, int Offset) {
2260 BitVector UndefElements;
2261 SDValue Splat = BV->getSplatValue(&UndefElements);
2262 if (!Splat)
2263 return;
2264
2265 for (int i = 0; i < NElts; ++i) {
2266 if (MaskVec[i] < Offset || MaskVec[i] >= (Offset + NElts))
2267 continue;
2268
2269 // If this input comes from undef, mark it as such.
2270 if (UndefElements[MaskVec[i] - Offset]) {
2271 MaskVec[i] = -1;
2272 continue;
2273 }
2274
2275 // If we can blend a non-undef lane, use that instead.
2276 if (!UndefElements[i])
2277 MaskVec[i] = i + Offset;
2278 }
2279 };
2280 if (auto *N1BV = dyn_cast<BuildVectorSDNode>(N1))
2281 BlendSplat(N1BV, 0);
2282 if (auto *N2BV = dyn_cast<BuildVectorSDNode>(N2))
2283 BlendSplat(N2BV, NElts);
2284 }
2285
2286 // Canonicalize all index into lhs, -> shuffle lhs, undef
2287 // Canonicalize all index into rhs, -> shuffle rhs, undef
2288 bool AllLHS = true, AllRHS = true;
2289 bool N2Undef = N2.isUndef();
2290 for (int i = 0; i != NElts; ++i) {
2291 if (MaskVec[i] >= NElts) {
2292 if (N2Undef)
2293 MaskVec[i] = -1;
2294 else
2295 AllLHS = false;
2296 } else if (MaskVec[i] >= 0) {
2297 AllRHS = false;
2298 }
2299 }
2300 if (AllLHS && AllRHS)
2301 return getUNDEF(VT);
2302 if (AllLHS && !N2Undef)
2303 N2 = getUNDEF(VT);
2304 if (AllRHS) {
2305 N1 = getUNDEF(VT);
2306 commuteShuffle(N1, N2, MaskVec);
2307 }
2308 // Reset our undef status after accounting for the mask.
2309 N2Undef = N2.isUndef();
2310 // Re-check whether both sides ended up undef.
2311 if (N1.isUndef() && N2Undef)
2312 return getUNDEF(VT);
2313
2314 // If Identity shuffle return that node.
2315 bool Identity = true, AllSame = true;
2316 for (int i = 0; i != NElts; ++i) {
2317 if (MaskVec[i] >= 0 && MaskVec[i] != i) Identity = false;
2318 if (MaskVec[i] != MaskVec[0]) AllSame = false;
2319 }
2320 if (Identity && NElts)
2321 return N1;
2322
2323 // Shuffling a constant splat doesn't change the result.
2324 if (N2Undef) {
2325 SDValue V = N1;
2326
2327 // Look through any bitcasts. We check that these don't change the number
2328 // (and size) of elements and just changes their types.
2329 while (V.getOpcode() == ISD::BITCAST)
2330 V = V->getOperand(0);
2331
2332 // A splat should always show up as a build vector node.
2333 if (auto *BV = dyn_cast<BuildVectorSDNode>(V)) {
2334 BitVector UndefElements;
2335 SDValue Splat = BV->getSplatValue(&UndefElements);
2336 // If this is a splat of an undef, shuffling it is also undef.
2337 if (Splat && Splat.isUndef())
2338 return getUNDEF(VT);
2339
2340 bool SameNumElts =
2341 V.getValueType().getVectorNumElements() == VT.getVectorNumElements();
2342
2343 // We only have a splat which can skip shuffles if there is a splatted
2344 // value and no undef lanes rearranged by the shuffle.
2345 if (Splat && UndefElements.none()) {
2346 // Splat of <x, x, ..., x>, return <x, x, ..., x>, provided that the
2347 // number of elements match or the value splatted is a zero constant.
2348 if (SameNumElts || isNullConstant(Splat))
2349 return N1;
2350 }
2351
2352 // If the shuffle itself creates a splat, build the vector directly.
2353 if (AllSame && SameNumElts) {
2354 EVT BuildVT = BV->getValueType(0);
2355 const SDValue &Splatted = BV->getOperand(MaskVec[0]);
2356 SDValue NewBV = getSplatBuildVector(BuildVT, dl, Splatted);
2357
2358 // We may have jumped through bitcasts, so the type of the
2359 // BUILD_VECTOR may not match the type of the shuffle.
2360 if (BuildVT != VT)
2361 NewBV = getNode(ISD::BITCAST, dl, VT, NewBV);
2362 return NewBV;
2363 }
2364 }
2365 }
2366
2367 SDVTList VTs = getVTList(VT);
2369 SDValue Ops[2] = { N1, N2 };
2371 for (int i = 0; i != NElts; ++i)
2372 ID.AddInteger(MaskVec[i]);
2373
2374 void* IP = nullptr;
2375 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
2376 return SDValue(E, 0);
2377
2378 // Allocate the mask array for the node out of the BumpPtrAllocator, since
2379 // SDNode doesn't have access to it. This memory will be "leaked" when
2380 // the node is deallocated, but recovered when the NodeAllocator is released.
2381 int *MaskAlloc = OperandAllocator.Allocate<int>(NElts);
2382 llvm::copy(MaskVec, MaskAlloc);
2383
2384 auto *N = newSDNode<ShuffleVectorSDNode>(VTs, dl.getIROrder(),
2385 dl.getDebugLoc(), MaskAlloc);
2386 createOperands(N, Ops);
2387
2388 CSEMap.InsertNode(N, IP);
2389 InsertNode(N);
2390 SDValue V = SDValue(N, 0);
2391 NewSDValueDbgMsg(V, "Creating new node: ", this);
2392 return V;
2393}
2394
2396 EVT VT = SV.getValueType(0);
2397 SmallVector<int, 8> MaskVec(SV.getMask());
2399
2400 SDValue Op0 = SV.getOperand(0);
2401 SDValue Op1 = SV.getOperand(1);
2402 return getVectorShuffle(VT, SDLoc(&SV), Op1, Op0, MaskVec);
2403}
2404
2406 SDVTList VTs = getVTList(VT);
2408 AddNodeIDNode(ID, ISD::Register, VTs, {});
2409 ID.AddInteger(Reg.id());
2410 void *IP = nullptr;
2411 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2412 return SDValue(E, 0);
2413
2414 auto *N = newSDNode<RegisterSDNode>(Reg, VTs);
2415 N->SDNodeBits.IsDivergent = TLI->isSDNodeSourceOfDivergence(N, FLI, UA);
2416 CSEMap.InsertNode(N, IP);
2417 InsertNode(N);
2418 return SDValue(N, 0);
2419}
2420
2423 AddNodeIDNode(ID, ISD::RegisterMask, getVTList(MVT::Untyped), {});
2424 ID.AddPointer(RegMask);
2425 void *IP = nullptr;
2426 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2427 return SDValue(E, 0);
2428
2429 auto *N = newSDNode<RegisterMaskSDNode>(RegMask);
2430 CSEMap.InsertNode(N, IP);
2431 InsertNode(N);
2432 return SDValue(N, 0);
2433}
2434
2436 MCSymbol *Label) {
2437 return getLabelNode(ISD::EH_LABEL, dl, Root, Label);
2438}
2439
2440SDValue SelectionDAG::getLabelNode(unsigned Opcode, const SDLoc &dl,
2441 SDValue Root, MCSymbol *Label) {
2443 SDValue Ops[] = { Root };
2444 AddNodeIDNode(ID, Opcode, getVTList(MVT::Other), Ops);
2445 ID.AddPointer(Label);
2446 void *IP = nullptr;
2447 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2448 return SDValue(E, 0);
2449
2450 auto *N =
2451 newSDNode<LabelSDNode>(Opcode, dl.getIROrder(), dl.getDebugLoc(), Label);
2452 createOperands(N, Ops);
2453
2454 CSEMap.InsertNode(N, IP);
2455 InsertNode(N);
2456 return SDValue(N, 0);
2457}
2458
2460 int64_t Offset, bool isTarget,
2461 unsigned TargetFlags) {
2462 unsigned Opc = isTarget ? ISD::TargetBlockAddress : ISD::BlockAddress;
2463 SDVTList VTs = getVTList(VT);
2464
2466 AddNodeIDNode(ID, Opc, VTs, {});
2467 ID.AddPointer(BA);
2468 ID.AddInteger(Offset);
2469 ID.AddInteger(TargetFlags);
2470 void *IP = nullptr;
2471 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2472 return SDValue(E, 0);
2473
2474 auto *N = newSDNode<BlockAddressSDNode>(Opc, VTs, BA, Offset, TargetFlags);
2475 CSEMap.InsertNode(N, IP);
2476 InsertNode(N);
2477 return SDValue(N, 0);
2478}
2479
2482 AddNodeIDNode(ID, ISD::SRCVALUE, getVTList(MVT::Other), {});
2483 ID.AddPointer(V);
2484
2485 void *IP = nullptr;
2486 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2487 return SDValue(E, 0);
2488
2489 auto *N = newSDNode<SrcValueSDNode>(V);
2490 CSEMap.InsertNode(N, IP);
2491 InsertNode(N);
2492 return SDValue(N, 0);
2493}
2494
2497 AddNodeIDNode(ID, ISD::MDNODE_SDNODE, getVTList(MVT::Other), {});
2498 ID.AddPointer(MD);
2499
2500 void *IP = nullptr;
2501 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2502 return SDValue(E, 0);
2503
2504 auto *N = newSDNode<MDNodeSDNode>(MD);
2505 CSEMap.InsertNode(N, IP);
2506 InsertNode(N);
2507 return SDValue(N, 0);
2508}
2509
2511 if (VT == V.getValueType())
2512 return V;
2513
2514 return getNode(ISD::BITCAST, SDLoc(V), VT, V);
2515}
2516
2518 unsigned SrcAS, unsigned DestAS) {
2519 SDVTList VTs = getVTList(VT);
2520 SDValue Ops[] = {Ptr};
2523 ID.AddInteger(SrcAS);
2524 ID.AddInteger(DestAS);
2525
2526 void *IP = nullptr;
2527 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
2528 return SDValue(E, 0);
2529
2530 auto *N = newSDNode<AddrSpaceCastSDNode>(dl.getIROrder(), dl.getDebugLoc(),
2531 VTs, SrcAS, DestAS);
2532 createOperands(N, Ops);
2533
2534 CSEMap.InsertNode(N, IP);
2535 InsertNode(N);
2536 return SDValue(N, 0);
2537}
2538
2540 return getNode(ISD::FREEZE, SDLoc(V), V.getValueType(), V);
2541}
2542
2544 bool PoisonOnly) {
2545 if (isGuaranteedNotToBeUndefOrPoison(V, DemandedElts, PoisonOnly))
2546 return V;
2547 return getFreeze(V);
2548}
2549
2550/// getShiftAmountOperand - Return the specified value casted to
2551/// the target's desired shift amount type.
2553 EVT OpTy = Op.getValueType();
2554 EVT ShTy = TLI->getShiftAmountTy(LHSTy, getDataLayout());
2555 if (OpTy == ShTy || OpTy.isVector()) return Op;
2556
2557 return getZExtOrTrunc(Op, SDLoc(Op), ShTy);
2558}
2559
2561 SDLoc dl(Node);
2563 const Value *V = cast<SrcValueSDNode>(Node->getOperand(2))->getValue();
2564 EVT VT = Node->getValueType(0);
2565 SDValue Tmp1 = Node->getOperand(0);
2566 SDValue Tmp2 = Node->getOperand(1);
2567 const MaybeAlign MA(Node->getConstantOperandVal(3));
2568
2569 SDValue VAListLoad = getLoad(TLI.getPointerTy(getDataLayout()), dl, Tmp1,
2570 Tmp2, MachinePointerInfo(V));
2571 SDValue VAList = VAListLoad;
2572
2573 if (MA && *MA > TLI.getMinStackArgumentAlignment()) {
2574 VAList = getNode(ISD::ADD, dl, VAList.getValueType(), VAList,
2575 getConstant(MA->value() - 1, dl, VAList.getValueType()));
2576
2577 VAList = getNode(
2578 ISD::AND, dl, VAList.getValueType(), VAList,
2579 getSignedConstant(-(int64_t)MA->value(), dl, VAList.getValueType()));
2580 }
2581
2582 // Increment the pointer, VAList, to the next vaarg
2583 Tmp1 = getNode(ISD::ADD, dl, VAList.getValueType(), VAList,
2584 getConstant(getDataLayout().getTypeAllocSize(
2585 VT.getTypeForEVT(*getContext())),
2586 dl, VAList.getValueType()));
2587 // Store the incremented VAList to the legalized pointer
2588 Tmp1 =
2589 getStore(VAListLoad.getValue(1), dl, Tmp1, Tmp2, MachinePointerInfo(V));
2590 // Load the actual argument out of the pointer VAList
2591 return getLoad(VT, dl, Tmp1, VAList, MachinePointerInfo());
2592}
2593
2595 SDLoc dl(Node);
2597 // This defaults to loading a pointer from the input and storing it to the
2598 // output, returning the chain.
2599 const Value *VD = cast<SrcValueSDNode>(Node->getOperand(3))->getValue();
2600 const Value *VS = cast<SrcValueSDNode>(Node->getOperand(4))->getValue();
2601 SDValue Tmp1 =
2602 getLoad(TLI.getPointerTy(getDataLayout()), dl, Node->getOperand(0),
2603 Node->getOperand(2), MachinePointerInfo(VS));
2604 return getStore(Tmp1.getValue(1), dl, Tmp1, Node->getOperand(1),
2605 MachinePointerInfo(VD));
2606}
2607
2609 const DataLayout &DL = getDataLayout();
2610 Type *Ty = VT.getTypeForEVT(*getContext());
2611 Align RedAlign = UseABI ? DL.getABITypeAlign(Ty) : DL.getPrefTypeAlign(Ty);
2612
2613 if (TLI->isTypeLegal(VT) || !VT.isVector())
2614 return RedAlign;
2615
2616 const TargetFrameLowering *TFI = MF->getSubtarget().getFrameLowering();
2617 const Align StackAlign = TFI->getStackAlign();
2618
2619 // See if we can choose a smaller ABI alignment in cases where it's an
2620 // illegal vector type that will get broken down.
2621 if (RedAlign > StackAlign) {
2622 EVT IntermediateVT;
2623 MVT RegisterVT;
2624 unsigned NumIntermediates;
2625 TLI->getVectorTypeBreakdown(*getContext(), VT, IntermediateVT,
2626 NumIntermediates, RegisterVT);
2627 Ty = IntermediateVT.getTypeForEVT(*getContext());
2628 Align RedAlign2 = UseABI ? DL.getABITypeAlign(Ty) : DL.getPrefTypeAlign(Ty);
2629 if (RedAlign2 < RedAlign)
2630 RedAlign = RedAlign2;
2631
2632 if (!getMachineFunction().getFrameInfo().isStackRealignable())
2633 // If the stack is not realignable, the alignment should be limited to the
2634 // StackAlignment
2635 RedAlign = std::min(RedAlign, StackAlign);
2636 }
2637
2638 return RedAlign;
2639}
2640
2642 MachineFrameInfo &MFI = MF->getFrameInfo();
2643 const TargetFrameLowering *TFI = MF->getSubtarget().getFrameLowering();
2644 int StackID = 0;
2645 if (Bytes.isScalable())
2646 StackID = TFI->getStackIDForScalableVectors();
2647 // The stack id gives an indication of whether the object is scalable or
2648 // not, so it's safe to pass in the minimum size here.
2649 int FrameIdx = MFI.CreateStackObject(Bytes.getKnownMinValue(), Alignment,
2650 false, nullptr, StackID);
2651 return getFrameIndex(FrameIdx, TLI->getFrameIndexTy(getDataLayout()));
2652}
2653
2655 Type *Ty = VT.getTypeForEVT(*getContext());
2656 Align StackAlign =
2657 std::max(getDataLayout().getPrefTypeAlign(Ty), Align(minAlign));
2658 return CreateStackTemporary(VT.getStoreSize(), StackAlign);
2659}
2660
2662 TypeSize VT1Size = VT1.getStoreSize();
2663 TypeSize VT2Size = VT2.getStoreSize();
2664 assert(VT1Size.isScalable() == VT2Size.isScalable() &&
2665 "Don't know how to choose the maximum size when creating a stack "
2666 "temporary");
2667 TypeSize Bytes = VT1Size.getKnownMinValue() > VT2Size.getKnownMinValue()
2668 ? VT1Size
2669 : VT2Size;
2670
2671 Type *Ty1 = VT1.getTypeForEVT(*getContext());
2672 Type *Ty2 = VT2.getTypeForEVT(*getContext());
2673 const DataLayout &DL = getDataLayout();
2674 Align Align = std::max(DL.getPrefTypeAlign(Ty1), DL.getPrefTypeAlign(Ty2));
2675 return CreateStackTemporary(Bytes, Align);
2676}
2677
2679 ISD::CondCode Cond, const SDLoc &dl,
2680 SDNodeFlags Flags) {
2681 EVT OpVT = N1.getValueType();
2682
2683 auto GetUndefBooleanConstant = [&]() {
2684 if (VT.getScalarType() == MVT::i1 ||
2685 TLI->getBooleanContents(OpVT) ==
2687 return getUNDEF(VT);
2688 // ZeroOrOne / ZeroOrNegative require specific values for the high bits,
2689 // so we cannot use getUNDEF(). Return zero instead.
2690 return getConstant(0, dl, VT);
2691 };
2692
2693 // These setcc operations always fold.
2694 switch (Cond) {
2695 default: break;
2696 case ISD::SETFALSE:
2697 case ISD::SETFALSE2: return getBoolConstant(false, dl, VT, OpVT);
2698 case ISD::SETTRUE:
2699 case ISD::SETTRUE2: return getBoolConstant(true, dl, VT, OpVT);
2700
2701 case ISD::SETOEQ:
2702 case ISD::SETOGT:
2703 case ISD::SETOGE:
2704 case ISD::SETOLT:
2705 case ISD::SETOLE:
2706 case ISD::SETONE:
2707 case ISD::SETO:
2708 case ISD::SETUO:
2709 case ISD::SETUEQ:
2710 case ISD::SETUNE:
2711 assert(!OpVT.isInteger() && "Illegal setcc for integer!");
2712 break;
2713 }
2714
2715 if (OpVT.isInteger()) {
2716 // For EQ and NE, we can always pick a value for the undef to make the
2717 // predicate pass or fail, so we can return undef.
2718 // Matches behavior in llvm::ConstantFoldCompareInstruction.
2719 // icmp eq/ne X, undef -> undef.
2720 if ((N1.isUndef() || N2.isUndef()) &&
2721 (Cond == ISD::SETEQ || Cond == ISD::SETNE))
2722 return GetUndefBooleanConstant();
2723
2724 // If both operands are undef, we can return undef for int comparison.
2725 // icmp undef, undef -> undef.
2726 if (N1.isUndef() && N2.isUndef())
2727 return GetUndefBooleanConstant();
2728
2729 // icmp X, X -> true/false
2730 // icmp X, undef -> true/false because undef could be X.
2731 if (N1.isUndef() || N2.isUndef() || N1 == N2)
2732 return getBoolConstant(ISD::isTrueWhenEqual(Cond), dl, VT, OpVT);
2733 }
2734
2736 const APInt &C2 = N2C->getAPIntValue();
2738 const APInt &C1 = N1C->getAPIntValue();
2739
2741 dl, VT, OpVT);
2742 }
2743 }
2744
2745 auto *N1CFP = dyn_cast<ConstantFPSDNode>(N1);
2746 auto *N2CFP = dyn_cast<ConstantFPSDNode>(N2);
2747
2748 if (N1CFP && N2CFP) {
2749 APFloat::cmpResult R = N1CFP->getValueAPF().compare(N2CFP->getValueAPF());
2750 switch (Cond) {
2751 default: break;
2752 case ISD::SETEQ: if (R==APFloat::cmpUnordered)
2753 return GetUndefBooleanConstant();
2754 [[fallthrough]];
2755 case ISD::SETOEQ: return getBoolConstant(R==APFloat::cmpEqual, dl, VT,
2756 OpVT);
2757 case ISD::SETNE: if (R==APFloat::cmpUnordered)
2758 return GetUndefBooleanConstant();
2759 [[fallthrough]];
2761 R==APFloat::cmpLessThan, dl, VT,
2762 OpVT);
2763 case ISD::SETLT: if (R==APFloat::cmpUnordered)
2764 return GetUndefBooleanConstant();
2765 [[fallthrough]];
2766 case ISD::SETOLT: return getBoolConstant(R==APFloat::cmpLessThan, dl, VT,
2767 OpVT);
2768 case ISD::SETGT: if (R==APFloat::cmpUnordered)
2769 return GetUndefBooleanConstant();
2770 [[fallthrough]];
2772 VT, OpVT);
2773 case ISD::SETLE: if (R==APFloat::cmpUnordered)
2774 return GetUndefBooleanConstant();
2775 [[fallthrough]];
2777 R==APFloat::cmpEqual, dl, VT,
2778 OpVT);
2779 case ISD::SETGE: if (R==APFloat::cmpUnordered)
2780 return GetUndefBooleanConstant();
2781 [[fallthrough]];
2783 R==APFloat::cmpEqual, dl, VT, OpVT);
2784 case ISD::SETO: return getBoolConstant(R!=APFloat::cmpUnordered, dl, VT,
2785 OpVT);
2786 case ISD::SETUO: return getBoolConstant(R==APFloat::cmpUnordered, dl, VT,
2787 OpVT);
2789 R==APFloat::cmpEqual, dl, VT,
2790 OpVT);
2791 case ISD::SETUNE: return getBoolConstant(R!=APFloat::cmpEqual, dl, VT,
2792 OpVT);
2794 R==APFloat::cmpLessThan, dl, VT,
2795 OpVT);
2797 R==APFloat::cmpUnordered, dl, VT,
2798 OpVT);
2800 VT, OpVT);
2801 case ISD::SETUGE: return getBoolConstant(R!=APFloat::cmpLessThan, dl, VT,
2802 OpVT);
2803 }
2804 } else if (N1CFP && OpVT.isSimple() && !N2.isUndef()) {
2805 // Ensure that the constant occurs on the RHS.
2807 if (!TLI->isCondCodeLegal(SwappedCond, OpVT.getSimpleVT()))
2808 return SDValue();
2809 return getSetCC(dl, VT, N2, N1, SwappedCond, /*Chian=*/{},
2810 /*IsSignaling=*/false, Flags);
2811 } else if ((N2CFP && N2CFP->getValueAPF().isNaN()) ||
2812 (OpVT.isFloatingPoint() && (N1.isUndef() || N2.isUndef()))) {
2813 // If an operand is known to be a nan (or undef that could be a nan), we can
2814 // fold it.
2815 // Choosing NaN for the undef will always make unordered comparison succeed
2816 // and ordered comparison fails.
2817 // Matches behavior in llvm::ConstantFoldCompareInstruction.
2818 switch (ISD::getUnorderedFlavor(Cond)) {
2819 default:
2820 llvm_unreachable("Unknown flavor!");
2821 case 0: // Known false.
2822 return getBoolConstant(false, dl, VT, OpVT);
2823 case 1: // Known true.
2824 return getBoolConstant(true, dl, VT, OpVT);
2825 case 2: // Undefined.
2826 return GetUndefBooleanConstant();
2827 }
2828 }
2829
2830 // Could not fold it.
2831 return SDValue();
2832}
2833
2834/// SignBitIsZero - Return true if the sign bit of Op is known to be zero. We
2835/// use this predicate to simplify operations downstream.
2837 unsigned BitWidth = Op.getScalarValueSizeInBits();
2839}
2840
2841// TODO: Should have argument to specify if sign bit of nan is ignorable.
2843 if (Depth >= MaxRecursionDepth)
2844 return false; // Limit search depth.
2845
2846 unsigned Opc = Op.getOpcode();
2847 switch (Opc) {
2848 case ISD::FABS:
2849 return true;
2850 case ISD::AssertNoFPClass: {
2851 FPClassTest NoFPClass =
2852 static_cast<FPClassTest>(Op.getConstantOperandVal(1));
2853
2854 const FPClassTest TestMask = fcNan | fcNegative;
2855 return (NoFPClass & TestMask) == TestMask;
2856 }
2857 case ISD::ARITH_FENCE:
2858 return SignBitIsZeroFP(Op.getOperand(0), Depth + 1);
2859 case ISD::FEXP:
2860 case ISD::FEXP2:
2861 case ISD::FEXP10:
2862 return Op->getFlags().hasNoNaNs();
2863 case ISD::FMINNUM:
2864 case ISD::FMINNUM_IEEE:
2865 case ISD::FMINIMUM:
2866 case ISD::FMINIMUMNUM:
2867 return SignBitIsZeroFP(Op.getOperand(1), Depth + 1) &&
2868 SignBitIsZeroFP(Op.getOperand(0), Depth + 1);
2869 case ISD::FMAXNUM:
2870 case ISD::FMAXNUM_IEEE:
2871 case ISD::FMAXIMUM:
2872 case ISD::FMAXIMUMNUM:
2873 // TODO: If we can ignore the sign bit of nans, only one side being known 0
2874 // is sufficient.
2875 return SignBitIsZeroFP(Op.getOperand(1), Depth + 1) &&
2876 SignBitIsZeroFP(Op.getOperand(0), Depth + 1);
2877 default:
2878 return false;
2879 }
2880
2881 llvm_unreachable("covered opcode switch");
2882}
2883
2884/// MaskedValueIsZero - Return true if 'V & Mask' is known to be zero. We use
2885/// this predicate to simplify operations downstream. Mask is known to be zero
2886/// for bits that V cannot have.
2888 unsigned Depth) const {
2889 return Mask.isSubsetOf(computeKnownBits(V, Depth).Zero);
2890}
2891
2892/// MaskedValueIsZero - Return true if 'V & Mask' is known to be zero in
2893/// DemandedElts. We use this predicate to simplify operations downstream.
2894/// Mask is known to be zero for bits that V cannot have.
2896 const APInt &DemandedElts,
2897 unsigned Depth) const {
2898 return Mask.isSubsetOf(computeKnownBits(V, DemandedElts, Depth).Zero);
2899}
2900
2901/// MaskedVectorIsZero - Return true if 'Op' is known to be zero in
2902/// DemandedElts. We use this predicate to simplify operations downstream.
2904 unsigned Depth /* = 0 */) const {
2905 return computeKnownBits(V, DemandedElts, Depth).isZero();
2906}
2907
2908/// MaskedValueIsAllOnes - Return true if '(Op & Mask) == Mask'.
2910 unsigned Depth) const {
2911 return Mask.isSubsetOf(computeKnownBits(V, Depth).One);
2912}
2913
2915 const APInt &DemandedElts,
2916 unsigned Depth) const {
2917 EVT VT = Op.getValueType();
2918 assert(VT.isVector() && !VT.isScalableVector() && "Only for fixed vectors!");
2919
2920 unsigned NumElts = VT.getVectorNumElements();
2921 assert(DemandedElts.getBitWidth() == NumElts && "Unexpected demanded mask.");
2922
2923 APInt KnownZeroElements = APInt::getZero(NumElts);
2924 for (unsigned EltIdx = 0; EltIdx != NumElts; ++EltIdx) {
2925 if (!DemandedElts[EltIdx])
2926 continue; // Don't query elements that are not demanded.
2927 APInt Mask = APInt::getOneBitSet(NumElts, EltIdx);
2928 if (MaskedVectorIsZero(Op, Mask, Depth))
2929 KnownZeroElements.setBit(EltIdx);
2930 }
2931 return KnownZeroElements;
2932}
2933
2934/// isSplatValue - Return true if the vector V has the same value
2935/// across all DemandedElts. For scalable vectors, we don't know the
2936/// number of lanes at compile time. Instead, we use a 1 bit APInt
2937/// to represent a conservative value for all lanes; that is, that
2938/// one bit value is implicitly splatted across all lanes.
2939bool SelectionDAG::isSplatValue(SDValue V, const APInt &DemandedElts,
2940 APInt &UndefElts, unsigned Depth) const {
2941 unsigned Opcode = V.getOpcode();
2942 EVT VT = V.getValueType();
2943 assert(VT.isVector() && "Vector type expected");
2944 assert((!VT.isScalableVector() || DemandedElts.getBitWidth() == 1) &&
2945 "scalable demanded bits are ignored");
2946
2947 if (!DemandedElts)
2948 return false; // No demanded elts, better to assume we don't know anything.
2949
2950 if (Depth >= MaxRecursionDepth)
2951 return false; // Limit search depth.
2952
2953 // Deal with some common cases here that work for both fixed and scalable
2954 // vector types.
2955 switch (Opcode) {
2956 case ISD::SPLAT_VECTOR:
2957 UndefElts = V.getOperand(0).isUndef()
2958 ? APInt::getAllOnes(DemandedElts.getBitWidth())
2959 : APInt(DemandedElts.getBitWidth(), 0);
2960 return true;
2961 case ISD::ADD:
2962 case ISD::SUB:
2963 case ISD::AND:
2964 case ISD::XOR:
2965 case ISD::OR: {
2966 APInt UndefLHS, UndefRHS;
2967 SDValue LHS = V.getOperand(0);
2968 SDValue RHS = V.getOperand(1);
2969 // Only recognize splats with the same demanded undef elements for both
2970 // operands, otherwise we might fail to handle binop-specific undef
2971 // handling.
2972 // e.g. (and undef, 0) -> 0 etc.
2973 if (isSplatValue(LHS, DemandedElts, UndefLHS, Depth + 1) &&
2974 isSplatValue(RHS, DemandedElts, UndefRHS, Depth + 1) &&
2975 (DemandedElts & UndefLHS) == (DemandedElts & UndefRHS)) {
2976 UndefElts = UndefLHS | UndefRHS;
2977 return true;
2978 }
2979 return false;
2980 }
2981 case ISD::ABS:
2982 case ISD::TRUNCATE:
2983 case ISD::SIGN_EXTEND:
2984 case ISD::ZERO_EXTEND:
2985 return isSplatValue(V.getOperand(0), DemandedElts, UndefElts, Depth + 1);
2986 default:
2987 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::INTRINSIC_WO_CHAIN ||
2988 Opcode == ISD::INTRINSIC_W_CHAIN || Opcode == ISD::INTRINSIC_VOID)
2989 return TLI->isSplatValueForTargetNode(V, DemandedElts, UndefElts, *this,
2990 Depth);
2991 break;
2992 }
2993
2994 // We don't support other cases than those above for scalable vectors at
2995 // the moment.
2996 if (VT.isScalableVector())
2997 return false;
2998
2999 unsigned NumElts = VT.getVectorNumElements();
3000 assert(NumElts == DemandedElts.getBitWidth() && "Vector size mismatch");
3001 UndefElts = APInt::getZero(NumElts);
3002
3003 switch (Opcode) {
3004 case ISD::BUILD_VECTOR: {
3005 SDValue Scl;
3006 for (unsigned i = 0; i != NumElts; ++i) {
3007 SDValue Op = V.getOperand(i);
3008 if (Op.isUndef()) {
3009 UndefElts.setBit(i);
3010 continue;
3011 }
3012 if (!DemandedElts[i])
3013 continue;
3014 if (Scl && Scl != Op)
3015 return false;
3016 Scl = Op;
3017 }
3018 return true;
3019 }
3020 case ISD::VECTOR_SHUFFLE: {
3021 // Check if this is a shuffle node doing a splat or a shuffle of a splat.
3022 APInt DemandedLHS = APInt::getZero(NumElts);
3023 APInt DemandedRHS = APInt::getZero(NumElts);
3024 ArrayRef<int> Mask = cast<ShuffleVectorSDNode>(V)->getMask();
3025 for (int i = 0; i != (int)NumElts; ++i) {
3026 int M = Mask[i];
3027 if (M < 0) {
3028 UndefElts.setBit(i);
3029 continue;
3030 }
3031 if (!DemandedElts[i])
3032 continue;
3033 if (M < (int)NumElts)
3034 DemandedLHS.setBit(M);
3035 else
3036 DemandedRHS.setBit(M - NumElts);
3037 }
3038
3039 // If we aren't demanding either op, assume there's no splat.
3040 // If we are demanding both ops, assume there's no splat.
3041 if ((DemandedLHS.isZero() && DemandedRHS.isZero()) ||
3042 (!DemandedLHS.isZero() && !DemandedRHS.isZero()))
3043 return false;
3044
3045 // See if the demanded elts of the source op is a splat or we only demand
3046 // one element, which should always be a splat.
3047 // TODO: Handle source ops splats with undefs.
3048 auto CheckSplatSrc = [&](SDValue Src, const APInt &SrcElts) {
3049 APInt SrcUndefs;
3050 return (SrcElts.popcount() == 1) ||
3051 (isSplatValue(Src, SrcElts, SrcUndefs, Depth + 1) &&
3052 (SrcElts & SrcUndefs).isZero());
3053 };
3054 if (!DemandedLHS.isZero())
3055 return CheckSplatSrc(V.getOperand(0), DemandedLHS);
3056 return CheckSplatSrc(V.getOperand(1), DemandedRHS);
3057 }
3059 // Offset the demanded elts by the subvector index.
3060 SDValue Src = V.getOperand(0);
3061 // We don't support scalable vectors at the moment.
3062 if (Src.getValueType().isScalableVector())
3063 return false;
3064 uint64_t Idx = V.getConstantOperandVal(1);
3065 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
3066 APInt UndefSrcElts;
3067 APInt DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
3068 if (isSplatValue(Src, DemandedSrcElts, UndefSrcElts, Depth + 1)) {
3069 UndefElts = UndefSrcElts.extractBits(NumElts, Idx);
3070 return true;
3071 }
3072 break;
3073 }
3077 // Widen the demanded elts by the src element count.
3078 SDValue Src = V.getOperand(0);
3079 // We don't support scalable vectors at the moment.
3080 if (Src.getValueType().isScalableVector())
3081 return false;
3082 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
3083 APInt UndefSrcElts;
3084 APInt DemandedSrcElts = DemandedElts.zext(NumSrcElts);
3085 if (isSplatValue(Src, DemandedSrcElts, UndefSrcElts, Depth + 1)) {
3086 UndefElts = UndefSrcElts.trunc(NumElts);
3087 return true;
3088 }
3089 break;
3090 }
3091 case ISD::BITCAST: {
3092 SDValue Src = V.getOperand(0);
3093 EVT SrcVT = Src.getValueType();
3094 unsigned SrcBitWidth = SrcVT.getScalarSizeInBits();
3095 unsigned BitWidth = VT.getScalarSizeInBits();
3096
3097 // Ignore bitcasts from unsupported types.
3098 // TODO: Add fp support?
3099 if (!SrcVT.isVector() || !SrcVT.isInteger() || !VT.isInteger())
3100 break;
3101
3102 // Bitcast 'small element' vector to 'large element' vector.
3103 if ((BitWidth % SrcBitWidth) == 0) {
3104 // See if each sub element is a splat.
3105 unsigned Scale = BitWidth / SrcBitWidth;
3106 unsigned NumSrcElts = SrcVT.getVectorNumElements();
3107 APInt ScaledDemandedElts =
3108 APIntOps::ScaleBitMask(DemandedElts, NumSrcElts);
3109 for (unsigned I = 0; I != Scale; ++I) {
3110 APInt SubUndefElts;
3111 APInt SubDemandedElt = APInt::getOneBitSet(Scale, I);
3112 APInt SubDemandedElts = APInt::getSplat(NumSrcElts, SubDemandedElt);
3113 SubDemandedElts &= ScaledDemandedElts;
3114 if (!isSplatValue(Src, SubDemandedElts, SubUndefElts, Depth + 1))
3115 return false;
3116 // TODO: Add support for merging sub undef elements.
3117 if (!SubUndefElts.isZero())
3118 return false;
3119 }
3120 return true;
3121 }
3122 break;
3123 }
3124 }
3125
3126 return false;
3127}
3128
3129/// Helper wrapper to main isSplatValue function.
3130bool SelectionDAG::isSplatValue(SDValue V, bool AllowUndefs) const {
3131 EVT VT = V.getValueType();
3132 assert(VT.isVector() && "Vector type expected");
3133
3134 APInt UndefElts;
3135 // Since the number of lanes in a scalable vector is unknown at compile time,
3136 // we track one bit which is implicitly broadcast to all lanes. This means
3137 // that all lanes in a scalable vector are considered demanded.
3138 APInt DemandedElts
3140 return isSplatValue(V, DemandedElts, UndefElts) &&
3141 (AllowUndefs || !UndefElts);
3142}
3143
3146
3147 EVT VT = V.getValueType();
3148 unsigned Opcode = V.getOpcode();
3149 switch (Opcode) {
3150 default: {
3151 APInt UndefElts;
3152 // Since the number of lanes in a scalable vector is unknown at compile time,
3153 // we track one bit which is implicitly broadcast to all lanes. This means
3154 // that all lanes in a scalable vector are considered demanded.
3155 APInt DemandedElts
3157
3158 if (isSplatValue(V, DemandedElts, UndefElts)) {
3159 if (VT.isScalableVector()) {
3160 // DemandedElts and UndefElts are ignored for scalable vectors, since
3161 // the only supported cases are SPLAT_VECTOR nodes.
3162 SplatIdx = 0;
3163 } else {
3164 // Handle case where all demanded elements are UNDEF.
3165 if (DemandedElts.isSubsetOf(UndefElts)) {
3166 SplatIdx = 0;
3167 return getUNDEF(VT);
3168 }
3169 SplatIdx = (UndefElts & DemandedElts).countr_one();
3170 }
3171 return V;
3172 }
3173 break;
3174 }
3175 case ISD::SPLAT_VECTOR:
3176 SplatIdx = 0;
3177 return V;
3178 case ISD::VECTOR_SHUFFLE: {
3179 assert(!VT.isScalableVector());
3180 // Check if this is a shuffle node doing a splat.
3181 // TODO - remove this and rely purely on SelectionDAG::isSplatValue,
3182 // getTargetVShiftNode currently struggles without the splat source.
3183 auto *SVN = cast<ShuffleVectorSDNode>(V);
3184 if (!SVN->isSplat())
3185 break;
3186 int Idx = SVN->getSplatIndex();
3187 int NumElts = V.getValueType().getVectorNumElements();
3188 SplatIdx = Idx % NumElts;
3189 return V.getOperand(Idx / NumElts);
3190 }
3191 }
3192
3193 return SDValue();
3194}
3195
3197 int SplatIdx;
3198 if (SDValue SrcVector = getSplatSourceVector(V, SplatIdx)) {
3199 EVT SVT = SrcVector.getValueType().getScalarType();
3200 EVT LegalSVT = SVT;
3201 if (LegalTypes && !TLI->isTypeLegal(SVT)) {
3202 if (!SVT.isInteger())
3203 return SDValue();
3204 LegalSVT = TLI->getTypeToTransformTo(*getContext(), LegalSVT);
3205 if (LegalSVT.bitsLT(SVT))
3206 return SDValue();
3207 }
3208 return getExtractVectorElt(SDLoc(V), LegalSVT, SrcVector, SplatIdx);
3209 }
3210 return SDValue();
3211}
3212
3213std::optional<ConstantRange>
3215 unsigned Depth) const {
3216 assert((V.getOpcode() == ISD::SHL || V.getOpcode() == ISD::SRL ||
3217 V.getOpcode() == ISD::SRA) &&
3218 "Unknown shift node");
3219 // Shifting more than the bitwidth is not valid.
3220 unsigned BitWidth = V.getScalarValueSizeInBits();
3221
3222 if (auto *Cst = dyn_cast<ConstantSDNode>(V.getOperand(1))) {
3223 const APInt &ShAmt = Cst->getAPIntValue();
3224 if (ShAmt.uge(BitWidth))
3225 return std::nullopt;
3226 return ConstantRange(ShAmt);
3227 }
3228
3229 if (auto *BV = dyn_cast<BuildVectorSDNode>(V.getOperand(1))) {
3230 const APInt *MinAmt = nullptr, *MaxAmt = nullptr;
3231 for (unsigned i = 0, e = BV->getNumOperands(); i != e; ++i) {
3232 if (!DemandedElts[i])
3233 continue;
3234 auto *SA = dyn_cast<ConstantSDNode>(BV->getOperand(i));
3235 if (!SA) {
3236 MinAmt = MaxAmt = nullptr;
3237 break;
3238 }
3239 const APInt &ShAmt = SA->getAPIntValue();
3240 if (ShAmt.uge(BitWidth))
3241 return std::nullopt;
3242 if (!MinAmt || MinAmt->ugt(ShAmt))
3243 MinAmt = &ShAmt;
3244 if (!MaxAmt || MaxAmt->ult(ShAmt))
3245 MaxAmt = &ShAmt;
3246 }
3247 assert(((!MinAmt && !MaxAmt) || (MinAmt && MaxAmt)) &&
3248 "Failed to find matching min/max shift amounts");
3249 if (MinAmt && MaxAmt)
3250 return ConstantRange(*MinAmt, *MaxAmt + 1);
3251 }
3252
3253 // Use computeKnownBits to find a hidden constant/knownbits (usually type
3254 // legalized). e.g. Hidden behind multiple bitcasts/build_vector/casts etc.
3255 KnownBits KnownAmt = computeKnownBits(V.getOperand(1), DemandedElts, Depth);
3256 if (KnownAmt.getMaxValue().ult(BitWidth))
3257 return ConstantRange::fromKnownBits(KnownAmt, /*IsSigned=*/false);
3258
3259 return std::nullopt;
3260}
3261
3262std::optional<unsigned>
3264 unsigned Depth) const {
3265 assert((V.getOpcode() == ISD::SHL || V.getOpcode() == ISD::SRL ||
3266 V.getOpcode() == ISD::SRA) &&
3267 "Unknown shift node");
3268 if (std::optional<ConstantRange> AmtRange =
3269 getValidShiftAmountRange(V, DemandedElts, Depth))
3270 if (const APInt *ShAmt = AmtRange->getSingleElement())
3271 return ShAmt->getZExtValue();
3272 return std::nullopt;
3273}
3274
3275std::optional<unsigned>
3277 EVT VT = V.getValueType();
3278 APInt DemandedElts = VT.isFixedLengthVector()
3280 : APInt(1, 1);
3281 return getValidShiftAmount(V, DemandedElts, Depth);
3282}
3283
3284std::optional<unsigned>
3286 unsigned Depth) const {
3287 assert((V.getOpcode() == ISD::SHL || V.getOpcode() == ISD::SRL ||
3288 V.getOpcode() == ISD::SRA) &&
3289 "Unknown shift node");
3290 if (std::optional<ConstantRange> AmtRange =
3291 getValidShiftAmountRange(V, DemandedElts, Depth))
3292 return AmtRange->getUnsignedMin().getZExtValue();
3293 return std::nullopt;
3294}
3295
3296std::optional<unsigned>
3298 EVT VT = V.getValueType();
3299 APInt DemandedElts = VT.isFixedLengthVector()
3301 : APInt(1, 1);
3302 return getValidMinimumShiftAmount(V, DemandedElts, Depth);
3303}
3304
3305std::optional<unsigned>
3307 unsigned Depth) const {
3308 assert((V.getOpcode() == ISD::SHL || V.getOpcode() == ISD::SRL ||
3309 V.getOpcode() == ISD::SRA) &&
3310 "Unknown shift node");
3311 if (std::optional<ConstantRange> AmtRange =
3312 getValidShiftAmountRange(V, DemandedElts, Depth))
3313 return AmtRange->getUnsignedMax().getZExtValue();
3314 return std::nullopt;
3315}
3316
3317std::optional<unsigned>
3319 EVT VT = V.getValueType();
3320 APInt DemandedElts = VT.isFixedLengthVector()
3322 : APInt(1, 1);
3323 return getValidMaximumShiftAmount(V, DemandedElts, Depth);
3324}
3325
3326/// Determine which bits of Op are known to be either zero or one and return
3327/// them in Known. For vectors, the known bits are those that are shared by
3328/// every vector element.
3330 EVT VT = Op.getValueType();
3331
3332 // Since the number of lanes in a scalable vector is unknown at compile time,
3333 // we track one bit which is implicitly broadcast to all lanes. This means
3334 // that all lanes in a scalable vector are considered demanded.
3335 APInt DemandedElts = VT.isFixedLengthVector()
3337 : APInt(1, 1);
3338 return computeKnownBits(Op, DemandedElts, Depth);
3339}
3340
3341/// Determine which bits of Op are known to be either zero or one and return
3342/// them in Known. The DemandedElts argument allows us to only collect the known
3343/// bits that are shared by the requested vector elements.
3345 unsigned Depth) const {
3346 unsigned BitWidth = Op.getScalarValueSizeInBits();
3347
3348 KnownBits Known(BitWidth); // Don't know anything.
3349
3350 if (auto OptAPInt = Op->bitcastToAPInt()) {
3351 // We know all of the bits for a constant!
3352 return KnownBits::makeConstant(*std::move(OptAPInt));
3353 }
3354
3355 if (Depth >= MaxRecursionDepth)
3356 return Known; // Limit search depth.
3357
3358 KnownBits Known2;
3359 unsigned NumElts = DemandedElts.getBitWidth();
3360 assert((!Op.getValueType().isScalableVector() || NumElts == 1) &&
3361 "DemandedElts for scalable vectors must be 1 to represent all lanes");
3362 assert((!Op.getValueType().isFixedLengthVector() ||
3363 NumElts == Op.getValueType().getVectorNumElements()) &&
3364 "Unexpected vector size");
3365
3366 if (!DemandedElts)
3367 return Known; // No demanded elts, better to assume we don't know anything.
3368
3369 unsigned Opcode = Op.getOpcode();
3370 switch (Opcode) {
3371 case ISD::MERGE_VALUES:
3372 return computeKnownBits(Op.getOperand(Op.getResNo()), DemandedElts,
3373 Depth + 1);
3374 case ISD::SPLAT_VECTOR: {
3375 SDValue SrcOp = Op.getOperand(0);
3376 assert(SrcOp.getValueSizeInBits() >= BitWidth &&
3377 "Expected SPLAT_VECTOR implicit truncation");
3378 // Implicitly truncate the bits to match the official semantics of
3379 // SPLAT_VECTOR.
3380 Known = computeKnownBits(SrcOp, Depth + 1).trunc(BitWidth);
3381 break;
3382 }
3384 unsigned ScalarSize = Op.getOperand(0).getScalarValueSizeInBits();
3385 assert(ScalarSize * Op.getNumOperands() == BitWidth &&
3386 "Expected SPLAT_VECTOR_PARTS scalars to cover element width");
3387 for (auto [I, SrcOp] : enumerate(Op->ops())) {
3388 Known.insertBits(computeKnownBits(SrcOp, Depth + 1), ScalarSize * I);
3389 }
3390 break;
3391 }
3392 case ISD::STEP_VECTOR: {
3393 const APInt &Step = Op.getConstantOperandAPInt(0);
3394
3395 if (Step.isPowerOf2())
3396 Known.Zero.setLowBits(Step.logBase2());
3397
3399
3400 if (!isUIntN(BitWidth, Op.getValueType().getVectorMinNumElements()))
3401 break;
3402 const APInt MinNumElts =
3403 APInt(BitWidth, Op.getValueType().getVectorMinNumElements());
3404
3405 bool Overflow;
3406 const APInt MaxNumElts = getVScaleRange(&F, BitWidth)
3408 .umul_ov(MinNumElts, Overflow);
3409 if (Overflow)
3410 break;
3411
3412 const APInt MaxValue = (MaxNumElts - 1).umul_ov(Step, Overflow);
3413 if (Overflow)
3414 break;
3415
3416 Known.Zero.setHighBits(MaxValue.countl_zero());
3417 break;
3418 }
3419 case ISD::BUILD_VECTOR:
3420 assert(!Op.getValueType().isScalableVector());
3421 // Collect the known bits that are shared by every demanded vector element.
3422 Known.setAllConflict();
3423 for (unsigned i = 0, e = Op.getNumOperands(); i != e; ++i) {
3424 if (!DemandedElts[i])
3425 continue;
3426
3427 SDValue SrcOp = Op.getOperand(i);
3428 Known2 = computeKnownBits(SrcOp, Depth + 1);
3429
3430 // BUILD_VECTOR can implicitly truncate sources, we must handle this.
3431 if (SrcOp.getValueSizeInBits() != BitWidth) {
3432 assert(SrcOp.getValueSizeInBits() > BitWidth &&
3433 "Expected BUILD_VECTOR implicit truncation");
3434 Known2 = Known2.trunc(BitWidth);
3435 }
3436
3437 // Known bits are the values that are shared by every demanded element.
3438 Known = Known.intersectWith(Known2);
3439
3440 // If we don't know any bits, early out.
3441 if (Known.isUnknown())
3442 break;
3443 }
3444 break;
3445 case ISD::VECTOR_COMPRESS: {
3446 SDValue Vec = Op.getOperand(0);
3447 SDValue PassThru = Op.getOperand(2);
3448 Known = computeKnownBits(PassThru, DemandedElts, Depth + 1);
3449 // If we don't know any bits, early out.
3450 if (Known.isUnknown())
3451 break;
3452 Known2 = computeKnownBits(Vec, Depth + 1);
3453 Known = Known.intersectWith(Known2);
3454 break;
3455 }
3456 case ISD::VECTOR_SHUFFLE: {
3457 assert(!Op.getValueType().isScalableVector());
3458 // Collect the known bits that are shared by every vector element referenced
3459 // by the shuffle.
3460 APInt DemandedLHS, DemandedRHS;
3462 assert(NumElts == SVN->getMask().size() && "Unexpected vector size");
3463 if (!getShuffleDemandedElts(NumElts, SVN->getMask(), DemandedElts,
3464 DemandedLHS, DemandedRHS))
3465 break;
3466
3467 // Known bits are the values that are shared by every demanded element.
3468 Known.setAllConflict();
3469 if (!!DemandedLHS) {
3470 SDValue LHS = Op.getOperand(0);
3471 Known2 = computeKnownBits(LHS, DemandedLHS, Depth + 1);
3472 Known = Known.intersectWith(Known2);
3473 }
3474 // If we don't know any bits, early out.
3475 if (Known.isUnknown())
3476 break;
3477 if (!!DemandedRHS) {
3478 SDValue RHS = Op.getOperand(1);
3479 Known2 = computeKnownBits(RHS, DemandedRHS, Depth + 1);
3480 Known = Known.intersectWith(Known2);
3481 }
3482 break;
3483 }
3484 case ISD::VSCALE: {
3486 const APInt &Multiplier = Op.getConstantOperandAPInt(0);
3487 Known = getVScaleRange(&F, BitWidth).multiply(Multiplier).toKnownBits();
3488 break;
3489 }
3490 case ISD::CONCAT_VECTORS: {
3491 if (Op.getValueType().isScalableVector())
3492 break;
3493 // Split DemandedElts and test each of the demanded subvectors.
3494 Known.setAllConflict();
3495 EVT SubVectorVT = Op.getOperand(0).getValueType();
3496 unsigned NumSubVectorElts = SubVectorVT.getVectorNumElements();
3497 unsigned NumSubVectors = Op.getNumOperands();
3498 for (unsigned i = 0; i != NumSubVectors; ++i) {
3499 APInt DemandedSub =
3500 DemandedElts.extractBits(NumSubVectorElts, i * NumSubVectorElts);
3501 if (!!DemandedSub) {
3502 SDValue Sub = Op.getOperand(i);
3503 Known2 = computeKnownBits(Sub, DemandedSub, Depth + 1);
3504 Known = Known.intersectWith(Known2);
3505 }
3506 // If we don't know any bits, early out.
3507 if (Known.isUnknown())
3508 break;
3509 }
3510 break;
3511 }
3512 case ISD::INSERT_SUBVECTOR: {
3513 if (Op.getValueType().isScalableVector())
3514 break;
3515 // Demand any elements from the subvector and the remainder from the src its
3516 // inserted into.
3517 SDValue Src = Op.getOperand(0);
3518 SDValue Sub = Op.getOperand(1);
3519 uint64_t Idx = Op.getConstantOperandVal(2);
3520 unsigned NumSubElts = Sub.getValueType().getVectorNumElements();
3521 APInt DemandedSubElts = DemandedElts.extractBits(NumSubElts, Idx);
3522 APInt DemandedSrcElts = DemandedElts;
3523 DemandedSrcElts.clearBits(Idx, Idx + NumSubElts);
3524
3525 Known.setAllConflict();
3526 if (!!DemandedSubElts) {
3527 Known = computeKnownBits(Sub, DemandedSubElts, Depth + 1);
3528 if (Known.isUnknown())
3529 break; // early-out.
3530 }
3531 if (!!DemandedSrcElts) {
3532 Known2 = computeKnownBits(Src, DemandedSrcElts, Depth + 1);
3533 Known = Known.intersectWith(Known2);
3534 }
3535 break;
3536 }
3538 // Offset the demanded elts by the subvector index.
3539 SDValue Src = Op.getOperand(0);
3540
3541 APInt DemandedSrcElts;
3542 if (Src.getValueType().isScalableVector())
3543 DemandedSrcElts = APInt(1, 1); // <=> 'demand all elements'
3544 else {
3545 uint64_t Idx = Op.getConstantOperandVal(1);
3546 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
3547 DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
3548 }
3549 Known = computeKnownBits(Src, DemandedSrcElts, Depth + 1);
3550 break;
3551 }
3552 case ISD::SCALAR_TO_VECTOR: {
3553 if (Op.getValueType().isScalableVector())
3554 break;
3555 // We know about scalar_to_vector as much as we know about it source,
3556 // which becomes the first element of otherwise unknown vector.
3557 if (DemandedElts != 1)
3558 break;
3559
3560 SDValue N0 = Op.getOperand(0);
3561 Known = computeKnownBits(N0, Depth + 1);
3562 if (N0.getValueSizeInBits() != BitWidth)
3563 Known = Known.trunc(BitWidth);
3564
3565 break;
3566 }
3567 case ISD::BITCAST: {
3568 if (Op.getValueType().isScalableVector())
3569 break;
3570
3571 SDValue N0 = Op.getOperand(0);
3572 EVT SubVT = N0.getValueType();
3573 unsigned SubBitWidth = SubVT.getScalarSizeInBits();
3574
3575 // Ignore bitcasts from unsupported types.
3576 if (!(SubVT.isInteger() || SubVT.isFloatingPoint()))
3577 break;
3578
3579 // Fast handling of 'identity' bitcasts.
3580 if (BitWidth == SubBitWidth) {
3581 Known = computeKnownBits(N0, DemandedElts, Depth + 1);
3582 break;
3583 }
3584
3585 bool IsLE = getDataLayout().isLittleEndian();
3586
3587 // Bitcast 'small element' vector to 'large element' scalar/vector.
3588 if ((BitWidth % SubBitWidth) == 0) {
3589 assert(N0.getValueType().isVector() && "Expected bitcast from vector");
3590
3591 // Collect known bits for the (larger) output by collecting the known
3592 // bits from each set of sub elements and shift these into place.
3593 // We need to separately call computeKnownBits for each set of
3594 // sub elements as the knownbits for each is likely to be different.
3595 unsigned SubScale = BitWidth / SubBitWidth;
3596 APInt SubDemandedElts(NumElts * SubScale, 0);
3597 for (unsigned i = 0; i != NumElts; ++i)
3598 if (DemandedElts[i])
3599 SubDemandedElts.setBit(i * SubScale);
3600
3601 for (unsigned i = 0; i != SubScale; ++i) {
3602 Known2 = computeKnownBits(N0, SubDemandedElts.shl(i),
3603 Depth + 1);
3604 unsigned Shifts = IsLE ? i : SubScale - 1 - i;
3605 Known.insertBits(Known2, SubBitWidth * Shifts);
3606 }
3607 }
3608
3609 // Bitcast 'large element' scalar/vector to 'small element' vector.
3610 if ((SubBitWidth % BitWidth) == 0) {
3611 assert(Op.getValueType().isVector() && "Expected bitcast to vector");
3612
3613 // Collect known bits for the (smaller) output by collecting the known
3614 // bits from the overlapping larger input elements and extracting the
3615 // sub sections we actually care about.
3616 unsigned SubScale = SubBitWidth / BitWidth;
3617 APInt SubDemandedElts =
3618 APIntOps::ScaleBitMask(DemandedElts, NumElts / SubScale);
3619 Known2 = computeKnownBits(N0, SubDemandedElts, Depth + 1);
3620
3621 Known.setAllConflict();
3622 for (unsigned i = 0; i != NumElts; ++i)
3623 if (DemandedElts[i]) {
3624 unsigned Shifts = IsLE ? i : NumElts - 1 - i;
3625 unsigned Offset = (Shifts % SubScale) * BitWidth;
3626 Known = Known.intersectWith(Known2.extractBits(BitWidth, Offset));
3627 // If we don't know any bits, early out.
3628 if (Known.isUnknown())
3629 break;
3630 }
3631 }
3632 break;
3633 }
3634 case ISD::AND:
3635 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3636 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3637
3638 Known &= Known2;
3639 break;
3640 case ISD::OR:
3641 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3642 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3643
3644 Known |= Known2;
3645 break;
3646 case ISD::XOR:
3647 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3648 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3649
3650 Known ^= Known2;
3651 break;
3652 case ISD::MUL: {
3653 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3654 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3655 bool SelfMultiply = Op.getOperand(0) == Op.getOperand(1);
3656 // TODO: SelfMultiply can be poison, but not undef.
3657 if (SelfMultiply)
3658 SelfMultiply &= isGuaranteedNotToBeUndefOrPoison(
3659 Op.getOperand(0), DemandedElts, false, Depth + 1);
3660 Known = KnownBits::mul(Known, Known2, SelfMultiply);
3661
3662 // If the multiplication is known not to overflow, the product of a number
3663 // with itself is non-negative. Only do this if we didn't already computed
3664 // the opposite value for the sign bit.
3665 if (Op->getFlags().hasNoSignedWrap() &&
3666 Op.getOperand(0) == Op.getOperand(1) &&
3667 !Known.isNegative())
3668 Known.makeNonNegative();
3669 break;
3670 }
3671 case ISD::MULHU: {
3672 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3673 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3674 Known = KnownBits::mulhu(Known, Known2);
3675 break;
3676 }
3677 case ISD::MULHS: {
3678 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3679 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3680 Known = KnownBits::mulhs(Known, Known2);
3681 break;
3682 }
3683 case ISD::ABDU: {
3684 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3685 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3686 Known = KnownBits::abdu(Known, Known2);
3687 break;
3688 }
3689 case ISD::ABDS: {
3690 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3691 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3692 Known = KnownBits::abds(Known, Known2);
3693 unsigned SignBits1 =
3694 ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
3695 if (SignBits1 == 1)
3696 break;
3697 unsigned SignBits0 =
3698 ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
3699 Known.Zero.setHighBits(std::min(SignBits0, SignBits1) - 1);
3700 break;
3701 }
3702 case ISD::UMUL_LOHI: {
3703 assert((Op.getResNo() == 0 || Op.getResNo() == 1) && "Unknown result");
3704 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3705 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3706 bool SelfMultiply = Op.getOperand(0) == Op.getOperand(1);
3707 if (Op.getResNo() == 0)
3708 Known = KnownBits::mul(Known, Known2, SelfMultiply);
3709 else
3710 Known = KnownBits::mulhu(Known, Known2);
3711 break;
3712 }
3713 case ISD::SMUL_LOHI: {
3714 assert((Op.getResNo() == 0 || Op.getResNo() == 1) && "Unknown result");
3715 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3716 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3717 bool SelfMultiply = Op.getOperand(0) == Op.getOperand(1);
3718 if (Op.getResNo() == 0)
3719 Known = KnownBits::mul(Known, Known2, SelfMultiply);
3720 else
3721 Known = KnownBits::mulhs(Known, Known2);
3722 break;
3723 }
3724 case ISD::AVGFLOORU: {
3725 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3726 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3727 Known = KnownBits::avgFloorU(Known, Known2);
3728 break;
3729 }
3730 case ISD::AVGCEILU: {
3731 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3732 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3733 Known = KnownBits::avgCeilU(Known, Known2);
3734 break;
3735 }
3736 case ISD::AVGFLOORS: {
3737 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3738 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3739 Known = KnownBits::avgFloorS(Known, Known2);
3740 break;
3741 }
3742 case ISD::AVGCEILS: {
3743 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3744 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3745 Known = KnownBits::avgCeilS(Known, Known2);
3746 break;
3747 }
3748 case ISD::SELECT:
3749 case ISD::VSELECT:
3750 Known = computeKnownBits(Op.getOperand(2), DemandedElts, Depth+1);
3751 // If we don't know any bits, early out.
3752 if (Known.isUnknown())
3753 break;
3754 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth+1);
3755
3756 // Only known if known in both the LHS and RHS.
3757 Known = Known.intersectWith(Known2);
3758 break;
3759 case ISD::SELECT_CC:
3760 Known = computeKnownBits(Op.getOperand(3), DemandedElts, Depth+1);
3761 // If we don't know any bits, early out.
3762 if (Known.isUnknown())
3763 break;
3764 Known2 = computeKnownBits(Op.getOperand(2), DemandedElts, Depth+1);
3765
3766 // Only known if known in both the LHS and RHS.
3767 Known = Known.intersectWith(Known2);
3768 break;
3769 case ISD::SMULO:
3770 case ISD::UMULO:
3771 if (Op.getResNo() != 1)
3772 break;
3773 // The boolean result conforms to getBooleanContents.
3774 // If we know the result of a setcc has the top bits zero, use this info.
3775 // We know that we have an integer-based boolean since these operations
3776 // are only available for integer.
3777 if (TLI->getBooleanContents(Op.getValueType().isVector(), false) ==
3779 BitWidth > 1)
3780 Known.Zero.setBitsFrom(1);
3781 break;
3782 case ISD::SETCC:
3783 case ISD::SETCCCARRY:
3784 case ISD::STRICT_FSETCC:
3785 case ISD::STRICT_FSETCCS: {
3786 unsigned OpNo = Op->isStrictFPOpcode() ? 1 : 0;
3787 // If we know the result of a setcc has the top bits zero, use this info.
3788 if (TLI->getBooleanContents(Op.getOperand(OpNo).getValueType()) ==
3790 BitWidth > 1)
3791 Known.Zero.setBitsFrom(1);
3792 break;
3793 }
3794 case ISD::SHL: {
3795 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3796 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3797
3798 bool NUW = Op->getFlags().hasNoUnsignedWrap();
3799 bool NSW = Op->getFlags().hasNoSignedWrap();
3800
3801 bool ShAmtNonZero = Known2.isNonZero();
3802
3803 Known = KnownBits::shl(Known, Known2, NUW, NSW, ShAmtNonZero);
3804
3805 // Minimum shift low bits are known zero.
3806 if (std::optional<unsigned> ShMinAmt =
3807 getValidMinimumShiftAmount(Op, DemandedElts, Depth + 1))
3808 Known.Zero.setLowBits(*ShMinAmt);
3809 break;
3810 }
3811 case ISD::SRL:
3812 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3813 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3814 Known = KnownBits::lshr(Known, Known2, /*ShAmtNonZero=*/false,
3815 Op->getFlags().hasExact());
3816
3817 // Minimum shift high bits are known zero.
3818 if (std::optional<unsigned> ShMinAmt =
3819 getValidMinimumShiftAmount(Op, DemandedElts, Depth + 1))
3820 Known.Zero.setHighBits(*ShMinAmt);
3821 break;
3822 case ISD::SRA:
3823 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3824 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3825 Known = KnownBits::ashr(Known, Known2, /*ShAmtNonZero=*/false,
3826 Op->getFlags().hasExact());
3827 break;
3828 case ISD::ROTL:
3829 case ISD::ROTR:
3830 if (ConstantSDNode *C =
3831 isConstOrConstSplat(Op.getOperand(1), DemandedElts)) {
3832 unsigned Amt = C->getAPIntValue().urem(BitWidth);
3833
3834 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3835
3836 // Canonicalize to ROTR.
3837 if (Opcode == ISD::ROTL && Amt != 0)
3838 Amt = BitWidth - Amt;
3839
3840 Known.Zero = Known.Zero.rotr(Amt);
3841 Known.One = Known.One.rotr(Amt);
3842 }
3843 break;
3844 case ISD::FSHL:
3845 case ISD::FSHR:
3846 if (ConstantSDNode *C = isConstOrConstSplat(Op.getOperand(2), DemandedElts)) {
3847 unsigned Amt = C->getAPIntValue().urem(BitWidth);
3848
3849 // For fshl, 0-shift returns the 1st arg.
3850 // For fshr, 0-shift returns the 2nd arg.
3851 if (Amt == 0) {
3852 Known = computeKnownBits(Op.getOperand(Opcode == ISD::FSHL ? 0 : 1),
3853 DemandedElts, Depth + 1);
3854 break;
3855 }
3856
3857 // fshl: (X << (Z % BW)) | (Y >> (BW - (Z % BW)))
3858 // fshr: (X << (BW - (Z % BW))) | (Y >> (Z % BW))
3859 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3860 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3861 if (Opcode == ISD::FSHL) {
3862 Known <<= Amt;
3863 Known2 >>= BitWidth - Amt;
3864 } else {
3865 Known <<= BitWidth - Amt;
3866 Known2 >>= Amt;
3867 }
3868 Known = Known.unionWith(Known2);
3869 }
3870 break;
3871 case ISD::SHL_PARTS:
3872 case ISD::SRA_PARTS:
3873 case ISD::SRL_PARTS: {
3874 assert((Op.getResNo() == 0 || Op.getResNo() == 1) && "Unknown result");
3875
3876 // Collect lo/hi source values and concatenate.
3877 unsigned LoBits = Op.getOperand(0).getScalarValueSizeInBits();
3878 unsigned HiBits = Op.getOperand(1).getScalarValueSizeInBits();
3879 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3880 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3881 Known = Known2.concat(Known);
3882
3883 // Collect shift amount.
3884 Known2 = computeKnownBits(Op.getOperand(2), DemandedElts, Depth + 1);
3885
3886 if (Opcode == ISD::SHL_PARTS)
3887 Known = KnownBits::shl(Known, Known2);
3888 else if (Opcode == ISD::SRA_PARTS)
3889 Known = KnownBits::ashr(Known, Known2);
3890 else // if (Opcode == ISD::SRL_PARTS)
3891 Known = KnownBits::lshr(Known, Known2);
3892
3893 // TODO: Minimum shift low/high bits are known zero.
3894
3895 if (Op.getResNo() == 0)
3896 Known = Known.extractBits(LoBits, 0);
3897 else
3898 Known = Known.extractBits(HiBits, LoBits);
3899 break;
3900 }
3902 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3903 EVT EVT = cast<VTSDNode>(Op.getOperand(1))->getVT();
3904 Known = Known.sextInReg(EVT.getScalarSizeInBits());
3905 break;
3906 }
3907 case ISD::CTTZ:
3908 case ISD::CTTZ_ZERO_UNDEF: {
3909 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3910 // If we have a known 1, its position is our upper bound.
3911 unsigned PossibleTZ = Known2.countMaxTrailingZeros();
3912 unsigned LowBits = llvm::bit_width(PossibleTZ);
3913 Known.Zero.setBitsFrom(LowBits);
3914 break;
3915 }
3916 case ISD::CTLZ:
3917 case ISD::CTLZ_ZERO_UNDEF: {
3918 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3919 // If we have a known 1, its position is our upper bound.
3920 unsigned PossibleLZ = Known2.countMaxLeadingZeros();
3921 unsigned LowBits = llvm::bit_width(PossibleLZ);
3922 Known.Zero.setBitsFrom(LowBits);
3923 break;
3924 }
3925 case ISD::CTLS: {
3926 unsigned MinRedundantSignBits =
3927 ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1) - 1;
3928 ConstantRange Range(APInt(BitWidth, MinRedundantSignBits),
3930 Known = Range.toKnownBits();
3931 break;
3932 }
3933 case ISD::CTPOP: {
3934 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3935 // If we know some of the bits are zero, they can't be one.
3936 unsigned PossibleOnes = Known2.countMaxPopulation();
3937 Known.Zero.setBitsFrom(llvm::bit_width(PossibleOnes));
3938 break;
3939 }
3940 case ISD::PARITY: {
3941 // Parity returns 0 everywhere but the LSB.
3942 Known.Zero.setBitsFrom(1);
3943 break;
3944 }
3945 case ISD::CLMUL: {
3946 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3947 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3948 Known = KnownBits::clmul(Known, Known2);
3949 break;
3950 }
3951 case ISD::MGATHER:
3952 case ISD::MLOAD: {
3953 ISD::LoadExtType ETy =
3954 (Opcode == ISD::MGATHER)
3955 ? cast<MaskedGatherSDNode>(Op)->getExtensionType()
3956 : cast<MaskedLoadSDNode>(Op)->getExtensionType();
3957 if (ETy == ISD::ZEXTLOAD) {
3958 EVT MemVT = cast<MemSDNode>(Op)->getMemoryVT();
3959 KnownBits Known0(MemVT.getScalarSizeInBits());
3960 return Known0.zext(BitWidth);
3961 }
3962 break;
3963 }
3964 case ISD::LOAD: {
3966 const Constant *Cst = TLI->getTargetConstantFromLoad(LD);
3967 if (ISD::isNON_EXTLoad(LD) && Cst) {
3968 // Determine any common known bits from the loaded constant pool value.
3969 Type *CstTy = Cst->getType();
3970 if ((NumElts * BitWidth) == CstTy->getPrimitiveSizeInBits() &&
3971 !Op.getValueType().isScalableVector()) {
3972 // If its a vector splat, then we can (quickly) reuse the scalar path.
3973 // NOTE: We assume all elements match and none are UNDEF.
3974 if (CstTy->isVectorTy()) {
3975 if (const Constant *Splat = Cst->getSplatValue()) {
3976 Cst = Splat;
3977 CstTy = Cst->getType();
3978 }
3979 }
3980 // TODO - do we need to handle different bitwidths?
3981 if (CstTy->isVectorTy() && BitWidth == CstTy->getScalarSizeInBits()) {
3982 // Iterate across all vector elements finding common known bits.
3983 Known.setAllConflict();
3984 for (unsigned i = 0; i != NumElts; ++i) {
3985 if (!DemandedElts[i])
3986 continue;
3987 if (Constant *Elt = Cst->getAggregateElement(i)) {
3988 if (auto *CInt = dyn_cast<ConstantInt>(Elt)) {
3989 const APInt &Value = CInt->getValue();
3990 Known.One &= Value;
3991 Known.Zero &= ~Value;
3992 continue;
3993 }
3994 if (auto *CFP = dyn_cast<ConstantFP>(Elt)) {
3995 APInt Value = CFP->getValueAPF().bitcastToAPInt();
3996 Known.One &= Value;
3997 Known.Zero &= ~Value;
3998 continue;
3999 }
4000 }
4001 Known.One.clearAllBits();
4002 Known.Zero.clearAllBits();
4003 break;
4004 }
4005 } else if (BitWidth == CstTy->getPrimitiveSizeInBits()) {
4006 if (auto *CInt = dyn_cast<ConstantInt>(Cst)) {
4007 Known = KnownBits::makeConstant(CInt->getValue());
4008 } else if (auto *CFP = dyn_cast<ConstantFP>(Cst)) {
4009 Known =
4010 KnownBits::makeConstant(CFP->getValueAPF().bitcastToAPInt());
4011 }
4012 }
4013 }
4014 } else if (Op.getResNo() == 0) {
4015 unsigned ScalarMemorySize = LD->getMemoryVT().getScalarSizeInBits();
4016 KnownBits KnownScalarMemory(ScalarMemorySize);
4017 if (const MDNode *MD = LD->getRanges())
4018 computeKnownBitsFromRangeMetadata(*MD, KnownScalarMemory);
4019
4020 // Extend the Known bits from memory to the size of the scalar result.
4021 if (ISD::isZEXTLoad(Op.getNode()))
4022 Known = KnownScalarMemory.zext(BitWidth);
4023 else if (ISD::isSEXTLoad(Op.getNode()))
4024 Known = KnownScalarMemory.sext(BitWidth);
4025 else if (ISD::isEXTLoad(Op.getNode()))
4026 Known = KnownScalarMemory.anyext(BitWidth);
4027 else
4028 Known = KnownScalarMemory;
4029 assert(Known.getBitWidth() == BitWidth);
4030 return Known;
4031 }
4032 break;
4033 }
4035 if (Op.getValueType().isScalableVector())
4036 break;
4037 EVT InVT = Op.getOperand(0).getValueType();
4038 APInt InDemandedElts = DemandedElts.zext(InVT.getVectorNumElements());
4039 Known = computeKnownBits(Op.getOperand(0), InDemandedElts, Depth + 1);
4040 Known = Known.zext(BitWidth);
4041 break;
4042 }
4043 case ISD::ZERO_EXTEND: {
4044 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4045 Known = Known.zext(BitWidth);
4046 break;
4047 }
4049 if (Op.getValueType().isScalableVector())
4050 break;
4051 EVT InVT = Op.getOperand(0).getValueType();
4052 APInt InDemandedElts = DemandedElts.zext(InVT.getVectorNumElements());
4053 Known = computeKnownBits(Op.getOperand(0), InDemandedElts, Depth + 1);
4054 // If the sign bit is known to be zero or one, then sext will extend
4055 // it to the top bits, else it will just zext.
4056 Known = Known.sext(BitWidth);
4057 break;
4058 }
4059 case ISD::SIGN_EXTEND: {
4060 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4061 // If the sign bit is known to be zero or one, then sext will extend
4062 // it to the top bits, else it will just zext.
4063 Known = Known.sext(BitWidth);
4064 break;
4065 }
4067 if (Op.getValueType().isScalableVector())
4068 break;
4069 EVT InVT = Op.getOperand(0).getValueType();
4070 APInt InDemandedElts = DemandedElts.zext(InVT.getVectorNumElements());
4071 Known = computeKnownBits(Op.getOperand(0), InDemandedElts, Depth + 1);
4072 Known = Known.anyext(BitWidth);
4073 break;
4074 }
4075 case ISD::ANY_EXTEND: {
4076 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4077 Known = Known.anyext(BitWidth);
4078 break;
4079 }
4080 case ISD::TRUNCATE: {
4081 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4082 Known = Known.trunc(BitWidth);
4083 break;
4084 }
4085 case ISD::TRUNCATE_SSAT_S: {
4086 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4087 Known = Known.truncSSat(BitWidth);
4088 break;
4089 }
4090 case ISD::TRUNCATE_SSAT_U: {
4091 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4092 Known = Known.truncSSatU(BitWidth);
4093 break;
4094 }
4095 case ISD::TRUNCATE_USAT_U: {
4096 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4097 Known = Known.truncUSat(BitWidth);
4098 break;
4099 }
4100 case ISD::AssertZext: {
4101 EVT VT = cast<VTSDNode>(Op.getOperand(1))->getVT();
4103 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4104 Known.Zero |= (~InMask);
4105 Known.One &= (~Known.Zero);
4106 break;
4107 }
4108 case ISD::AssertAlign: {
4109 unsigned LogOfAlign = Log2(cast<AssertAlignSDNode>(Op)->getAlign());
4110 assert(LogOfAlign != 0);
4111
4112 // TODO: Should use maximum with source
4113 // If a node is guaranteed to be aligned, set low zero bits accordingly as
4114 // well as clearing one bits.
4115 Known.Zero.setLowBits(LogOfAlign);
4116 Known.One.clearLowBits(LogOfAlign);
4117 break;
4118 }
4119 case ISD::AssertNoFPClass: {
4120 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4121
4122 FPClassTest NoFPClass =
4123 static_cast<FPClassTest>(Op.getConstantOperandVal(1));
4124 const FPClassTest NegativeTestMask = fcNan | fcNegative;
4125 if ((NoFPClass & NegativeTestMask) == NegativeTestMask) {
4126 // Cannot be negative.
4127 Known.makeNonNegative();
4128 }
4129
4130 const FPClassTest PositiveTestMask = fcNan | fcPositive;
4131 if ((NoFPClass & PositiveTestMask) == PositiveTestMask) {
4132 // Cannot be positive.
4133 Known.makeNegative();
4134 }
4135
4136 break;
4137 }
4138 case ISD::FGETSIGN:
4139 // All bits are zero except the low bit.
4140 Known.Zero.setBitsFrom(1);
4141 break;
4142 case ISD::ADD:
4143 case ISD::SUB: {
4144 SDNodeFlags Flags = Op.getNode()->getFlags();
4145 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4146 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4148 Op.getOpcode() == ISD::ADD, Flags.hasNoSignedWrap(),
4149 Flags.hasNoUnsignedWrap(), Known, Known2);
4150 break;
4151 }
4152 case ISD::USUBO:
4153 case ISD::SSUBO:
4154 case ISD::USUBO_CARRY:
4155 case ISD::SSUBO_CARRY:
4156 if (Op.getResNo() == 1) {
4157 // If we know the result of a setcc has the top bits zero, use this info.
4158 if (TLI->getBooleanContents(Op.getOperand(0).getValueType()) ==
4160 BitWidth > 1)
4161 Known.Zero.setBitsFrom(1);
4162 break;
4163 }
4164 [[fallthrough]];
4165 case ISD::SUBC: {
4166 assert(Op.getResNo() == 0 &&
4167 "We only compute knownbits for the difference here.");
4168
4169 // With USUBO_CARRY and SSUBO_CARRY a borrow bit may be added in.
4170 KnownBits Borrow(1);
4171 if (Opcode == ISD::USUBO_CARRY || Opcode == ISD::SSUBO_CARRY) {
4172 Borrow = computeKnownBits(Op.getOperand(2), DemandedElts, Depth + 1);
4173 // Borrow has bit width 1
4174 Borrow = Borrow.trunc(1);
4175 } else {
4176 Borrow.setAllZero();
4177 }
4178
4179 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4180 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4181 Known = KnownBits::computeForSubBorrow(Known, Known2, Borrow);
4182 break;
4183 }
4184 case ISD::UADDO:
4185 case ISD::SADDO:
4186 case ISD::UADDO_CARRY:
4187 case ISD::SADDO_CARRY:
4188 if (Op.getResNo() == 1) {
4189 // If we know the result of a setcc has the top bits zero, use this info.
4190 if (TLI->getBooleanContents(Op.getOperand(0).getValueType()) ==
4192 BitWidth > 1)
4193 Known.Zero.setBitsFrom(1);
4194 break;
4195 }
4196 [[fallthrough]];
4197 case ISD::ADDC:
4198 case ISD::ADDE: {
4199 assert(Op.getResNo() == 0 && "We only compute knownbits for the sum here.");
4200
4201 // With ADDE and UADDO_CARRY, a carry bit may be added in.
4202 KnownBits Carry(1);
4203 if (Opcode == ISD::ADDE)
4204 // Can't track carry from glue, set carry to unknown.
4205 Carry.resetAll();
4206 else if (Opcode == ISD::UADDO_CARRY || Opcode == ISD::SADDO_CARRY) {
4207 Carry = computeKnownBits(Op.getOperand(2), DemandedElts, Depth + 1);
4208 // Carry has bit width 1
4209 Carry = Carry.trunc(1);
4210 } else {
4211 Carry.setAllZero();
4212 }
4213
4214 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4215 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4216 Known = KnownBits::computeForAddCarry(Known, Known2, Carry);
4217 break;
4218 }
4219 case ISD::UDIV: {
4220 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4221 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4222 Known = KnownBits::udiv(Known, Known2, Op->getFlags().hasExact());
4223 break;
4224 }
4225 case ISD::SDIV: {
4226 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4227 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4228 Known = KnownBits::sdiv(Known, Known2, Op->getFlags().hasExact());
4229 break;
4230 }
4231 case ISD::SREM: {
4232 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4233 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4234 Known = KnownBits::srem(Known, Known2);
4235 break;
4236 }
4237 case ISD::UREM: {
4238 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4239 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4240 Known = KnownBits::urem(Known, Known2);
4241 break;
4242 }
4243 case ISD::EXTRACT_ELEMENT: {
4244 Known = computeKnownBits(Op.getOperand(0), Depth+1);
4245 const unsigned Index = Op.getConstantOperandVal(1);
4246 const unsigned EltBitWidth = Op.getValueSizeInBits();
4247
4248 // Remove low part of known bits mask
4249 Known.Zero = Known.Zero.getHiBits(Known.getBitWidth() - Index * EltBitWidth);
4250 Known.One = Known.One.getHiBits(Known.getBitWidth() - Index * EltBitWidth);
4251
4252 // Remove high part of known bit mask
4253 Known = Known.trunc(EltBitWidth);
4254 break;
4255 }
4257 SDValue InVec = Op.getOperand(0);
4258 SDValue EltNo = Op.getOperand(1);
4259 EVT VecVT = InVec.getValueType();
4260 // computeKnownBits not yet implemented for scalable vectors.
4261 if (VecVT.isScalableVector())
4262 break;
4263 const unsigned EltBitWidth = VecVT.getScalarSizeInBits();
4264 const unsigned NumSrcElts = VecVT.getVectorNumElements();
4265
4266 // If BitWidth > EltBitWidth the value is anyext:ed. So we do not know
4267 // anything about the extended bits.
4268 if (BitWidth > EltBitWidth)
4269 Known = Known.trunc(EltBitWidth);
4270
4271 // If we know the element index, just demand that vector element, else for
4272 // an unknown element index, ignore DemandedElts and demand them all.
4273 APInt DemandedSrcElts = APInt::getAllOnes(NumSrcElts);
4274 auto *ConstEltNo = dyn_cast<ConstantSDNode>(EltNo);
4275 if (ConstEltNo && ConstEltNo->getAPIntValue().ult(NumSrcElts))
4276 DemandedSrcElts =
4277 APInt::getOneBitSet(NumSrcElts, ConstEltNo->getZExtValue());
4278
4279 Known = computeKnownBits(InVec, DemandedSrcElts, Depth + 1);
4280 if (BitWidth > EltBitWidth)
4281 Known = Known.anyext(BitWidth);
4282 break;
4283 }
4285 if (Op.getValueType().isScalableVector())
4286 break;
4287
4288 // If we know the element index, split the demand between the
4289 // source vector and the inserted element, otherwise assume we need
4290 // the original demanded vector elements and the value.
4291 SDValue InVec = Op.getOperand(0);
4292 SDValue InVal = Op.getOperand(1);
4293 SDValue EltNo = Op.getOperand(2);
4294 bool DemandedVal = true;
4295 APInt DemandedVecElts = DemandedElts;
4296 auto *CEltNo = dyn_cast<ConstantSDNode>(EltNo);
4297 if (CEltNo && CEltNo->getAPIntValue().ult(NumElts)) {
4298 unsigned EltIdx = CEltNo->getZExtValue();
4299 DemandedVal = !!DemandedElts[EltIdx];
4300 DemandedVecElts.clearBit(EltIdx);
4301 }
4302 Known.setAllConflict();
4303 if (DemandedVal) {
4304 Known2 = computeKnownBits(InVal, Depth + 1);
4305 Known = Known.intersectWith(Known2.zextOrTrunc(BitWidth));
4306 }
4307 if (!!DemandedVecElts) {
4308 Known2 = computeKnownBits(InVec, DemandedVecElts, Depth + 1);
4309 Known = Known.intersectWith(Known2);
4310 }
4311 break;
4312 }
4313 case ISD::BITREVERSE: {
4314 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4315 Known = Known2.reverseBits();
4316 break;
4317 }
4318 case ISD::BSWAP: {
4319 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4320 Known = Known2.byteSwap();
4321 break;
4322 }
4323 case ISD::ABS: {
4324 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4325 Known = Known2.abs();
4326 Known.Zero.setHighBits(
4327 ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1) - 1);
4328 break;
4329 }
4330 case ISD::USUBSAT: {
4331 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4332 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4333 Known = KnownBits::usub_sat(Known, Known2);
4334 break;
4335 }
4336 case ISD::UMIN: {
4337 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4338 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4339 Known = KnownBits::umin(Known, Known2);
4340 break;
4341 }
4342 case ISD::UMAX: {
4343 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4344 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4345 Known = KnownBits::umax(Known, Known2);
4346 break;
4347 }
4348 case ISD::SMIN:
4349 case ISD::SMAX: {
4350 // If we have a clamp pattern, we know that the number of sign bits will be
4351 // the minimum of the clamp min/max range.
4352 bool IsMax = (Opcode == ISD::SMAX);
4353 ConstantSDNode *CstLow = nullptr, *CstHigh = nullptr;
4354 if ((CstLow = isConstOrConstSplat(Op.getOperand(1), DemandedElts)))
4355 if (Op.getOperand(0).getOpcode() == (IsMax ? ISD::SMIN : ISD::SMAX))
4356 CstHigh =
4357 isConstOrConstSplat(Op.getOperand(0).getOperand(1), DemandedElts);
4358 if (CstLow && CstHigh) {
4359 if (!IsMax)
4360 std::swap(CstLow, CstHigh);
4361
4362 const APInt &ValueLow = CstLow->getAPIntValue();
4363 const APInt &ValueHigh = CstHigh->getAPIntValue();
4364 if (ValueLow.sle(ValueHigh)) {
4365 unsigned LowSignBits = ValueLow.getNumSignBits();
4366 unsigned HighSignBits = ValueHigh.getNumSignBits();
4367 unsigned MinSignBits = std::min(LowSignBits, HighSignBits);
4368 if (ValueLow.isNegative() && ValueHigh.isNegative()) {
4369 Known.One.setHighBits(MinSignBits);
4370 break;
4371 }
4372 if (ValueLow.isNonNegative() && ValueHigh.isNonNegative()) {
4373 Known.Zero.setHighBits(MinSignBits);
4374 break;
4375 }
4376 }
4377 }
4378
4379 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4380 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4381 if (IsMax)
4382 Known = KnownBits::smax(Known, Known2);
4383 else
4384 Known = KnownBits::smin(Known, Known2);
4385
4386 // For SMAX, if CstLow is non-negative we know the result will be
4387 // non-negative and thus all sign bits are 0.
4388 // TODO: There's an equivalent of this for smin with negative constant for
4389 // known ones.
4390 if (IsMax && CstLow) {
4391 const APInt &ValueLow = CstLow->getAPIntValue();
4392 if (ValueLow.isNonNegative()) {
4393 unsigned SignBits = ComputeNumSignBits(Op.getOperand(0), Depth + 1);
4394 Known.Zero.setHighBits(std::min(SignBits, ValueLow.getNumSignBits()));
4395 }
4396 }
4397
4398 break;
4399 }
4400 case ISD::UINT_TO_FP: {
4401 Known.makeNonNegative();
4402 break;
4403 }
4404 case ISD::SINT_TO_FP: {
4405 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4406 if (Known2.isNonNegative())
4407 Known.makeNonNegative();
4408 else if (Known2.isNegative())
4409 Known.makeNegative();
4410 break;
4411 }
4412 case ISD::FP_TO_UINT_SAT: {
4413 // FP_TO_UINT_SAT produces an unsigned value that fits in the saturating VT.
4414 EVT VT = cast<VTSDNode>(Op.getOperand(1))->getVT();
4416 break;
4417 }
4418 case ISD::ATOMIC_LOAD: {
4419 // If we are looking at the loaded value.
4420 if (Op.getResNo() == 0) {
4421 auto *AT = cast<AtomicSDNode>(Op);
4422 unsigned ScalarMemorySize = AT->getMemoryVT().getScalarSizeInBits();
4423 KnownBits KnownScalarMemory(ScalarMemorySize);
4424 if (const MDNode *MD = AT->getRanges())
4425 computeKnownBitsFromRangeMetadata(*MD, KnownScalarMemory);
4426
4427 switch (AT->getExtensionType()) {
4428 case ISD::ZEXTLOAD:
4429 Known = KnownScalarMemory.zext(BitWidth);
4430 break;
4431 case ISD::SEXTLOAD:
4432 Known = KnownScalarMemory.sext(BitWidth);
4433 break;
4434 case ISD::EXTLOAD:
4435 switch (TLI->getExtendForAtomicOps()) {
4436 case ISD::ZERO_EXTEND:
4437 Known = KnownScalarMemory.zext(BitWidth);
4438 break;
4439 case ISD::SIGN_EXTEND:
4440 Known = KnownScalarMemory.sext(BitWidth);
4441 break;
4442 default:
4443 Known = KnownScalarMemory.anyext(BitWidth);
4444 break;
4445 }
4446 break;
4447 case ISD::NON_EXTLOAD:
4448 Known = KnownScalarMemory;
4449 break;
4450 }
4451 assert(Known.getBitWidth() == BitWidth);
4452 }
4453 break;
4454 }
4456 if (Op.getResNo() == 1) {
4457 // The boolean result conforms to getBooleanContents.
4458 // If we know the result of a setcc has the top bits zero, use this info.
4459 // We know that we have an integer-based boolean since these operations
4460 // are only available for integer.
4461 if (TLI->getBooleanContents(Op.getValueType().isVector(), false) ==
4463 BitWidth > 1)
4464 Known.Zero.setBitsFrom(1);
4465 break;
4466 }
4467 [[fallthrough]];
4469 case ISD::ATOMIC_SWAP:
4480 case ISD::ATOMIC_LOAD_UMAX: {
4481 // If we are looking at the loaded value.
4482 if (Op.getResNo() == 0) {
4483 auto *AT = cast<AtomicSDNode>(Op);
4484 unsigned MemBits = AT->getMemoryVT().getScalarSizeInBits();
4485
4486 if (TLI->getExtendForAtomicOps() == ISD::ZERO_EXTEND)
4487 Known.Zero.setBitsFrom(MemBits);
4488 }
4489 break;
4490 }
4491 case ISD::FrameIndex:
4493 TLI->computeKnownBitsForFrameIndex(cast<FrameIndexSDNode>(Op)->getIndex(),
4494 Known, getMachineFunction());
4495 break;
4496
4497 default:
4498 if (Opcode < ISD::BUILTIN_OP_END)
4499 break;
4500 [[fallthrough]];
4504 // TODO: Probably okay to remove after audit; here to reduce change size
4505 // in initial enablement patch for scalable vectors
4506 if (Op.getValueType().isScalableVector())
4507 break;
4508
4509 // Allow the target to implement this method for its nodes.
4510 TLI->computeKnownBitsForTargetNode(Op, Known, DemandedElts, *this, Depth);
4511 break;
4512 }
4513
4514 return Known;
4515}
4516
4517/// Convert ConstantRange OverflowResult into SelectionDAG::OverflowKind.
4530
4533 // X + 0 never overflow
4534 if (isNullConstant(N1))
4535 return OFK_Never;
4536
4537 // If both operands each have at least two sign bits, the addition
4538 // cannot overflow.
4539 if (ComputeNumSignBits(N0) > 1 && ComputeNumSignBits(N1) > 1)
4540 return OFK_Never;
4541
4542 // TODO: Add ConstantRange::signedAddMayOverflow handling.
4543 return OFK_Sometime;
4544}
4545
4548 // X + 0 never overflow
4549 if (isNullConstant(N1))
4550 return OFK_Never;
4551
4552 // mulhi + 1 never overflow
4553 KnownBits N1Known = computeKnownBits(N1);
4554 if (N0.getOpcode() == ISD::UMUL_LOHI && N0.getResNo() == 1 &&
4555 N1Known.getMaxValue().ult(2))
4556 return OFK_Never;
4557
4558 KnownBits N0Known = computeKnownBits(N0);
4559 if (N1.getOpcode() == ISD::UMUL_LOHI && N1.getResNo() == 1 &&
4560 N0Known.getMaxValue().ult(2))
4561 return OFK_Never;
4562
4563 // Fallback to ConstantRange::unsignedAddMayOverflow handling.
4564 ConstantRange N0Range = ConstantRange::fromKnownBits(N0Known, false);
4565 ConstantRange N1Range = ConstantRange::fromKnownBits(N1Known, false);
4566 return mapOverflowResult(N0Range.unsignedAddMayOverflow(N1Range));
4567}
4568
4571 // X - 0 never overflow
4572 if (isNullConstant(N1))
4573 return OFK_Never;
4574
4575 // If both operands each have at least two sign bits, the subtraction
4576 // cannot overflow.
4577 if (ComputeNumSignBits(N0) > 1 && ComputeNumSignBits(N1) > 1)
4578 return OFK_Never;
4579
4580 KnownBits N0Known = computeKnownBits(N0);
4581 KnownBits N1Known = computeKnownBits(N1);
4582 ConstantRange N0Range = ConstantRange::fromKnownBits(N0Known, true);
4583 ConstantRange N1Range = ConstantRange::fromKnownBits(N1Known, true);
4584 return mapOverflowResult(N0Range.signedSubMayOverflow(N1Range));
4585}
4586
4589 // X - 0 never overflow
4590 if (isNullConstant(N1))
4591 return OFK_Never;
4592
4593 KnownBits N0Known = computeKnownBits(N0);
4594 KnownBits N1Known = computeKnownBits(N1);
4595 ConstantRange N0Range = ConstantRange::fromKnownBits(N0Known, false);
4596 ConstantRange N1Range = ConstantRange::fromKnownBits(N1Known, false);
4597 return mapOverflowResult(N0Range.unsignedSubMayOverflow(N1Range));
4598}
4599
4602 // X * 0 and X * 1 never overflow.
4603 if (isNullConstant(N1) || isOneConstant(N1))
4604 return OFK_Never;
4605
4606 KnownBits N0Known = computeKnownBits(N0);
4607 KnownBits N1Known = computeKnownBits(N1);
4608 ConstantRange N0Range = ConstantRange::fromKnownBits(N0Known, false);
4609 ConstantRange N1Range = ConstantRange::fromKnownBits(N1Known, false);
4610 return mapOverflowResult(N0Range.unsignedMulMayOverflow(N1Range));
4611}
4612
4615 // X * 0 and X * 1 never overflow.
4616 if (isNullConstant(N1) || isOneConstant(N1))
4617 return OFK_Never;
4618
4619 // Get the size of the result.
4620 unsigned BitWidth = N0.getScalarValueSizeInBits();
4621
4622 // Sum of the sign bits.
4623 unsigned SignBits = ComputeNumSignBits(N0) + ComputeNumSignBits(N1);
4624
4625 // If we have enough sign bits, then there's no overflow.
4626 if (SignBits > BitWidth + 1)
4627 return OFK_Never;
4628
4629 if (SignBits == BitWidth + 1) {
4630 // The overflow occurs when the true multiplication of the
4631 // the operands is the minimum negative number.
4632 KnownBits N0Known = computeKnownBits(N0);
4633 KnownBits N1Known = computeKnownBits(N1);
4634 // If one of the operands is non-negative, then there's no
4635 // overflow.
4636 if (N0Known.isNonNegative() || N1Known.isNonNegative())
4637 return OFK_Never;
4638 }
4639
4640 return OFK_Sometime;
4641}
4642
4644 unsigned Depth) const {
4645 EVT VT = Val.getValueType();
4646
4647 // Since the number of lanes in a scalable vector is unknown at compile time,
4648 // we track one bit which is implicitly broadcast to all lanes. This means
4649 // that all lanes in a scalable vector are considered demanded.
4650 APInt DemandedElts = VT.isFixedLengthVector()
4652 : APInt(1, 1);
4653
4654 return isKnownToBeAPowerOfTwo(Val, DemandedElts, OrZero, Depth);
4655}
4656
4658 const APInt &DemandedElts,
4659 bool OrZero, unsigned Depth) const {
4660 if (Depth >= MaxRecursionDepth)
4661 return false; // Limit search depth.
4662
4663 EVT OpVT = Val.getValueType();
4664 unsigned BitWidth = OpVT.getScalarSizeInBits();
4665 [[maybe_unused]] unsigned NumElts = DemandedElts.getBitWidth();
4666 assert((!OpVT.isScalableVector() || NumElts == 1) &&
4667 "DemandedElts for scalable vectors must be 1 to represent all lanes");
4668 assert(
4669 (!OpVT.isFixedLengthVector() || NumElts == OpVT.getVectorNumElements()) &&
4670 "Unexpected vector size");
4671
4672 auto IsPowerOfTwoOrZero = [BitWidth, OrZero](const ConstantSDNode *C) {
4673 APInt V = C->getAPIntValue().zextOrTrunc(BitWidth);
4674 return (OrZero && V.isZero()) || V.isPowerOf2();
4675 };
4676
4677 // Is the constant a known power of 2 or zero?
4678 if (ISD::matchUnaryPredicate(Val, IsPowerOfTwoOrZero))
4679 return true;
4680
4681 switch (Val.getOpcode()) {
4682 case ISD::BUILD_VECTOR:
4683 // Are all operands of a build vector constant powers of two or zero?
4684 if (all_of(enumerate(Val->ops()), [&](auto P) {
4685 auto *C = dyn_cast<ConstantSDNode>(P.value());
4686 return !DemandedElts[P.index()] || (C && IsPowerOfTwoOrZero(C));
4687 }))
4688 return true;
4689 break;
4690
4691 case ISD::SPLAT_VECTOR:
4692 // Is the operand of a splat vector a constant power of two?
4693 if (auto *C = dyn_cast<ConstantSDNode>(Val->getOperand(0)))
4694 if (IsPowerOfTwoOrZero(C))
4695 return true;
4696 break;
4697
4699 SDValue InVec = Val.getOperand(0);
4700 SDValue EltNo = Val.getOperand(1);
4701 EVT VecVT = InVec.getValueType();
4702
4703 // Skip scalable vectors or implicit extensions.
4704 if (VecVT.isScalableVector() ||
4705 OpVT.getScalarSizeInBits() != VecVT.getScalarSizeInBits())
4706 break;
4707
4708 // If we know the element index, just demand that vector element, else for
4709 // an unknown element index, ignore DemandedElts and demand them all.
4710 const unsigned NumSrcElts = VecVT.getVectorNumElements();
4711 auto *ConstEltNo = dyn_cast<ConstantSDNode>(EltNo);
4712 APInt DemandedSrcElts =
4713 ConstEltNo && ConstEltNo->getAPIntValue().ult(NumSrcElts)
4714 ? APInt::getOneBitSet(NumSrcElts, ConstEltNo->getZExtValue())
4715 : APInt::getAllOnes(NumSrcElts);
4716 return isKnownToBeAPowerOfTwo(InVec, DemandedSrcElts, OrZero, Depth + 1);
4717 }
4718
4719 case ISD::AND: {
4720 // Looking for `x & -x` pattern:
4721 // If x == 0:
4722 // x & -x -> 0
4723 // If x != 0:
4724 // x & -x -> non-zero pow2
4725 // so if we find the pattern return whether we know `x` is non-zero.
4726 SDValue X;
4727 if (sd_match(Val, m_And(m_Value(X), m_Neg(m_Deferred(X)))))
4728 return OrZero || isKnownNeverZero(X, DemandedElts, Depth);
4729 break;
4730 }
4731
4732 case ISD::SHL: {
4733 // A left-shift of a constant one will have exactly one bit set because
4734 // shifting the bit off the end is undefined.
4735 auto *C = isConstOrConstSplat(Val.getOperand(0), DemandedElts);
4736 if (C && C->getAPIntValue() == 1)
4737 return true;
4738 return (OrZero || isKnownNeverZero(Val, DemandedElts, Depth)) &&
4739 isKnownToBeAPowerOfTwo(Val.getOperand(0), DemandedElts, OrZero,
4740 Depth + 1);
4741 }
4742
4743 case ISD::SRL: {
4744 // A logical right-shift of a constant sign-bit will have exactly
4745 // one bit set.
4746 auto *C = isConstOrConstSplat(Val.getOperand(0));
4747 if (C && C->getAPIntValue().isSignMask())
4748 return true;
4749 return isKnownToBeAPowerOfTwo(Val.getOperand(0), /*OrZero=*/false,
4750 Depth + 1) &&
4751 isKnownNeverZero(Val, Depth);
4752 }
4753
4754 case ISD::ROTL:
4755 case ISD::ROTR:
4756 return isKnownToBeAPowerOfTwo(Val.getOperand(0), DemandedElts, OrZero,
4757 Depth + 1);
4758 case ISD::BSWAP:
4759 case ISD::BITREVERSE:
4760 return isKnownToBeAPowerOfTwo(Val.getOperand(0), DemandedElts, OrZero,
4761 Depth + 1);
4762
4763 case ISD::SMIN:
4764 case ISD::SMAX:
4765 case ISD::UMIN:
4766 case ISD::UMAX:
4767 return isKnownToBeAPowerOfTwo(Val.getOperand(1), /*OrZero=*/false,
4768 Depth + 1) &&
4769 isKnownToBeAPowerOfTwo(Val.getOperand(0), /*OrZero=*/false,
4770 Depth + 1);
4771
4772 case ISD::SELECT:
4773 case ISD::VSELECT:
4774 return isKnownToBeAPowerOfTwo(Val.getOperand(2), DemandedElts, OrZero,
4775 Depth + 1) &&
4776 isKnownToBeAPowerOfTwo(Val.getOperand(1), DemandedElts, OrZero,
4777 Depth + 1);
4778
4779 case ISD::ZERO_EXTEND:
4780 return isKnownToBeAPowerOfTwo(Val.getOperand(0), /*OrZero=*/false,
4781 Depth + 1);
4782
4783 case ISD::VSCALE:
4784 // vscale(power-of-two) is a power-of-two
4785 return isKnownToBeAPowerOfTwo(Val.getOperand(0), /*OrZero=*/false,
4786 Depth + 1);
4787
4788 case ISD::VECTOR_SHUFFLE: {
4790 // Demanded elements with undef shuffle mask elements are unknown
4791 // - we cannot guarantee they are a power of two, so return false.
4792 APInt DemandedLHS, DemandedRHS;
4794 assert(NumElts == SVN->getMask().size() && "Unexpected vector size");
4795 if (!getShuffleDemandedElts(NumElts, SVN->getMask(), DemandedElts,
4796 DemandedLHS, DemandedRHS))
4797 return false;
4798
4799 // All demanded elements from LHS must be known power of two.
4800 if (!!DemandedLHS && !isKnownToBeAPowerOfTwo(Val.getOperand(0), DemandedLHS,
4801 OrZero, Depth + 1))
4802 return false;
4803
4804 // All demanded elements from RHS must be known power of two.
4805 if (!!DemandedRHS && !isKnownToBeAPowerOfTwo(Val.getOperand(1), DemandedRHS,
4806 OrZero, Depth + 1))
4807 return false;
4808
4809 return true;
4810 }
4811 }
4812
4813 // More could be done here, though the above checks are enough
4814 // to handle some common cases.
4815 return false;
4816}
4817
4819 if (ConstantFPSDNode *C1 = isConstOrConstSplatFP(Val, true))
4820 return C1->getValueAPF().getExactLog2Abs() >= 0;
4821
4822 if (Val.getOpcode() == ISD::UINT_TO_FP || Val.getOpcode() == ISD::SINT_TO_FP)
4823 return isKnownToBeAPowerOfTwo(Val.getOperand(0), Depth + 1);
4824
4825 return false;
4826}
4827
4829 EVT VT = Op.getValueType();
4830
4831 // Since the number of lanes in a scalable vector is unknown at compile time,
4832 // we track one bit which is implicitly broadcast to all lanes. This means
4833 // that all lanes in a scalable vector are considered demanded.
4834 APInt DemandedElts = VT.isFixedLengthVector()
4836 : APInt(1, 1);
4837 return ComputeNumSignBits(Op, DemandedElts, Depth);
4838}
4839
4840unsigned SelectionDAG::ComputeNumSignBits(SDValue Op, const APInt &DemandedElts,
4841 unsigned Depth) const {
4842 EVT VT = Op.getValueType();
4843 assert((VT.isInteger() || VT.isFloatingPoint()) && "Invalid VT!");
4844 unsigned VTBits = VT.getScalarSizeInBits();
4845 unsigned NumElts = DemandedElts.getBitWidth();
4846 unsigned Tmp, Tmp2;
4847 unsigned FirstAnswer = 1;
4848
4849 assert((!VT.isScalableVector() || NumElts == 1) &&
4850 "DemandedElts for scalable vectors must be 1 to represent all lanes");
4851
4852 if (auto *C = dyn_cast<ConstantSDNode>(Op)) {
4853 const APInt &Val = C->getAPIntValue();
4854 return Val.getNumSignBits();
4855 }
4856
4857 if (Depth >= MaxRecursionDepth)
4858 return 1; // Limit search depth.
4859
4860 if (!DemandedElts)
4861 return 1; // No demanded elts, better to assume we don't know anything.
4862
4863 unsigned Opcode = Op.getOpcode();
4864 switch (Opcode) {
4865 default: break;
4866 case ISD::AssertSext:
4867 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getSizeInBits();
4868 return VTBits-Tmp+1;
4869 case ISD::AssertZext:
4870 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getSizeInBits();
4871 return VTBits-Tmp;
4872 case ISD::FREEZE:
4873 if (isGuaranteedNotToBeUndefOrPoison(Op.getOperand(0), DemandedElts,
4874 /*PoisonOnly=*/false))
4875 return ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
4876 break;
4877 case ISD::MERGE_VALUES:
4878 return ComputeNumSignBits(Op.getOperand(Op.getResNo()), DemandedElts,
4879 Depth + 1);
4880 case ISD::SPLAT_VECTOR: {
4881 // Check if the sign bits of source go down as far as the truncated value.
4882 unsigned NumSrcBits = Op.getOperand(0).getValueSizeInBits();
4883 unsigned NumSrcSignBits = ComputeNumSignBits(Op.getOperand(0), Depth + 1);
4884 if (NumSrcSignBits > (NumSrcBits - VTBits))
4885 return NumSrcSignBits - (NumSrcBits - VTBits);
4886 break;
4887 }
4888 case ISD::BUILD_VECTOR:
4889 assert(!VT.isScalableVector());
4890 Tmp = VTBits;
4891 for (unsigned i = 0, e = Op.getNumOperands(); (i < e) && (Tmp > 1); ++i) {
4892 if (!DemandedElts[i])
4893 continue;
4894
4895 SDValue SrcOp = Op.getOperand(i);
4896 // BUILD_VECTOR can implicitly truncate sources, we handle this specially
4897 // for constant nodes to ensure we only look at the sign bits.
4899 APInt T = C->getAPIntValue().trunc(VTBits);
4900 Tmp2 = T.getNumSignBits();
4901 } else {
4902 Tmp2 = ComputeNumSignBits(SrcOp, Depth + 1);
4903
4904 if (SrcOp.getValueSizeInBits() != VTBits) {
4905 assert(SrcOp.getValueSizeInBits() > VTBits &&
4906 "Expected BUILD_VECTOR implicit truncation");
4907 unsigned ExtraBits = SrcOp.getValueSizeInBits() - VTBits;
4908 Tmp2 = (Tmp2 > ExtraBits ? Tmp2 - ExtraBits : 1);
4909 }
4910 }
4911 Tmp = std::min(Tmp, Tmp2);
4912 }
4913 return Tmp;
4914
4915 case ISD::VECTOR_COMPRESS: {
4916 SDValue Vec = Op.getOperand(0);
4917 SDValue PassThru = Op.getOperand(2);
4918 Tmp = ComputeNumSignBits(PassThru, DemandedElts, Depth + 1);
4919 if (Tmp == 1)
4920 return 1;
4921 Tmp2 = ComputeNumSignBits(Vec, Depth + 1);
4922 Tmp = std::min(Tmp, Tmp2);
4923 return Tmp;
4924 }
4925
4926 case ISD::VECTOR_SHUFFLE: {
4927 // Collect the minimum number of sign bits that are shared by every vector
4928 // element referenced by the shuffle.
4929 APInt DemandedLHS, DemandedRHS;
4931 assert(NumElts == SVN->getMask().size() && "Unexpected vector size");
4932 if (!getShuffleDemandedElts(NumElts, SVN->getMask(), DemandedElts,
4933 DemandedLHS, DemandedRHS))
4934 return 1;
4935
4936 Tmp = std::numeric_limits<unsigned>::max();
4937 if (!!DemandedLHS)
4938 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedLHS, Depth + 1);
4939 if (!!DemandedRHS) {
4940 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedRHS, Depth + 1);
4941 Tmp = std::min(Tmp, Tmp2);
4942 }
4943 // If we don't know anything, early out and try computeKnownBits fall-back.
4944 if (Tmp == 1)
4945 break;
4946 assert(Tmp <= VTBits && "Failed to determine minimum sign bits");
4947 return Tmp;
4948 }
4949
4950 case ISD::BITCAST: {
4951 if (VT.isScalableVector())
4952 break;
4953 SDValue N0 = Op.getOperand(0);
4954 EVT SrcVT = N0.getValueType();
4955 unsigned SrcBits = SrcVT.getScalarSizeInBits();
4956
4957 // Ignore bitcasts from unsupported types..
4958 if (!(SrcVT.isInteger() || SrcVT.isFloatingPoint()))
4959 break;
4960
4961 // Fast handling of 'identity' bitcasts.
4962 if (VTBits == SrcBits)
4963 return ComputeNumSignBits(N0, DemandedElts, Depth + 1);
4964
4965 bool IsLE = getDataLayout().isLittleEndian();
4966
4967 // Bitcast 'large element' scalar/vector to 'small element' vector.
4968 if ((SrcBits % VTBits) == 0) {
4969 assert(VT.isVector() && "Expected bitcast to vector");
4970
4971 unsigned Scale = SrcBits / VTBits;
4972 APInt SrcDemandedElts =
4973 APIntOps::ScaleBitMask(DemandedElts, NumElts / Scale);
4974
4975 // Fast case - sign splat can be simply split across the small elements.
4976 Tmp = ComputeNumSignBits(N0, SrcDemandedElts, Depth + 1);
4977 if (Tmp == SrcBits)
4978 return VTBits;
4979
4980 // Slow case - determine how far the sign extends into each sub-element.
4981 Tmp2 = VTBits;
4982 for (unsigned i = 0; i != NumElts; ++i)
4983 if (DemandedElts[i]) {
4984 unsigned SubOffset = i % Scale;
4985 SubOffset = (IsLE ? ((Scale - 1) - SubOffset) : SubOffset);
4986 SubOffset = SubOffset * VTBits;
4987 if (Tmp <= SubOffset)
4988 return 1;
4989 Tmp2 = std::min(Tmp2, Tmp - SubOffset);
4990 }
4991 return Tmp2;
4992 }
4993 break;
4994 }
4995
4997 // FP_TO_SINT_SAT produces a signed value that fits in the saturating VT.
4998 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getScalarSizeInBits();
4999 return VTBits - Tmp + 1;
5000 case ISD::SIGN_EXTEND:
5001 Tmp = VTBits - Op.getOperand(0).getScalarValueSizeInBits();
5002 return ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth+1) + Tmp;
5004 // Max of the input and what this extends.
5005 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getScalarSizeInBits();
5006 Tmp = VTBits-Tmp+1;
5007 Tmp2 = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth+1);
5008 return std::max(Tmp, Tmp2);
5010 if (VT.isScalableVector())
5011 break;
5012 SDValue Src = Op.getOperand(0);
5013 EVT SrcVT = Src.getValueType();
5014 APInt DemandedSrcElts = DemandedElts.zext(SrcVT.getVectorNumElements());
5015 Tmp = VTBits - SrcVT.getScalarSizeInBits();
5016 return ComputeNumSignBits(Src, DemandedSrcElts, Depth+1) + Tmp;
5017 }
5018 case ISD::SRA:
5019 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5020 // SRA X, C -> adds C sign bits.
5021 if (std::optional<unsigned> ShAmt =
5022 getValidMinimumShiftAmount(Op, DemandedElts, Depth + 1))
5023 Tmp = std::min(Tmp + *ShAmt, VTBits);
5024 return Tmp;
5025 case ISD::SHL:
5026 if (std::optional<ConstantRange> ShAmtRange =
5027 getValidShiftAmountRange(Op, DemandedElts, Depth + 1)) {
5028 unsigned MaxShAmt = ShAmtRange->getUnsignedMax().getZExtValue();
5029 unsigned MinShAmt = ShAmtRange->getUnsignedMin().getZExtValue();
5030 // Try to look through ZERO/SIGN/ANY_EXTEND. If all extended bits are
5031 // shifted out, then we can compute the number of sign bits for the
5032 // operand being extended. A future improvement could be to pass along the
5033 // "shifted left by" information in the recursive calls to
5034 // ComputeKnownSignBits. Allowing us to handle this more generically.
5035 if (ISD::isExtOpcode(Op.getOperand(0).getOpcode())) {
5036 SDValue Ext = Op.getOperand(0);
5037 EVT ExtVT = Ext.getValueType();
5038 SDValue Extendee = Ext.getOperand(0);
5039 EVT ExtendeeVT = Extendee.getValueType();
5040 unsigned SizeDifference =
5041 ExtVT.getScalarSizeInBits() - ExtendeeVT.getScalarSizeInBits();
5042 if (SizeDifference <= MinShAmt) {
5043 Tmp = SizeDifference +
5044 ComputeNumSignBits(Extendee, DemandedElts, Depth + 1);
5045 if (MaxShAmt < Tmp)
5046 return Tmp - MaxShAmt;
5047 }
5048 }
5049 // shl destroys sign bits, ensure it doesn't shift out all sign bits.
5050 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5051 if (MaxShAmt < Tmp)
5052 return Tmp - MaxShAmt;
5053 }
5054 break;
5055 case ISD::AND:
5056 case ISD::OR:
5057 case ISD::XOR: // NOT is handled here.
5058 // Logical binary ops preserve the number of sign bits at the worst.
5059 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth+1);
5060 if (Tmp != 1) {
5061 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth+1);
5062 FirstAnswer = std::min(Tmp, Tmp2);
5063 // We computed what we know about the sign bits as our first
5064 // answer. Now proceed to the generic code that uses
5065 // computeKnownBits, and pick whichever answer is better.
5066 }
5067 break;
5068
5069 case ISD::SELECT:
5070 case ISD::VSELECT:
5071 Tmp = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth+1);
5072 if (Tmp == 1) return 1; // Early out.
5073 Tmp2 = ComputeNumSignBits(Op.getOperand(2), DemandedElts, Depth+1);
5074 return std::min(Tmp, Tmp2);
5075 case ISD::SELECT_CC:
5076 Tmp = ComputeNumSignBits(Op.getOperand(2), DemandedElts, Depth+1);
5077 if (Tmp == 1) return 1; // Early out.
5078 Tmp2 = ComputeNumSignBits(Op.getOperand(3), DemandedElts, Depth+1);
5079 return std::min(Tmp, Tmp2);
5080
5081 case ISD::SMIN:
5082 case ISD::SMAX: {
5083 // If we have a clamp pattern, we know that the number of sign bits will be
5084 // the minimum of the clamp min/max range.
5085 bool IsMax = (Opcode == ISD::SMAX);
5086 ConstantSDNode *CstLow = nullptr, *CstHigh = nullptr;
5087 if ((CstLow = isConstOrConstSplat(Op.getOperand(1), DemandedElts)))
5088 if (Op.getOperand(0).getOpcode() == (IsMax ? ISD::SMIN : ISD::SMAX))
5089 CstHigh =
5090 isConstOrConstSplat(Op.getOperand(0).getOperand(1), DemandedElts);
5091 if (CstLow && CstHigh) {
5092 if (!IsMax)
5093 std::swap(CstLow, CstHigh);
5094 if (CstLow->getAPIntValue().sle(CstHigh->getAPIntValue())) {
5095 Tmp = CstLow->getAPIntValue().getNumSignBits();
5096 Tmp2 = CstHigh->getAPIntValue().getNumSignBits();
5097 return std::min(Tmp, Tmp2);
5098 }
5099 }
5100
5101 // Fallback - just get the minimum number of sign bits of the operands.
5102 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5103 if (Tmp == 1)
5104 return 1; // Early out.
5105 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5106 return std::min(Tmp, Tmp2);
5107 }
5108 case ISD::UMIN:
5109 case ISD::UMAX:
5110 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5111 if (Tmp == 1)
5112 return 1; // Early out.
5113 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5114 return std::min(Tmp, Tmp2);
5115 case ISD::SSUBO_CARRY:
5116 case ISD::USUBO_CARRY:
5117 // sub_carry(x,x,c) -> 0/-1 (sext carry)
5118 if (Op.getResNo() == 0 && Op.getOperand(0) == Op.getOperand(1))
5119 return VTBits;
5120 [[fallthrough]];
5121 case ISD::SADDO:
5122 case ISD::UADDO:
5123 case ISD::SADDO_CARRY:
5124 case ISD::UADDO_CARRY:
5125 case ISD::SSUBO:
5126 case ISD::USUBO:
5127 case ISD::SMULO:
5128 case ISD::UMULO:
5129 if (Op.getResNo() != 1)
5130 break;
5131 // The boolean result conforms to getBooleanContents. Fall through.
5132 // If setcc returns 0/-1, all bits are sign bits.
5133 // We know that we have an integer-based boolean since these operations
5134 // are only available for integer.
5135 if (TLI->getBooleanContents(VT.isVector(), false) ==
5137 return VTBits;
5138 break;
5139 case ISD::SETCC:
5140 case ISD::SETCCCARRY:
5141 case ISD::STRICT_FSETCC:
5142 case ISD::STRICT_FSETCCS: {
5143 unsigned OpNo = Op->isStrictFPOpcode() ? 1 : 0;
5144 // If setcc returns 0/-1, all bits are sign bits.
5145 if (TLI->getBooleanContents(Op.getOperand(OpNo).getValueType()) ==
5147 return VTBits;
5148 break;
5149 }
5150 case ISD::ROTL:
5151 case ISD::ROTR:
5152 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5153
5154 // If we're rotating an 0/-1 value, then it stays an 0/-1 value.
5155 if (Tmp == VTBits)
5156 return VTBits;
5157
5158 if (ConstantSDNode *C =
5159 isConstOrConstSplat(Op.getOperand(1), DemandedElts)) {
5160 unsigned RotAmt = C->getAPIntValue().urem(VTBits);
5161
5162 // Handle rotate right by N like a rotate left by 32-N.
5163 if (Opcode == ISD::ROTR)
5164 RotAmt = (VTBits - RotAmt) % VTBits;
5165
5166 // If we aren't rotating out all of the known-in sign bits, return the
5167 // number that are left. This handles rotl(sext(x), 1) for example.
5168 if (Tmp > (RotAmt + 1)) return (Tmp - RotAmt);
5169 }
5170 break;
5171 case ISD::ADD:
5172 case ISD::ADDC:
5173 // TODO: Move Operand 1 check before Operand 0 check
5174 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5175 if (Tmp == 1) return 1; // Early out.
5176
5177 // Special case decrementing a value (ADD X, -1):
5178 if (ConstantSDNode *CRHS =
5179 isConstOrConstSplat(Op.getOperand(1), DemandedElts))
5180 if (CRHS->isAllOnes()) {
5181 KnownBits Known =
5182 computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
5183
5184 // If the input is known to be 0 or 1, the output is 0/-1, which is all
5185 // sign bits set.
5186 if ((Known.Zero | 1).isAllOnes())
5187 return VTBits;
5188
5189 // If we are subtracting one from a positive number, there is no carry
5190 // out of the result.
5191 if (Known.isNonNegative())
5192 return Tmp;
5193 }
5194
5195 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5196 if (Tmp2 == 1) return 1; // Early out.
5197
5198 // Add can have at most one carry bit. Thus we know that the output
5199 // is, at worst, one more bit than the inputs.
5200 return std::min(Tmp, Tmp2) - 1;
5201 case ISD::SUB:
5202 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5203 if (Tmp2 == 1) return 1; // Early out.
5204
5205 // Handle NEG.
5206 if (ConstantSDNode *CLHS =
5207 isConstOrConstSplat(Op.getOperand(0), DemandedElts))
5208 if (CLHS->isZero()) {
5209 KnownBits Known =
5210 computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
5211 // If the input is known to be 0 or 1, the output is 0/-1, which is all
5212 // sign bits set.
5213 if ((Known.Zero | 1).isAllOnes())
5214 return VTBits;
5215
5216 // If the input is known to be positive (the sign bit is known clear),
5217 // the output of the NEG has the same number of sign bits as the input.
5218 if (Known.isNonNegative())
5219 return Tmp2;
5220
5221 // Otherwise, we treat this like a SUB.
5222 }
5223
5224 // Sub can have at most one carry bit. Thus we know that the output
5225 // is, at worst, one more bit than the inputs.
5226 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5227 if (Tmp == 1) return 1; // Early out.
5228 return std::min(Tmp, Tmp2) - 1;
5229 case ISD::MUL: {
5230 // The output of the Mul can be at most twice the valid bits in the inputs.
5231 unsigned SignBitsOp0 = ComputeNumSignBits(Op.getOperand(0), Depth + 1);
5232 if (SignBitsOp0 == 1)
5233 break;
5234 unsigned SignBitsOp1 = ComputeNumSignBits(Op.getOperand(1), Depth + 1);
5235 if (SignBitsOp1 == 1)
5236 break;
5237 unsigned OutValidBits =
5238 (VTBits - SignBitsOp0 + 1) + (VTBits - SignBitsOp1 + 1);
5239 return OutValidBits > VTBits ? 1 : VTBits - OutValidBits + 1;
5240 }
5241 case ISD::AVGCEILS:
5242 case ISD::AVGFLOORS:
5243 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5244 if (Tmp == 1)
5245 return 1; // Early out.
5246 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5247 return std::min(Tmp, Tmp2);
5248 case ISD::SREM:
5249 // The sign bit is the LHS's sign bit, except when the result of the
5250 // remainder is zero. The magnitude of the result should be less than or
5251 // equal to the magnitude of the LHS. Therefore, the result should have
5252 // at least as many sign bits as the left hand side.
5253 return ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5254 case ISD::TRUNCATE: {
5255 // Check if the sign bits of source go down as far as the truncated value.
5256 unsigned NumSrcBits = Op.getOperand(0).getScalarValueSizeInBits();
5257 unsigned NumSrcSignBits = ComputeNumSignBits(Op.getOperand(0), Depth + 1);
5258 if (NumSrcSignBits > (NumSrcBits - VTBits))
5259 return NumSrcSignBits - (NumSrcBits - VTBits);
5260 break;
5261 }
5262 case ISD::EXTRACT_ELEMENT: {
5263 if (VT.isScalableVector())
5264 break;
5265 const int KnownSign = ComputeNumSignBits(Op.getOperand(0), Depth+1);
5266 const int BitWidth = Op.getValueSizeInBits();
5267 const int Items = Op.getOperand(0).getValueSizeInBits() / BitWidth;
5268
5269 // Get reverse index (starting from 1), Op1 value indexes elements from
5270 // little end. Sign starts at big end.
5271 const int rIndex = Items - 1 - Op.getConstantOperandVal(1);
5272
5273 // If the sign portion ends in our element the subtraction gives correct
5274 // result. Otherwise it gives either negative or > bitwidth result
5275 return std::clamp(KnownSign - rIndex * BitWidth, 1, BitWidth);
5276 }
5278 if (VT.isScalableVector())
5279 break;
5280 // If we know the element index, split the demand between the
5281 // source vector and the inserted element, otherwise assume we need
5282 // the original demanded vector elements and the value.
5283 SDValue InVec = Op.getOperand(0);
5284 SDValue InVal = Op.getOperand(1);
5285 SDValue EltNo = Op.getOperand(2);
5286 bool DemandedVal = true;
5287 APInt DemandedVecElts = DemandedElts;
5288 auto *CEltNo = dyn_cast<ConstantSDNode>(EltNo);
5289 if (CEltNo && CEltNo->getAPIntValue().ult(NumElts)) {
5290 unsigned EltIdx = CEltNo->getZExtValue();
5291 DemandedVal = !!DemandedElts[EltIdx];
5292 DemandedVecElts.clearBit(EltIdx);
5293 }
5294 Tmp = std::numeric_limits<unsigned>::max();
5295 if (DemandedVal) {
5296 // TODO - handle implicit truncation of inserted elements.
5297 if (InVal.getScalarValueSizeInBits() != VTBits)
5298 break;
5299 Tmp2 = ComputeNumSignBits(InVal, Depth + 1);
5300 Tmp = std::min(Tmp, Tmp2);
5301 }
5302 if (!!DemandedVecElts) {
5303 Tmp2 = ComputeNumSignBits(InVec, DemandedVecElts, Depth + 1);
5304 Tmp = std::min(Tmp, Tmp2);
5305 }
5306 assert(Tmp <= VTBits && "Failed to determine minimum sign bits");
5307 return Tmp;
5308 }
5310 SDValue InVec = Op.getOperand(0);
5311 SDValue EltNo = Op.getOperand(1);
5312 EVT VecVT = InVec.getValueType();
5313 // ComputeNumSignBits not yet implemented for scalable vectors.
5314 if (VecVT.isScalableVector())
5315 break;
5316 const unsigned BitWidth = Op.getValueSizeInBits();
5317 const unsigned EltBitWidth = Op.getOperand(0).getScalarValueSizeInBits();
5318 const unsigned NumSrcElts = VecVT.getVectorNumElements();
5319
5320 // If BitWidth > EltBitWidth the value is anyext:ed, and we do not know
5321 // anything about sign bits. But if the sizes match we can derive knowledge
5322 // about sign bits from the vector operand.
5323 if (BitWidth != EltBitWidth)
5324 break;
5325
5326 // If we know the element index, just demand that vector element, else for
5327 // an unknown element index, ignore DemandedElts and demand them all.
5328 APInt DemandedSrcElts = APInt::getAllOnes(NumSrcElts);
5329 auto *ConstEltNo = dyn_cast<ConstantSDNode>(EltNo);
5330 if (ConstEltNo && ConstEltNo->getAPIntValue().ult(NumSrcElts))
5331 DemandedSrcElts =
5332 APInt::getOneBitSet(NumSrcElts, ConstEltNo->getZExtValue());
5333
5334 return ComputeNumSignBits(InVec, DemandedSrcElts, Depth + 1);
5335 }
5337 // Offset the demanded elts by the subvector index.
5338 SDValue Src = Op.getOperand(0);
5339
5340 APInt DemandedSrcElts;
5341 if (Src.getValueType().isScalableVector())
5342 DemandedSrcElts = APInt(1, 1);
5343 else {
5344 uint64_t Idx = Op.getConstantOperandVal(1);
5345 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
5346 DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
5347 }
5348 return ComputeNumSignBits(Src, DemandedSrcElts, Depth + 1);
5349 }
5350 case ISD::CONCAT_VECTORS: {
5351 if (VT.isScalableVector())
5352 break;
5353 // Determine the minimum number of sign bits across all demanded
5354 // elts of the input vectors. Early out if the result is already 1.
5355 Tmp = std::numeric_limits<unsigned>::max();
5356 EVT SubVectorVT = Op.getOperand(0).getValueType();
5357 unsigned NumSubVectorElts = SubVectorVT.getVectorNumElements();
5358 unsigned NumSubVectors = Op.getNumOperands();
5359 for (unsigned i = 0; (i < NumSubVectors) && (Tmp > 1); ++i) {
5360 APInt DemandedSub =
5361 DemandedElts.extractBits(NumSubVectorElts, i * NumSubVectorElts);
5362 if (!DemandedSub)
5363 continue;
5364 Tmp2 = ComputeNumSignBits(Op.getOperand(i), DemandedSub, Depth + 1);
5365 Tmp = std::min(Tmp, Tmp2);
5366 }
5367 assert(Tmp <= VTBits && "Failed to determine minimum sign bits");
5368 return Tmp;
5369 }
5370 case ISD::INSERT_SUBVECTOR: {
5371 if (VT.isScalableVector())
5372 break;
5373 // Demand any elements from the subvector and the remainder from the src its
5374 // inserted into.
5375 SDValue Src = Op.getOperand(0);
5376 SDValue Sub = Op.getOperand(1);
5377 uint64_t Idx = Op.getConstantOperandVal(2);
5378 unsigned NumSubElts = Sub.getValueType().getVectorNumElements();
5379 APInt DemandedSubElts = DemandedElts.extractBits(NumSubElts, Idx);
5380 APInt DemandedSrcElts = DemandedElts;
5381 DemandedSrcElts.clearBits(Idx, Idx + NumSubElts);
5382
5383 Tmp = std::numeric_limits<unsigned>::max();
5384 if (!!DemandedSubElts) {
5385 Tmp = ComputeNumSignBits(Sub, DemandedSubElts, Depth + 1);
5386 if (Tmp == 1)
5387 return 1; // early-out
5388 }
5389 if (!!DemandedSrcElts) {
5390 Tmp2 = ComputeNumSignBits(Src, DemandedSrcElts, Depth + 1);
5391 Tmp = std::min(Tmp, Tmp2);
5392 }
5393 assert(Tmp <= VTBits && "Failed to determine minimum sign bits");
5394 return Tmp;
5395 }
5396 case ISD::LOAD: {
5397 // If we are looking at the loaded value of the SDNode.
5398 if (Op.getResNo() != 0)
5399 break;
5400
5402 if (const MDNode *Ranges = LD->getRanges()) {
5403 if (DemandedElts != 1)
5404 break;
5405
5407 if (VTBits > CR.getBitWidth()) {
5408 switch (LD->getExtensionType()) {
5409 case ISD::SEXTLOAD:
5410 CR = CR.signExtend(VTBits);
5411 break;
5412 case ISD::ZEXTLOAD:
5413 CR = CR.zeroExtend(VTBits);
5414 break;
5415 default:
5416 break;
5417 }
5418 }
5419
5420 if (VTBits != CR.getBitWidth())
5421 break;
5422 return std::min(CR.getSignedMin().getNumSignBits(),
5424 }
5425
5426 unsigned ExtType = LD->getExtensionType();
5427 switch (ExtType) {
5428 default:
5429 break;
5430 case ISD::SEXTLOAD: // e.g. i16->i32 = '17' bits known.
5431 Tmp = LD->getMemoryVT().getScalarSizeInBits();
5432 return VTBits - Tmp + 1;
5433 case ISD::ZEXTLOAD: // e.g. i16->i32 = '16' bits known.
5434 Tmp = LD->getMemoryVT().getScalarSizeInBits();
5435 return VTBits - Tmp;
5436 case ISD::NON_EXTLOAD:
5437 if (const Constant *Cst = TLI->getTargetConstantFromLoad(LD)) {
5438 // We only need to handle vectors - computeKnownBits should handle
5439 // scalar cases.
5440 Type *CstTy = Cst->getType();
5441 if (CstTy->isVectorTy() && !VT.isScalableVector() &&
5442 (NumElts * VTBits) == CstTy->getPrimitiveSizeInBits() &&
5443 VTBits == CstTy->getScalarSizeInBits()) {
5444 Tmp = VTBits;
5445 for (unsigned i = 0; i != NumElts; ++i) {
5446 if (!DemandedElts[i])
5447 continue;
5448 if (Constant *Elt = Cst->getAggregateElement(i)) {
5449 if (auto *CInt = dyn_cast<ConstantInt>(Elt)) {
5450 const APInt &Value = CInt->getValue();
5451 Tmp = std::min(Tmp, Value.getNumSignBits());
5452 continue;
5453 }
5454 if (auto *CFP = dyn_cast<ConstantFP>(Elt)) {
5455 APInt Value = CFP->getValueAPF().bitcastToAPInt();
5456 Tmp = std::min(Tmp, Value.getNumSignBits());
5457 continue;
5458 }
5459 }
5460 // Unknown type. Conservatively assume no bits match sign bit.
5461 return 1;
5462 }
5463 return Tmp;
5464 }
5465 }
5466 break;
5467 }
5468
5469 break;
5470 }
5473 case ISD::ATOMIC_SWAP:
5485 case ISD::ATOMIC_LOAD: {
5486 auto *AT = cast<AtomicSDNode>(Op);
5487 // If we are looking at the loaded value.
5488 if (Op.getResNo() == 0) {
5489 Tmp = AT->getMemoryVT().getScalarSizeInBits();
5490 if (Tmp == VTBits)
5491 return 1; // early-out
5492
5493 // For atomic_load, prefer to use the extension type.
5494 if (Op->getOpcode() == ISD::ATOMIC_LOAD) {
5495 switch (AT->getExtensionType()) {
5496 default:
5497 break;
5498 case ISD::SEXTLOAD:
5499 return VTBits - Tmp + 1;
5500 case ISD::ZEXTLOAD:
5501 return VTBits - Tmp;
5502 }
5503 }
5504
5505 if (TLI->getExtendForAtomicOps() == ISD::SIGN_EXTEND)
5506 return VTBits - Tmp + 1;
5507 if (TLI->getExtendForAtomicOps() == ISD::ZERO_EXTEND)
5508 return VTBits - Tmp;
5509 }
5510 break;
5511 }
5512 }
5513
5514 // Allow the target to implement this method for its nodes.
5515 if (Opcode >= ISD::BUILTIN_OP_END ||
5516 Opcode == ISD::INTRINSIC_WO_CHAIN ||
5517 Opcode == ISD::INTRINSIC_W_CHAIN ||
5518 Opcode == ISD::INTRINSIC_VOID) {
5519 // TODO: This can probably be removed once target code is audited. This
5520 // is here purely to reduce patch size and review complexity.
5521 if (!VT.isScalableVector()) {
5522 unsigned NumBits =
5523 TLI->ComputeNumSignBitsForTargetNode(Op, DemandedElts, *this, Depth);
5524 if (NumBits > 1)
5525 FirstAnswer = std::max(FirstAnswer, NumBits);
5526 }
5527 }
5528
5529 // Finally, if we can prove that the top bits of the result are 0's or 1's,
5530 // use this information.
5531 KnownBits Known = computeKnownBits(Op, DemandedElts, Depth);
5532 return std::max(FirstAnswer, Known.countMinSignBits());
5533}
5534
5536 unsigned Depth) const {
5537 unsigned SignBits = ComputeNumSignBits(Op, Depth);
5538 return Op.getScalarValueSizeInBits() - SignBits + 1;
5539}
5540
5542 const APInt &DemandedElts,
5543 unsigned Depth) const {
5544 unsigned SignBits = ComputeNumSignBits(Op, DemandedElts, Depth);
5545 return Op.getScalarValueSizeInBits() - SignBits + 1;
5546}
5547
5549 unsigned Depth) const {
5550 // Early out for FREEZE.
5551 if (Op.getOpcode() == ISD::FREEZE)
5552 return true;
5553
5554 EVT VT = Op.getValueType();
5555 APInt DemandedElts = VT.isFixedLengthVector()
5557 : APInt(1, 1);
5558 return isGuaranteedNotToBeUndefOrPoison(Op, DemandedElts, PoisonOnly, Depth);
5559}
5560
5562 const APInt &DemandedElts,
5563 bool PoisonOnly,
5564 unsigned Depth) const {
5565 unsigned Opcode = Op.getOpcode();
5566
5567 // Early out for FREEZE.
5568 if (Opcode == ISD::FREEZE)
5569 return true;
5570
5571 if (Depth >= MaxRecursionDepth)
5572 return false; // Limit search depth.
5573
5574 if (isIntOrFPConstant(Op))
5575 return true;
5576
5577 switch (Opcode) {
5578 case ISD::CONDCODE:
5579 case ISD::VALUETYPE:
5580 case ISD::FrameIndex:
5582 case ISD::CopyFromReg:
5583 return true;
5584
5585 case ISD::POISON:
5586 return false;
5587
5588 case ISD::UNDEF:
5589 return PoisonOnly;
5590
5591 case ISD::BUILD_VECTOR:
5592 // NOTE: BUILD_VECTOR has implicit truncation of wider scalar elements -
5593 // this shouldn't affect the result.
5594 for (unsigned i = 0, e = Op.getNumOperands(); i < e; ++i) {
5595 if (!DemandedElts[i])
5596 continue;
5598 Depth + 1))
5599 return false;
5600 }
5601 return true;
5602
5604 SDValue Src = Op.getOperand(0);
5605 if (Src.getValueType().isScalableVector())
5606 break;
5607 uint64_t Idx = Op.getConstantOperandVal(1);
5608 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
5609 APInt DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
5610 return isGuaranteedNotToBeUndefOrPoison(Src, DemandedSrcElts, PoisonOnly,
5611 Depth + 1);
5612 }
5613
5614 case ISD::INSERT_SUBVECTOR: {
5615 if (Op.getValueType().isScalableVector())
5616 break;
5617 SDValue Src = Op.getOperand(0);
5618 SDValue Sub = Op.getOperand(1);
5619 uint64_t Idx = Op.getConstantOperandVal(2);
5620 unsigned NumSubElts = Sub.getValueType().getVectorNumElements();
5621 APInt DemandedSubElts = DemandedElts.extractBits(NumSubElts, Idx);
5622 APInt DemandedSrcElts = DemandedElts;
5623 DemandedSrcElts.clearBits(Idx, Idx + NumSubElts);
5624
5625 if (!!DemandedSubElts && !isGuaranteedNotToBeUndefOrPoison(
5626 Sub, DemandedSubElts, PoisonOnly, Depth + 1))
5627 return false;
5628 if (!!DemandedSrcElts && !isGuaranteedNotToBeUndefOrPoison(
5629 Src, DemandedSrcElts, PoisonOnly, Depth + 1))
5630 return false;
5631 return true;
5632 }
5633
5635 SDValue Src = Op.getOperand(0);
5636 auto *IndexC = dyn_cast<ConstantSDNode>(Op.getOperand(1));
5637 EVT SrcVT = Src.getValueType();
5638 if (SrcVT.isFixedLengthVector() && IndexC &&
5639 IndexC->getAPIntValue().ult(SrcVT.getVectorNumElements())) {
5640 APInt DemandedSrcElts = APInt::getOneBitSet(SrcVT.getVectorNumElements(),
5641 IndexC->getZExtValue());
5642 return isGuaranteedNotToBeUndefOrPoison(Src, DemandedSrcElts, PoisonOnly,
5643 Depth + 1);
5644 }
5645 break;
5646 }
5647
5649 SDValue InVec = Op.getOperand(0);
5650 SDValue InVal = Op.getOperand(1);
5651 SDValue EltNo = Op.getOperand(2);
5652 EVT VT = InVec.getValueType();
5653 auto *IndexC = dyn_cast<ConstantSDNode>(EltNo);
5654 if (IndexC && VT.isFixedLengthVector() &&
5655 IndexC->getAPIntValue().ult(VT.getVectorNumElements())) {
5656 if (DemandedElts[IndexC->getZExtValue()] &&
5658 return false;
5659 APInt InVecDemandedElts = DemandedElts;
5660 InVecDemandedElts.clearBit(IndexC->getZExtValue());
5661 if (!!InVecDemandedElts &&
5663 peekThroughInsertVectorElt(InVec, InVecDemandedElts),
5664 InVecDemandedElts, PoisonOnly, Depth + 1))
5665 return false;
5666 return true;
5667 }
5668 break;
5669 }
5670
5672 // Check upper (known undef) elements.
5673 if (DemandedElts.ugt(1) && !PoisonOnly)
5674 return false;
5675 // Check element zero.
5676 if (DemandedElts[0] && !isGuaranteedNotToBeUndefOrPoison(
5677 Op.getOperand(0), PoisonOnly, Depth + 1))
5678 return false;
5679 return true;
5680
5681 case ISD::SPLAT_VECTOR:
5682 return isGuaranteedNotToBeUndefOrPoison(Op.getOperand(0), PoisonOnly,
5683 Depth + 1);
5684
5685 case ISD::VECTOR_SHUFFLE: {
5686 APInt DemandedLHS, DemandedRHS;
5687 auto *SVN = cast<ShuffleVectorSDNode>(Op);
5688 if (!getShuffleDemandedElts(DemandedElts.getBitWidth(), SVN->getMask(),
5689 DemandedElts, DemandedLHS, DemandedRHS,
5690 /*AllowUndefElts=*/false))
5691 return false;
5692 if (!DemandedLHS.isZero() &&
5693 !isGuaranteedNotToBeUndefOrPoison(Op.getOperand(0), DemandedLHS,
5694 PoisonOnly, Depth + 1))
5695 return false;
5696 if (!DemandedRHS.isZero() &&
5697 !isGuaranteedNotToBeUndefOrPoison(Op.getOperand(1), DemandedRHS,
5698 PoisonOnly, Depth + 1))
5699 return false;
5700 return true;
5701 }
5702
5703 case ISD::SHL:
5704 case ISD::SRL:
5705 case ISD::SRA:
5706 // Shift amount operand is checked by canCreateUndefOrPoison. So it is
5707 // enough to check operand 0 if Op can't create undef/poison.
5708 return !canCreateUndefOrPoison(Op, DemandedElts, PoisonOnly,
5709 /*ConsiderFlags*/ true, Depth) &&
5710 isGuaranteedNotToBeUndefOrPoison(Op.getOperand(0), DemandedElts,
5711 PoisonOnly, Depth + 1);
5712
5713 case ISD::BSWAP:
5714 case ISD::CTPOP:
5715 case ISD::BITREVERSE:
5716 case ISD::AND:
5717 case ISD::OR:
5718 case ISD::XOR:
5719 case ISD::ADD:
5720 case ISD::SUB:
5721 case ISD::MUL:
5722 case ISD::SADDSAT:
5723 case ISD::UADDSAT:
5724 case ISD::SSUBSAT:
5725 case ISD::USUBSAT:
5726 case ISD::SSHLSAT:
5727 case ISD::USHLSAT:
5728 case ISD::SMIN:
5729 case ISD::SMAX:
5730 case ISD::UMIN:
5731 case ISD::UMAX:
5732 case ISD::ZERO_EXTEND:
5733 case ISD::SIGN_EXTEND:
5734 case ISD::ANY_EXTEND:
5735 case ISD::TRUNCATE:
5736 case ISD::VSELECT: {
5737 // If Op can't create undef/poison and none of its operands are undef/poison
5738 // then Op is never undef/poison. A difference from the more common check
5739 // below, outside the switch, is that we handle elementwise operations for
5740 // which the DemandedElts mask is valid for all operands here.
5741 return !canCreateUndefOrPoison(Op, DemandedElts, PoisonOnly,
5742 /*ConsiderFlags*/ true, Depth) &&
5743 all_of(Op->ops(), [&](SDValue V) {
5744 return isGuaranteedNotToBeUndefOrPoison(V, DemandedElts,
5745 PoisonOnly, Depth + 1);
5746 });
5747 }
5748
5749 // TODO: Search for noundef attributes from library functions.
5750
5751 // TODO: Pointers dereferenced by ISD::LOAD/STORE ops are noundef.
5752
5753 default:
5754 // Allow the target to implement this method for its nodes.
5755 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::INTRINSIC_WO_CHAIN ||
5756 Opcode == ISD::INTRINSIC_W_CHAIN || Opcode == ISD::INTRINSIC_VOID)
5757 return TLI->isGuaranteedNotToBeUndefOrPoisonForTargetNode(
5758 Op, DemandedElts, *this, PoisonOnly, Depth);
5759 break;
5760 }
5761
5762 // If Op can't create undef/poison and none of its operands are undef/poison
5763 // then Op is never undef/poison.
5764 // NOTE: TargetNodes can handle this in themselves in
5765 // isGuaranteedNotToBeUndefOrPoisonForTargetNode or let
5766 // TargetLowering::isGuaranteedNotToBeUndefOrPoisonForTargetNode handle it.
5767 return !canCreateUndefOrPoison(Op, PoisonOnly, /*ConsiderFlags*/ true,
5768 Depth) &&
5769 all_of(Op->ops(), [&](SDValue V) {
5770 return isGuaranteedNotToBeUndefOrPoison(V, PoisonOnly, Depth + 1);
5771 });
5772}
5773
5775 bool ConsiderFlags,
5776 unsigned Depth) const {
5777 EVT VT = Op.getValueType();
5778 APInt DemandedElts = VT.isFixedLengthVector()
5780 : APInt(1, 1);
5781 return canCreateUndefOrPoison(Op, DemandedElts, PoisonOnly, ConsiderFlags,
5782 Depth);
5783}
5784
5786 bool PoisonOnly, bool ConsiderFlags,
5787 unsigned Depth) const {
5788 if (ConsiderFlags && Op->hasPoisonGeneratingFlags())
5789 return true;
5790
5791 unsigned Opcode = Op.getOpcode();
5792 switch (Opcode) {
5793 case ISD::AssertSext:
5794 case ISD::AssertZext:
5795 case ISD::AssertAlign:
5797 // Assertion nodes can create poison if the assertion fails.
5798 return true;
5799
5800 case ISD::FREEZE:
5804 case ISD::SADDSAT:
5805 case ISD::UADDSAT:
5806 case ISD::SSUBSAT:
5807 case ISD::USUBSAT:
5808 case ISD::MULHU:
5809 case ISD::MULHS:
5810 case ISD::AVGFLOORS:
5811 case ISD::AVGFLOORU:
5812 case ISD::AVGCEILS:
5813 case ISD::AVGCEILU:
5814 case ISD::ABDU:
5815 case ISD::ABDS:
5816 case ISD::SMIN:
5817 case ISD::SMAX:
5818 case ISD::SCMP:
5819 case ISD::UMIN:
5820 case ISD::UMAX:
5821 case ISD::UCMP:
5822 case ISD::AND:
5823 case ISD::XOR:
5824 case ISD::ROTL:
5825 case ISD::ROTR:
5826 case ISD::FSHL:
5827 case ISD::FSHR:
5828 case ISD::BSWAP:
5829 case ISD::CTTZ:
5830 case ISD::CTLZ:
5831 case ISD::CTLS:
5832 case ISD::CTPOP:
5833 case ISD::BITREVERSE:
5834 case ISD::PARITY:
5835 case ISD::SIGN_EXTEND:
5836 case ISD::TRUNCATE:
5840 case ISD::BITCAST:
5841 case ISD::BUILD_VECTOR:
5842 case ISD::BUILD_PAIR:
5843 case ISD::SPLAT_VECTOR:
5844 case ISD::FABS:
5845 return false;
5846
5847 case ISD::ABS:
5848 // ISD::ABS defines abs(INT_MIN) -> INT_MIN and never generates poison.
5849 // Different to Intrinsic::abs.
5850 return false;
5851
5852 case ISD::ADDC:
5853 case ISD::SUBC:
5854 case ISD::ADDE:
5855 case ISD::SUBE:
5856 case ISD::SADDO:
5857 case ISD::SSUBO:
5858 case ISD::SMULO:
5859 case ISD::SADDO_CARRY:
5860 case ISD::SSUBO_CARRY:
5861 case ISD::UADDO:
5862 case ISD::USUBO:
5863 case ISD::UMULO:
5864 case ISD::UADDO_CARRY:
5865 case ISD::USUBO_CARRY:
5866 // No poison on result or overflow flags.
5867 return false;
5868
5869 case ISD::SELECT_CC:
5870 case ISD::SETCC: {
5871 // Integer setcc cannot create undef or poison.
5872 if (Op.getOperand(0).getValueType().isInteger())
5873 return false;
5874
5875 // FP compares are more complicated. They can create poison for nan/infinity
5876 // based on options and flags. The options and flags also cause special
5877 // nonan condition codes to be used. Those condition codes may be preserved
5878 // even if the nonan flag is dropped somewhere.
5879 unsigned CCOp = Opcode == ISD::SETCC ? 2 : 4;
5880 ISD::CondCode CCCode = cast<CondCodeSDNode>(Op.getOperand(CCOp))->get();
5881 return (unsigned)CCCode & 0x10U;
5882 }
5883
5884 case ISD::OR:
5885 case ISD::ZERO_EXTEND:
5886 case ISD::SELECT:
5887 case ISD::VSELECT:
5888 case ISD::ADD:
5889 case ISD::SUB:
5890 case ISD::MUL:
5891 case ISD::FNEG:
5892 case ISD::FADD:
5893 case ISD::FSUB:
5894 case ISD::FMUL:
5895 case ISD::FDIV:
5896 case ISD::FREM:
5897 case ISD::FCOPYSIGN:
5898 case ISD::FMA:
5899 case ISD::FMAD:
5900 case ISD::FMULADD:
5901 case ISD::FP_EXTEND:
5907 // No poison except from flags (which is handled above)
5908 return false;
5909
5910 case ISD::SHL:
5911 case ISD::SRL:
5912 case ISD::SRA:
5913 // If the max shift amount isn't in range, then the shift can
5914 // create poison.
5915 return !getValidMaximumShiftAmount(Op, DemandedElts, Depth + 1);
5916
5919 // If the amount is zero then the result will be poison.
5920 // TODO: Add isKnownNeverZero DemandedElts handling.
5921 return !isKnownNeverZero(Op.getOperand(0), Depth + 1);
5922
5924 // Check if we demand any upper (undef) elements.
5925 return !PoisonOnly && DemandedElts.ugt(1);
5926
5929 // Ensure that the element index is in bounds.
5930 EVT VecVT = Op.getOperand(0).getValueType();
5931 SDValue Idx = Op.getOperand(Opcode == ISD::INSERT_VECTOR_ELT ? 2 : 1);
5932 KnownBits KnownIdx = computeKnownBits(Idx, Depth + 1);
5933 return KnownIdx.getMaxValue().uge(VecVT.getVectorMinNumElements());
5934 }
5935
5936 case ISD::VECTOR_SHUFFLE: {
5937 // Check for any demanded shuffle element that is undef.
5938 auto *SVN = cast<ShuffleVectorSDNode>(Op);
5939 for (auto [Idx, Elt] : enumerate(SVN->getMask()))
5940 if (Elt < 0 && DemandedElts[Idx])
5941 return true;
5942 return false;
5943 }
5944
5946 return false;
5947
5948 default:
5949 // Allow the target to implement this method for its nodes.
5950 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::INTRINSIC_WO_CHAIN ||
5951 Opcode == ISD::INTRINSIC_W_CHAIN || Opcode == ISD::INTRINSIC_VOID)
5952 return TLI->canCreateUndefOrPoisonForTargetNode(
5953 Op, DemandedElts, *this, PoisonOnly, ConsiderFlags, Depth);
5954 break;
5955 }
5956
5957 // Be conservative and return true.
5958 return true;
5959}
5960
5961bool SelectionDAG::isADDLike(SDValue Op, bool NoWrap) const {
5962 unsigned Opcode = Op.getOpcode();
5963 if (Opcode == ISD::OR)
5964 return Op->getFlags().hasDisjoint() ||
5965 haveNoCommonBitsSet(Op.getOperand(0), Op.getOperand(1));
5966 if (Opcode == ISD::XOR)
5967 return !NoWrap && isMinSignedConstant(Op.getOperand(1));
5968 return false;
5969}
5970
5972 return Op.getNumOperands() == 2 && isa<ConstantSDNode>(Op.getOperand(1)) &&
5973 (Op.isAnyAdd() || isADDLike(Op));
5974}
5975
5977 unsigned Depth) const {
5978 EVT VT = Op.getValueType();
5979
5980 // Since the number of lanes in a scalable vector is unknown at compile time,
5981 // we track one bit which is implicitly broadcast to all lanes. This means
5982 // that all lanes in a scalable vector are considered demanded.
5983 APInt DemandedElts = VT.isFixedLengthVector()
5985 : APInt(1, 1);
5986
5987 return isKnownNeverNaN(Op, DemandedElts, SNaN, Depth);
5988}
5989
5991 bool SNaN, unsigned Depth) const {
5992 assert(!DemandedElts.isZero() && "No demanded elements");
5993
5994 // If we're told that NaNs won't happen, assume they won't.
5995 if (Op->getFlags().hasNoNaNs())
5996 return true;
5997
5998 if (Depth >= MaxRecursionDepth)
5999 return false; // Limit search depth.
6000
6001 // If the value is a constant, we can obviously see if it is a NaN or not.
6003 return !C->getValueAPF().isNaN() ||
6004 (SNaN && !C->getValueAPF().isSignaling());
6005 }
6006
6007 unsigned Opcode = Op.getOpcode();
6008 switch (Opcode) {
6009 case ISD::FADD:
6010 case ISD::FSUB:
6011 case ISD::FMUL:
6012 case ISD::FDIV:
6013 case ISD::FREM:
6014 case ISD::FSIN:
6015 case ISD::FCOS:
6016 case ISD::FTAN:
6017 case ISD::FASIN:
6018 case ISD::FACOS:
6019 case ISD::FATAN:
6020 case ISD::FATAN2:
6021 case ISD::FSINH:
6022 case ISD::FCOSH:
6023 case ISD::FTANH:
6024 case ISD::FMA:
6025 case ISD::FMULADD:
6026 case ISD::FMAD: {
6027 if (SNaN)
6028 return true;
6029 // TODO: Need isKnownNeverInfinity
6030 return false;
6031 }
6032 case ISD::FCANONICALIZE:
6033 case ISD::FEXP:
6034 case ISD::FEXP2:
6035 case ISD::FEXP10:
6036 case ISD::FTRUNC:
6037 case ISD::FFLOOR:
6038 case ISD::FCEIL:
6039 case ISD::FROUND:
6040 case ISD::FROUNDEVEN:
6041 case ISD::LROUND:
6042 case ISD::LLROUND:
6043 case ISD::FRINT:
6044 case ISD::LRINT:
6045 case ISD::LLRINT:
6046 case ISD::FNEARBYINT:
6047 case ISD::FLDEXP: {
6048 if (SNaN)
6049 return true;
6050 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1);
6051 }
6052 case ISD::FABS:
6053 case ISD::FNEG:
6054 case ISD::FCOPYSIGN: {
6055 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1);
6056 }
6057 case ISD::SELECT:
6058 return isKnownNeverNaN(Op.getOperand(1), DemandedElts, SNaN, Depth + 1) &&
6059 isKnownNeverNaN(Op.getOperand(2), DemandedElts, SNaN, Depth + 1);
6060 case ISD::FP_EXTEND:
6061 case ISD::FP_ROUND: {
6062 if (SNaN)
6063 return true;
6064 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1);
6065 }
6066 case ISD::SINT_TO_FP:
6067 case ISD::UINT_TO_FP:
6068 return true;
6069 case ISD::FSQRT: // Need is known positive
6070 case ISD::FLOG:
6071 case ISD::FLOG2:
6072 case ISD::FLOG10:
6073 case ISD::FPOWI:
6074 case ISD::FPOW: {
6075 if (SNaN)
6076 return true;
6077 // TODO: Refine on operand
6078 return false;
6079 }
6080 case ISD::FMINNUM:
6081 case ISD::FMAXNUM:
6082 case ISD::FMINIMUMNUM:
6083 case ISD::FMAXIMUMNUM: {
6084 // Only one needs to be known not-nan, since it will be returned if the
6085 // other ends up being one.
6086 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1) ||
6087 isKnownNeverNaN(Op.getOperand(1), DemandedElts, SNaN, Depth + 1);
6088 }
6089 case ISD::FMINNUM_IEEE:
6090 case ISD::FMAXNUM_IEEE: {
6091 if (SNaN)
6092 return true;
6093 // This can return a NaN if either operand is an sNaN, or if both operands
6094 // are NaN.
6095 return (isKnownNeverNaN(Op.getOperand(0), DemandedElts, false, Depth + 1) &&
6096 isKnownNeverSNaN(Op.getOperand(1), DemandedElts, Depth + 1)) ||
6097 (isKnownNeverNaN(Op.getOperand(1), DemandedElts, false, Depth + 1) &&
6098 isKnownNeverSNaN(Op.getOperand(0), DemandedElts, Depth + 1));
6099 }
6100 case ISD::FMINIMUM:
6101 case ISD::FMAXIMUM: {
6102 // TODO: Does this quiet or return the origina NaN as-is?
6103 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1) &&
6104 isKnownNeverNaN(Op.getOperand(1), DemandedElts, SNaN, Depth + 1);
6105 }
6107 SDValue Src = Op.getOperand(0);
6108 auto *Idx = dyn_cast<ConstantSDNode>(Op.getOperand(1));
6109 EVT SrcVT = Src.getValueType();
6110 if (SrcVT.isFixedLengthVector() && Idx &&
6111 Idx->getAPIntValue().ult(SrcVT.getVectorNumElements())) {
6112 APInt DemandedSrcElts = APInt::getOneBitSet(SrcVT.getVectorNumElements(),
6113 Idx->getZExtValue());
6114 return isKnownNeverNaN(Src, DemandedSrcElts, SNaN, Depth + 1);
6115 }
6116 return isKnownNeverNaN(Src, SNaN, Depth + 1);
6117 }
6119 SDValue Src = Op.getOperand(0);
6120 if (Src.getValueType().isFixedLengthVector()) {
6121 unsigned Idx = Op.getConstantOperandVal(1);
6122 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
6123 APInt DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
6124 return isKnownNeverNaN(Src, DemandedSrcElts, SNaN, Depth + 1);
6125 }
6126 return isKnownNeverNaN(Src, SNaN, Depth + 1);
6127 }
6128 case ISD::INSERT_SUBVECTOR: {
6129 SDValue BaseVector = Op.getOperand(0);
6130 SDValue SubVector = Op.getOperand(1);
6131 EVT BaseVectorVT = BaseVector.getValueType();
6132 if (BaseVectorVT.isFixedLengthVector()) {
6133 unsigned Idx = Op.getConstantOperandVal(2);
6134 unsigned NumBaseElts = BaseVectorVT.getVectorNumElements();
6135 unsigned NumSubElts = SubVector.getValueType().getVectorNumElements();
6136
6137 // Clear/Extract the bits at the position where the subvector will be
6138 // inserted.
6139 APInt DemandedMask =
6140 APInt::getBitsSet(NumBaseElts, Idx, Idx + NumSubElts);
6141 APInt DemandedSrcElts = DemandedElts & ~DemandedMask;
6142 APInt DemandedSubElts = DemandedElts.extractBits(NumSubElts, Idx);
6143
6144 bool NeverNaN = true;
6145 if (!DemandedSrcElts.isZero())
6146 NeverNaN &=
6147 isKnownNeverNaN(BaseVector, DemandedSrcElts, SNaN, Depth + 1);
6148 if (NeverNaN && !DemandedSubElts.isZero())
6149 NeverNaN &=
6150 isKnownNeverNaN(SubVector, DemandedSubElts, SNaN, Depth + 1);
6151 return NeverNaN;
6152 }
6153 return isKnownNeverNaN(BaseVector, SNaN, Depth + 1) &&
6154 isKnownNeverNaN(SubVector, SNaN, Depth + 1);
6155 }
6156 case ISD::BUILD_VECTOR: {
6157 unsigned NumElts = Op.getNumOperands();
6158 for (unsigned I = 0; I != NumElts; ++I)
6159 if (DemandedElts[I] &&
6160 !isKnownNeverNaN(Op.getOperand(I), SNaN, Depth + 1))
6161 return false;
6162 return true;
6163 }
6164 case ISD::SPLAT_VECTOR:
6165 return isKnownNeverNaN(Op.getOperand(0), SNaN, Depth + 1);
6166 case ISD::AssertNoFPClass: {
6167 FPClassTest NoFPClass =
6168 static_cast<FPClassTest>(Op.getConstantOperandVal(1));
6169 if ((NoFPClass & fcNan) == fcNan)
6170 return true;
6171 if (SNaN && (NoFPClass & fcSNan) == fcSNan)
6172 return true;
6173 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1);
6174 }
6175 default:
6176 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::INTRINSIC_WO_CHAIN ||
6177 Opcode == ISD::INTRINSIC_W_CHAIN || Opcode == ISD::INTRINSIC_VOID) {
6178 return TLI->isKnownNeverNaNForTargetNode(Op, DemandedElts, *this, SNaN,
6179 Depth);
6180 }
6181
6182 return false;
6183 }
6184}
6185
6187 assert(Op.getValueType().isFloatingPoint() &&
6188 "Floating point type expected");
6189
6190 // If the value is a constant, we can obviously see if it is a zero or not.
6192 Op, [](ConstantFPSDNode *C) { return !C->isZero(); });
6193}
6194
6196 EVT VT = Op.getValueType();
6197
6198 // Since the number of lanes in a scalable vector is unknown at compile time,
6199 // we track one bit which is implicitly broadcast to all lanes. This means
6200 // that all lanes in a scalable vector are considered demanded.
6201 APInt DemandedElts = VT.isFixedLengthVector()
6203 : APInt(1, 1);
6204
6205 return isKnownNeverZero(Op, DemandedElts, Depth);
6206}
6207
6209 unsigned Depth) const {
6210 if (Depth >= MaxRecursionDepth)
6211 return false; // Limit search depth.
6212
6213 EVT OpVT = Op.getValueType();
6214 unsigned BitWidth = OpVT.getScalarSizeInBits();
6215
6216 assert(!Op.getValueType().isFloatingPoint() &&
6217 "Floating point types unsupported - use isKnownNeverZeroFloat");
6218
6219 // If the value is a constant, we can obviously see if it is a zero or not.
6220 auto IsNeverZero = [BitWidth](const ConstantSDNode *C) {
6221 APInt V = C->getAPIntValue().zextOrTrunc(BitWidth);
6222 return !V.isZero();
6223 };
6224
6225 if (ISD::matchUnaryPredicate(Op, IsNeverZero))
6226 return true;
6227
6228 // TODO: Recognize more cases here. Most of the cases are also incomplete to
6229 // some degree.
6230 switch (Op.getOpcode()) {
6231 default:
6232 break;
6233
6234 case ISD::BUILD_VECTOR:
6235 // Are all operands of a build vector constant non-zero?
6236 if (all_of(enumerate(Op->ops()), [&](auto P) {
6237 auto *C = dyn_cast<ConstantSDNode>(P.value());
6238 return !DemandedElts[P.index()] || (C && IsNeverZero(C));
6239 }))
6240 return true;
6241 break;
6242
6243 case ISD::SPLAT_VECTOR:
6244 // Is the operand of a splat vector a constant non-zero?
6245 if (auto *C = dyn_cast<ConstantSDNode>(Op->getOperand(0)))
6246 if (IsNeverZero(C))
6247 return true;
6248 break;
6249
6251 SDValue InVec = Op.getOperand(0);
6252 SDValue EltNo = Op.getOperand(1);
6253 EVT VecVT = InVec.getValueType();
6254
6255 // Skip scalable vectors or implicit extensions.
6256 if (VecVT.isScalableVector() ||
6257 OpVT.getScalarSizeInBits() != VecVT.getScalarSizeInBits())
6258 break;
6259
6260 // If we know the element index, just demand that vector element, else for
6261 // an unknown element index, ignore DemandedElts and demand them all.
6262 const unsigned NumSrcElts = VecVT.getVectorNumElements();
6263 APInt DemandedSrcElts = APInt::getAllOnes(NumSrcElts);
6264 auto *ConstEltNo = dyn_cast<ConstantSDNode>(EltNo);
6265 if (ConstEltNo && ConstEltNo->getAPIntValue().ult(NumSrcElts))
6266 DemandedSrcElts =
6267 APInt::getOneBitSet(NumSrcElts, ConstEltNo->getZExtValue());
6268
6269 return isKnownNeverZero(InVec, DemandedSrcElts, Depth + 1);
6270 }
6271
6272 case ISD::OR:
6273 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) ||
6274 isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6275
6276 case ISD::VSELECT:
6277 case ISD::SELECT:
6278 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) &&
6279 isKnownNeverZero(Op.getOperand(2), DemandedElts, Depth + 1);
6280
6281 case ISD::SHL: {
6282 if (Op->getFlags().hasNoSignedWrap() || Op->getFlags().hasNoUnsignedWrap())
6283 return isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6284 KnownBits ValKnown =
6285 computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
6286 // 1 << X is never zero.
6287 if (ValKnown.One[0])
6288 return true;
6289 // If max shift cnt of known ones is non-zero, result is non-zero.
6290 APInt MaxCnt = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1)
6291 .getMaxValue();
6292 if (MaxCnt.ult(ValKnown.getBitWidth()) &&
6293 !ValKnown.One.shl(MaxCnt).isZero())
6294 return true;
6295 break;
6296 }
6297
6298 case ISD::UADDSAT:
6299 case ISD::UMAX:
6300 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) ||
6301 isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6302
6303 case ISD::UMIN:
6304 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) &&
6305 isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6306
6307 // For smin/smax: If either operand is known negative/positive
6308 // respectively we don't need the other to be known at all.
6309 case ISD::SMAX: {
6310 KnownBits Op1 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
6311 if (Op1.isStrictlyPositive())
6312 return true;
6313
6314 KnownBits Op0 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
6315 if (Op0.isStrictlyPositive())
6316 return true;
6317
6318 if (Op1.isNonZero() && Op0.isNonZero())
6319 return true;
6320
6321 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) &&
6322 isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6323 }
6324 case ISD::SMIN: {
6325 KnownBits Op1 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
6326 if (Op1.isNegative())
6327 return true;
6328
6329 KnownBits Op0 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
6330 if (Op0.isNegative())
6331 return true;
6332
6333 if (Op1.isNonZero() && Op0.isNonZero())
6334 return true;
6335
6336 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) &&
6337 isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6338 }
6339
6340 case ISD::ROTL:
6341 case ISD::ROTR:
6342 case ISD::BITREVERSE:
6343 case ISD::BSWAP:
6344 case ISD::CTPOP:
6345 case ISD::ABS:
6346 return isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6347
6348 case ISD::SRA:
6349 case ISD::SRL: {
6350 if (Op->getFlags().hasExact())
6351 return isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6352 KnownBits ValKnown =
6353 computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
6354 if (ValKnown.isNegative())
6355 return true;
6356 // If max shift cnt of known ones is non-zero, result is non-zero.
6357 APInt MaxCnt = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1)
6358 .getMaxValue();
6359 if (MaxCnt.ult(ValKnown.getBitWidth()) &&
6360 !ValKnown.One.lshr(MaxCnt).isZero())
6361 return true;
6362 break;
6363 }
6364 case ISD::UDIV:
6365 case ISD::SDIV:
6366 // div exact can only produce a zero if the dividend is zero.
6367 // TODO: For udiv this is also true if Op1 u<= Op0
6368 if (Op->getFlags().hasExact())
6369 return isKnownNeverZero(Op.getOperand(0), Depth + 1);
6370 break;
6371
6372 case ISD::ADD:
6373 if (Op->getFlags().hasNoUnsignedWrap())
6374 if (isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) ||
6375 isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1))
6376 return true;
6377 // TODO: There are a lot more cases we can prove for add.
6378 break;
6379
6380 case ISD::SUB: {
6381 if (isNullConstant(Op.getOperand(0)))
6382 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1);
6383
6384 std::optional<bool> ne = KnownBits::ne(
6385 computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1),
6386 computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1));
6387 return ne && *ne;
6388 }
6389
6390 case ISD::MUL:
6391 if (Op->getFlags().hasNoSignedWrap() || Op->getFlags().hasNoUnsignedWrap())
6392 if (isKnownNeverZero(Op.getOperand(1), Depth + 1) &&
6393 isKnownNeverZero(Op.getOperand(0), Depth + 1))
6394 return true;
6395 break;
6396
6397 case ISD::ZERO_EXTEND:
6398 case ISD::SIGN_EXTEND:
6399 return isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6400 case ISD::VSCALE: {
6402 const APInt &Multiplier = Op.getConstantOperandAPInt(0);
6403 ConstantRange CR =
6404 getVScaleRange(&F, Op.getScalarValueSizeInBits()).multiply(Multiplier);
6405 if (!CR.contains(APInt(CR.getBitWidth(), 0)))
6406 return true;
6407 break;
6408 }
6409 }
6410
6412}
6413
6415 if (ConstantFPSDNode *C1 = isConstOrConstSplatFP(Op, true))
6416 return !C1->isNegative();
6417
6418 switch (Op.getOpcode()) {
6419 case ISD::FABS:
6420 case ISD::FEXP:
6421 case ISD::FEXP2:
6422 case ISD::FEXP10:
6423 return true;
6424 default:
6425 return false;
6426 }
6427
6428 llvm_unreachable("covered opcode switch");
6429}
6430
6432 assert(Use.getValueType().isFloatingPoint());
6433 const SDNode *User = Use.getUser();
6434 if (User->getFlags().hasNoSignedZeros())
6435 return true;
6436
6437 unsigned OperandNo = Use.getOperandNo();
6438 // Check if this use is insensitive to the sign of zero
6439 switch (User->getOpcode()) {
6440 case ISD::SETCC:
6441 // Comparisons: IEEE-754 specifies +0.0 == -0.0.
6442 case ISD::FABS:
6443 // fabs always produces +0.0.
6444 return true;
6445 case ISD::FCOPYSIGN:
6446 // copysign overwrites the sign bit of the first operand.
6447 return OperandNo == 0;
6448 case ISD::FADD:
6449 case ISD::FSUB: {
6450 // Arithmetic with non-zero constants fixes the uncertainty around the
6451 // sign bit.
6452 SDValue Other = User->getOperand(1 - OperandNo);
6454 }
6455 case ISD::FP_TO_SINT:
6456 case ISD::FP_TO_UINT:
6457 // fp-to-int conversions normalize signed zeros.
6458 return true;
6459 default:
6460 return false;
6461 }
6462}
6463
6465 if (Op->getFlags().hasNoSignedZeros())
6466 return true;
6467 // FIXME: Limit the amount of checked uses to not introduce a compile-time
6468 // regression. Ideally, this should be implemented as a demanded-bits
6469 // optimization that stems from the users.
6470 if (Op->use_size() > 2)
6471 return false;
6472 return all_of(Op->uses(),
6473 [&](const SDUse &Use) { return canIgnoreSignBitOfZero(Use); });
6474}
6475
6477 // Check the obvious case.
6478 if (A == B) return true;
6479
6480 // For negative and positive zero.
6483 if (CA->isZero() && CB->isZero()) return true;
6484
6485 // Otherwise they may not be equal.
6486 return false;
6487}
6488
6489// Only bits set in Mask must be negated, other bits may be arbitrary.
6491 if (isBitwiseNot(V, AllowUndefs))
6492 return V.getOperand(0);
6493
6494 // Handle any_extend (not (truncate X)) pattern, where Mask only sets
6495 // bits in the non-extended part.
6496 ConstantSDNode *MaskC = isConstOrConstSplat(Mask);
6497 if (!MaskC || V.getOpcode() != ISD::ANY_EXTEND)
6498 return SDValue();
6499 SDValue ExtArg = V.getOperand(0);
6500 if (ExtArg.getScalarValueSizeInBits() >=
6501 MaskC->getAPIntValue().getActiveBits() &&
6502 isBitwiseNot(ExtArg, AllowUndefs) &&
6503 ExtArg.getOperand(0).getOpcode() == ISD::TRUNCATE &&
6504 ExtArg.getOperand(0).getOperand(0).getValueType() == V.getValueType())
6505 return ExtArg.getOperand(0).getOperand(0);
6506 return SDValue();
6507}
6508
6510 // Match masked merge pattern (X & ~M) op (Y & M)
6511 // Including degenerate case (X & ~M) op M
6512 auto MatchNoCommonBitsPattern = [&](SDValue Not, SDValue Mask,
6513 SDValue Other) {
6514 if (SDValue NotOperand =
6515 getBitwiseNotOperand(Not, Mask, /* AllowUndefs */ true)) {
6516 if (NotOperand->getOpcode() == ISD::ZERO_EXTEND ||
6517 NotOperand->getOpcode() == ISD::TRUNCATE)
6518 NotOperand = NotOperand->getOperand(0);
6519
6520 if (Other == NotOperand)
6521 return true;
6522 if (Other->getOpcode() == ISD::AND)
6523 return NotOperand == Other->getOperand(0) ||
6524 NotOperand == Other->getOperand(1);
6525 }
6526 return false;
6527 };
6528
6529 if (A->getOpcode() == ISD::ZERO_EXTEND || A->getOpcode() == ISD::TRUNCATE)
6530 A = A->getOperand(0);
6531
6532 if (B->getOpcode() == ISD::ZERO_EXTEND || B->getOpcode() == ISD::TRUNCATE)
6533 B = B->getOperand(0);
6534
6535 if (A->getOpcode() == ISD::AND)
6536 return MatchNoCommonBitsPattern(A->getOperand(0), A->getOperand(1), B) ||
6537 MatchNoCommonBitsPattern(A->getOperand(1), A->getOperand(0), B);
6538 return false;
6539}
6540
6541// FIXME: unify with llvm::haveNoCommonBitsSet.
6543 assert(A.getValueType() == B.getValueType() &&
6544 "Values must have the same type");
6547 return true;
6550}
6551
6552static SDValue FoldSTEP_VECTOR(const SDLoc &DL, EVT VT, SDValue Step,
6553 SelectionDAG &DAG) {
6554 if (cast<ConstantSDNode>(Step)->isZero())
6555 return DAG.getConstant(0, DL, VT);
6556
6557 return SDValue();
6558}
6559
6562 SelectionDAG &DAG) {
6563 int NumOps = Ops.size();
6564 assert(NumOps != 0 && "Can't build an empty vector!");
6565 assert(!VT.isScalableVector() &&
6566 "BUILD_VECTOR cannot be used with scalable types");
6567 assert(VT.getVectorNumElements() == (unsigned)NumOps &&
6568 "Incorrect element count in BUILD_VECTOR!");
6569
6570 // BUILD_VECTOR of UNDEFs is UNDEF.
6571 if (llvm::all_of(Ops, [](SDValue Op) { return Op.isUndef(); }))
6572 return DAG.getUNDEF(VT);
6573
6574 // BUILD_VECTOR of seq extract/insert from the same vector + type is Identity.
6575 SDValue IdentitySrc;
6576 bool IsIdentity = true;
6577 for (int i = 0; i != NumOps; ++i) {
6579 Ops[i].getOperand(0).getValueType() != VT ||
6580 (IdentitySrc && Ops[i].getOperand(0) != IdentitySrc) ||
6581 !isa<ConstantSDNode>(Ops[i].getOperand(1)) ||
6582 Ops[i].getConstantOperandAPInt(1) != i) {
6583 IsIdentity = false;
6584 break;
6585 }
6586 IdentitySrc = Ops[i].getOperand(0);
6587 }
6588 if (IsIdentity)
6589 return IdentitySrc;
6590
6591 return SDValue();
6592}
6593
6594/// Try to simplify vector concatenation to an input value, undef, or build
6595/// vector.
6598 SelectionDAG &DAG) {
6599 assert(!Ops.empty() && "Can't concatenate an empty list of vectors!");
6601 [Ops](SDValue Op) {
6602 return Ops[0].getValueType() == Op.getValueType();
6603 }) &&
6604 "Concatenation of vectors with inconsistent value types!");
6605 assert((Ops[0].getValueType().getVectorElementCount() * Ops.size()) ==
6606 VT.getVectorElementCount() &&
6607 "Incorrect element count in vector concatenation!");
6608
6609 if (Ops.size() == 1)
6610 return Ops[0];
6611
6612 // Concat of UNDEFs is UNDEF.
6613 if (llvm::all_of(Ops, [](SDValue Op) { return Op.isUndef(); }))
6614 return DAG.getUNDEF(VT);
6615
6616 // Scan the operands and look for extract operations from a single source
6617 // that correspond to insertion at the same location via this concatenation:
6618 // concat (extract X, 0*subvec_elts), (extract X, 1*subvec_elts), ...
6619 SDValue IdentitySrc;
6620 bool IsIdentity = true;
6621 for (unsigned i = 0, e = Ops.size(); i != e; ++i) {
6622 SDValue Op = Ops[i];
6623 unsigned IdentityIndex = i * Op.getValueType().getVectorMinNumElements();
6624 if (Op.getOpcode() != ISD::EXTRACT_SUBVECTOR ||
6625 Op.getOperand(0).getValueType() != VT ||
6626 (IdentitySrc && Op.getOperand(0) != IdentitySrc) ||
6627 Op.getConstantOperandVal(1) != IdentityIndex) {
6628 IsIdentity = false;
6629 break;
6630 }
6631 assert((!IdentitySrc || IdentitySrc == Op.getOperand(0)) &&
6632 "Unexpected identity source vector for concat of extracts");
6633 IdentitySrc = Op.getOperand(0);
6634 }
6635 if (IsIdentity) {
6636 assert(IdentitySrc && "Failed to set source vector of extracts");
6637 return IdentitySrc;
6638 }
6639
6640 // The code below this point is only designed to work for fixed width
6641 // vectors, so we bail out for now.
6642 if (VT.isScalableVector())
6643 return SDValue();
6644
6645 // A CONCAT_VECTOR of scalar sources, such as UNDEF, BUILD_VECTOR and
6646 // single-element INSERT_VECTOR_ELT operands can be simplified to one big
6647 // BUILD_VECTOR.
6648 // FIXME: Add support for SCALAR_TO_VECTOR as well.
6649 EVT SVT = VT.getScalarType();
6651 for (SDValue Op : Ops) {
6652 EVT OpVT = Op.getValueType();
6653 if (Op.isUndef())
6654 Elts.append(OpVT.getVectorNumElements(), DAG.getUNDEF(SVT));
6655 else if (Op.getOpcode() == ISD::BUILD_VECTOR)
6656 Elts.append(Op->op_begin(), Op->op_end());
6657 else if (Op.getOpcode() == ISD::INSERT_VECTOR_ELT &&
6658 OpVT.getVectorNumElements() == 1 &&
6659 isNullConstant(Op.getOperand(2)))
6660 Elts.push_back(Op.getOperand(1));
6661 else
6662 return SDValue();
6663 }
6664
6665 // BUILD_VECTOR requires all inputs to be of the same type, find the
6666 // maximum type and extend them all.
6667 for (SDValue Op : Elts)
6668 SVT = (SVT.bitsLT(Op.getValueType()) ? Op.getValueType() : SVT);
6669
6670 if (SVT.bitsGT(VT.getScalarType())) {
6671 for (SDValue &Op : Elts) {
6672 if (Op.isUndef())
6673 Op = DAG.getUNDEF(SVT);
6674 else
6675 Op = DAG.getTargetLoweringInfo().isZExtFree(Op.getValueType(), SVT)
6676 ? DAG.getZExtOrTrunc(Op, DL, SVT)
6677 : DAG.getSExtOrTrunc(Op, DL, SVT);
6678 }
6679 }
6680
6681 SDValue V = DAG.getBuildVector(VT, DL, Elts);
6682 NewSDValueDbgMsg(V, "New node fold concat vectors: ", &DAG);
6683 return V;
6684}
6685
6686/// Gets or creates the specified node.
6687SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT) {
6688 SDVTList VTs = getVTList(VT);
6690 AddNodeIDNode(ID, Opcode, VTs, {});
6691 void *IP = nullptr;
6692 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP))
6693 return SDValue(E, 0);
6694
6695 auto *N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
6696 CSEMap.InsertNode(N, IP);
6697
6698 InsertNode(N);
6699 SDValue V = SDValue(N, 0);
6700 NewSDValueDbgMsg(V, "Creating new node: ", this);
6701 return V;
6702}
6703
6704SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
6705 SDValue N1) {
6706 SDNodeFlags Flags;
6707 if (Inserter)
6708 Flags = Inserter->getFlags();
6709 return getNode(Opcode, DL, VT, N1, Flags);
6710}
6711
6712SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
6713 SDValue N1, const SDNodeFlags Flags) {
6714 assert(N1.getOpcode() != ISD::DELETED_NODE && "Operand is DELETED_NODE!");
6715
6716 // Constant fold unary operations with a vector integer or float operand.
6717 switch (Opcode) {
6718 default:
6719 // FIXME: Entirely reasonable to perform folding of other unary
6720 // operations here as the need arises.
6721 break;
6722 case ISD::FNEG:
6723 case ISD::FABS:
6724 case ISD::FCEIL:
6725 case ISD::FTRUNC:
6726 case ISD::FFLOOR:
6727 case ISD::FP_EXTEND:
6728 case ISD::FP_TO_SINT:
6729 case ISD::FP_TO_UINT:
6730 case ISD::FP_TO_FP16:
6731 case ISD::FP_TO_BF16:
6732 case ISD::TRUNCATE:
6733 case ISD::ANY_EXTEND:
6734 case ISD::ZERO_EXTEND:
6735 case ISD::SIGN_EXTEND:
6736 case ISD::UINT_TO_FP:
6737 case ISD::SINT_TO_FP:
6738 case ISD::FP16_TO_FP:
6739 case ISD::BF16_TO_FP:
6740 case ISD::BITCAST:
6741 case ISD::ABS:
6742 case ISD::BITREVERSE:
6743 case ISD::BSWAP:
6744 case ISD::CTLZ:
6746 case ISD::CTTZ:
6748 case ISD::CTPOP:
6749 case ISD::CTLS:
6750 case ISD::STEP_VECTOR: {
6751 SDValue Ops = {N1};
6752 if (SDValue Fold = FoldConstantArithmetic(Opcode, DL, VT, Ops))
6753 return Fold;
6754 }
6755 }
6756
6757 unsigned OpOpcode = N1.getNode()->getOpcode();
6758 switch (Opcode) {
6759 case ISD::STEP_VECTOR:
6760 assert(VT.isScalableVector() &&
6761 "STEP_VECTOR can only be used with scalable types");
6762 assert(OpOpcode == ISD::TargetConstant &&
6763 VT.getVectorElementType() == N1.getValueType() &&
6764 "Unexpected step operand");
6765 break;
6766 case ISD::FREEZE:
6767 assert(VT == N1.getValueType() && "Unexpected VT!");
6768 if (isGuaranteedNotToBeUndefOrPoison(N1, /*PoisonOnly=*/false))
6769 return N1;
6770 break;
6771 case ISD::TokenFactor:
6772 case ISD::MERGE_VALUES:
6774 return N1; // Factor, merge or concat of one node? No need.
6775 case ISD::BUILD_VECTOR: {
6776 // Attempt to simplify BUILD_VECTOR.
6777 SDValue Ops[] = {N1};
6778 if (SDValue V = FoldBUILD_VECTOR(DL, VT, Ops, *this))
6779 return V;
6780 break;
6781 }
6782 case ISD::FP_ROUND: llvm_unreachable("Invalid method to make FP_ROUND node");
6783 case ISD::FP_EXTEND:
6785 "Invalid FP cast!");
6786 if (N1.getValueType() == VT) return N1; // noop conversion.
6787 assert((!VT.isVector() || VT.getVectorElementCount() ==
6789 "Vector element count mismatch!");
6790 assert(N1.getValueType().bitsLT(VT) && "Invalid fpext node, dst < src!");
6791 if (N1.isUndef())
6792 return getUNDEF(VT);
6793 break;
6794 case ISD::FP_TO_SINT:
6795 case ISD::FP_TO_UINT:
6796 if (N1.isUndef())
6797 return getUNDEF(VT);
6798 break;
6799 case ISD::SINT_TO_FP:
6800 case ISD::UINT_TO_FP:
6801 // [us]itofp(undef) = 0, because the result value is bounded.
6802 if (N1.isUndef())
6803 return getConstantFP(0.0, DL, VT);
6804 break;
6805 case ISD::SIGN_EXTEND:
6806 assert(VT.isInteger() && N1.getValueType().isInteger() &&
6807 "Invalid SIGN_EXTEND!");
6808 assert(VT.isVector() == N1.getValueType().isVector() &&
6809 "SIGN_EXTEND result type type should be vector iff the operand "
6810 "type is vector!");
6811 if (N1.getValueType() == VT) return N1; // noop extension
6812 assert((!VT.isVector() || VT.getVectorElementCount() ==
6814 "Vector element count mismatch!");
6815 assert(N1.getValueType().bitsLT(VT) && "Invalid sext node, dst < src!");
6816 if (OpOpcode == ISD::SIGN_EXTEND || OpOpcode == ISD::ZERO_EXTEND) {
6817 SDNodeFlags Flags;
6818 if (OpOpcode == ISD::ZERO_EXTEND)
6819 Flags.setNonNeg(N1->getFlags().hasNonNeg());
6820 SDValue NewVal = getNode(OpOpcode, DL, VT, N1.getOperand(0), Flags);
6821 transferDbgValues(N1, NewVal);
6822 return NewVal;
6823 }
6824
6825 if (OpOpcode == ISD::POISON)
6826 return getPOISON(VT);
6827
6828 if (N1.isUndef())
6829 // sext(undef) = 0, because the top bits will all be the same.
6830 return getConstant(0, DL, VT);
6831
6832 // Skip unnecessary sext_inreg pattern:
6833 // (sext (trunc x)) -> x iff the upper bits are all signbits.
6834 if (OpOpcode == ISD::TRUNCATE) {
6835 SDValue OpOp = N1.getOperand(0);
6836 if (OpOp.getValueType() == VT) {
6837 unsigned NumSignExtBits =
6839 if (ComputeNumSignBits(OpOp) > NumSignExtBits) {
6840 transferDbgValues(N1, OpOp);
6841 return OpOp;
6842 }
6843 }
6844 }
6845 break;
6846 case ISD::ZERO_EXTEND:
6847 assert(VT.isInteger() && N1.getValueType().isInteger() &&
6848 "Invalid ZERO_EXTEND!");
6849 assert(VT.isVector() == N1.getValueType().isVector() &&
6850 "ZERO_EXTEND result type type should be vector iff the operand "
6851 "type is vector!");
6852 if (N1.getValueType() == VT) return N1; // noop extension
6853 assert((!VT.isVector() || VT.getVectorElementCount() ==
6855 "Vector element count mismatch!");
6856 assert(N1.getValueType().bitsLT(VT) && "Invalid zext node, dst < src!");
6857 if (OpOpcode == ISD::ZERO_EXTEND) { // (zext (zext x)) -> (zext x)
6858 SDNodeFlags Flags;
6859 Flags.setNonNeg(N1->getFlags().hasNonNeg());
6860 SDValue NewVal =
6861 getNode(ISD::ZERO_EXTEND, DL, VT, N1.getOperand(0), Flags);
6862 transferDbgValues(N1, NewVal);
6863 return NewVal;
6864 }
6865
6866 if (OpOpcode == ISD::POISON)
6867 return getPOISON(VT);
6868
6869 if (N1.isUndef())
6870 // zext(undef) = 0, because the top bits will be zero.
6871 return getConstant(0, DL, VT);
6872
6873 // Skip unnecessary zext_inreg pattern:
6874 // (zext (trunc x)) -> x iff the upper bits are known zero.
6875 // TODO: Remove (zext (trunc (and x, c))) exception which some targets
6876 // use to recognise zext_inreg patterns.
6877 if (OpOpcode == ISD::TRUNCATE) {
6878 SDValue OpOp = N1.getOperand(0);
6879 if (OpOp.getValueType() == VT) {
6880 if (OpOp.getOpcode() != ISD::AND) {
6883 if (MaskedValueIsZero(OpOp, HiBits)) {
6884 transferDbgValues(N1, OpOp);
6885 return OpOp;
6886 }
6887 }
6888 }
6889 }
6890 break;
6891 case ISD::ANY_EXTEND:
6892 assert(VT.isInteger() && N1.getValueType().isInteger() &&
6893 "Invalid ANY_EXTEND!");
6894 assert(VT.isVector() == N1.getValueType().isVector() &&
6895 "ANY_EXTEND result type type should be vector iff the operand "
6896 "type is vector!");
6897 if (N1.getValueType() == VT) return N1; // noop extension
6898 assert((!VT.isVector() || VT.getVectorElementCount() ==
6900 "Vector element count mismatch!");
6901 assert(N1.getValueType().bitsLT(VT) && "Invalid anyext node, dst < src!");
6902
6903 if (OpOpcode == ISD::ZERO_EXTEND || OpOpcode == ISD::SIGN_EXTEND ||
6904 OpOpcode == ISD::ANY_EXTEND) {
6905 SDNodeFlags Flags;
6906 if (OpOpcode == ISD::ZERO_EXTEND)
6907 Flags.setNonNeg(N1->getFlags().hasNonNeg());
6908 // (ext (zext x)) -> (zext x) and (ext (sext x)) -> (sext x)
6909 return getNode(OpOpcode, DL, VT, N1.getOperand(0), Flags);
6910 }
6911 if (N1.isUndef())
6912 return getUNDEF(VT);
6913
6914 // (ext (trunc x)) -> x
6915 if (OpOpcode == ISD::TRUNCATE) {
6916 SDValue OpOp = N1.getOperand(0);
6917 if (OpOp.getValueType() == VT) {
6918 transferDbgValues(N1, OpOp);
6919 return OpOp;
6920 }
6921 }
6922 break;
6923 case ISD::TRUNCATE:
6924 assert(VT.isInteger() && N1.getValueType().isInteger() &&
6925 "Invalid TRUNCATE!");
6926 assert(VT.isVector() == N1.getValueType().isVector() &&
6927 "TRUNCATE result type type should be vector iff the operand "
6928 "type is vector!");
6929 if (N1.getValueType() == VT) return N1; // noop truncate
6930 assert((!VT.isVector() || VT.getVectorElementCount() ==
6932 "Vector element count mismatch!");
6933 assert(N1.getValueType().bitsGT(VT) && "Invalid truncate node, src < dst!");
6934 if (OpOpcode == ISD::TRUNCATE)
6935 return getNode(ISD::TRUNCATE, DL, VT, N1.getOperand(0));
6936 if (OpOpcode == ISD::ZERO_EXTEND || OpOpcode == ISD::SIGN_EXTEND ||
6937 OpOpcode == ISD::ANY_EXTEND) {
6938 // If the source is smaller than the dest, we still need an extend.
6940 VT.getScalarType())) {
6941 SDNodeFlags Flags;
6942 if (OpOpcode == ISD::ZERO_EXTEND)
6943 Flags.setNonNeg(N1->getFlags().hasNonNeg());
6944 return getNode(OpOpcode, DL, VT, N1.getOperand(0), Flags);
6945 }
6946 if (N1.getOperand(0).getValueType().bitsGT(VT))
6947 return getNode(ISD::TRUNCATE, DL, VT, N1.getOperand(0));
6948 return N1.getOperand(0);
6949 }
6950 if (N1.isUndef())
6951 return getUNDEF(VT);
6952 if (OpOpcode == ISD::VSCALE && !NewNodesMustHaveLegalTypes)
6953 return getVScale(DL, VT,
6955 break;
6959 assert(VT.isVector() && "This DAG node is restricted to vector types.");
6960 assert(N1.getValueType().bitsLE(VT) &&
6961 "The input must be the same size or smaller than the result.");
6964 "The destination vector type must have fewer lanes than the input.");
6965 break;
6966 case ISD::ABS:
6967 assert(VT.isInteger() && VT == N1.getValueType() && "Invalid ABS!");
6968 if (N1.isUndef())
6969 return getConstant(0, DL, VT);
6970 break;
6971 case ISD::BSWAP:
6972 assert(VT.isInteger() && VT == N1.getValueType() && "Invalid BSWAP!");
6973 assert((VT.getScalarSizeInBits() % 16 == 0) &&
6974 "BSWAP types must be a multiple of 16 bits!");
6975 if (N1.isUndef())
6976 return getUNDEF(VT);
6977 // bswap(bswap(X)) -> X.
6978 if (OpOpcode == ISD::BSWAP)
6979 return N1.getOperand(0);
6980 break;
6981 case ISD::BITREVERSE:
6982 assert(VT.isInteger() && VT == N1.getValueType() && "Invalid BITREVERSE!");
6983 if (N1.isUndef())
6984 return getUNDEF(VT);
6985 break;
6986 case ISD::BITCAST:
6988 "Cannot BITCAST between types of different sizes!");
6989 if (VT == N1.getValueType()) return N1; // noop conversion.
6990 if (OpOpcode == ISD::BITCAST) // bitconv(bitconv(x)) -> bitconv(x)
6991 return getNode(ISD::BITCAST, DL, VT, N1.getOperand(0));
6992 if (N1.isUndef())
6993 return getUNDEF(VT);
6994 break;
6996 assert(VT.isVector() && !N1.getValueType().isVector() &&
6997 (VT.getVectorElementType() == N1.getValueType() ||
6999 N1.getValueType().isInteger() &&
7001 "Illegal SCALAR_TO_VECTOR node!");
7002 if (N1.isUndef())
7003 return getUNDEF(VT);
7004 // scalar_to_vector(extract_vector_elt V, 0) -> V, top bits are undefined.
7005 if (OpOpcode == ISD::EXTRACT_VECTOR_ELT &&
7007 N1.getConstantOperandVal(1) == 0 &&
7008 N1.getOperand(0).getValueType() == VT)
7009 return N1.getOperand(0);
7010 break;
7011 case ISD::FNEG:
7012 // Negation of an unknown bag of bits is still completely undefined.
7013 if (N1.isUndef())
7014 return getUNDEF(VT);
7015
7016 if (OpOpcode == ISD::FNEG) // --X -> X
7017 return N1.getOperand(0);
7018 break;
7019 case ISD::FABS:
7020 if (OpOpcode == ISD::FNEG) // abs(-X) -> abs(X)
7021 return getNode(ISD::FABS, DL, VT, N1.getOperand(0));
7022 break;
7023 case ISD::VSCALE:
7024 assert(VT == N1.getValueType() && "Unexpected VT!");
7025 break;
7026 case ISD::CTPOP:
7027 if (N1.getValueType().getScalarType() == MVT::i1)
7028 return N1;
7029 break;
7030 case ISD::CTLZ:
7031 case ISD::CTTZ:
7032 if (N1.getValueType().getScalarType() == MVT::i1)
7033 return getNOT(DL, N1, N1.getValueType());
7034 break;
7035 case ISD::CTLS:
7036 if (N1.getValueType().getScalarType() == MVT::i1)
7037 return getConstant(0, DL, VT);
7038 break;
7039 case ISD::VECREDUCE_ADD:
7040 if (N1.getValueType().getScalarType() == MVT::i1)
7041 return getNode(ISD::VECREDUCE_XOR, DL, VT, N1);
7042 break;
7045 if (N1.getValueType().getScalarType() == MVT::i1)
7046 return getNode(ISD::VECREDUCE_OR, DL, VT, N1);
7047 break;
7050 if (N1.getValueType().getScalarType() == MVT::i1)
7051 return getNode(ISD::VECREDUCE_AND, DL, VT, N1);
7052 break;
7053 case ISD::SPLAT_VECTOR:
7054 assert(VT.isVector() && "Wrong return type!");
7055 // FIXME: Hexagon uses i32 scalar for a floating point zero vector so allow
7056 // that for now.
7058 (VT.isFloatingPoint() && N1.getValueType() == MVT::i32) ||
7060 N1.getValueType().isInteger() &&
7062 "Wrong operand type!");
7063 break;
7064 }
7065
7066 SDNode *N;
7067 SDVTList VTs = getVTList(VT);
7068 SDValue Ops[] = {N1};
7069 if (VT != MVT::Glue) { // Don't CSE glue producing nodes
7071 AddNodeIDNode(ID, Opcode, VTs, Ops);
7072 void *IP = nullptr;
7073 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
7074 E->intersectFlagsWith(Flags);
7075 return SDValue(E, 0);
7076 }
7077
7078 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
7079 N->setFlags(Flags);
7080 createOperands(N, Ops);
7081 CSEMap.InsertNode(N, IP);
7082 } else {
7083 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
7084 createOperands(N, Ops);
7085 }
7086
7087 InsertNode(N);
7088 SDValue V = SDValue(N, 0);
7089 NewSDValueDbgMsg(V, "Creating new node: ", this);
7090 return V;
7091}
7092
7093static std::optional<APInt> FoldValue(unsigned Opcode, const APInt &C1,
7094 const APInt &C2) {
7095 switch (Opcode) {
7096 case ISD::ADD: return C1 + C2;
7097 case ISD::SUB: return C1 - C2;
7098 case ISD::MUL: return C1 * C2;
7099 case ISD::AND: return C1 & C2;
7100 case ISD::OR: return C1 | C2;
7101 case ISD::XOR: return C1 ^ C2;
7102 case ISD::SHL: return C1 << C2;
7103 case ISD::SRL: return C1.lshr(C2);
7104 case ISD::SRA: return C1.ashr(C2);
7105 case ISD::ROTL: return C1.rotl(C2);
7106 case ISD::ROTR: return C1.rotr(C2);
7107 case ISD::SMIN: return C1.sle(C2) ? C1 : C2;
7108 case ISD::SMAX: return C1.sge(C2) ? C1 : C2;
7109 case ISD::UMIN: return C1.ule(C2) ? C1 : C2;
7110 case ISD::UMAX: return C1.uge(C2) ? C1 : C2;
7111 case ISD::SADDSAT: return C1.sadd_sat(C2);
7112 case ISD::UADDSAT: return C1.uadd_sat(C2);
7113 case ISD::SSUBSAT: return C1.ssub_sat(C2);
7114 case ISD::USUBSAT: return C1.usub_sat(C2);
7115 case ISD::SSHLSAT: return C1.sshl_sat(C2);
7116 case ISD::USHLSAT: return C1.ushl_sat(C2);
7117 case ISD::UDIV:
7118 if (!C2.getBoolValue())
7119 break;
7120 return C1.udiv(C2);
7121 case ISD::UREM:
7122 if (!C2.getBoolValue())
7123 break;
7124 return C1.urem(C2);
7125 case ISD::SDIV:
7126 if (!C2.getBoolValue())
7127 break;
7128 return C1.sdiv(C2);
7129 case ISD::SREM:
7130 if (!C2.getBoolValue())
7131 break;
7132 return C1.srem(C2);
7133 case ISD::AVGFLOORS:
7134 return APIntOps::avgFloorS(C1, C2);
7135 case ISD::AVGFLOORU:
7136 return APIntOps::avgFloorU(C1, C2);
7137 case ISD::AVGCEILS:
7138 return APIntOps::avgCeilS(C1, C2);
7139 case ISD::AVGCEILU:
7140 return APIntOps::avgCeilU(C1, C2);
7141 case ISD::ABDS:
7142 return APIntOps::abds(C1, C2);
7143 case ISD::ABDU:
7144 return APIntOps::abdu(C1, C2);
7145 case ISD::MULHS:
7146 return APIntOps::mulhs(C1, C2);
7147 case ISD::MULHU:
7148 return APIntOps::mulhu(C1, C2);
7149 case ISD::CLMUL:
7150 return APIntOps::clmul(C1, C2);
7151 case ISD::CLMULR:
7152 return APIntOps::clmulr(C1, C2);
7153 case ISD::CLMULH:
7154 return APIntOps::clmulh(C1, C2);
7155 }
7156 return std::nullopt;
7157}
7158// Handle constant folding with UNDEF.
7159// TODO: Handle more cases.
7160static std::optional<APInt> FoldValueWithUndef(unsigned Opcode, const APInt &C1,
7161 bool IsUndef1, const APInt &C2,
7162 bool IsUndef2) {
7163 if (!(IsUndef1 || IsUndef2))
7164 return FoldValue(Opcode, C1, C2);
7165
7166 // Fold and(x, undef) -> 0
7167 // Fold mul(x, undef) -> 0
7168 if (Opcode == ISD::AND || Opcode == ISD::MUL)
7169 return APInt::getZero(C1.getBitWidth());
7170
7171 return std::nullopt;
7172}
7173
7175 const GlobalAddressSDNode *GA,
7176 const SDNode *N2) {
7177 if (GA->getOpcode() != ISD::GlobalAddress)
7178 return SDValue();
7179 if (!TLI->isOffsetFoldingLegal(GA))
7180 return SDValue();
7181 auto *C2 = dyn_cast<ConstantSDNode>(N2);
7182 if (!C2)
7183 return SDValue();
7184 int64_t Offset = C2->getSExtValue();
7185 switch (Opcode) {
7186 case ISD::ADD:
7187 case ISD::PTRADD:
7188 break;
7189 case ISD::SUB: Offset = -uint64_t(Offset); break;
7190 default: return SDValue();
7191 }
7192 return getGlobalAddress(GA->getGlobal(), SDLoc(C2), VT,
7193 GA->getOffset() + uint64_t(Offset));
7194}
7195
7197 switch (Opcode) {
7198 case ISD::SDIV:
7199 case ISD::UDIV:
7200 case ISD::SREM:
7201 case ISD::UREM: {
7202 // If a divisor is zero/undef or any element of a divisor vector is
7203 // zero/undef, the whole op is undef.
7204 assert(Ops.size() == 2 && "Div/rem should have 2 operands");
7205 SDValue Divisor = Ops[1];
7206 if (Divisor.isUndef() || isNullConstant(Divisor))
7207 return true;
7208
7209 return ISD::isBuildVectorOfConstantSDNodes(Divisor.getNode()) &&
7210 llvm::any_of(Divisor->op_values(),
7211 [](SDValue V) { return V.isUndef() ||
7212 isNullConstant(V); });
7213 // TODO: Handle signed overflow.
7214 }
7215 // TODO: Handle oversized shifts.
7216 default:
7217 return false;
7218 }
7219}
7220
7223 SDNodeFlags Flags) {
7224 // If the opcode is a target-specific ISD node, there's nothing we can
7225 // do here and the operand rules may not line up with the below, so
7226 // bail early.
7227 // We can't create a scalar CONCAT_VECTORS so skip it. It will break
7228 // for concats involving SPLAT_VECTOR. Concats of BUILD_VECTORS are handled by
7229 // foldCONCAT_VECTORS in getNode before this is called.
7230 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::CONCAT_VECTORS)
7231 return SDValue();
7232
7233 unsigned NumOps = Ops.size();
7234 if (NumOps == 0)
7235 return SDValue();
7236
7237 if (isUndef(Opcode, Ops))
7238 return getUNDEF(VT);
7239
7240 // Handle unary special cases.
7241 if (NumOps == 1) {
7242 SDValue N1 = Ops[0];
7243
7244 // Constant fold unary operations with an integer constant operand. Even
7245 // opaque constant will be folded, because the folding of unary operations
7246 // doesn't create new constants with different values. Nevertheless, the
7247 // opaque flag is preserved during folding to prevent future folding with
7248 // other constants.
7249 if (auto *C = dyn_cast<ConstantSDNode>(N1)) {
7250 const APInt &Val = C->getAPIntValue();
7251 switch (Opcode) {
7252 case ISD::SIGN_EXTEND:
7253 return getConstant(Val.sextOrTrunc(VT.getSizeInBits()), DL, VT,
7254 C->isTargetOpcode(), C->isOpaque());
7255 case ISD::TRUNCATE:
7256 if (C->isOpaque())
7257 break;
7258 [[fallthrough]];
7259 case ISD::ZERO_EXTEND:
7260 return getConstant(Val.zextOrTrunc(VT.getSizeInBits()), DL, VT,
7261 C->isTargetOpcode(), C->isOpaque());
7262 case ISD::ANY_EXTEND:
7263 // Some targets like RISCV prefer to sign extend some types.
7264 if (TLI->isSExtCheaperThanZExt(N1.getValueType(), VT))
7265 return getConstant(Val.sextOrTrunc(VT.getSizeInBits()), DL, VT,
7266 C->isTargetOpcode(), C->isOpaque());
7267 return getConstant(Val.zextOrTrunc(VT.getSizeInBits()), DL, VT,
7268 C->isTargetOpcode(), C->isOpaque());
7269 case ISD::ABS:
7270 return getConstant(Val.abs(), DL, VT, C->isTargetOpcode(),
7271 C->isOpaque());
7272 case ISD::BITREVERSE:
7273 return getConstant(Val.reverseBits(), DL, VT, C->isTargetOpcode(),
7274 C->isOpaque());
7275 case ISD::BSWAP:
7276 return getConstant(Val.byteSwap(), DL, VT, C->isTargetOpcode(),
7277 C->isOpaque());
7278 case ISD::CTPOP:
7279 return getConstant(Val.popcount(), DL, VT, C->isTargetOpcode(),
7280 C->isOpaque());
7281 case ISD::CTLZ:
7283 return getConstant(Val.countl_zero(), DL, VT, C->isTargetOpcode(),
7284 C->isOpaque());
7285 case ISD::CTTZ:
7287 return getConstant(Val.countr_zero(), DL, VT, C->isTargetOpcode(),
7288 C->isOpaque());
7289 case ISD::CTLS:
7290 // CTLS returns the number of extra sign bits so subtract one.
7291 return getConstant(Val.getNumSignBits() - 1, DL, VT,
7292 C->isTargetOpcode(), C->isOpaque());
7293 case ISD::UINT_TO_FP:
7294 case ISD::SINT_TO_FP: {
7296 (void)FPV.convertFromAPInt(Val, Opcode == ISD::SINT_TO_FP,
7298 return getConstantFP(FPV, DL, VT);
7299 }
7300 case ISD::FP16_TO_FP:
7301 case ISD::BF16_TO_FP: {
7302 bool Ignored;
7303 APFloat FPV(Opcode == ISD::FP16_TO_FP ? APFloat::IEEEhalf()
7304 : APFloat::BFloat(),
7305 (Val.getBitWidth() == 16) ? Val : Val.trunc(16));
7306
7307 // This can return overflow, underflow, or inexact; we don't care.
7308 // FIXME need to be more flexible about rounding mode.
7310 &Ignored);
7311 return getConstantFP(FPV, DL, VT);
7312 }
7313 case ISD::STEP_VECTOR:
7314 if (SDValue V = FoldSTEP_VECTOR(DL, VT, N1, *this))
7315 return V;
7316 break;
7317 case ISD::BITCAST:
7318 if (VT == MVT::f16 && C->getValueType(0) == MVT::i16)
7319 return getConstantFP(APFloat(APFloat::IEEEhalf(), Val), DL, VT);
7320 if (VT == MVT::f32 && C->getValueType(0) == MVT::i32)
7321 return getConstantFP(APFloat(APFloat::IEEEsingle(), Val), DL, VT);
7322 if (VT == MVT::f64 && C->getValueType(0) == MVT::i64)
7323 return getConstantFP(APFloat(APFloat::IEEEdouble(), Val), DL, VT);
7324 if (VT == MVT::f128 && C->getValueType(0) == MVT::i128)
7325 return getConstantFP(APFloat(APFloat::IEEEquad(), Val), DL, VT);
7326 break;
7327 }
7328 }
7329
7330 // Constant fold unary operations with a floating point constant operand.
7331 if (auto *C = dyn_cast<ConstantFPSDNode>(N1)) {
7332 APFloat V = C->getValueAPF(); // make copy
7333 switch (Opcode) {
7334 case ISD::FNEG:
7335 V.changeSign();
7336 return getConstantFP(V, DL, VT);
7337 case ISD::FABS:
7338 V.clearSign();
7339 return getConstantFP(V, DL, VT);
7340 case ISD::FCEIL: {
7341 APFloat::opStatus fs = V.roundToIntegral(APFloat::rmTowardPositive);
7343 return getConstantFP(V, DL, VT);
7344 return SDValue();
7345 }
7346 case ISD::FTRUNC: {
7347 APFloat::opStatus fs = V.roundToIntegral(APFloat::rmTowardZero);
7349 return getConstantFP(V, DL, VT);
7350 return SDValue();
7351 }
7352 case ISD::FFLOOR: {
7353 APFloat::opStatus fs = V.roundToIntegral(APFloat::rmTowardNegative);
7355 return getConstantFP(V, DL, VT);
7356 return SDValue();
7357 }
7358 case ISD::FP_EXTEND: {
7359 bool ignored;
7360 // This can return overflow, underflow, or inexact; we don't care.
7361 // FIXME need to be more flexible about rounding mode.
7362 (void)V.convert(VT.getFltSemantics(), APFloat::rmNearestTiesToEven,
7363 &ignored);
7364 return getConstantFP(V, DL, VT);
7365 }
7366 case ISD::FP_TO_SINT:
7367 case ISD::FP_TO_UINT: {
7368 bool ignored;
7369 APSInt IntVal(VT.getSizeInBits(), Opcode == ISD::FP_TO_UINT);
7370 // FIXME need to be more flexible about rounding mode.
7372 V.convertToInteger(IntVal, APFloat::rmTowardZero, &ignored);
7373 if (s == APFloat::opInvalidOp) // inexact is OK, in fact usual
7374 break;
7375 return getConstant(IntVal, DL, VT);
7376 }
7377 case ISD::FP_TO_FP16:
7378 case ISD::FP_TO_BF16: {
7379 bool Ignored;
7380 // This can return overflow, underflow, or inexact; we don't care.
7381 // FIXME need to be more flexible about rounding mode.
7382 (void)V.convert(Opcode == ISD::FP_TO_FP16 ? APFloat::IEEEhalf()
7383 : APFloat::BFloat(),
7385 return getConstant(V.bitcastToAPInt().getZExtValue(), DL, VT);
7386 }
7387 case ISD::BITCAST:
7388 if (VT == MVT::i16 && C->getValueType(0) == MVT::f16)
7389 return getConstant((uint16_t)V.bitcastToAPInt().getZExtValue(), DL,
7390 VT);
7391 if (VT == MVT::i16 && C->getValueType(0) == MVT::bf16)
7392 return getConstant((uint16_t)V.bitcastToAPInt().getZExtValue(), DL,
7393 VT);
7394 if (VT == MVT::i32 && C->getValueType(0) == MVT::f32)
7395 return getConstant((uint32_t)V.bitcastToAPInt().getZExtValue(), DL,
7396 VT);
7397 if (VT == MVT::i64 && C->getValueType(0) == MVT::f64)
7398 return getConstant(V.bitcastToAPInt().getZExtValue(), DL, VT);
7399 break;
7400 }
7401 }
7402
7403 // Early-out if we failed to constant fold a bitcast.
7404 if (Opcode == ISD::BITCAST)
7405 return SDValue();
7406 }
7407
7408 // Handle binops special cases.
7409 if (NumOps == 2) {
7410 if (SDValue CFP = foldConstantFPMath(Opcode, DL, VT, Ops))
7411 return CFP;
7412
7413 if (auto *C1 = dyn_cast<ConstantSDNode>(Ops[0])) {
7414 if (auto *C2 = dyn_cast<ConstantSDNode>(Ops[1])) {
7415 if (C1->isOpaque() || C2->isOpaque())
7416 return SDValue();
7417
7418 std::optional<APInt> FoldAttempt =
7419 FoldValue(Opcode, C1->getAPIntValue(), C2->getAPIntValue());
7420 if (!FoldAttempt)
7421 return SDValue();
7422
7423 SDValue Folded = getConstant(*FoldAttempt, DL, VT);
7424 assert((!Folded || !VT.isVector()) &&
7425 "Can't fold vectors ops with scalar operands");
7426 return Folded;
7427 }
7428 }
7429
7430 // fold (add Sym, c) -> Sym+c
7432 return FoldSymbolOffset(Opcode, VT, GA, Ops[1].getNode());
7433 if (TLI->isCommutativeBinOp(Opcode))
7435 return FoldSymbolOffset(Opcode, VT, GA, Ops[0].getNode());
7436
7437 // fold (sext_in_reg c1) -> c2
7438 if (Opcode == ISD::SIGN_EXTEND_INREG) {
7439 EVT EVT = cast<VTSDNode>(Ops[1])->getVT();
7440
7441 auto SignExtendInReg = [&](APInt Val, llvm::EVT ConstantVT) {
7442 unsigned FromBits = EVT.getScalarSizeInBits();
7443 Val <<= Val.getBitWidth() - FromBits;
7444 Val.ashrInPlace(Val.getBitWidth() - FromBits);
7445 return getConstant(Val, DL, ConstantVT);
7446 };
7447
7448 if (auto *C1 = dyn_cast<ConstantSDNode>(Ops[0])) {
7449 const APInt &Val = C1->getAPIntValue();
7450 return SignExtendInReg(Val, VT);
7451 }
7452
7454 SmallVector<SDValue, 8> ScalarOps;
7455 llvm::EVT OpVT = Ops[0].getOperand(0).getValueType();
7456 for (int I = 0, E = VT.getVectorNumElements(); I != E; ++I) {
7457 SDValue Op = Ops[0].getOperand(I);
7458 if (Op.isUndef()) {
7459 ScalarOps.push_back(getUNDEF(OpVT));
7460 continue;
7461 }
7462 const APInt &Val = cast<ConstantSDNode>(Op)->getAPIntValue();
7463 ScalarOps.push_back(SignExtendInReg(Val, OpVT));
7464 }
7465 return getBuildVector(VT, DL, ScalarOps);
7466 }
7467
7468 if (Ops[0].getOpcode() == ISD::SPLAT_VECTOR &&
7469 isa<ConstantSDNode>(Ops[0].getOperand(0)))
7470 return getNode(ISD::SPLAT_VECTOR, DL, VT,
7471 SignExtendInReg(Ops[0].getConstantOperandAPInt(0),
7472 Ops[0].getOperand(0).getValueType()));
7473 }
7474 }
7475
7476 // Handle fshl/fshr special cases.
7477 if (Opcode == ISD::FSHL || Opcode == ISD::FSHR) {
7478 auto *C1 = dyn_cast<ConstantSDNode>(Ops[0]);
7479 auto *C2 = dyn_cast<ConstantSDNode>(Ops[1]);
7480 auto *C3 = dyn_cast<ConstantSDNode>(Ops[2]);
7481
7482 if (C1 && C2 && C3) {
7483 if (C1->isOpaque() || C2->isOpaque() || C3->isOpaque())
7484 return SDValue();
7485 const APInt &V1 = C1->getAPIntValue(), &V2 = C2->getAPIntValue(),
7486 &V3 = C3->getAPIntValue();
7487
7488 APInt FoldedVal = Opcode == ISD::FSHL ? APIntOps::fshl(V1, V2, V3)
7489 : APIntOps::fshr(V1, V2, V3);
7490 return getConstant(FoldedVal, DL, VT);
7491 }
7492 }
7493
7494 // Handle fma/fmad special cases.
7495 if (Opcode == ISD::FMA || Opcode == ISD::FMAD || Opcode == ISD::FMULADD) {
7496 assert(VT.isFloatingPoint() && "This operator only applies to FP types!");
7497 assert(Ops[0].getValueType() == VT && Ops[1].getValueType() == VT &&
7498 Ops[2].getValueType() == VT && "FMA types must match!");
7502 if (C1 && C2 && C3) {
7503 APFloat V1 = C1->getValueAPF();
7504 const APFloat &V2 = C2->getValueAPF();
7505 const APFloat &V3 = C3->getValueAPF();
7506 if (Opcode == ISD::FMAD || Opcode == ISD::FMULADD) {
7509 } else
7511 return getConstantFP(V1, DL, VT);
7512 }
7513 }
7514
7515 // This is for vector folding only from here on.
7516 if (!VT.isVector())
7517 return SDValue();
7518
7519 ElementCount NumElts = VT.getVectorElementCount();
7520
7521 // See if we can fold through any bitcasted integer ops.
7522 if (NumOps == 2 && VT.isFixedLengthVector() && VT.isInteger() &&
7523 Ops[0].getValueType() == VT && Ops[1].getValueType() == VT &&
7524 (Ops[0].getOpcode() == ISD::BITCAST ||
7525 Ops[1].getOpcode() == ISD::BITCAST)) {
7528 auto *BV1 = dyn_cast<BuildVectorSDNode>(N1);
7529 auto *BV2 = dyn_cast<BuildVectorSDNode>(N2);
7530 if (BV1 && BV2 && N1.getValueType().isInteger() &&
7531 N2.getValueType().isInteger()) {
7532 bool IsLE = getDataLayout().isLittleEndian();
7533 unsigned EltBits = VT.getScalarSizeInBits();
7534 SmallVector<APInt> RawBits1, RawBits2;
7535 BitVector UndefElts1, UndefElts2;
7536 if (BV1->getConstantRawBits(IsLE, EltBits, RawBits1, UndefElts1) &&
7537 BV2->getConstantRawBits(IsLE, EltBits, RawBits2, UndefElts2)) {
7538 SmallVector<APInt> RawBits;
7539 for (unsigned I = 0, E = NumElts.getFixedValue(); I != E; ++I) {
7540 std::optional<APInt> Fold = FoldValueWithUndef(
7541 Opcode, RawBits1[I], UndefElts1[I], RawBits2[I], UndefElts2[I]);
7542 if (!Fold)
7543 break;
7544 RawBits.push_back(*Fold);
7545 }
7546 if (RawBits.size() == NumElts.getFixedValue()) {
7547 // We have constant folded, but we might need to cast this again back
7548 // to the original (possibly legalized) type.
7549 EVT BVVT, BVEltVT;
7550 if (N1.getValueType() == VT) {
7551 BVVT = N1.getValueType();
7552 BVEltVT = BV1->getOperand(0).getValueType();
7553 } else {
7554 BVVT = N2.getValueType();
7555 BVEltVT = BV2->getOperand(0).getValueType();
7556 }
7557 unsigned BVEltBits = BVEltVT.getSizeInBits();
7558 SmallVector<APInt> DstBits;
7559 BitVector DstUndefs;
7561 DstBits, RawBits, DstUndefs,
7562 BitVector(RawBits.size(), false));
7563 SmallVector<SDValue> Ops(DstBits.size(), getUNDEF(BVEltVT));
7564 for (unsigned I = 0, E = DstBits.size(); I != E; ++I) {
7565 if (DstUndefs[I])
7566 continue;
7567 Ops[I] = getConstant(DstBits[I].sext(BVEltBits), DL, BVEltVT);
7568 }
7569 return getBitcast(VT, getBuildVector(BVVT, DL, Ops));
7570 }
7571 }
7572 }
7573 }
7574
7575 // Fold (mul step_vector(C0), C1) to (step_vector(C0 * C1)).
7576 // (shl step_vector(C0), C1) -> (step_vector(C0 << C1))
7577 if ((Opcode == ISD::MUL || Opcode == ISD::SHL) &&
7578 Ops[0].getOpcode() == ISD::STEP_VECTOR) {
7579 APInt RHSVal;
7580 if (ISD::isConstantSplatVector(Ops[1].getNode(), RHSVal)) {
7581 APInt NewStep = Opcode == ISD::MUL
7582 ? Ops[0].getConstantOperandAPInt(0) * RHSVal
7583 : Ops[0].getConstantOperandAPInt(0) << RHSVal;
7584 return getStepVector(DL, VT, NewStep);
7585 }
7586 }
7587
7588 auto IsScalarOrSameVectorSize = [NumElts](const SDValue &Op) {
7589 return !Op.getValueType().isVector() ||
7590 Op.getValueType().getVectorElementCount() == NumElts;
7591 };
7592
7593 auto IsBuildVectorSplatVectorOrUndef = [](const SDValue &Op) {
7594 return Op.isUndef() || Op.getOpcode() == ISD::CONDCODE ||
7595 Op.getOpcode() == ISD::BUILD_VECTOR ||
7596 Op.getOpcode() == ISD::SPLAT_VECTOR;
7597 };
7598
7599 // All operands must be vector types with the same number of elements as
7600 // the result type and must be either UNDEF or a build/splat vector
7601 // or UNDEF scalars.
7602 if (!llvm::all_of(Ops, IsBuildVectorSplatVectorOrUndef) ||
7603 !llvm::all_of(Ops, IsScalarOrSameVectorSize))
7604 return SDValue();
7605
7606 // If we are comparing vectors, then the result needs to be a i1 boolean that
7607 // is then extended back to the legal result type depending on how booleans
7608 // are represented.
7609 EVT SVT = (Opcode == ISD::SETCC ? MVT::i1 : VT.getScalarType());
7610 ISD::NodeType ExtendCode =
7611 (Opcode == ISD::SETCC && SVT != VT.getScalarType())
7612 ? TargetLowering::getExtendForContent(TLI->getBooleanContents(VT))
7614
7615 // Find legal integer scalar type for constant promotion and
7616 // ensure that its scalar size is at least as large as source.
7617 EVT LegalSVT = VT.getScalarType();
7618 if (NewNodesMustHaveLegalTypes && LegalSVT.isInteger()) {
7619 LegalSVT = TLI->getTypeToTransformTo(*getContext(), LegalSVT);
7620 if (LegalSVT.bitsLT(VT.getScalarType()))
7621 return SDValue();
7622 }
7623
7624 // For scalable vector types we know we're dealing with SPLAT_VECTORs. We
7625 // only have one operand to check. For fixed-length vector types we may have
7626 // a combination of BUILD_VECTOR and SPLAT_VECTOR.
7627 unsigned NumVectorElts = NumElts.isScalable() ? 1 : NumElts.getFixedValue();
7628
7629 // Constant fold each scalar lane separately.
7630 SmallVector<SDValue, 4> ScalarResults;
7631 for (unsigned I = 0; I != NumVectorElts; I++) {
7632 SmallVector<SDValue, 4> ScalarOps;
7633 for (SDValue Op : Ops) {
7634 EVT InSVT = Op.getValueType().getScalarType();
7635 if (Op.getOpcode() != ISD::BUILD_VECTOR &&
7636 Op.getOpcode() != ISD::SPLAT_VECTOR) {
7637 if (Op.isUndef())
7638 ScalarOps.push_back(getUNDEF(InSVT));
7639 else
7640 ScalarOps.push_back(Op);
7641 continue;
7642 }
7643
7644 SDValue ScalarOp =
7645 Op.getOperand(Op.getOpcode() == ISD::SPLAT_VECTOR ? 0 : I);
7646 EVT ScalarVT = ScalarOp.getValueType();
7647
7648 // Build vector (integer) scalar operands may need implicit
7649 // truncation - do this before constant folding.
7650 if (ScalarVT.isInteger() && ScalarVT.bitsGT(InSVT)) {
7651 // Don't create illegally-typed nodes unless they're constants or undef
7652 // - if we fail to constant fold we can't guarantee the (dead) nodes
7653 // we're creating will be cleaned up before being visited for
7654 // legalization.
7655 if (NewNodesMustHaveLegalTypes && !ScalarOp.isUndef() &&
7656 !isa<ConstantSDNode>(ScalarOp) &&
7657 TLI->getTypeAction(*getContext(), InSVT) !=
7659 return SDValue();
7660 ScalarOp = getNode(ISD::TRUNCATE, DL, InSVT, ScalarOp);
7661 }
7662
7663 ScalarOps.push_back(ScalarOp);
7664 }
7665
7666 // Constant fold the scalar operands.
7667 SDValue ScalarResult = getNode(Opcode, DL, SVT, ScalarOps, Flags);
7668
7669 // Scalar folding only succeeded if the result is a constant or UNDEF.
7670 if (!ScalarResult.isUndef() && ScalarResult.getOpcode() != ISD::Constant &&
7671 ScalarResult.getOpcode() != ISD::ConstantFP)
7672 return SDValue();
7673
7674 // Legalize the (integer) scalar constant if necessary. We only do
7675 // this once we know the folding succeeded, since otherwise we would
7676 // get a node with illegal type which has a user.
7677 if (LegalSVT != SVT)
7678 ScalarResult = getNode(ExtendCode, DL, LegalSVT, ScalarResult);
7679
7680 ScalarResults.push_back(ScalarResult);
7681 }
7682
7683 SDValue V = NumElts.isScalable() ? getSplatVector(VT, DL, ScalarResults[0])
7684 : getBuildVector(VT, DL, ScalarResults);
7685 NewSDValueDbgMsg(V, "New node fold constant vector: ", this);
7686 return V;
7687}
7688
7691 // TODO: Add support for unary/ternary fp opcodes.
7692 if (Ops.size() != 2)
7693 return SDValue();
7694
7695 // TODO: We don't do any constant folding for strict FP opcodes here, but we
7696 // should. That will require dealing with a potentially non-default
7697 // rounding mode, checking the "opStatus" return value from the APFloat
7698 // math calculations, and possibly other variations.
7699 SDValue N1 = Ops[0];
7700 SDValue N2 = Ops[1];
7701 ConstantFPSDNode *N1CFP = isConstOrConstSplatFP(N1, /*AllowUndefs*/ false);
7702 ConstantFPSDNode *N2CFP = isConstOrConstSplatFP(N2, /*AllowUndefs*/ false);
7703 if (N1CFP && N2CFP) {
7704 APFloat C1 = N1CFP->getValueAPF(); // make copy
7705 const APFloat &C2 = N2CFP->getValueAPF();
7706 switch (Opcode) {
7707 case ISD::FADD:
7709 return getConstantFP(C1, DL, VT);
7710 case ISD::FSUB:
7712 return getConstantFP(C1, DL, VT);
7713 case ISD::FMUL:
7715 return getConstantFP(C1, DL, VT);
7716 case ISD::FDIV:
7718 return getConstantFP(C1, DL, VT);
7719 case ISD::FREM:
7720 C1.mod(C2);
7721 return getConstantFP(C1, DL, VT);
7722 case ISD::FCOPYSIGN:
7723 C1.copySign(C2);
7724 return getConstantFP(C1, DL, VT);
7725 case ISD::FMINNUM:
7726 return getConstantFP(minnum(C1, C2), DL, VT);
7727 case ISD::FMAXNUM:
7728 return getConstantFP(maxnum(C1, C2), DL, VT);
7729 case ISD::FMINIMUM:
7730 return getConstantFP(minimum(C1, C2), DL, VT);
7731 case ISD::FMAXIMUM:
7732 return getConstantFP(maximum(C1, C2), DL, VT);
7733 case ISD::FMINIMUMNUM:
7734 return getConstantFP(minimumnum(C1, C2), DL, VT);
7735 case ISD::FMAXIMUMNUM:
7736 return getConstantFP(maximumnum(C1, C2), DL, VT);
7737 default: break;
7738 }
7739 }
7740 if (N1CFP && Opcode == ISD::FP_ROUND) {
7741 APFloat C1 = N1CFP->getValueAPF(); // make copy
7742 bool Unused;
7743 // This can return overflow, underflow, or inexact; we don't care.
7744 // FIXME need to be more flexible about rounding mode.
7746 &Unused);
7747 return getConstantFP(C1, DL, VT);
7748 }
7749
7750 switch (Opcode) {
7751 case ISD::FSUB:
7752 // -0.0 - undef --> undef (consistent with "fneg undef")
7753 if (ConstantFPSDNode *N1C = isConstOrConstSplatFP(N1, /*AllowUndefs*/ true))
7754 if (N1C && N1C->getValueAPF().isNegZero() && N2.isUndef())
7755 return getUNDEF(VT);
7756 [[fallthrough]];
7757
7758 case ISD::FADD:
7759 case ISD::FMUL:
7760 case ISD::FDIV:
7761 case ISD::FREM:
7762 // If both operands are undef, the result is undef. If 1 operand is undef,
7763 // the result is NaN. This should match the behavior of the IR optimizer.
7764 if (N1.isUndef() && N2.isUndef())
7765 return getUNDEF(VT);
7766 if (N1.isUndef() || N2.isUndef())
7768 }
7769 return SDValue();
7770}
7771
7773 const SDLoc &DL, EVT DstEltVT) {
7774 EVT SrcEltVT = BV->getValueType(0).getVectorElementType();
7775
7776 // If this is already the right type, we're done.
7777 if (SrcEltVT == DstEltVT)
7778 return SDValue(BV, 0);
7779
7780 unsigned SrcBitSize = SrcEltVT.getSizeInBits();
7781 unsigned DstBitSize = DstEltVT.getSizeInBits();
7782
7783 // If this is a conversion of N elements of one type to N elements of another
7784 // type, convert each element. This handles FP<->INT cases.
7785 if (SrcBitSize == DstBitSize) {
7787 for (SDValue Op : BV->op_values()) {
7788 // If the vector element type is not legal, the BUILD_VECTOR operands
7789 // are promoted and implicitly truncated. Make that explicit here.
7790 if (Op.getValueType() != SrcEltVT)
7791 Op = getNode(ISD::TRUNCATE, DL, SrcEltVT, Op);
7792 Ops.push_back(getBitcast(DstEltVT, Op));
7793 }
7794 EVT VT = EVT::getVectorVT(*getContext(), DstEltVT,
7796 return getBuildVector(VT, DL, Ops);
7797 }
7798
7799 // Otherwise, we're growing or shrinking the elements. To avoid having to
7800 // handle annoying details of growing/shrinking FP values, we convert them to
7801 // int first.
7802 if (SrcEltVT.isFloatingPoint()) {
7803 // Convert the input float vector to a int vector where the elements are the
7804 // same sizes.
7805 EVT IntEltVT = EVT::getIntegerVT(*getContext(), SrcEltVT.getSizeInBits());
7806 if (SDValue Tmp = FoldConstantBuildVector(BV, DL, IntEltVT))
7808 DstEltVT);
7809 return SDValue();
7810 }
7811
7812 // Now we know the input is an integer vector. If the output is a FP type,
7813 // convert to integer first, then to FP of the right size.
7814 if (DstEltVT.isFloatingPoint()) {
7815 EVT IntEltVT = EVT::getIntegerVT(*getContext(), DstEltVT.getSizeInBits());
7816 if (SDValue Tmp = FoldConstantBuildVector(BV, DL, IntEltVT))
7818 DstEltVT);
7819 return SDValue();
7820 }
7821
7822 // Okay, we know the src/dst types are both integers of differing types.
7823 assert(SrcEltVT.isInteger() && DstEltVT.isInteger());
7824
7825 // Extract the constant raw bit data.
7826 BitVector UndefElements;
7827 SmallVector<APInt> RawBits;
7828 bool IsLE = getDataLayout().isLittleEndian();
7829 if (!BV->getConstantRawBits(IsLE, DstBitSize, RawBits, UndefElements))
7830 return SDValue();
7831
7833 for (unsigned I = 0, E = RawBits.size(); I != E; ++I) {
7834 if (UndefElements[I])
7835 Ops.push_back(getUNDEF(DstEltVT));
7836 else
7837 Ops.push_back(getConstant(RawBits[I], DL, DstEltVT));
7838 }
7839
7840 EVT VT = EVT::getVectorVT(*getContext(), DstEltVT, Ops.size());
7841 return getBuildVector(VT, DL, Ops);
7842}
7843
7845 assert(Val.getValueType().isInteger() && "Invalid AssertAlign!");
7846
7847 // There's no need to assert on a byte-aligned pointer. All pointers are at
7848 // least byte aligned.
7849 if (A == Align(1))
7850 return Val;
7851
7852 SDVTList VTs = getVTList(Val.getValueType());
7854 AddNodeIDNode(ID, ISD::AssertAlign, VTs, {Val});
7855 ID.AddInteger(A.value());
7856
7857 void *IP = nullptr;
7858 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP))
7859 return SDValue(E, 0);
7860
7861 auto *N =
7862 newSDNode<AssertAlignSDNode>(DL.getIROrder(), DL.getDebugLoc(), VTs, A);
7863 createOperands(N, {Val});
7864
7865 CSEMap.InsertNode(N, IP);
7866 InsertNode(N);
7867
7868 SDValue V(N, 0);
7869 NewSDValueDbgMsg(V, "Creating new node: ", this);
7870 return V;
7871}
7872
7873SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
7874 SDValue N1, SDValue N2) {
7875 SDNodeFlags Flags;
7876 if (Inserter)
7877 Flags = Inserter->getFlags();
7878 return getNode(Opcode, DL, VT, N1, N2, Flags);
7879}
7880
7882 SDValue &N2) const {
7883 if (!TLI->isCommutativeBinOp(Opcode))
7884 return;
7885
7886 // Canonicalize:
7887 // binop(const, nonconst) -> binop(nonconst, const)
7890 bool N1CFP = isConstantFPBuildVectorOrConstantFP(N1);
7891 bool N2CFP = isConstantFPBuildVectorOrConstantFP(N2);
7892 if ((N1C && !N2C) || (N1CFP && !N2CFP))
7893 std::swap(N1, N2);
7894
7895 // Canonicalize:
7896 // binop(splat(x), step_vector) -> binop(step_vector, splat(x))
7897 else if (N1.getOpcode() == ISD::SPLAT_VECTOR &&
7899 std::swap(N1, N2);
7900}
7901
7902SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
7903 SDValue N1, SDValue N2, const SDNodeFlags Flags) {
7905 N2.getOpcode() != ISD::DELETED_NODE &&
7906 "Operand is DELETED_NODE!");
7907
7908 canonicalizeCommutativeBinop(Opcode, N1, N2);
7909
7910 auto *N1C = dyn_cast<ConstantSDNode>(N1);
7911 auto *N2C = dyn_cast<ConstantSDNode>(N2);
7912
7913 // Don't allow undefs in vector splats - we might be returning N2 when folding
7914 // to zero etc.
7915 ConstantSDNode *N2CV =
7916 isConstOrConstSplat(N2, /*AllowUndefs*/ false, /*AllowTruncation*/ true);
7917
7918 switch (Opcode) {
7919 default: break;
7920 case ISD::TokenFactor:
7921 assert(VT == MVT::Other && N1.getValueType() == MVT::Other &&
7922 N2.getValueType() == MVT::Other && "Invalid token factor!");
7923 // Fold trivial token factors.
7924 if (N1.getOpcode() == ISD::EntryToken) return N2;
7925 if (N2.getOpcode() == ISD::EntryToken) return N1;
7926 if (N1 == N2) return N1;
7927 break;
7928 case ISD::BUILD_VECTOR: {
7929 // Attempt to simplify BUILD_VECTOR.
7930 SDValue Ops[] = {N1, N2};
7931 if (SDValue V = FoldBUILD_VECTOR(DL, VT, Ops, *this))
7932 return V;
7933 break;
7934 }
7935 case ISD::CONCAT_VECTORS: {
7936 SDValue Ops[] = {N1, N2};
7937 if (SDValue V = foldCONCAT_VECTORS(DL, VT, Ops, *this))
7938 return V;
7939 break;
7940 }
7941 case ISD::AND:
7942 assert(VT.isInteger() && "This operator does not apply to FP types!");
7943 assert(N1.getValueType() == N2.getValueType() &&
7944 N1.getValueType() == VT && "Binary operator types must match!");
7945 // (X & 0) -> 0. This commonly occurs when legalizing i64 values, so it's
7946 // worth handling here.
7947 if (N2CV && N2CV->isZero())
7948 return N2;
7949 if (N2CV && N2CV->isAllOnes()) // X & -1 -> X
7950 return N1;
7951 break;
7952 case ISD::OR:
7953 case ISD::XOR:
7954 case ISD::ADD:
7955 case ISD::PTRADD:
7956 case ISD::SUB:
7957 assert(VT.isInteger() && "This operator does not apply to FP types!");
7958 assert(N1.getValueType() == N2.getValueType() &&
7959 N1.getValueType() == VT && "Binary operator types must match!");
7960 // The equal operand types requirement is unnecessarily strong for PTRADD.
7961 // However, the SelectionDAGBuilder does not generate PTRADDs with different
7962 // operand types, and we'd need to re-implement GEP's non-standard wrapping
7963 // logic everywhere where PTRADDs may be folded or combined to properly
7964 // support them. If/when we introduce pointer types to the SDAG, we will
7965 // need to relax this constraint.
7966
7967 // (X ^|+- 0) -> X. This commonly occurs when legalizing i64 values, so
7968 // it's worth handling here.
7969 if (N2CV && N2CV->isZero())
7970 return N1;
7971 if ((Opcode == ISD::ADD || Opcode == ISD::SUB) &&
7972 VT.getScalarType() == MVT::i1)
7973 return getNode(ISD::XOR, DL, VT, N1, N2);
7974 // Fold (add (vscale * C0), (vscale * C1)) to (vscale * (C0 + C1)).
7975 if (Opcode == ISD::ADD && N1.getOpcode() == ISD::VSCALE &&
7976 N2.getOpcode() == ISD::VSCALE) {
7977 const APInt &C1 = N1->getConstantOperandAPInt(0);
7978 const APInt &C2 = N2->getConstantOperandAPInt(0);
7979 return getVScale(DL, VT, C1 + C2);
7980 }
7981 break;
7982 case ISD::MUL:
7983 assert(VT.isInteger() && "This operator does not apply to FP types!");
7984 assert(N1.getValueType() == N2.getValueType() &&
7985 N1.getValueType() == VT && "Binary operator types must match!");
7986 if (VT.getScalarType() == MVT::i1)
7987 return getNode(ISD::AND, DL, VT, N1, N2);
7988 if (N2CV && N2CV->isZero())
7989 return N2;
7990 if (N2C && (N1.getOpcode() == ISD::VSCALE) && Flags.hasNoSignedWrap()) {
7991 const APInt &MulImm = N1->getConstantOperandAPInt(0);
7992 const APInt &N2CImm = N2C->getAPIntValue();
7993 return getVScale(DL, VT, MulImm * N2CImm);
7994 }
7995 break;
7996 case ISD::UDIV:
7997 case ISD::UREM:
7998 case ISD::MULHU:
7999 case ISD::MULHS:
8000 case ISD::SDIV:
8001 case ISD::SREM:
8002 case ISD::SADDSAT:
8003 case ISD::SSUBSAT:
8004 case ISD::UADDSAT:
8005 case ISD::USUBSAT:
8006 assert(VT.isInteger() && "This operator does not apply to FP types!");
8007 assert(N1.getValueType() == N2.getValueType() &&
8008 N1.getValueType() == VT && "Binary operator types must match!");
8009 if (VT.getScalarType() == MVT::i1) {
8010 // fold (add_sat x, y) -> (or x, y) for bool types.
8011 if (Opcode == ISD::SADDSAT || Opcode == ISD::UADDSAT)
8012 return getNode(ISD::OR, DL, VT, N1, N2);
8013 // fold (sub_sat x, y) -> (and x, ~y) for bool types.
8014 if (Opcode == ISD::SSUBSAT || Opcode == ISD::USUBSAT)
8015 return getNode(ISD::AND, DL, VT, N1, getNOT(DL, N2, VT));
8016 }
8017 break;
8018 case ISD::SCMP:
8019 case ISD::UCMP:
8020 assert(N1.getValueType() == N2.getValueType() &&
8021 "Types of operands of UCMP/SCMP must match");
8022 assert(N1.getValueType().isVector() == VT.isVector() &&
8023 "Operands and return type of must both be scalars or vectors");
8024 if (VT.isVector())
8027 "Result and operands must have the same number of elements");
8028 break;
8029 case ISD::AVGFLOORS:
8030 case ISD::AVGFLOORU:
8031 case ISD::AVGCEILS:
8032 case ISD::AVGCEILU:
8033 assert(VT.isInteger() && "This operator does not apply to FP types!");
8034 assert(N1.getValueType() == N2.getValueType() &&
8035 N1.getValueType() == VT && "Binary operator types must match!");
8036 break;
8037 case ISD::ABDS:
8038 case ISD::ABDU:
8039 assert(VT.isInteger() && "This operator does not apply to FP types!");
8040 assert(N1.getValueType() == N2.getValueType() &&
8041 N1.getValueType() == VT && "Binary operator types must match!");
8042 if (VT.getScalarType() == MVT::i1)
8043 return getNode(ISD::XOR, DL, VT, N1, N2);
8044 break;
8045 case ISD::SMIN:
8046 case ISD::UMAX:
8047 assert(VT.isInteger() && "This operator does not apply to FP types!");
8048 assert(N1.getValueType() == N2.getValueType() &&
8049 N1.getValueType() == VT && "Binary operator types must match!");
8050 if (VT.getScalarType() == MVT::i1)
8051 return getNode(ISD::OR, DL, VT, N1, N2);
8052 break;
8053 case ISD::SMAX:
8054 case ISD::UMIN:
8055 assert(VT.isInteger() && "This operator does not apply to FP types!");
8056 assert(N1.getValueType() == N2.getValueType() &&
8057 N1.getValueType() == VT && "Binary operator types must match!");
8058 if (VT.getScalarType() == MVT::i1)
8059 return getNode(ISD::AND, DL, VT, N1, N2);
8060 break;
8061 case ISD::FADD:
8062 case ISD::FSUB:
8063 case ISD::FMUL:
8064 case ISD::FDIV:
8065 case ISD::FREM:
8066 assert(VT.isFloatingPoint() && "This operator only applies to FP types!");
8067 assert(N1.getValueType() == N2.getValueType() &&
8068 N1.getValueType() == VT && "Binary operator types must match!");
8069 if (SDValue V = simplifyFPBinop(Opcode, N1, N2, Flags))
8070 return V;
8071 break;
8072 case ISD::FCOPYSIGN: // N1 and result must match. N1/N2 need not match.
8073 assert(N1.getValueType() == VT &&
8076 "Invalid FCOPYSIGN!");
8077 break;
8078 case ISD::SHL:
8079 if (N2C && (N1.getOpcode() == ISD::VSCALE) && Flags.hasNoSignedWrap()) {
8080 const APInt &MulImm = N1->getConstantOperandAPInt(0);
8081 const APInt &ShiftImm = N2C->getAPIntValue();
8082 return getVScale(DL, VT, MulImm << ShiftImm);
8083 }
8084 [[fallthrough]];
8085 case ISD::SRA:
8086 case ISD::SRL:
8087 if (SDValue V = simplifyShift(N1, N2))
8088 return V;
8089 [[fallthrough]];
8090 case ISD::ROTL:
8091 case ISD::ROTR:
8092 case ISD::SSHLSAT:
8093 case ISD::USHLSAT:
8094 assert(VT == N1.getValueType() &&
8095 "Shift operators return type must be the same as their first arg");
8096 assert(VT.isInteger() && N2.getValueType().isInteger() &&
8097 "Shifts only work on integers");
8098 assert((!VT.isVector() || VT == N2.getValueType()) &&
8099 "Vector shift amounts must be in the same as their first arg");
8100 // Verify that the shift amount VT is big enough to hold valid shift
8101 // amounts. This catches things like trying to shift an i1024 value by an
8102 // i8, which is easy to fall into in generic code that uses
8103 // TLI.getShiftAmount().
8106 "Invalid use of small shift amount with oversized value!");
8107
8108 // Always fold shifts of i1 values so the code generator doesn't need to
8109 // handle them. Since we know the size of the shift has to be less than the
8110 // size of the value, the shift/rotate count is guaranteed to be zero.
8111 if (VT == MVT::i1)
8112 return N1;
8113 if (N2CV && N2CV->isZero())
8114 return N1;
8115 break;
8116 case ISD::FP_ROUND:
8118 VT.bitsLE(N1.getValueType()) && N2C &&
8119 (N2C->getZExtValue() == 0 || N2C->getZExtValue() == 1) &&
8120 N2.getOpcode() == ISD::TargetConstant && "Invalid FP_ROUND!");
8121 if (N1.getValueType() == VT) return N1; // noop conversion.
8122 break;
8123 case ISD::AssertNoFPClass: {
8125 "AssertNoFPClass is used for a non-floating type");
8126 assert(isa<ConstantSDNode>(N2) && "NoFPClass is not Constant");
8127 FPClassTest NoFPClass = static_cast<FPClassTest>(N2->getAsZExtVal());
8128 assert(llvm::to_underlying(NoFPClass) <=
8130 "FPClassTest value too large");
8131 (void)NoFPClass;
8132 break;
8133 }
8134 case ISD::AssertSext:
8135 case ISD::AssertZext: {
8136 EVT EVT = cast<VTSDNode>(N2)->getVT();
8137 assert(VT == N1.getValueType() && "Not an inreg extend!");
8138 assert(VT.isInteger() && EVT.isInteger() &&
8139 "Cannot *_EXTEND_INREG FP types");
8140 assert(!EVT.isVector() &&
8141 "AssertSExt/AssertZExt type should be the vector element type "
8142 "rather than the vector type!");
8143 assert(EVT.bitsLE(VT.getScalarType()) && "Not extending!");
8144 if (VT.getScalarType() == EVT) return N1; // noop assertion.
8145 break;
8146 }
8148 EVT EVT = cast<VTSDNode>(N2)->getVT();
8149 assert(VT == N1.getValueType() && "Not an inreg extend!");
8150 assert(VT.isInteger() && EVT.isInteger() &&
8151 "Cannot *_EXTEND_INREG FP types");
8152 assert(EVT.isVector() == VT.isVector() &&
8153 "SIGN_EXTEND_INREG type should be vector iff the operand "
8154 "type is vector!");
8155 assert((!EVT.isVector() ||
8157 "Vector element counts must match in SIGN_EXTEND_INREG");
8158 assert(EVT.getScalarType().bitsLE(VT.getScalarType()) && "Not extending!");
8159 if (EVT == VT) return N1; // Not actually extending
8160 break;
8161 }
8163 case ISD::FP_TO_UINT_SAT: {
8164 assert(VT.isInteger() && cast<VTSDNode>(N2)->getVT().isInteger() &&
8165 N1.getValueType().isFloatingPoint() && "Invalid FP_TO_*INT_SAT");
8166 assert(N1.getValueType().isVector() == VT.isVector() &&
8167 "FP_TO_*INT_SAT type should be vector iff the operand type is "
8168 "vector!");
8169 assert((!VT.isVector() || VT.getVectorElementCount() ==
8171 "Vector element counts must match in FP_TO_*INT_SAT");
8172 assert(!cast<VTSDNode>(N2)->getVT().isVector() &&
8173 "Type to saturate to must be a scalar.");
8174 assert(cast<VTSDNode>(N2)->getVT().bitsLE(VT.getScalarType()) &&
8175 "Not extending!");
8176 break;
8177 }
8180 "The result of EXTRACT_VECTOR_ELT must be at least as wide as the \
8181 element type of the vector.");
8182
8183 // Extract from an undefined value or using an undefined index is undefined.
8184 if (N1.isUndef() || N2.isUndef())
8185 return getUNDEF(VT);
8186
8187 // EXTRACT_VECTOR_ELT of out-of-bounds element is an UNDEF for fixed length
8188 // vectors. For scalable vectors we will provide appropriate support for
8189 // dealing with arbitrary indices.
8190 if (N2C && N1.getValueType().isFixedLengthVector() &&
8191 N2C->getAPIntValue().uge(N1.getValueType().getVectorNumElements()))
8192 return getUNDEF(VT);
8193
8194 // EXTRACT_VECTOR_ELT of CONCAT_VECTORS is often formed while lowering is
8195 // expanding copies of large vectors from registers. This only works for
8196 // fixed length vectors, since we need to know the exact number of
8197 // elements.
8198 if (N2C && N1.getOpcode() == ISD::CONCAT_VECTORS &&
8200 unsigned Factor = N1.getOperand(0).getValueType().getVectorNumElements();
8201 return getExtractVectorElt(DL, VT,
8202 N1.getOperand(N2C->getZExtValue() / Factor),
8203 N2C->getZExtValue() % Factor);
8204 }
8205
8206 // EXTRACT_VECTOR_ELT of BUILD_VECTOR or SPLAT_VECTOR is often formed while
8207 // lowering is expanding large vector constants.
8208 if (N2C && (N1.getOpcode() == ISD::BUILD_VECTOR ||
8209 N1.getOpcode() == ISD::SPLAT_VECTOR)) {
8212 "BUILD_VECTOR used for scalable vectors");
8213 unsigned Index =
8214 N1.getOpcode() == ISD::BUILD_VECTOR ? N2C->getZExtValue() : 0;
8215 SDValue Elt = N1.getOperand(Index);
8216
8217 if (VT != Elt.getValueType())
8218 // If the vector element type is not legal, the BUILD_VECTOR operands
8219 // are promoted and implicitly truncated, and the result implicitly
8220 // extended. Make that explicit here.
8221 Elt = getAnyExtOrTrunc(Elt, DL, VT);
8222
8223 return Elt;
8224 }
8225
8226 // EXTRACT_VECTOR_ELT of INSERT_VECTOR_ELT is often formed when vector
8227 // operations are lowered to scalars.
8228 if (N1.getOpcode() == ISD::INSERT_VECTOR_ELT) {
8229 // If the indices are the same, return the inserted element else
8230 // if the indices are known different, extract the element from
8231 // the original vector.
8232 SDValue N1Op2 = N1.getOperand(2);
8234
8235 if (N1Op2C && N2C) {
8236 if (N1Op2C->getZExtValue() == N2C->getZExtValue()) {
8237 if (VT == N1.getOperand(1).getValueType())
8238 return N1.getOperand(1);
8239 if (VT.isFloatingPoint()) {
8241 return getFPExtendOrRound(N1.getOperand(1), DL, VT);
8242 }
8243 return getSExtOrTrunc(N1.getOperand(1), DL, VT);
8244 }
8245 return getNode(ISD::EXTRACT_VECTOR_ELT, DL, VT, N1.getOperand(0), N2);
8246 }
8247 }
8248
8249 // EXTRACT_VECTOR_ELT of v1iX EXTRACT_SUBVECTOR could be formed
8250 // when vector types are scalarized and v1iX is legal.
8251 // vextract (v1iX extract_subvector(vNiX, Idx)) -> vextract(vNiX,Idx).
8252 // Here we are completely ignoring the extract element index (N2),
8253 // which is fine for fixed width vectors, since any index other than 0
8254 // is undefined anyway. However, this cannot be ignored for scalable
8255 // vectors - in theory we could support this, but we don't want to do this
8256 // without a profitability check.
8257 if (N1.getOpcode() == ISD::EXTRACT_SUBVECTOR &&
8259 N1.getValueType().getVectorNumElements() == 1) {
8260 return getNode(ISD::EXTRACT_VECTOR_ELT, DL, VT, N1.getOperand(0),
8261 N1.getOperand(1));
8262 }
8263 break;
8265 assert(N2C && (unsigned)N2C->getZExtValue() < 2 && "Bad EXTRACT_ELEMENT!");
8266 assert(!N1.getValueType().isVector() && !VT.isVector() &&
8267 (N1.getValueType().isInteger() == VT.isInteger()) &&
8268 N1.getValueType() != VT &&
8269 "Wrong types for EXTRACT_ELEMENT!");
8270
8271 // EXTRACT_ELEMENT of BUILD_PAIR is often formed while legalize is expanding
8272 // 64-bit integers into 32-bit parts. Instead of building the extract of
8273 // the BUILD_PAIR, only to have legalize rip it apart, just do it now.
8274 if (N1.getOpcode() == ISD::BUILD_PAIR)
8275 return N1.getOperand(N2C->getZExtValue());
8276
8277 // EXTRACT_ELEMENT of a constant int is also very common.
8278 if (N1C) {
8279 unsigned ElementSize = VT.getSizeInBits();
8280 unsigned Shift = ElementSize * N2C->getZExtValue();
8281 const APInt &Val = N1C->getAPIntValue();
8282 return getConstant(Val.extractBits(ElementSize, Shift), DL, VT);
8283 }
8284 break;
8286 EVT N1VT = N1.getValueType();
8287 assert(VT.isVector() && N1VT.isVector() &&
8288 "Extract subvector VTs must be vectors!");
8290 "Extract subvector VTs must have the same element type!");
8291 assert((VT.isFixedLengthVector() || N1VT.isScalableVector()) &&
8292 "Cannot extract a scalable vector from a fixed length vector!");
8293 assert((VT.isScalableVector() != N1VT.isScalableVector() ||
8295 "Extract subvector must be from larger vector to smaller vector!");
8296 assert(N2C && "Extract subvector index must be a constant");
8297 assert((VT.isScalableVector() != N1VT.isScalableVector() ||
8298 (VT.getVectorMinNumElements() + N2C->getZExtValue()) <=
8299 N1VT.getVectorMinNumElements()) &&
8300 "Extract subvector overflow!");
8301 assert(N2C->getAPIntValue().getBitWidth() ==
8302 TLI->getVectorIdxWidth(getDataLayout()) &&
8303 "Constant index for EXTRACT_SUBVECTOR has an invalid size");
8304 assert(N2C->getZExtValue() % VT.getVectorMinNumElements() == 0 &&
8305 "Extract index is not a multiple of the output vector length");
8306
8307 // Trivial extraction.
8308 if (VT == N1VT)
8309 return N1;
8310
8311 // EXTRACT_SUBVECTOR of an UNDEF is an UNDEF.
8312 if (N1.isUndef())
8313 return getUNDEF(VT);
8314
8315 // EXTRACT_SUBVECTOR of CONCAT_VECTOR can be simplified if the pieces of
8316 // the concat have the same type as the extract.
8317 if (N1.getOpcode() == ISD::CONCAT_VECTORS &&
8318 VT == N1.getOperand(0).getValueType()) {
8319 unsigned Factor = VT.getVectorMinNumElements();
8320 return N1.getOperand(N2C->getZExtValue() / Factor);
8321 }
8322
8323 // EXTRACT_SUBVECTOR of INSERT_SUBVECTOR is often created
8324 // during shuffle legalization.
8325 if (N1.getOpcode() == ISD::INSERT_SUBVECTOR && N2 == N1.getOperand(2) &&
8326 VT == N1.getOperand(1).getValueType())
8327 return N1.getOperand(1);
8328 break;
8329 }
8330 }
8331
8332 if (N1.getOpcode() == ISD::POISON || N2.getOpcode() == ISD::POISON) {
8333 switch (Opcode) {
8334 case ISD::XOR:
8335 case ISD::ADD:
8336 case ISD::PTRADD:
8337 case ISD::SUB:
8339 case ISD::UDIV:
8340 case ISD::SDIV:
8341 case ISD::UREM:
8342 case ISD::SREM:
8343 case ISD::MUL:
8344 case ISD::AND:
8345 case ISD::SSUBSAT:
8346 case ISD::USUBSAT:
8347 case ISD::UMIN:
8348 case ISD::OR:
8349 case ISD::SADDSAT:
8350 case ISD::UADDSAT:
8351 case ISD::UMAX:
8352 case ISD::SMAX:
8353 case ISD::SMIN:
8354 // fold op(arg1, poison) -> poison, fold op(poison, arg2) -> poison.
8355 return N2.getOpcode() == ISD::POISON ? N2 : N1;
8356 }
8357 }
8358
8359 // Canonicalize an UNDEF to the RHS, even over a constant.
8360 if (N1.getOpcode() == ISD::UNDEF && N2.getOpcode() != ISD::UNDEF) {
8361 if (TLI->isCommutativeBinOp(Opcode)) {
8362 std::swap(N1, N2);
8363 } else {
8364 switch (Opcode) {
8365 case ISD::PTRADD:
8366 case ISD::SUB:
8367 // fold op(undef, non_undef_arg2) -> undef.
8368 return N1;
8370 case ISD::UDIV:
8371 case ISD::SDIV:
8372 case ISD::UREM:
8373 case ISD::SREM:
8374 case ISD::SSUBSAT:
8375 case ISD::USUBSAT:
8376 // fold op(undef, non_undef_arg2) -> 0.
8377 return getConstant(0, DL, VT);
8378 }
8379 }
8380 }
8381
8382 // Fold a bunch of operators when the RHS is undef.
8383 if (N2.getOpcode() == ISD::UNDEF) {
8384 switch (Opcode) {
8385 case ISD::XOR:
8386 if (N1.getOpcode() == ISD::UNDEF)
8387 // Handle undef ^ undef -> 0 special case. This is a common
8388 // idiom (misuse).
8389 return getConstant(0, DL, VT);
8390 [[fallthrough]];
8391 case ISD::ADD:
8392 case ISD::PTRADD:
8393 case ISD::SUB:
8394 // fold op(arg1, undef) -> undef.
8395 return N2;
8396 case ISD::UDIV:
8397 case ISD::SDIV:
8398 case ISD::UREM:
8399 case ISD::SREM:
8400 // fold op(arg1, undef) -> poison.
8401 return getPOISON(VT);
8402 case ISD::MUL:
8403 case ISD::AND:
8404 case ISD::SSUBSAT:
8405 case ISD::USUBSAT:
8406 case ISD::UMIN:
8407 // fold op(undef, undef) -> undef, fold op(arg1, undef) -> 0.
8408 return N1.getOpcode() == ISD::UNDEF ? N2 : getConstant(0, DL, VT);
8409 case ISD::OR:
8410 case ISD::SADDSAT:
8411 case ISD::UADDSAT:
8412 case ISD::UMAX:
8413 // fold op(undef, undef) -> undef, fold op(arg1, undef) -> -1.
8414 return N1.getOpcode() == ISD::UNDEF ? N2 : getAllOnesConstant(DL, VT);
8415 case ISD::SMAX:
8416 // fold op(undef, undef) -> undef, fold op(arg1, undef) -> MAX_INT.
8417 return N1.getOpcode() == ISD::UNDEF
8418 ? N2
8419 : getConstant(
8421 VT);
8422 case ISD::SMIN:
8423 // fold op(undef, undef) -> undef, fold op(arg1, undef) -> MIN_INT.
8424 return N1.getOpcode() == ISD::UNDEF
8425 ? N2
8426 : getConstant(
8428 VT);
8429 }
8430 }
8431
8432 // Perform trivial constant folding.
8433 if (SDValue SV = FoldConstantArithmetic(Opcode, DL, VT, {N1, N2}, Flags))
8434 return SV;
8435
8436 // Memoize this node if possible.
8437 SDNode *N;
8438 SDVTList VTs = getVTList(VT);
8439 SDValue Ops[] = {N1, N2};
8440 if (VT != MVT::Glue) {
8442 AddNodeIDNode(ID, Opcode, VTs, Ops);
8443 void *IP = nullptr;
8444 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
8445 E->intersectFlagsWith(Flags);
8446 return SDValue(E, 0);
8447 }
8448
8449 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
8450 N->setFlags(Flags);
8451 createOperands(N, Ops);
8452 CSEMap.InsertNode(N, IP);
8453 } else {
8454 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
8455 createOperands(N, Ops);
8456 }
8457
8458 InsertNode(N);
8459 SDValue V = SDValue(N, 0);
8460 NewSDValueDbgMsg(V, "Creating new node: ", this);
8461 return V;
8462}
8463
8464SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8465 SDValue N1, SDValue N2, SDValue N3) {
8466 SDNodeFlags Flags;
8467 if (Inserter)
8468 Flags = Inserter->getFlags();
8469 return getNode(Opcode, DL, VT, N1, N2, N3, Flags);
8470}
8471
8472SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8473 SDValue N1, SDValue N2, SDValue N3,
8474 const SDNodeFlags Flags) {
8476 N2.getOpcode() != ISD::DELETED_NODE &&
8477 N3.getOpcode() != ISD::DELETED_NODE &&
8478 "Operand is DELETED_NODE!");
8479 // Perform various simplifications.
8480 switch (Opcode) {
8481 case ISD::BUILD_VECTOR: {
8482 // Attempt to simplify BUILD_VECTOR.
8483 SDValue Ops[] = {N1, N2, N3};
8484 if (SDValue V = FoldBUILD_VECTOR(DL, VT, Ops, *this))
8485 return V;
8486 break;
8487 }
8488 case ISD::CONCAT_VECTORS: {
8489 SDValue Ops[] = {N1, N2, N3};
8490 if (SDValue V = foldCONCAT_VECTORS(DL, VT, Ops, *this))
8491 return V;
8492 break;
8493 }
8494 case ISD::SETCC: {
8495 assert(VT.isInteger() && "SETCC result type must be an integer!");
8496 assert(N1.getValueType() == N2.getValueType() &&
8497 "SETCC operands must have the same type!");
8498 assert(VT.isVector() == N1.getValueType().isVector() &&
8499 "SETCC type should be vector iff the operand type is vector!");
8500 assert((!VT.isVector() || VT.getVectorElementCount() ==
8502 "SETCC vector element counts must match!");
8503 // Use FoldSetCC to simplify SETCC's.
8504 if (SDValue V =
8505 FoldSetCC(VT, N1, N2, cast<CondCodeSDNode>(N3)->get(), DL, Flags))
8506 return V;
8507 break;
8508 }
8509 case ISD::SELECT:
8510 case ISD::VSELECT:
8511 if (SDValue V = simplifySelect(N1, N2, N3))
8512 return V;
8513 break;
8515 llvm_unreachable("should use getVectorShuffle constructor!");
8517 if (isNullConstant(N3))
8518 return N1;
8519 break;
8521 if (isNullConstant(N3))
8522 return N2;
8523 break;
8525 assert(VT.isVector() && VT == N1.getValueType() &&
8526 "INSERT_VECTOR_ELT vector type mismatch");
8528 "INSERT_VECTOR_ELT scalar fp/int mismatch");
8529 assert((!VT.isFloatingPoint() ||
8530 VT.getVectorElementType() == N2.getValueType()) &&
8531 "INSERT_VECTOR_ELT fp scalar type mismatch");
8532 assert((!VT.isInteger() ||
8534 "INSERT_VECTOR_ELT int scalar size mismatch");
8535
8536 auto *N3C = dyn_cast<ConstantSDNode>(N3);
8537 // INSERT_VECTOR_ELT into out-of-bounds element is an UNDEF, except
8538 // for scalable vectors where we will generate appropriate code to
8539 // deal with out-of-bounds cases correctly.
8540 if (N3C && VT.isFixedLengthVector() &&
8541 N3C->getZExtValue() >= VT.getVectorNumElements())
8542 return getUNDEF(VT);
8543
8544 // Undefined index can be assumed out-of-bounds, so that's UNDEF too.
8545 if (N3.isUndef())
8546 return getUNDEF(VT);
8547
8548 // If inserting poison, just use the input vector.
8549 if (N2.getOpcode() == ISD::POISON)
8550 return N1;
8551
8552 // Inserting undef into undef/poison is still undef.
8553 if (N2.getOpcode() == ISD::UNDEF && N1.isUndef())
8554 return getUNDEF(VT);
8555
8556 // If the inserted element is an UNDEF, just use the input vector.
8557 // But not if skipping the insert could make the result more poisonous.
8558 if (N2.isUndef()) {
8559 if (N3C && VT.isFixedLengthVector()) {
8560 APInt EltMask =
8561 APInt::getOneBitSet(VT.getVectorNumElements(), N3C->getZExtValue());
8562 if (isGuaranteedNotToBePoison(N1, EltMask))
8563 return N1;
8564 } else if (isGuaranteedNotToBePoison(N1))
8565 return N1;
8566 }
8567 break;
8568 }
8569 case ISD::INSERT_SUBVECTOR: {
8570 // If inserting poison, just use the input vector,
8571 if (N2.getOpcode() == ISD::POISON)
8572 return N1;
8573
8574 // Inserting undef into undef/poison is still undef.
8575 if (N2.getOpcode() == ISD::UNDEF && N1.isUndef())
8576 return getUNDEF(VT);
8577
8578 EVT N2VT = N2.getValueType();
8579 assert(VT == N1.getValueType() &&
8580 "Dest and insert subvector source types must match!");
8581 assert(VT.isVector() && N2VT.isVector() &&
8582 "Insert subvector VTs must be vectors!");
8584 "Insert subvector VTs must have the same element type!");
8585 assert((VT.isScalableVector() || N2VT.isFixedLengthVector()) &&
8586 "Cannot insert a scalable vector into a fixed length vector!");
8587 assert((VT.isScalableVector() != N2VT.isScalableVector() ||
8589 "Insert subvector must be from smaller vector to larger vector!");
8591 "Insert subvector index must be constant");
8592 assert((VT.isScalableVector() != N2VT.isScalableVector() ||
8593 (N2VT.getVectorMinNumElements() + N3->getAsZExtVal()) <=
8595 "Insert subvector overflow!");
8597 TLI->getVectorIdxWidth(getDataLayout()) &&
8598 "Constant index for INSERT_SUBVECTOR has an invalid size");
8599
8600 // Trivial insertion.
8601 if (VT == N2VT)
8602 return N2;
8603
8604 // If this is an insert of an extracted vector into an undef/poison vector,
8605 // we can just use the input to the extract. But not if skipping the
8606 // extract+insert could make the result more poisonous.
8607 if (N1.isUndef() && N2.getOpcode() == ISD::EXTRACT_SUBVECTOR &&
8608 N2.getOperand(1) == N3 && N2.getOperand(0).getValueType() == VT) {
8609 if (N1.getOpcode() == ISD::POISON)
8610 return N2.getOperand(0);
8611 if (VT.isFixedLengthVector() && N2VT.isFixedLengthVector()) {
8612 unsigned LoBit = N3->getAsZExtVal();
8613 unsigned HiBit = LoBit + N2VT.getVectorNumElements();
8614 APInt EltMask =
8615 APInt::getBitsSet(VT.getVectorNumElements(), LoBit, HiBit);
8616 if (isGuaranteedNotToBePoison(N2.getOperand(0), ~EltMask))
8617 return N2.getOperand(0);
8618 } else if (isGuaranteedNotToBePoison(N2.getOperand(0)))
8619 return N2.getOperand(0);
8620 }
8621
8622 // If the inserted subvector is UNDEF, just use the input vector.
8623 // But not if skipping the insert could make the result more poisonous.
8624 if (N2.isUndef()) {
8625 if (VT.isFixedLengthVector()) {
8626 unsigned LoBit = N3->getAsZExtVal();
8627 unsigned HiBit = LoBit + N2VT.getVectorNumElements();
8628 APInt EltMask =
8629 APInt::getBitsSet(VT.getVectorNumElements(), LoBit, HiBit);
8630 if (isGuaranteedNotToBePoison(N1, EltMask))
8631 return N1;
8632 } else if (isGuaranteedNotToBePoison(N1))
8633 return N1;
8634 }
8635 break;
8636 }
8637 case ISD::BITCAST:
8638 // Fold bit_convert nodes from a type to themselves.
8639 if (N1.getValueType() == VT)
8640 return N1;
8641 break;
8642 case ISD::VP_TRUNCATE:
8643 case ISD::VP_SIGN_EXTEND:
8644 case ISD::VP_ZERO_EXTEND:
8645 // Don't create noop casts.
8646 if (N1.getValueType() == VT)
8647 return N1;
8648 break;
8649 case ISD::VECTOR_COMPRESS: {
8650 [[maybe_unused]] EVT VecVT = N1.getValueType();
8651 [[maybe_unused]] EVT MaskVT = N2.getValueType();
8652 [[maybe_unused]] EVT PassthruVT = N3.getValueType();
8653 assert(VT == VecVT && "Vector and result type don't match.");
8654 assert(VecVT.isVector() && MaskVT.isVector() && PassthruVT.isVector() &&
8655 "All inputs must be vectors.");
8656 assert(VecVT == PassthruVT && "Vector and passthru types don't match.");
8658 "Vector and mask must have same number of elements.");
8659
8660 if (N1.isUndef() || N2.isUndef())
8661 return N3;
8662
8663 break;
8664 }
8669 [[maybe_unused]] EVT AccVT = N1.getValueType();
8670 [[maybe_unused]] EVT Input1VT = N2.getValueType();
8671 [[maybe_unused]] EVT Input2VT = N3.getValueType();
8672 assert(Input1VT.isVector() && Input1VT == Input2VT &&
8673 "Expected the second and third operands of the PARTIAL_REDUCE_MLA "
8674 "node to have the same type!");
8675 assert(VT.isVector() && VT == AccVT &&
8676 "Expected the first operand of the PARTIAL_REDUCE_MLA node to have "
8677 "the same type as its result!");
8679 AccVT.getVectorElementCount()) &&
8680 "Expected the element count of the second and third operands of the "
8681 "PARTIAL_REDUCE_MLA node to be a positive integer multiple of the "
8682 "element count of the first operand and the result!");
8684 "Expected the second and third operands of the PARTIAL_REDUCE_MLA "
8685 "node to have an element type which is the same as or smaller than "
8686 "the element type of the first operand and result!");
8687 break;
8688 }
8689 }
8690
8691 // Perform trivial constant folding for arithmetic operators.
8692 switch (Opcode) {
8693 case ISD::FMA:
8694 case ISD::FMAD:
8695 case ISD::SETCC:
8696 case ISD::FSHL:
8697 case ISD::FSHR:
8698 if (SDValue SV =
8699 FoldConstantArithmetic(Opcode, DL, VT, {N1, N2, N3}, Flags))
8700 return SV;
8701 break;
8702 }
8703
8704 // Memoize node if it doesn't produce a glue result.
8705 SDNode *N;
8706 SDVTList VTs = getVTList(VT);
8707 SDValue Ops[] = {N1, N2, N3};
8708 if (VT != MVT::Glue) {
8710 AddNodeIDNode(ID, Opcode, VTs, Ops);
8711 void *IP = nullptr;
8712 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
8713 E->intersectFlagsWith(Flags);
8714 return SDValue(E, 0);
8715 }
8716
8717 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
8718 N->setFlags(Flags);
8719 createOperands(N, Ops);
8720 CSEMap.InsertNode(N, IP);
8721 } else {
8722 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
8723 createOperands(N, Ops);
8724 }
8725
8726 InsertNode(N);
8727 SDValue V = SDValue(N, 0);
8728 NewSDValueDbgMsg(V, "Creating new node: ", this);
8729 return V;
8730}
8731
8732SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8733 SDValue N1, SDValue N2, SDValue N3, SDValue N4,
8734 const SDNodeFlags Flags) {
8735 SDValue Ops[] = { N1, N2, N3, N4 };
8736 return getNode(Opcode, DL, VT, Ops, Flags);
8737}
8738
8739SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8740 SDValue N1, SDValue N2, SDValue N3, SDValue N4) {
8741 SDNodeFlags Flags;
8742 if (Inserter)
8743 Flags = Inserter->getFlags();
8744 return getNode(Opcode, DL, VT, N1, N2, N3, N4, Flags);
8745}
8746
8747SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8748 SDValue N1, SDValue N2, SDValue N3, SDValue N4,
8749 SDValue N5, const SDNodeFlags Flags) {
8750 SDValue Ops[] = { N1, N2, N3, N4, N5 };
8751 return getNode(Opcode, DL, VT, Ops, Flags);
8752}
8753
8754SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8755 SDValue N1, SDValue N2, SDValue N3, SDValue N4,
8756 SDValue N5) {
8757 SDNodeFlags Flags;
8758 if (Inserter)
8759 Flags = Inserter->getFlags();
8760 return getNode(Opcode, DL, VT, N1, N2, N3, N4, N5, Flags);
8761}
8762
8763/// getStackArgumentTokenFactor - Compute a TokenFactor to force all
8764/// the incoming stack arguments to be loaded from the stack.
8766 SmallVector<SDValue, 8> ArgChains;
8767
8768 // Include the original chain at the beginning of the list. When this is
8769 // used by target LowerCall hooks, this helps legalize find the
8770 // CALLSEQ_BEGIN node.
8771 ArgChains.push_back(Chain);
8772
8773 // Add a chain value for each stack argument.
8774 for (SDNode *U : getEntryNode().getNode()->users())
8775 if (LoadSDNode *L = dyn_cast<LoadSDNode>(U))
8776 if (FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(L->getBasePtr()))
8777 if (FI->getIndex() < 0)
8778 ArgChains.push_back(SDValue(L, 1));
8779
8780 // Build a tokenfactor for all the chains.
8781 return getNode(ISD::TokenFactor, SDLoc(Chain), MVT::Other, ArgChains);
8782}
8783
8784/// getMemsetValue - Vectorized representation of the memset value
8785/// operand.
8787 const SDLoc &dl) {
8788 assert(!Value.isUndef());
8789
8790 unsigned NumBits = VT.getScalarSizeInBits();
8792 assert(C->getAPIntValue().getBitWidth() == 8);
8793 APInt Val = APInt::getSplat(NumBits, C->getAPIntValue());
8794 if (VT.isInteger()) {
8795 bool IsOpaque = VT.getSizeInBits() > 64 ||
8796 !DAG.getTargetLoweringInfo().isLegalStoreImmediate(C->getSExtValue());
8797 return DAG.getConstant(Val, dl, VT, false, IsOpaque);
8798 }
8799 return DAG.getConstantFP(APFloat(VT.getFltSemantics(), Val), dl, VT);
8800 }
8801
8802 assert(Value.getValueType() == MVT::i8 && "memset with non-byte fill value?");
8803 EVT IntVT = VT.getScalarType();
8804 if (!IntVT.isInteger())
8805 IntVT = EVT::getIntegerVT(*DAG.getContext(), IntVT.getSizeInBits());
8806
8807 Value = DAG.getNode(ISD::ZERO_EXTEND, dl, IntVT, Value);
8808 if (NumBits > 8) {
8809 // Use a multiplication with 0x010101... to extend the input to the
8810 // required length.
8811 APInt Magic = APInt::getSplat(NumBits, APInt(8, 0x01));
8812 Value = DAG.getNode(ISD::MUL, dl, IntVT, Value,
8813 DAG.getConstant(Magic, dl, IntVT));
8814 }
8815
8816 if (VT != Value.getValueType() && !VT.isInteger())
8817 Value = DAG.getBitcast(VT.getScalarType(), Value);
8818 if (VT != Value.getValueType())
8819 Value = DAG.getSplatBuildVector(VT, dl, Value);
8820
8821 return Value;
8822}
8823
8824/// getMemsetStringVal - Similar to getMemsetValue. Except this is only
8825/// used when a memcpy is turned into a memset when the source is a constant
8826/// string ptr.
8828 const TargetLowering &TLI,
8829 const ConstantDataArraySlice &Slice) {
8830 // Handle vector with all elements zero.
8831 if (Slice.Array == nullptr) {
8832 if (VT.isInteger())
8833 return DAG.getConstant(0, dl, VT);
8834 return DAG.getNode(ISD::BITCAST, dl, VT,
8835 DAG.getConstant(0, dl, VT.changeTypeToInteger()));
8836 }
8837
8838 assert(!VT.isVector() && "Can't handle vector type here!");
8839 unsigned NumVTBits = VT.getSizeInBits();
8840 unsigned NumVTBytes = NumVTBits / 8;
8841 unsigned NumBytes = std::min(NumVTBytes, unsigned(Slice.Length));
8842
8843 APInt Val(NumVTBits, 0);
8844 if (DAG.getDataLayout().isLittleEndian()) {
8845 for (unsigned i = 0; i != NumBytes; ++i)
8846 Val |= (uint64_t)(unsigned char)Slice[i] << i*8;
8847 } else {
8848 for (unsigned i = 0; i != NumBytes; ++i)
8849 Val |= (uint64_t)(unsigned char)Slice[i] << (NumVTBytes-i-1)*8;
8850 }
8851
8852 // If the "cost" of materializing the integer immediate is less than the cost
8853 // of a load, then it is cost effective to turn the load into the immediate.
8854 Type *Ty = VT.getTypeForEVT(*DAG.getContext());
8855 if (TLI.shouldConvertConstantLoadToIntImm(Val, Ty))
8856 return DAG.getConstant(Val, dl, VT);
8857 return SDValue();
8858}
8859
8861 const SDLoc &DL,
8862 const SDNodeFlags Flags) {
8863 SDValue Index = getTypeSize(DL, Base.getValueType(), Offset);
8864 return getMemBasePlusOffset(Base, Index, DL, Flags);
8865}
8866
8868 const SDLoc &DL,
8869 const SDNodeFlags Flags) {
8870 assert(Offset.getValueType().isInteger());
8871 EVT BasePtrVT = Ptr.getValueType();
8872 if (TLI->shouldPreservePtrArith(this->getMachineFunction().getFunction(),
8873 BasePtrVT))
8874 return getNode(ISD::PTRADD, DL, BasePtrVT, Ptr, Offset, Flags);
8875 // InBounds only applies to PTRADD, don't set it if we generate ADD.
8876 SDNodeFlags AddFlags = Flags;
8877 AddFlags.setInBounds(false);
8878 return getNode(ISD::ADD, DL, BasePtrVT, Ptr, Offset, AddFlags);
8879}
8880
8881/// Returns true if memcpy source is constant data.
8883 uint64_t SrcDelta = 0;
8884 GlobalAddressSDNode *G = nullptr;
8885 if (Src.getOpcode() == ISD::GlobalAddress)
8887 else if (Src->isAnyAdd() &&
8888 Src.getOperand(0).getOpcode() == ISD::GlobalAddress &&
8889 Src.getOperand(1).getOpcode() == ISD::Constant) {
8890 G = cast<GlobalAddressSDNode>(Src.getOperand(0));
8891 SrcDelta = Src.getConstantOperandVal(1);
8892 }
8893 if (!G)
8894 return false;
8895
8896 return getConstantDataArrayInfo(G->getGlobal(), Slice, 8,
8897 SrcDelta + G->getOffset());
8898}
8899
8901 SelectionDAG &DAG) {
8902 // On Darwin, -Os means optimize for size without hurting performance, so
8903 // only really optimize for size when -Oz (MinSize) is used.
8905 return MF.getFunction().hasMinSize();
8906 return DAG.shouldOptForSize();
8907}
8908
8910 SmallVector<SDValue, 32> &OutChains, unsigned From,
8911 unsigned To, SmallVector<SDValue, 16> &OutLoadChains,
8912 SmallVector<SDValue, 16> &OutStoreChains) {
8913 assert(OutLoadChains.size() && "Missing loads in memcpy inlining");
8914 assert(OutStoreChains.size() && "Missing stores in memcpy inlining");
8915 SmallVector<SDValue, 16> GluedLoadChains;
8916 for (unsigned i = From; i < To; ++i) {
8917 OutChains.push_back(OutLoadChains[i]);
8918 GluedLoadChains.push_back(OutLoadChains[i]);
8919 }
8920
8921 // Chain for all loads.
8922 SDValue LoadToken = DAG.getNode(ISD::TokenFactor, dl, MVT::Other,
8923 GluedLoadChains);
8924
8925 for (unsigned i = From; i < To; ++i) {
8926 StoreSDNode *ST = dyn_cast<StoreSDNode>(OutStoreChains[i]);
8927 SDValue NewStore = DAG.getTruncStore(LoadToken, dl, ST->getValue(),
8928 ST->getBasePtr(), ST->getMemoryVT(),
8929 ST->getMemOperand());
8930 OutChains.push_back(NewStore);
8931 }
8932}
8933
8935 SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Dst, SDValue Src,
8936 uint64_t Size, Align Alignment, bool isVol, bool AlwaysInline,
8937 MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo,
8938 const AAMDNodes &AAInfo, BatchAAResults *BatchAA) {
8939 // Turn a memcpy of undef to nop.
8940 // FIXME: We need to honor volatile even is Src is undef.
8941 if (Src.isUndef())
8942 return Chain;
8943
8944 // Expand memcpy to a series of load and store ops if the size operand falls
8945 // below a certain threshold.
8946 // TODO: In the AlwaysInline case, if the size is big then generate a loop
8947 // rather than maybe a humongous number of loads and stores.
8948 const TargetLowering &TLI = DAG.getTargetLoweringInfo();
8949 const DataLayout &DL = DAG.getDataLayout();
8950 LLVMContext &C = *DAG.getContext();
8951 std::vector<EVT> MemOps;
8952 bool DstAlignCanChange = false;
8954 MachineFrameInfo &MFI = MF.getFrameInfo();
8955 bool OptSize = shouldLowerMemFuncForSize(MF, DAG);
8957 if (FI && !MFI.isFixedObjectIndex(FI->getIndex()))
8958 DstAlignCanChange = true;
8959 MaybeAlign SrcAlign = DAG.InferPtrAlign(Src);
8960 if (!SrcAlign || Alignment > *SrcAlign)
8961 SrcAlign = Alignment;
8962 assert(SrcAlign && "SrcAlign must be set");
8964 // If marked as volatile, perform a copy even when marked as constant.
8965 bool CopyFromConstant = !isVol && isMemSrcFromConstant(Src, Slice);
8966 bool isZeroConstant = CopyFromConstant && Slice.Array == nullptr;
8967 unsigned Limit = AlwaysInline ? ~0U : TLI.getMaxStoresPerMemcpy(OptSize);
8968 const MemOp Op = isZeroConstant
8969 ? MemOp::Set(Size, DstAlignCanChange, Alignment,
8970 /*IsZeroMemset*/ true, isVol)
8971 : MemOp::Copy(Size, DstAlignCanChange, Alignment,
8972 *SrcAlign, isVol, CopyFromConstant);
8973 if (!TLI.findOptimalMemOpLowering(
8974 C, MemOps, Limit, Op, DstPtrInfo.getAddrSpace(),
8975 SrcPtrInfo.getAddrSpace(), MF.getFunction().getAttributes(), nullptr))
8976 return SDValue();
8977
8978 if (DstAlignCanChange) {
8979 Type *Ty = MemOps[0].getTypeForEVT(C);
8980 Align NewAlign = DL.getABITypeAlign(Ty);
8981
8982 // Don't promote to an alignment that would require dynamic stack
8983 // realignment which may conflict with optimizations such as tail call
8984 // optimization.
8986 if (!TRI->hasStackRealignment(MF))
8987 if (MaybeAlign StackAlign = DL.getStackAlignment())
8988 NewAlign = std::min(NewAlign, *StackAlign);
8989
8990 if (NewAlign > Alignment) {
8991 // Give the stack frame object a larger alignment if needed.
8992 if (MFI.getObjectAlign(FI->getIndex()) < NewAlign)
8993 MFI.setObjectAlignment(FI->getIndex(), NewAlign);
8994 Alignment = NewAlign;
8995 }
8996 }
8997
8998 // Prepare AAInfo for loads/stores after lowering this memcpy.
8999 AAMDNodes NewAAInfo = AAInfo;
9000 NewAAInfo.TBAA = NewAAInfo.TBAAStruct = nullptr;
9001
9002 const Value *SrcVal = dyn_cast_if_present<const Value *>(SrcPtrInfo.V);
9003 bool isConstant =
9004 BatchAA && SrcVal &&
9005 BatchAA->pointsToConstantMemory(MemoryLocation(SrcVal, Size, AAInfo));
9006
9007 MachineMemOperand::Flags MMOFlags =
9009 SmallVector<SDValue, 16> OutLoadChains;
9010 SmallVector<SDValue, 16> OutStoreChains;
9011 SmallVector<SDValue, 32> OutChains;
9012 unsigned NumMemOps = MemOps.size();
9013 uint64_t SrcOff = 0, DstOff = 0;
9014 for (unsigned i = 0; i != NumMemOps; ++i) {
9015 EVT VT = MemOps[i];
9016 unsigned VTSize = VT.getSizeInBits() / 8;
9017 SDValue Value, Store;
9018
9019 if (VTSize > Size) {
9020 // Issuing an unaligned load / store pair that overlaps with the previous
9021 // pair. Adjust the offset accordingly.
9022 assert(i == NumMemOps-1 && i != 0);
9023 SrcOff -= VTSize - Size;
9024 DstOff -= VTSize - Size;
9025 }
9026
9027 if (CopyFromConstant &&
9028 (isZeroConstant || (VT.isInteger() && !VT.isVector()))) {
9029 // It's unlikely a store of a vector immediate can be done in a single
9030 // instruction. It would require a load from a constantpool first.
9031 // We only handle zero vectors here.
9032 // FIXME: Handle other cases where store of vector immediate is done in
9033 // a single instruction.
9034 ConstantDataArraySlice SubSlice;
9035 if (SrcOff < Slice.Length) {
9036 SubSlice = Slice;
9037 SubSlice.move(SrcOff);
9038 } else {
9039 // This is an out-of-bounds access and hence UB. Pretend we read zero.
9040 SubSlice.Array = nullptr;
9041 SubSlice.Offset = 0;
9042 SubSlice.Length = VTSize;
9043 }
9044 Value = getMemsetStringVal(VT, dl, DAG, TLI, SubSlice);
9045 if (Value.getNode()) {
9046 Store = DAG.getStore(
9047 Chain, dl, Value,
9048 DAG.getObjectPtrOffset(dl, Dst, TypeSize::getFixed(DstOff)),
9049 DstPtrInfo.getWithOffset(DstOff), Alignment, MMOFlags, NewAAInfo);
9050 OutChains.push_back(Store);
9051 }
9052 }
9053
9054 if (!Store.getNode()) {
9055 // The type might not be legal for the target. This should only happen
9056 // if the type is smaller than a legal type, as on PPC, so the right
9057 // thing to do is generate a LoadExt/StoreTrunc pair. These simplify
9058 // to Load/Store if NVT==VT.
9059 // FIXME does the case above also need this?
9060 EVT NVT = TLI.getTypeToTransformTo(C, VT);
9061 assert(NVT.bitsGE(VT));
9062
9063 bool isDereferenceable =
9064 SrcPtrInfo.getWithOffset(SrcOff).isDereferenceable(VTSize, C, DL);
9065 MachineMemOperand::Flags SrcMMOFlags = MMOFlags;
9066 if (isDereferenceable)
9068 if (isConstant)
9069 SrcMMOFlags |= MachineMemOperand::MOInvariant;
9070
9071 Value = DAG.getExtLoad(
9072 ISD::EXTLOAD, dl, NVT, Chain,
9073 DAG.getObjectPtrOffset(dl, Src, TypeSize::getFixed(SrcOff)),
9074 SrcPtrInfo.getWithOffset(SrcOff), VT,
9075 commonAlignment(*SrcAlign, SrcOff), SrcMMOFlags, NewAAInfo);
9076 OutLoadChains.push_back(Value.getValue(1));
9077
9078 Store = DAG.getTruncStore(
9079 Chain, dl, Value,
9080 DAG.getObjectPtrOffset(dl, Dst, TypeSize::getFixed(DstOff)),
9081 DstPtrInfo.getWithOffset(DstOff), VT, Alignment, MMOFlags, NewAAInfo);
9082 OutStoreChains.push_back(Store);
9083 }
9084 SrcOff += VTSize;
9085 DstOff += VTSize;
9086 Size -= VTSize;
9087 }
9088
9089 unsigned GluedLdStLimit = MaxLdStGlue == 0 ?
9091 unsigned NumLdStInMemcpy = OutStoreChains.size();
9092
9093 if (NumLdStInMemcpy) {
9094 // It may be that memcpy might be converted to memset if it's memcpy
9095 // of constants. In such a case, we won't have loads and stores, but
9096 // just stores. In the absence of loads, there is nothing to gang up.
9097 if ((GluedLdStLimit <= 1) || !EnableMemCpyDAGOpt) {
9098 // If target does not care, just leave as it.
9099 for (unsigned i = 0; i < NumLdStInMemcpy; ++i) {
9100 OutChains.push_back(OutLoadChains[i]);
9101 OutChains.push_back(OutStoreChains[i]);
9102 }
9103 } else {
9104 // Ld/St less than/equal limit set by target.
9105 if (NumLdStInMemcpy <= GluedLdStLimit) {
9106 chainLoadsAndStoresForMemcpy(DAG, dl, OutChains, 0,
9107 NumLdStInMemcpy, OutLoadChains,
9108 OutStoreChains);
9109 } else {
9110 unsigned NumberLdChain = NumLdStInMemcpy / GluedLdStLimit;
9111 unsigned RemainingLdStInMemcpy = NumLdStInMemcpy % GluedLdStLimit;
9112 unsigned GlueIter = 0;
9113
9114 // Residual ld/st.
9115 if (RemainingLdStInMemcpy) {
9117 DAG, dl, OutChains, NumLdStInMemcpy - RemainingLdStInMemcpy,
9118 NumLdStInMemcpy, OutLoadChains, OutStoreChains);
9119 }
9120
9121 for (unsigned cnt = 0; cnt < NumberLdChain; ++cnt) {
9122 unsigned IndexFrom = NumLdStInMemcpy - RemainingLdStInMemcpy -
9123 GlueIter - GluedLdStLimit;
9124 unsigned IndexTo = NumLdStInMemcpy - RemainingLdStInMemcpy - GlueIter;
9125 chainLoadsAndStoresForMemcpy(DAG, dl, OutChains, IndexFrom, IndexTo,
9126 OutLoadChains, OutStoreChains);
9127 GlueIter += GluedLdStLimit;
9128 }
9129 }
9130 }
9131 }
9132 return DAG.getNode(ISD::TokenFactor, dl, MVT::Other, OutChains);
9133}
9134
9136 SDValue Chain, SDValue Dst, SDValue Src,
9137 uint64_t Size, Align Alignment,
9138 bool isVol, bool AlwaysInline,
9139 MachinePointerInfo DstPtrInfo,
9140 MachinePointerInfo SrcPtrInfo,
9141 const AAMDNodes &AAInfo) {
9142 // Turn a memmove of undef to nop.
9143 // FIXME: We need to honor volatile even is Src is undef.
9144 if (Src.isUndef())
9145 return Chain;
9146
9147 // Expand memmove to a series of load and store ops if the size operand falls
9148 // below a certain threshold.
9149 const TargetLowering &TLI = DAG.getTargetLoweringInfo();
9150 const DataLayout &DL = DAG.getDataLayout();
9151 LLVMContext &C = *DAG.getContext();
9152 std::vector<EVT> MemOps;
9153 bool DstAlignCanChange = false;
9155 MachineFrameInfo &MFI = MF.getFrameInfo();
9156 bool OptSize = shouldLowerMemFuncForSize(MF, DAG);
9158 if (FI && !MFI.isFixedObjectIndex(FI->getIndex()))
9159 DstAlignCanChange = true;
9160 MaybeAlign SrcAlign = DAG.InferPtrAlign(Src);
9161 if (!SrcAlign || Alignment > *SrcAlign)
9162 SrcAlign = Alignment;
9163 assert(SrcAlign && "SrcAlign must be set");
9164 unsigned Limit = AlwaysInline ? ~0U : TLI.getMaxStoresPerMemmove(OptSize);
9165 if (!TLI.findOptimalMemOpLowering(
9166 C, MemOps, Limit,
9167 MemOp::Copy(Size, DstAlignCanChange, Alignment, *SrcAlign,
9168 /*IsVolatile*/ true),
9169 DstPtrInfo.getAddrSpace(), SrcPtrInfo.getAddrSpace(),
9170 MF.getFunction().getAttributes(), nullptr))
9171 return SDValue();
9172
9173 if (DstAlignCanChange) {
9174 Type *Ty = MemOps[0].getTypeForEVT(C);
9175 Align NewAlign = DL.getABITypeAlign(Ty);
9176
9177 // Don't promote to an alignment that would require dynamic stack
9178 // realignment which may conflict with optimizations such as tail call
9179 // optimization.
9181 if (!TRI->hasStackRealignment(MF))
9182 if (MaybeAlign StackAlign = DL.getStackAlignment())
9183 NewAlign = std::min(NewAlign, *StackAlign);
9184
9185 if (NewAlign > Alignment) {
9186 // Give the stack frame object a larger alignment if needed.
9187 if (MFI.getObjectAlign(FI->getIndex()) < NewAlign)
9188 MFI.setObjectAlignment(FI->getIndex(), NewAlign);
9189 Alignment = NewAlign;
9190 }
9191 }
9192
9193 // Prepare AAInfo for loads/stores after lowering this memmove.
9194 AAMDNodes NewAAInfo = AAInfo;
9195 NewAAInfo.TBAA = NewAAInfo.TBAAStruct = nullptr;
9196
9197 MachineMemOperand::Flags MMOFlags =
9199 uint64_t SrcOff = 0, DstOff = 0;
9200 SmallVector<SDValue, 8> LoadValues;
9201 SmallVector<SDValue, 8> LoadChains;
9202 SmallVector<SDValue, 8> OutChains;
9203 unsigned NumMemOps = MemOps.size();
9204 for (unsigned i = 0; i < NumMemOps; i++) {
9205 EVT VT = MemOps[i];
9206 unsigned VTSize = VT.getSizeInBits() / 8;
9207 SDValue Value;
9208
9209 bool isDereferenceable =
9210 SrcPtrInfo.getWithOffset(SrcOff).isDereferenceable(VTSize, C, DL);
9211 MachineMemOperand::Flags SrcMMOFlags = MMOFlags;
9212 if (isDereferenceable)
9214
9215 Value = DAG.getLoad(
9216 VT, dl, Chain,
9217 DAG.getObjectPtrOffset(dl, Src, TypeSize::getFixed(SrcOff)),
9218 SrcPtrInfo.getWithOffset(SrcOff), *SrcAlign, SrcMMOFlags, NewAAInfo);
9219 LoadValues.push_back(Value);
9220 LoadChains.push_back(Value.getValue(1));
9221 SrcOff += VTSize;
9222 }
9223 Chain = DAG.getNode(ISD::TokenFactor, dl, MVT::Other, LoadChains);
9224 OutChains.clear();
9225 for (unsigned i = 0; i < NumMemOps; i++) {
9226 EVT VT = MemOps[i];
9227 unsigned VTSize = VT.getSizeInBits() / 8;
9228 SDValue Store;
9229
9230 Store = DAG.getStore(
9231 Chain, dl, LoadValues[i],
9232 DAG.getObjectPtrOffset(dl, Dst, TypeSize::getFixed(DstOff)),
9233 DstPtrInfo.getWithOffset(DstOff), Alignment, MMOFlags, NewAAInfo);
9234 OutChains.push_back(Store);
9235 DstOff += VTSize;
9236 }
9237
9238 return DAG.getNode(ISD::TokenFactor, dl, MVT::Other, OutChains);
9239}
9240
9241/// Lower the call to 'memset' intrinsic function into a series of store
9242/// operations.
9243///
9244/// \param DAG Selection DAG where lowered code is placed.
9245/// \param dl Link to corresponding IR location.
9246/// \param Chain Control flow dependency.
9247/// \param Dst Pointer to destination memory location.
9248/// \param Src Value of byte to write into the memory.
9249/// \param Size Number of bytes to write.
9250/// \param Alignment Alignment of the destination in bytes.
9251/// \param isVol True if destination is volatile.
9252/// \param AlwaysInline Makes sure no function call is generated.
9253/// \param DstPtrInfo IR information on the memory pointer.
9254/// \returns New head in the control flow, if lowering was successful, empty
9255/// SDValue otherwise.
9256///
9257/// The function tries to replace 'llvm.memset' intrinsic with several store
9258/// operations and value calculation code. This is usually profitable for small
9259/// memory size or when the semantic requires inlining.
9261 SDValue Chain, SDValue Dst, SDValue Src,
9262 uint64_t Size, Align Alignment, bool isVol,
9263 bool AlwaysInline, MachinePointerInfo DstPtrInfo,
9264 const AAMDNodes &AAInfo) {
9265 // Turn a memset of undef to nop.
9266 // FIXME: We need to honor volatile even is Src is undef.
9267 if (Src.isUndef())
9268 return Chain;
9269
9270 // Expand memset to a series of load/store ops if the size operand
9271 // falls below a certain threshold.
9272 const TargetLowering &TLI = DAG.getTargetLoweringInfo();
9273 std::vector<EVT> MemOps;
9274 bool DstAlignCanChange = false;
9275 LLVMContext &C = *DAG.getContext();
9277 MachineFrameInfo &MFI = MF.getFrameInfo();
9278 bool OptSize = shouldLowerMemFuncForSize(MF, DAG);
9280 if (FI && !MFI.isFixedObjectIndex(FI->getIndex()))
9281 DstAlignCanChange = true;
9282 bool IsZeroVal = isNullConstant(Src);
9283 unsigned Limit = AlwaysInline ? ~0 : TLI.getMaxStoresPerMemset(OptSize);
9284
9285 EVT LargestVT;
9286 if (!TLI.findOptimalMemOpLowering(
9287 C, MemOps, Limit,
9288 MemOp::Set(Size, DstAlignCanChange, Alignment, IsZeroVal, isVol),
9289 DstPtrInfo.getAddrSpace(), ~0u, MF.getFunction().getAttributes(),
9290 &LargestVT))
9291 return SDValue();
9292
9293 if (DstAlignCanChange) {
9294 Type *Ty = MemOps[0].getTypeForEVT(*DAG.getContext());
9295 const DataLayout &DL = DAG.getDataLayout();
9296 Align NewAlign = DL.getABITypeAlign(Ty);
9297
9298 // Don't promote to an alignment that would require dynamic stack
9299 // realignment which may conflict with optimizations such as tail call
9300 // optimization.
9302 if (!TRI->hasStackRealignment(MF))
9303 if (MaybeAlign StackAlign = DL.getStackAlignment())
9304 NewAlign = std::min(NewAlign, *StackAlign);
9305
9306 if (NewAlign > Alignment) {
9307 // Give the stack frame object a larger alignment if needed.
9308 if (MFI.getObjectAlign(FI->getIndex()) < NewAlign)
9309 MFI.setObjectAlignment(FI->getIndex(), NewAlign);
9310 Alignment = NewAlign;
9311 }
9312 }
9313
9314 SmallVector<SDValue, 8> OutChains;
9315 uint64_t DstOff = 0;
9316 unsigned NumMemOps = MemOps.size();
9317
9318 // Find the largest store and generate the bit pattern for it.
9319 // If target didn't set LargestVT, compute it from MemOps.
9320 if (!LargestVT.isSimple()) {
9321 LargestVT = MemOps[0];
9322 for (unsigned i = 1; i < NumMemOps; i++)
9323 if (MemOps[i].bitsGT(LargestVT))
9324 LargestVT = MemOps[i];
9325 }
9326 SDValue MemSetValue = getMemsetValue(Src, LargestVT, DAG, dl);
9327
9328 // Prepare AAInfo for loads/stores after lowering this memset.
9329 AAMDNodes NewAAInfo = AAInfo;
9330 NewAAInfo.TBAA = NewAAInfo.TBAAStruct = nullptr;
9331
9332 for (unsigned i = 0; i < NumMemOps; i++) {
9333 EVT VT = MemOps[i];
9334 unsigned VTSize = VT.getSizeInBits() / 8;
9335 // The target should specify store types that exactly cover the memset size
9336 // (with the last store potentially being oversized for overlapping stores).
9337 assert(Size > 0 && "Target specified more stores than needed in "
9338 "findOptimalMemOpLowering");
9339 if (VTSize > Size) {
9340 // Issuing an unaligned load / store pair that overlaps with the previous
9341 // pair. Adjust the offset accordingly.
9342 assert(i == NumMemOps-1 && i != 0);
9343 DstOff -= VTSize - Size;
9344 }
9345
9346 // If this store is smaller than the largest store see whether we can get
9347 // the smaller value for free with a truncate or extract vector element and
9348 // then store.
9349 SDValue Value = MemSetValue;
9350 if (VT.bitsLT(LargestVT)) {
9351 unsigned Index;
9352 unsigned NElts = LargestVT.getSizeInBits() / VT.getSizeInBits();
9353 EVT SVT = EVT::getVectorVT(*DAG.getContext(), VT.getScalarType(), NElts);
9354 if (!LargestVT.isVector() && !VT.isVector() &&
9355 TLI.isTruncateFree(LargestVT, VT))
9356 Value = DAG.getNode(ISD::TRUNCATE, dl, VT, MemSetValue);
9357 else if (LargestVT.isVector() && !VT.isVector() &&
9359 LargestVT.getTypeForEVT(*DAG.getContext()),
9360 VT.getSizeInBits(), Index) &&
9361 TLI.isTypeLegal(SVT) &&
9362 LargestVT.getSizeInBits() == SVT.getSizeInBits()) {
9363 // Target which can combine store(extractelement VectorTy, Idx) can get
9364 // the smaller value for free.
9365 SDValue TailValue = DAG.getNode(ISD::BITCAST, dl, SVT, MemSetValue);
9366 Value = DAG.getExtractVectorElt(dl, VT, TailValue, Index);
9367 } else
9368 Value = getMemsetValue(Src, VT, DAG, dl);
9369 }
9370 assert(Value.getValueType() == VT && "Value with wrong type.");
9371 SDValue Store = DAG.getStore(
9372 Chain, dl, Value,
9373 DAG.getObjectPtrOffset(dl, Dst, TypeSize::getFixed(DstOff)),
9374 DstPtrInfo.getWithOffset(DstOff), Alignment,
9376 NewAAInfo);
9377 OutChains.push_back(Store);
9378 DstOff += VT.getSizeInBits() / 8;
9379 // For oversized overlapping stores, only subtract the remaining bytes.
9380 // For normal stores, subtract the full store size.
9381 if (VTSize > Size) {
9382 Size = 0;
9383 } else {
9384 Size -= VTSize;
9385 }
9386 }
9387
9388 // After processing all stores, Size should be exactly 0. Any remaining bytes
9389 // indicate a bug in the target's findOptimalMemOpLowering implementation.
9390 assert(Size == 0 && "Target's findOptimalMemOpLowering did not specify "
9391 "stores that exactly cover the memset size");
9392
9393 return DAG.getNode(ISD::TokenFactor, dl, MVT::Other, OutChains);
9394}
9395
9397 unsigned AS) {
9398 // Lowering memcpy / memset / memmove intrinsics to calls is only valid if all
9399 // pointer operands can be losslessly bitcasted to pointers of address space 0
9400 if (AS != 0 && !TLI->getTargetMachine().isNoopAddrSpaceCast(AS, 0)) {
9401 report_fatal_error("cannot lower memory intrinsic in address space " +
9402 Twine(AS));
9403 }
9404}
9405
9407 const SelectionDAG *SelDAG,
9408 bool AllowReturnsFirstArg) {
9409 if (!CI || !CI->isTailCall())
9410 return false;
9411 // TODO: Fix "returns-first-arg" determination so it doesn't depend on which
9412 // helper symbol we lower to.
9413 return isInTailCallPosition(*CI, SelDAG->getTarget(),
9414 AllowReturnsFirstArg &&
9416}
9417
9418static std::pair<SDValue, SDValue>
9421 const CallInst *CI, RTLIB::Libcall Call,
9422 SelectionDAG *DAG, const TargetLowering *TLI) {
9423 RTLIB::LibcallImpl LCImpl = DAG->getLibcalls().getLibcallImpl(Call);
9424
9425 if (LCImpl == RTLIB::Unsupported)
9426 return {};
9427
9429 bool IsTailCall =
9430 isInTailCallPositionWrapper(CI, DAG, /*AllowReturnsFirstArg=*/true);
9431 SDValue Callee =
9432 DAG->getExternalSymbol(LCImpl, TLI->getPointerTy(DAG->getDataLayout()));
9433
9434 CLI.setDebugLoc(dl)
9435 .setChain(Chain)
9437 CI->getType(), Callee, std::move(Args))
9438 .setTailCall(IsTailCall);
9439
9440 return TLI->LowerCallTo(CLI);
9441}
9442
9443std::pair<SDValue, SDValue> SelectionDAG::getStrcmp(SDValue Chain,
9444 const SDLoc &dl, SDValue S1,
9445 SDValue S2,
9446 const CallInst *CI) {
9448 TargetLowering::ArgListTy Args = {{S1, PT}, {S2, PT}};
9449 return getRuntimeCallSDValueHelper(Chain, dl, std::move(Args), CI,
9450 RTLIB::STRCMP, this, TLI);
9451}
9452
9453std::pair<SDValue, SDValue> SelectionDAG::getStrstr(SDValue Chain,
9454 const SDLoc &dl, SDValue S1,
9455 SDValue S2,
9456 const CallInst *CI) {
9458 TargetLowering::ArgListTy Args = {{S1, PT}, {S2, PT}};
9459 return getRuntimeCallSDValueHelper(Chain, dl, std::move(Args), CI,
9460 RTLIB::STRSTR, this, TLI);
9461}
9462
9463std::pair<SDValue, SDValue> SelectionDAG::getMemccpy(SDValue Chain,
9464 const SDLoc &dl,
9465 SDValue Dst, SDValue Src,
9467 const CallInst *CI) {
9469
9471 {Dst, PT},
9472 {Src, PT},
9475 return getRuntimeCallSDValueHelper(Chain, dl, std::move(Args), CI,
9476 RTLIB::MEMCCPY, this, TLI);
9477}
9478
9479std::pair<SDValue, SDValue>
9481 SDValue Mem1, SDValue Size, const CallInst *CI) {
9484 {Mem0, PT},
9485 {Mem1, PT},
9487 return getRuntimeCallSDValueHelper(Chain, dl, std::move(Args), CI,
9488 RTLIB::MEMCMP, this, TLI);
9489}
9490
9491std::pair<SDValue, SDValue> SelectionDAG::getStrcpy(SDValue Chain,
9492 const SDLoc &dl,
9493 SDValue Dst, SDValue Src,
9494 const CallInst *CI) {
9496 TargetLowering::ArgListTy Args = {{Dst, PT}, {Src, PT}};
9497 return getRuntimeCallSDValueHelper(Chain, dl, std::move(Args), CI,
9498 RTLIB::STRCPY, this, TLI);
9499}
9500
9501std::pair<SDValue, SDValue> SelectionDAG::getStrlen(SDValue Chain,
9502 const SDLoc &dl,
9503 SDValue Src,
9504 const CallInst *CI) {
9505 // Emit a library call.
9508 return getRuntimeCallSDValueHelper(Chain, dl, std::move(Args), CI,
9509 RTLIB::STRLEN, this, TLI);
9510}
9511
9513 SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size,
9514 Align Alignment, bool isVol, bool AlwaysInline, const CallInst *CI,
9515 std::optional<bool> OverrideTailCall, MachinePointerInfo DstPtrInfo,
9516 MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo,
9517 BatchAAResults *BatchAA) {
9518 // Check to see if we should lower the memcpy to loads and stores first.
9519 // For cases within the target-specified limits, this is the best choice.
9521 if (ConstantSize) {
9522 // Memcpy with size zero? Just return the original chain.
9523 if (ConstantSize->isZero())
9524 return Chain;
9525
9527 *this, dl, Chain, Dst, Src, ConstantSize->getZExtValue(), Alignment,
9528 isVol, false, DstPtrInfo, SrcPtrInfo, AAInfo, BatchAA);
9529 if (Result.getNode())
9530 return Result;
9531 }
9532
9533 // Then check to see if we should lower the memcpy with target-specific
9534 // code. If the target chooses to do this, this is the next best.
9535 if (TSI) {
9536 SDValue Result = TSI->EmitTargetCodeForMemcpy(
9537 *this, dl, Chain, Dst, Src, Size, Alignment, isVol, AlwaysInline,
9538 DstPtrInfo, SrcPtrInfo);
9539 if (Result.getNode())
9540 return Result;
9541 }
9542
9543 // If we really need inline code and the target declined to provide it,
9544 // use a (potentially long) sequence of loads and stores.
9545 if (AlwaysInline) {
9546 assert(ConstantSize && "AlwaysInline requires a constant size!");
9548 *this, dl, Chain, Dst, Src, ConstantSize->getZExtValue(), Alignment,
9549 isVol, true, DstPtrInfo, SrcPtrInfo, AAInfo, BatchAA);
9550 }
9551
9554
9555 // FIXME: If the memcpy is volatile (isVol), lowering it to a plain libc
9556 // memcpy is not guaranteed to be safe. libc memcpys aren't required to
9557 // respect volatile, so they may do things like read or write memory
9558 // beyond the given memory regions. But fixing this isn't easy, and most
9559 // people don't care.
9560
9561 // Emit a library call.
9564 Args.emplace_back(Dst, PtrTy);
9565 Args.emplace_back(Src, PtrTy);
9566 Args.emplace_back(Size, getDataLayout().getIntPtrType(*getContext()));
9567 // FIXME: pass in SDLoc
9569 bool IsTailCall = false;
9570 RTLIB::LibcallImpl MemCpyImpl = TLI->getMemcpyImpl();
9571
9572 if (OverrideTailCall.has_value()) {
9573 IsTailCall = *OverrideTailCall;
9574 } else {
9575 bool LowersToMemcpy = MemCpyImpl == RTLIB::impl_memcpy;
9576 IsTailCall = isInTailCallPositionWrapper(CI, this, LowersToMemcpy);
9577 }
9578
9579 CLI.setDebugLoc(dl)
9580 .setChain(Chain)
9581 .setLibCallee(
9582 Libcalls->getLibcallImplCallingConv(MemCpyImpl),
9583 Dst.getValueType().getTypeForEVT(*getContext()),
9584 getExternalSymbol(MemCpyImpl, TLI->getPointerTy(getDataLayout())),
9585 std::move(Args))
9587 .setTailCall(IsTailCall);
9588
9589 std::pair<SDValue,SDValue> CallResult = TLI->LowerCallTo(CLI);
9590 return CallResult.second;
9591}
9592
9594 SDValue Dst, SDValue Src, SDValue Size,
9595 Type *SizeTy, unsigned ElemSz,
9596 bool isTailCall,
9597 MachinePointerInfo DstPtrInfo,
9598 MachinePointerInfo SrcPtrInfo) {
9599 // Emit a library call.
9602 Args.emplace_back(Dst, ArgTy);
9603 Args.emplace_back(Src, ArgTy);
9604 Args.emplace_back(Size, SizeTy);
9605
9606 RTLIB::Libcall LibraryCall =
9608 RTLIB::LibcallImpl LibcallImpl = Libcalls->getLibcallImpl(LibraryCall);
9609 if (LibcallImpl == RTLIB::Unsupported)
9610 report_fatal_error("Unsupported element size");
9611
9613 CLI.setDebugLoc(dl)
9614 .setChain(Chain)
9615 .setLibCallee(
9616 Libcalls->getLibcallImplCallingConv(LibcallImpl),
9618 getExternalSymbol(LibcallImpl, TLI->getPointerTy(getDataLayout())),
9619 std::move(Args))
9621 .setTailCall(isTailCall);
9622
9623 std::pair<SDValue, SDValue> CallResult = TLI->LowerCallTo(CLI);
9624 return CallResult.second;
9625}
9626
9628 SDValue Src, SDValue Size, Align Alignment,
9629 bool isVol, const CallInst *CI,
9630 std::optional<bool> OverrideTailCall,
9631 MachinePointerInfo DstPtrInfo,
9632 MachinePointerInfo SrcPtrInfo,
9633 const AAMDNodes &AAInfo,
9634 BatchAAResults *BatchAA) {
9635 // Check to see if we should lower the memmove to loads and stores first.
9636 // For cases within the target-specified limits, this is the best choice.
9638 if (ConstantSize) {
9639 // Memmove with size zero? Just return the original chain.
9640 if (ConstantSize->isZero())
9641 return Chain;
9642
9644 *this, dl, Chain, Dst, Src, ConstantSize->getZExtValue(), Alignment,
9645 isVol, false, DstPtrInfo, SrcPtrInfo, AAInfo);
9646 if (Result.getNode())
9647 return Result;
9648 }
9649
9650 // Then check to see if we should lower the memmove with target-specific
9651 // code. If the target chooses to do this, this is the next best.
9652 if (TSI) {
9653 SDValue Result =
9654 TSI->EmitTargetCodeForMemmove(*this, dl, Chain, Dst, Src, Size,
9655 Alignment, isVol, DstPtrInfo, SrcPtrInfo);
9656 if (Result.getNode())
9657 return Result;
9658 }
9659
9662
9663 // FIXME: If the memmove is volatile, lowering it to plain libc memmove may
9664 // not be safe. See memcpy above for more details.
9665
9666 // Emit a library call.
9669 Args.emplace_back(Dst, PtrTy);
9670 Args.emplace_back(Src, PtrTy);
9671 Args.emplace_back(Size, getDataLayout().getIntPtrType(*getContext()));
9672 // FIXME: pass in SDLoc
9674
9675 RTLIB::LibcallImpl MemmoveImpl = Libcalls->getLibcallImpl(RTLIB::MEMMOVE);
9676
9677 bool IsTailCall = false;
9678 if (OverrideTailCall.has_value()) {
9679 IsTailCall = *OverrideTailCall;
9680 } else {
9681 bool LowersToMemmove = MemmoveImpl == RTLIB::impl_memmove;
9682 IsTailCall = isInTailCallPositionWrapper(CI, this, LowersToMemmove);
9683 }
9684
9685 CLI.setDebugLoc(dl)
9686 .setChain(Chain)
9687 .setLibCallee(
9688 Libcalls->getLibcallImplCallingConv(MemmoveImpl),
9689 Dst.getValueType().getTypeForEVT(*getContext()),
9690 getExternalSymbol(MemmoveImpl, TLI->getPointerTy(getDataLayout())),
9691 std::move(Args))
9693 .setTailCall(IsTailCall);
9694
9695 std::pair<SDValue,SDValue> CallResult = TLI->LowerCallTo(CLI);
9696 return CallResult.second;
9697}
9698
9700 SDValue Dst, SDValue Src, SDValue Size,
9701 Type *SizeTy, unsigned ElemSz,
9702 bool isTailCall,
9703 MachinePointerInfo DstPtrInfo,
9704 MachinePointerInfo SrcPtrInfo) {
9705 // Emit a library call.
9707 Type *IntPtrTy = getDataLayout().getIntPtrType(*getContext());
9708 Args.emplace_back(Dst, IntPtrTy);
9709 Args.emplace_back(Src, IntPtrTy);
9710 Args.emplace_back(Size, SizeTy);
9711
9712 RTLIB::Libcall LibraryCall =
9714 RTLIB::LibcallImpl LibcallImpl = Libcalls->getLibcallImpl(LibraryCall);
9715 if (LibcallImpl == RTLIB::Unsupported)
9716 report_fatal_error("Unsupported element size");
9717
9719 CLI.setDebugLoc(dl)
9720 .setChain(Chain)
9721 .setLibCallee(
9722 Libcalls->getLibcallImplCallingConv(LibcallImpl),
9724 getExternalSymbol(LibcallImpl, TLI->getPointerTy(getDataLayout())),
9725 std::move(Args))
9727 .setTailCall(isTailCall);
9728
9729 std::pair<SDValue, SDValue> CallResult = TLI->LowerCallTo(CLI);
9730 return CallResult.second;
9731}
9732
9734 SDValue Src, SDValue Size, Align Alignment,
9735 bool isVol, bool AlwaysInline,
9736 const CallInst *CI,
9737 MachinePointerInfo DstPtrInfo,
9738 const AAMDNodes &AAInfo) {
9739 // Check to see if we should lower the memset to stores first.
9740 // For cases within the target-specified limits, this is the best choice.
9742 if (ConstantSize) {
9743 // Memset with size zero? Just return the original chain.
9744 if (ConstantSize->isZero())
9745 return Chain;
9746
9747 SDValue Result = getMemsetStores(*this, dl, Chain, Dst, Src,
9748 ConstantSize->getZExtValue(), Alignment,
9749 isVol, false, DstPtrInfo, AAInfo);
9750
9751 if (Result.getNode())
9752 return Result;
9753 }
9754
9755 // Then check to see if we should lower the memset with target-specific
9756 // code. If the target chooses to do this, this is the next best.
9757 if (TSI) {
9758 SDValue Result = TSI->EmitTargetCodeForMemset(
9759 *this, dl, Chain, Dst, Src, Size, Alignment, isVol, AlwaysInline, DstPtrInfo);
9760 if (Result.getNode())
9761 return Result;
9762 }
9763
9764 // If we really need inline code and the target declined to provide it,
9765 // use a (potentially long) sequence of loads and stores.
9766 if (AlwaysInline) {
9767 assert(ConstantSize && "AlwaysInline requires a constant size!");
9768 SDValue Result = getMemsetStores(*this, dl, Chain, Dst, Src,
9769 ConstantSize->getZExtValue(), Alignment,
9770 isVol, true, DstPtrInfo, AAInfo);
9771 assert(Result &&
9772 "getMemsetStores must return a valid sequence when AlwaysInline");
9773 return Result;
9774 }
9775
9777
9778 // Emit a library call.
9779 auto &Ctx = *getContext();
9780 const auto& DL = getDataLayout();
9781
9783 // FIXME: pass in SDLoc
9784 CLI.setDebugLoc(dl).setChain(Chain);
9785
9786 RTLIB::LibcallImpl BzeroImpl = Libcalls->getLibcallImpl(RTLIB::BZERO);
9787 bool UseBZero = BzeroImpl != RTLIB::Unsupported && isNullConstant(Src);
9788
9789 // If zeroing out and bzero is present, use it.
9790 if (UseBZero) {
9792 Args.emplace_back(Dst, PointerType::getUnqual(Ctx));
9793 Args.emplace_back(Size, DL.getIntPtrType(Ctx));
9794 CLI.setLibCallee(
9795 Libcalls->getLibcallImplCallingConv(BzeroImpl), Type::getVoidTy(Ctx),
9796 getExternalSymbol(BzeroImpl, TLI->getPointerTy(DL)), std::move(Args));
9797 } else {
9798 RTLIB::LibcallImpl MemsetImpl = Libcalls->getLibcallImpl(RTLIB::MEMSET);
9799
9801 Args.emplace_back(Dst, PointerType::getUnqual(Ctx));
9802 Args.emplace_back(Src, Src.getValueType().getTypeForEVT(Ctx));
9803 Args.emplace_back(Size, DL.getIntPtrType(Ctx));
9804 CLI.setLibCallee(Libcalls->getLibcallImplCallingConv(MemsetImpl),
9805 Dst.getValueType().getTypeForEVT(Ctx),
9806 getExternalSymbol(MemsetImpl, TLI->getPointerTy(DL)),
9807 std::move(Args));
9808 }
9809
9810 RTLIB::LibcallImpl MemsetImpl = Libcalls->getLibcallImpl(RTLIB::MEMSET);
9811 bool LowersToMemset = MemsetImpl == RTLIB::impl_memset;
9812
9813 // If we're going to use bzero, make sure not to tail call unless the
9814 // subsequent return doesn't need a value, as bzero doesn't return the first
9815 // arg unlike memset.
9816 bool ReturnsFirstArg = CI && funcReturnsFirstArgOfCall(*CI) && !UseBZero;
9817 bool IsTailCall =
9818 CI && CI->isTailCall() &&
9819 isInTailCallPosition(*CI, getTarget(), ReturnsFirstArg && LowersToMemset);
9820 CLI.setDiscardResult().setTailCall(IsTailCall);
9821
9822 std::pair<SDValue, SDValue> CallResult = TLI->LowerCallTo(CLI);
9823 return CallResult.second;
9824}
9825
9828 Type *SizeTy, unsigned ElemSz,
9829 bool isTailCall,
9830 MachinePointerInfo DstPtrInfo) {
9831 // Emit a library call.
9833 Args.emplace_back(Dst, getDataLayout().getIntPtrType(*getContext()));
9834 Args.emplace_back(Value, Type::getInt8Ty(*getContext()));
9835 Args.emplace_back(Size, SizeTy);
9836
9837 RTLIB::Libcall LibraryCall =
9839 RTLIB::LibcallImpl LibcallImpl = Libcalls->getLibcallImpl(LibraryCall);
9840 if (LibcallImpl == RTLIB::Unsupported)
9841 report_fatal_error("Unsupported element size");
9842
9844 CLI.setDebugLoc(dl)
9845 .setChain(Chain)
9846 .setLibCallee(
9847 Libcalls->getLibcallImplCallingConv(LibcallImpl),
9849 getExternalSymbol(LibcallImpl, TLI->getPointerTy(getDataLayout())),
9850 std::move(Args))
9852 .setTailCall(isTailCall);
9853
9854 std::pair<SDValue, SDValue> CallResult = TLI->LowerCallTo(CLI);
9855 return CallResult.second;
9856}
9857
9858SDValue SelectionDAG::getAtomic(unsigned Opcode, const SDLoc &dl, EVT MemVT,
9860 MachineMemOperand *MMO,
9861 ISD::LoadExtType ExtType) {
9863 AddNodeIDNode(ID, Opcode, VTList, Ops);
9864 ID.AddInteger(MemVT.getRawBits());
9865 ID.AddInteger(getSyntheticNodeSubclassData<AtomicSDNode>(
9866 dl.getIROrder(), Opcode, VTList, MemVT, MMO, ExtType));
9867 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
9868 ID.AddInteger(MMO->getFlags());
9869 void* IP = nullptr;
9870 if (auto *E = cast_or_null<AtomicSDNode>(FindNodeOrInsertPos(ID, dl, IP))) {
9871 E->refineAlignment(MMO);
9872 E->refineRanges(MMO);
9873 return SDValue(E, 0);
9874 }
9875
9876 auto *N = newSDNode<AtomicSDNode>(dl.getIROrder(), dl.getDebugLoc(), Opcode,
9877 VTList, MemVT, MMO, ExtType);
9878 createOperands(N, Ops);
9879
9880 CSEMap.InsertNode(N, IP);
9881 InsertNode(N);
9882 SDValue V(N, 0);
9883 NewSDValueDbgMsg(V, "Creating new node: ", this);
9884 return V;
9885}
9886
9888 EVT MemVT, SDVTList VTs, SDValue Chain,
9889 SDValue Ptr, SDValue Cmp, SDValue Swp,
9890 MachineMemOperand *MMO) {
9891 assert(Opcode == ISD::ATOMIC_CMP_SWAP ||
9893 assert(Cmp.getValueType() == Swp.getValueType() && "Invalid Atomic Op Types");
9894
9895 SDValue Ops[] = {Chain, Ptr, Cmp, Swp};
9896 return getAtomic(Opcode, dl, MemVT, VTs, Ops, MMO);
9897}
9898
9899SDValue SelectionDAG::getAtomic(unsigned Opcode, const SDLoc &dl, EVT MemVT,
9900 SDValue Chain, SDValue Ptr, SDValue Val,
9901 MachineMemOperand *MMO) {
9902 assert((Opcode == ISD::ATOMIC_LOAD_ADD || Opcode == ISD::ATOMIC_LOAD_SUB ||
9903 Opcode == ISD::ATOMIC_LOAD_AND || Opcode == ISD::ATOMIC_LOAD_CLR ||
9904 Opcode == ISD::ATOMIC_LOAD_OR || Opcode == ISD::ATOMIC_LOAD_XOR ||
9905 Opcode == ISD::ATOMIC_LOAD_NAND || Opcode == ISD::ATOMIC_LOAD_MIN ||
9906 Opcode == ISD::ATOMIC_LOAD_MAX || Opcode == ISD::ATOMIC_LOAD_UMIN ||
9907 Opcode == ISD::ATOMIC_LOAD_UMAX || Opcode == ISD::ATOMIC_LOAD_FADD ||
9908 Opcode == ISD::ATOMIC_LOAD_FSUB || Opcode == ISD::ATOMIC_LOAD_FMAX ||
9909 Opcode == ISD::ATOMIC_LOAD_FMIN ||
9910 Opcode == ISD::ATOMIC_LOAD_FMINIMUM ||
9911 Opcode == ISD::ATOMIC_LOAD_FMAXIMUM ||
9912 Opcode == ISD::ATOMIC_LOAD_UINC_WRAP ||
9913 Opcode == ISD::ATOMIC_LOAD_UDEC_WRAP ||
9914 Opcode == ISD::ATOMIC_LOAD_USUB_COND ||
9915 Opcode == ISD::ATOMIC_LOAD_USUB_SAT || Opcode == ISD::ATOMIC_SWAP ||
9916 Opcode == ISD::ATOMIC_STORE) &&
9917 "Invalid Atomic Op");
9918
9919 EVT VT = Val.getValueType();
9920
9921 SDVTList VTs = Opcode == ISD::ATOMIC_STORE ? getVTList(MVT::Other) :
9922 getVTList(VT, MVT::Other);
9923 SDValue Ops[] = {Chain, Ptr, Val};
9924 return getAtomic(Opcode, dl, MemVT, VTs, Ops, MMO);
9925}
9926
9928 EVT MemVT, EVT VT, SDValue Chain,
9929 SDValue Ptr, MachineMemOperand *MMO) {
9930 SDVTList VTs = getVTList(VT, MVT::Other);
9931 SDValue Ops[] = {Chain, Ptr};
9932 return getAtomic(ISD::ATOMIC_LOAD, dl, MemVT, VTs, Ops, MMO, ExtType);
9933}
9934
9935/// getMergeValues - Create a MERGE_VALUES node from the given operands.
9937 if (Ops.size() == 1)
9938 return Ops[0];
9939
9941 VTs.reserve(Ops.size());
9942 for (const SDValue &Op : Ops)
9943 VTs.push_back(Op.getValueType());
9944 return getNode(ISD::MERGE_VALUES, dl, getVTList(VTs), Ops);
9945}
9946
9948 unsigned Opcode, const SDLoc &dl, SDVTList VTList, ArrayRef<SDValue> Ops,
9949 EVT MemVT, MachinePointerInfo PtrInfo, Align Alignment,
9951 const AAMDNodes &AAInfo) {
9952 if (Size.hasValue() && !Size.getValue())
9954
9956 MachineMemOperand *MMO =
9957 MF.getMachineMemOperand(PtrInfo, Flags, Size, Alignment, AAInfo);
9958
9959 return getMemIntrinsicNode(Opcode, dl, VTList, Ops, MemVT, MMO);
9960}
9961
9963 SDVTList VTList,
9964 ArrayRef<SDValue> Ops, EVT MemVT,
9965 MachineMemOperand *MMO) {
9966 return getMemIntrinsicNode(Opcode, dl, VTList, Ops, MemVT, ArrayRef(MMO));
9967}
9968
9970 SDVTList VTList,
9971 ArrayRef<SDValue> Ops, EVT MemVT,
9973 assert(!MMOs.empty() && "Must have at least one MMO");
9974 assert(
9975 (Opcode == ISD::INTRINSIC_VOID || Opcode == ISD::INTRINSIC_W_CHAIN ||
9976 Opcode == ISD::PREFETCH ||
9977 (Opcode <= (unsigned)std::numeric_limits<int>::max() &&
9978 Opcode >= ISD::BUILTIN_OP_END && TSI->isTargetMemoryOpcode(Opcode))) &&
9979 "Opcode is not a memory-accessing opcode!");
9980
9982 if (MMOs.size() == 1) {
9983 MemRefs = MMOs[0];
9984 } else {
9985 // Allocate: [size_t count][MMO*][MMO*]...
9986 size_t AllocSize =
9987 sizeof(size_t) + MMOs.size() * sizeof(MachineMemOperand *);
9988 void *Buffer = Allocator.Allocate(AllocSize, alignof(size_t));
9989 size_t *CountPtr = static_cast<size_t *>(Buffer);
9990 *CountPtr = MMOs.size();
9991 MachineMemOperand **Array =
9992 reinterpret_cast<MachineMemOperand **>(CountPtr + 1);
9993 llvm::copy(MMOs, Array);
9994 MemRefs = Array;
9995 }
9996
9997 // Memoize the node unless it returns a glue result.
9999 if (VTList.VTs[VTList.NumVTs-1] != MVT::Glue) {
10001 AddNodeIDNode(ID, Opcode, VTList, Ops);
10002 ID.AddInteger(getSyntheticNodeSubclassData<MemIntrinsicSDNode>(
10003 Opcode, dl.getIROrder(), VTList, MemVT, MemRefs));
10004 ID.AddInteger(MemVT.getRawBits());
10005 for (const MachineMemOperand *MMO : MMOs) {
10006 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10007 ID.AddInteger(MMO->getFlags());
10008 }
10009 void *IP = nullptr;
10010 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10011 cast<MemIntrinsicSDNode>(E)->refineAlignment(MMOs);
10012 return SDValue(E, 0);
10013 }
10014
10015 N = newSDNode<MemIntrinsicSDNode>(Opcode, dl.getIROrder(), dl.getDebugLoc(),
10016 VTList, MemVT, MemRefs);
10017 createOperands(N, Ops);
10018 CSEMap.InsertNode(N, IP);
10019 } else {
10020 N = newSDNode<MemIntrinsicSDNode>(Opcode, dl.getIROrder(), dl.getDebugLoc(),
10021 VTList, MemVT, MemRefs);
10022 createOperands(N, Ops);
10023 }
10024 InsertNode(N);
10025 SDValue V(N, 0);
10026 NewSDValueDbgMsg(V, "Creating new node: ", this);
10027 return V;
10028}
10029
10031 SDValue Chain, int FrameIndex) {
10032 const unsigned Opcode = IsStart ? ISD::LIFETIME_START : ISD::LIFETIME_END;
10033 const auto VTs = getVTList(MVT::Other);
10034 SDValue Ops[2] = {
10035 Chain,
10036 getFrameIndex(FrameIndex,
10037 getTargetLoweringInfo().getFrameIndexTy(getDataLayout()),
10038 true)};
10039
10041 AddNodeIDNode(ID, Opcode, VTs, Ops);
10042 ID.AddInteger(FrameIndex);
10043 void *IP = nullptr;
10044 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
10045 return SDValue(E, 0);
10046
10047 LifetimeSDNode *N =
10048 newSDNode<LifetimeSDNode>(Opcode, dl.getIROrder(), dl.getDebugLoc(), VTs);
10049 createOperands(N, Ops);
10050 CSEMap.InsertNode(N, IP);
10051 InsertNode(N);
10052 SDValue V(N, 0);
10053 NewSDValueDbgMsg(V, "Creating new node: ", this);
10054 return V;
10055}
10056
10058 uint64_t Guid, uint64_t Index,
10059 uint32_t Attr) {
10060 const unsigned Opcode = ISD::PSEUDO_PROBE;
10061 const auto VTs = getVTList(MVT::Other);
10062 SDValue Ops[] = {Chain};
10064 AddNodeIDNode(ID, Opcode, VTs, Ops);
10065 ID.AddInteger(Guid);
10066 ID.AddInteger(Index);
10067 void *IP = nullptr;
10068 if (SDNode *E = FindNodeOrInsertPos(ID, Dl, IP))
10069 return SDValue(E, 0);
10070
10071 auto *N = newSDNode<PseudoProbeSDNode>(
10072 Opcode, Dl.getIROrder(), Dl.getDebugLoc(), VTs, Guid, Index, Attr);
10073 createOperands(N, Ops);
10074 CSEMap.InsertNode(N, IP);
10075 InsertNode(N);
10076 SDValue V(N, 0);
10077 NewSDValueDbgMsg(V, "Creating new node: ", this);
10078 return V;
10079}
10080
10081/// InferPointerInfo - If the specified ptr/offset is a frame index, infer a
10082/// MachinePointerInfo record from it. This is particularly useful because the
10083/// code generator has many cases where it doesn't bother passing in a
10084/// MachinePointerInfo to getLoad or getStore when it has "FI+Cst".
10086 SelectionDAG &DAG, SDValue Ptr,
10087 int64_t Offset = 0) {
10088 // If this is FI+Offset, we can model it.
10089 if (const FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(Ptr))
10091 FI->getIndex(), Offset);
10092
10093 // If this is (FI+Offset1)+Offset2, we can model it.
10094 if (Ptr.getOpcode() != ISD::ADD ||
10097 return Info;
10098
10099 int FI = cast<FrameIndexSDNode>(Ptr.getOperand(0))->getIndex();
10101 DAG.getMachineFunction(), FI,
10102 Offset + cast<ConstantSDNode>(Ptr.getOperand(1))->getSExtValue());
10103}
10104
10105/// InferPointerInfo - If the specified ptr/offset is a frame index, infer a
10106/// MachinePointerInfo record from it. This is particularly useful because the
10107/// code generator has many cases where it doesn't bother passing in a
10108/// MachinePointerInfo to getLoad or getStore when it has "FI+Cst".
10110 SelectionDAG &DAG, SDValue Ptr,
10111 SDValue OffsetOp) {
10112 // If the 'Offset' value isn't a constant, we can't handle this.
10114 return InferPointerInfo(Info, DAG, Ptr, OffsetNode->getSExtValue());
10115 if (OffsetOp.isUndef())
10116 return InferPointerInfo(Info, DAG, Ptr);
10117 return Info;
10118}
10119
10121 EVT VT, const SDLoc &dl, SDValue Chain,
10122 SDValue Ptr, SDValue Offset,
10123 MachinePointerInfo PtrInfo, EVT MemVT,
10124 Align Alignment,
10125 MachineMemOperand::Flags MMOFlags,
10126 const AAMDNodes &AAInfo, const MDNode *Ranges) {
10127 assert(Chain.getValueType() == MVT::Other &&
10128 "Invalid chain type");
10129
10130 MMOFlags |= MachineMemOperand::MOLoad;
10131 assert((MMOFlags & MachineMemOperand::MOStore) == 0);
10132 // If we don't have a PtrInfo, infer the trivial frame index case to simplify
10133 // clients.
10134 if (PtrInfo.V.isNull())
10135 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr, Offset);
10136
10137 TypeSize Size = MemVT.getStoreSize();
10139 MachineMemOperand *MMO = MF.getMachineMemOperand(PtrInfo, MMOFlags, Size,
10140 Alignment, AAInfo, Ranges);
10141 return getLoad(AM, ExtType, VT, dl, Chain, Ptr, Offset, MemVT, MMO);
10142}
10143
10145 EVT VT, const SDLoc &dl, SDValue Chain,
10146 SDValue Ptr, SDValue Offset, EVT MemVT,
10147 MachineMemOperand *MMO) {
10148 if (VT == MemVT) {
10149 ExtType = ISD::NON_EXTLOAD;
10150 } else if (ExtType == ISD::NON_EXTLOAD) {
10151 assert(VT == MemVT && "Non-extending load from different memory type!");
10152 } else {
10153 // Extending load.
10154 assert(MemVT.getScalarType().bitsLT(VT.getScalarType()) &&
10155 "Should only be an extending load, not truncating!");
10156 assert(VT.isInteger() == MemVT.isInteger() &&
10157 "Cannot convert from FP to Int or Int -> FP!");
10158 assert(VT.isVector() == MemVT.isVector() &&
10159 "Cannot use an ext load to convert to or from a vector!");
10160 assert((!VT.isVector() ||
10162 "Cannot use an ext load to change the number of vector elements!");
10163 }
10164
10165 assert((!MMO->getRanges() ||
10167 ->getBitWidth() == MemVT.getScalarSizeInBits() &&
10168 MemVT.isInteger())) &&
10169 "Range metadata and load type must match!");
10170
10171 bool Indexed = AM != ISD::UNINDEXED;
10172 assert((Indexed || Offset.isUndef()) && "Unindexed load with an offset!");
10173
10174 SDVTList VTs = Indexed ?
10175 getVTList(VT, Ptr.getValueType(), MVT::Other) : getVTList(VT, MVT::Other);
10176 SDValue Ops[] = { Chain, Ptr, Offset };
10178 AddNodeIDNode(ID, ISD::LOAD, VTs, Ops);
10179 ID.AddInteger(MemVT.getRawBits());
10180 ID.AddInteger(getSyntheticNodeSubclassData<LoadSDNode>(
10181 dl.getIROrder(), VTs, AM, ExtType, MemVT, MMO));
10182 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10183 ID.AddInteger(MMO->getFlags());
10184 void *IP = nullptr;
10185 if (auto *E = cast_or_null<LoadSDNode>(FindNodeOrInsertPos(ID, dl, IP))) {
10186 E->refineAlignment(MMO);
10187 E->refineRanges(MMO);
10188 return SDValue(E, 0);
10189 }
10190 auto *N = newSDNode<LoadSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10191 ExtType, MemVT, MMO);
10192 createOperands(N, Ops);
10193
10194 CSEMap.InsertNode(N, IP);
10195 InsertNode(N);
10196 SDValue V(N, 0);
10197 NewSDValueDbgMsg(V, "Creating new node: ", this);
10198 return V;
10199}
10200
10202 SDValue Ptr, MachinePointerInfo PtrInfo,
10203 MaybeAlign Alignment,
10204 MachineMemOperand::Flags MMOFlags,
10205 const AAMDNodes &AAInfo, const MDNode *Ranges) {
10207 return getLoad(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, dl, Chain, Ptr, Undef,
10208 PtrInfo, VT, Alignment, MMOFlags, AAInfo, Ranges);
10209}
10210
10212 SDValue Ptr, MachineMemOperand *MMO) {
10214 return getLoad(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, dl, Chain, Ptr, Undef,
10215 VT, MMO);
10216}
10217
10219 EVT VT, SDValue Chain, SDValue Ptr,
10220 MachinePointerInfo PtrInfo, EVT MemVT,
10221 MaybeAlign Alignment,
10222 MachineMemOperand::Flags MMOFlags,
10223 const AAMDNodes &AAInfo) {
10225 return getLoad(ISD::UNINDEXED, ExtType, VT, dl, Chain, Ptr, Undef, PtrInfo,
10226 MemVT, Alignment, MMOFlags, AAInfo);
10227}
10228
10230 EVT VT, SDValue Chain, SDValue Ptr, EVT MemVT,
10231 MachineMemOperand *MMO) {
10233 return getLoad(ISD::UNINDEXED, ExtType, VT, dl, Chain, Ptr, Undef,
10234 MemVT, MMO);
10235}
10236
10240 LoadSDNode *LD = cast<LoadSDNode>(OrigLoad);
10241 assert(LD->getOffset().isUndef() && "Load is already a indexed load!");
10242 // Don't propagate the invariant or dereferenceable flags.
10243 auto MMOFlags =
10244 LD->getMemOperand()->getFlags() &
10246 return getLoad(AM, LD->getExtensionType(), OrigLoad.getValueType(), dl,
10247 LD->getChain(), Base, Offset, LD->getPointerInfo(),
10248 LD->getMemoryVT(), LD->getAlign(), MMOFlags, LD->getAAInfo());
10249}
10250
10252 SDValue Ptr, MachinePointerInfo PtrInfo,
10253 Align Alignment,
10254 MachineMemOperand::Flags MMOFlags,
10255 const AAMDNodes &AAInfo) {
10256 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10257
10258 MMOFlags |= MachineMemOperand::MOStore;
10259 assert((MMOFlags & MachineMemOperand::MOLoad) == 0);
10260
10261 if (PtrInfo.V.isNull())
10262 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr);
10263
10266 MachineMemOperand *MMO =
10267 MF.getMachineMemOperand(PtrInfo, MMOFlags, Size, Alignment, AAInfo);
10268 return getStore(Chain, dl, Val, Ptr, MMO);
10269}
10270
10272 SDValue Ptr, MachineMemOperand *MMO) {
10274 return getStore(Chain, dl, Val, Ptr, Undef, Val.getValueType(), MMO,
10276}
10277
10279 SDValue Ptr, SDValue Offset, EVT SVT,
10281 bool IsTruncating) {
10282 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10283 EVT VT = Val.getValueType();
10284 if (VT == SVT) {
10285 IsTruncating = false;
10286 } else if (!IsTruncating) {
10287 assert(VT == SVT && "No-truncating store from different memory type!");
10288 } else {
10290 "Should only be a truncating store, not extending!");
10291 assert(VT.isInteger() == SVT.isInteger() && "Can't do FP-INT conversion!");
10292 assert(VT.isVector() == SVT.isVector() &&
10293 "Cannot use trunc store to convert to or from a vector!");
10294 assert((!VT.isVector() ||
10296 "Cannot use trunc store to change the number of vector elements!");
10297 }
10298
10299 bool Indexed = AM != ISD::UNINDEXED;
10300 assert((Indexed || Offset.isUndef()) && "Unindexed store with an offset!");
10301 SDVTList VTs = Indexed ? getVTList(Ptr.getValueType(), MVT::Other)
10302 : getVTList(MVT::Other);
10303 SDValue Ops[] = {Chain, Val, Ptr, Offset};
10306 ID.AddInteger(SVT.getRawBits());
10307 ID.AddInteger(getSyntheticNodeSubclassData<StoreSDNode>(
10308 dl.getIROrder(), VTs, AM, IsTruncating, SVT, MMO));
10309 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10310 ID.AddInteger(MMO->getFlags());
10311 void *IP = nullptr;
10312 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10313 cast<StoreSDNode>(E)->refineAlignment(MMO);
10314 return SDValue(E, 0);
10315 }
10316 auto *N = newSDNode<StoreSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10317 IsTruncating, SVT, MMO);
10318 createOperands(N, Ops);
10319
10320 CSEMap.InsertNode(N, IP);
10321 InsertNode(N);
10322 SDValue V(N, 0);
10323 NewSDValueDbgMsg(V, "Creating new node: ", this);
10324 return V;
10325}
10326
10328 SDValue Ptr, MachinePointerInfo PtrInfo,
10329 EVT SVT, Align Alignment,
10330 MachineMemOperand::Flags MMOFlags,
10331 const AAMDNodes &AAInfo) {
10332 assert(Chain.getValueType() == MVT::Other &&
10333 "Invalid chain type");
10334
10335 MMOFlags |= MachineMemOperand::MOStore;
10336 assert((MMOFlags & MachineMemOperand::MOLoad) == 0);
10337
10338 if (PtrInfo.V.isNull())
10339 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr);
10340
10342 MachineMemOperand *MMO = MF.getMachineMemOperand(
10343 PtrInfo, MMOFlags, SVT.getStoreSize(), Alignment, AAInfo);
10344 return getTruncStore(Chain, dl, Val, Ptr, SVT, MMO);
10345}
10346
10348 SDValue Ptr, EVT SVT,
10349 MachineMemOperand *MMO) {
10351 return getStore(Chain, dl, Val, Ptr, Undef, SVT, MMO, ISD::UNINDEXED, true);
10352}
10353
10357 StoreSDNode *ST = cast<StoreSDNode>(OrigStore);
10358 assert(ST->getOffset().isUndef() && "Store is already a indexed store!");
10359 return getStore(ST->getChain(), dl, ST->getValue(), Base, Offset,
10360 ST->getMemoryVT(), ST->getMemOperand(), AM,
10361 ST->isTruncatingStore());
10362}
10363
10365 ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, EVT VT, const SDLoc &dl,
10366 SDValue Chain, SDValue Ptr, SDValue Offset, SDValue Mask, SDValue EVL,
10367 MachinePointerInfo PtrInfo, EVT MemVT, Align Alignment,
10368 MachineMemOperand::Flags MMOFlags, const AAMDNodes &AAInfo,
10369 const MDNode *Ranges, bool IsExpanding) {
10370 MMOFlags |= MachineMemOperand::MOLoad;
10371 assert((MMOFlags & MachineMemOperand::MOStore) == 0);
10372 // If we don't have a PtrInfo, infer the trivial frame index case to simplify
10373 // clients.
10374 if (PtrInfo.V.isNull())
10375 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr, Offset);
10376
10377 TypeSize Size = MemVT.getStoreSize();
10379 MachineMemOperand *MMO = MF.getMachineMemOperand(PtrInfo, MMOFlags, Size,
10380 Alignment, AAInfo, Ranges);
10381 return getLoadVP(AM, ExtType, VT, dl, Chain, Ptr, Offset, Mask, EVL, MemVT,
10382 MMO, IsExpanding);
10383}
10384
10386 ISD::LoadExtType ExtType, EVT VT,
10387 const SDLoc &dl, SDValue Chain, SDValue Ptr,
10388 SDValue Offset, SDValue Mask, SDValue EVL,
10389 EVT MemVT, MachineMemOperand *MMO,
10390 bool IsExpanding) {
10391 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10392 assert(Mask.getValueType().getVectorElementCount() ==
10393 VT.getVectorElementCount() &&
10394 "Vector width mismatch between mask and data");
10395
10396 bool Indexed = AM != ISD::UNINDEXED;
10397 assert((Indexed || Offset.isUndef()) && "Unindexed load with an offset!");
10398
10399 SDVTList VTs = Indexed ? getVTList(VT, Ptr.getValueType(), MVT::Other)
10400 : getVTList(VT, MVT::Other);
10401 SDValue Ops[] = {Chain, Ptr, Offset, Mask, EVL};
10403 AddNodeIDNode(ID, ISD::VP_LOAD, VTs, Ops);
10404 ID.AddInteger(MemVT.getRawBits());
10405 ID.AddInteger(getSyntheticNodeSubclassData<VPLoadSDNode>(
10406 dl.getIROrder(), VTs, AM, ExtType, IsExpanding, MemVT, MMO));
10407 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10408 ID.AddInteger(MMO->getFlags());
10409 void *IP = nullptr;
10410 if (auto *E = cast_or_null<VPLoadSDNode>(FindNodeOrInsertPos(ID, dl, IP))) {
10411 E->refineAlignment(MMO);
10412 E->refineRanges(MMO);
10413 return SDValue(E, 0);
10414 }
10415 auto *N = newSDNode<VPLoadSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10416 ExtType, IsExpanding, MemVT, MMO);
10417 createOperands(N, Ops);
10418
10419 CSEMap.InsertNode(N, IP);
10420 InsertNode(N);
10421 SDValue V(N, 0);
10422 NewSDValueDbgMsg(V, "Creating new node: ", this);
10423 return V;
10424}
10425
10427 SDValue Ptr, SDValue Mask, SDValue EVL,
10428 MachinePointerInfo PtrInfo,
10429 MaybeAlign Alignment,
10430 MachineMemOperand::Flags MMOFlags,
10431 const AAMDNodes &AAInfo, const MDNode *Ranges,
10432 bool IsExpanding) {
10434 return getLoadVP(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, dl, Chain, Ptr, Undef,
10435 Mask, EVL, PtrInfo, VT, Alignment, MMOFlags, AAInfo, Ranges,
10436 IsExpanding);
10437}
10438
10440 SDValue Ptr, SDValue Mask, SDValue EVL,
10441 MachineMemOperand *MMO, bool IsExpanding) {
10443 return getLoadVP(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, dl, Chain, Ptr, Undef,
10444 Mask, EVL, VT, MMO, IsExpanding);
10445}
10446
10448 EVT VT, SDValue Chain, SDValue Ptr,
10449 SDValue Mask, SDValue EVL,
10450 MachinePointerInfo PtrInfo, EVT MemVT,
10451 MaybeAlign Alignment,
10452 MachineMemOperand::Flags MMOFlags,
10453 const AAMDNodes &AAInfo, bool IsExpanding) {
10455 return getLoadVP(ISD::UNINDEXED, ExtType, VT, dl, Chain, Ptr, Undef, Mask,
10456 EVL, PtrInfo, MemVT, Alignment, MMOFlags, AAInfo, nullptr,
10457 IsExpanding);
10458}
10459
10461 EVT VT, SDValue Chain, SDValue Ptr,
10462 SDValue Mask, SDValue EVL, EVT MemVT,
10463 MachineMemOperand *MMO, bool IsExpanding) {
10465 return getLoadVP(ISD::UNINDEXED, ExtType, VT, dl, Chain, Ptr, Undef, Mask,
10466 EVL, MemVT, MMO, IsExpanding);
10467}
10468
10472 auto *LD = cast<VPLoadSDNode>(OrigLoad);
10473 assert(LD->getOffset().isUndef() && "Load is already a indexed load!");
10474 // Don't propagate the invariant or dereferenceable flags.
10475 auto MMOFlags =
10476 LD->getMemOperand()->getFlags() &
10478 return getLoadVP(AM, LD->getExtensionType(), OrigLoad.getValueType(), dl,
10479 LD->getChain(), Base, Offset, LD->getMask(),
10480 LD->getVectorLength(), LD->getPointerInfo(),
10481 LD->getMemoryVT(), LD->getAlign(), MMOFlags, LD->getAAInfo(),
10482 nullptr, LD->isExpandingLoad());
10483}
10484
10486 SDValue Ptr, SDValue Offset, SDValue Mask,
10487 SDValue EVL, EVT MemVT, MachineMemOperand *MMO,
10488 ISD::MemIndexedMode AM, bool IsTruncating,
10489 bool IsCompressing) {
10490 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10491 assert(Mask.getValueType().getVectorElementCount() ==
10493 "Vector width mismatch between mask and data");
10494
10495 bool Indexed = AM != ISD::UNINDEXED;
10496 assert((Indexed || Offset.isUndef()) && "Unindexed vp_store with an offset!");
10497 SDVTList VTs = Indexed ? getVTList(Ptr.getValueType(), MVT::Other)
10498 : getVTList(MVT::Other);
10499 SDValue Ops[] = {Chain, Val, Ptr, Offset, Mask, EVL};
10501 AddNodeIDNode(ID, ISD::VP_STORE, VTs, Ops);
10502 ID.AddInteger(MemVT.getRawBits());
10503 ID.AddInteger(getSyntheticNodeSubclassData<VPStoreSDNode>(
10504 dl.getIROrder(), VTs, AM, IsTruncating, IsCompressing, MemVT, MMO));
10505 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10506 ID.AddInteger(MMO->getFlags());
10507 void *IP = nullptr;
10508 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10509 cast<VPStoreSDNode>(E)->refineAlignment(MMO);
10510 return SDValue(E, 0);
10511 }
10512 auto *N = newSDNode<VPStoreSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10513 IsTruncating, IsCompressing, MemVT, MMO);
10514 createOperands(N, Ops);
10515
10516 CSEMap.InsertNode(N, IP);
10517 InsertNode(N);
10518 SDValue V(N, 0);
10519 NewSDValueDbgMsg(V, "Creating new node: ", this);
10520 return V;
10521}
10522
10524 SDValue Val, SDValue Ptr, SDValue Mask,
10525 SDValue EVL, MachinePointerInfo PtrInfo,
10526 EVT SVT, Align Alignment,
10527 MachineMemOperand::Flags MMOFlags,
10528 const AAMDNodes &AAInfo,
10529 bool IsCompressing) {
10530 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10531
10532 MMOFlags |= MachineMemOperand::MOStore;
10533 assert((MMOFlags & MachineMemOperand::MOLoad) == 0);
10534
10535 if (PtrInfo.V.isNull())
10536 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr);
10537
10539 MachineMemOperand *MMO = MF.getMachineMemOperand(
10540 PtrInfo, MMOFlags, SVT.getStoreSize(), Alignment, AAInfo);
10541 return getTruncStoreVP(Chain, dl, Val, Ptr, Mask, EVL, SVT, MMO,
10542 IsCompressing);
10543}
10544
10546 SDValue Val, SDValue Ptr, SDValue Mask,
10547 SDValue EVL, EVT SVT,
10548 MachineMemOperand *MMO,
10549 bool IsCompressing) {
10550 EVT VT = Val.getValueType();
10551
10552 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10553 if (VT == SVT)
10554 return getStoreVP(Chain, dl, Val, Ptr, getUNDEF(Ptr.getValueType()), Mask,
10555 EVL, VT, MMO, ISD::UNINDEXED,
10556 /*IsTruncating*/ false, IsCompressing);
10557
10559 "Should only be a truncating store, not extending!");
10560 assert(VT.isInteger() == SVT.isInteger() && "Can't do FP-INT conversion!");
10561 assert(VT.isVector() == SVT.isVector() &&
10562 "Cannot use trunc store to convert to or from a vector!");
10563 assert((!VT.isVector() ||
10565 "Cannot use trunc store to change the number of vector elements!");
10566
10567 SDVTList VTs = getVTList(MVT::Other);
10569 SDValue Ops[] = {Chain, Val, Ptr, Undef, Mask, EVL};
10571 AddNodeIDNode(ID, ISD::VP_STORE, VTs, Ops);
10572 ID.AddInteger(SVT.getRawBits());
10573 ID.AddInteger(getSyntheticNodeSubclassData<VPStoreSDNode>(
10574 dl.getIROrder(), VTs, ISD::UNINDEXED, true, IsCompressing, SVT, MMO));
10575 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10576 ID.AddInteger(MMO->getFlags());
10577 void *IP = nullptr;
10578 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10579 cast<VPStoreSDNode>(E)->refineAlignment(MMO);
10580 return SDValue(E, 0);
10581 }
10582 auto *N =
10583 newSDNode<VPStoreSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs,
10584 ISD::UNINDEXED, true, IsCompressing, SVT, MMO);
10585 createOperands(N, Ops);
10586
10587 CSEMap.InsertNode(N, IP);
10588 InsertNode(N);
10589 SDValue V(N, 0);
10590 NewSDValueDbgMsg(V, "Creating new node: ", this);
10591 return V;
10592}
10593
10597 auto *ST = cast<VPStoreSDNode>(OrigStore);
10598 assert(ST->getOffset().isUndef() && "Store is already an indexed store!");
10599 SDVTList VTs = getVTList(Base.getValueType(), MVT::Other);
10600 SDValue Ops[] = {ST->getChain(), ST->getValue(), Base,
10601 Offset, ST->getMask(), ST->getVectorLength()};
10603 AddNodeIDNode(ID, ISD::VP_STORE, VTs, Ops);
10604 ID.AddInteger(ST->getMemoryVT().getRawBits());
10605 ID.AddInteger(ST->getRawSubclassData());
10606 ID.AddInteger(ST->getPointerInfo().getAddrSpace());
10607 ID.AddInteger(ST->getMemOperand()->getFlags());
10608 void *IP = nullptr;
10609 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
10610 return SDValue(E, 0);
10611
10612 auto *N = newSDNode<VPStoreSDNode>(
10613 dl.getIROrder(), dl.getDebugLoc(), VTs, AM, ST->isTruncatingStore(),
10614 ST->isCompressingStore(), ST->getMemoryVT(), ST->getMemOperand());
10615 createOperands(N, Ops);
10616
10617 CSEMap.InsertNode(N, IP);
10618 InsertNode(N);
10619 SDValue V(N, 0);
10620 NewSDValueDbgMsg(V, "Creating new node: ", this);
10621 return V;
10622}
10623
10625 ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, EVT VT, const SDLoc &DL,
10626 SDValue Chain, SDValue Ptr, SDValue Offset, SDValue Stride, SDValue Mask,
10627 SDValue EVL, EVT MemVT, MachineMemOperand *MMO, bool IsExpanding) {
10628 bool Indexed = AM != ISD::UNINDEXED;
10629 assert((Indexed || Offset.isUndef()) && "Unindexed load with an offset!");
10630
10631 SDValue Ops[] = {Chain, Ptr, Offset, Stride, Mask, EVL};
10632 SDVTList VTs = Indexed ? getVTList(VT, Ptr.getValueType(), MVT::Other)
10633 : getVTList(VT, MVT::Other);
10635 AddNodeIDNode(ID, ISD::EXPERIMENTAL_VP_STRIDED_LOAD, VTs, Ops);
10636 ID.AddInteger(VT.getRawBits());
10637 ID.AddInteger(getSyntheticNodeSubclassData<VPStridedLoadSDNode>(
10638 DL.getIROrder(), VTs, AM, ExtType, IsExpanding, MemVT, MMO));
10639 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10640
10641 void *IP = nullptr;
10642 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
10643 cast<VPStridedLoadSDNode>(E)->refineAlignment(MMO);
10644 return SDValue(E, 0);
10645 }
10646
10647 auto *N =
10648 newSDNode<VPStridedLoadSDNode>(DL.getIROrder(), DL.getDebugLoc(), VTs, AM,
10649 ExtType, IsExpanding, MemVT, MMO);
10650 createOperands(N, Ops);
10651 CSEMap.InsertNode(N, IP);
10652 InsertNode(N);
10653 SDValue V(N, 0);
10654 NewSDValueDbgMsg(V, "Creating new node: ", this);
10655 return V;
10656}
10657
10659 SDValue Ptr, SDValue Stride,
10660 SDValue Mask, SDValue EVL,
10661 MachineMemOperand *MMO,
10662 bool IsExpanding) {
10664 return getStridedLoadVP(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, DL, Chain, Ptr,
10665 Undef, Stride, Mask, EVL, VT, MMO, IsExpanding);
10666}
10667
10669 ISD::LoadExtType ExtType, const SDLoc &DL, EVT VT, SDValue Chain,
10670 SDValue Ptr, SDValue Stride, SDValue Mask, SDValue EVL, EVT MemVT,
10671 MachineMemOperand *MMO, bool IsExpanding) {
10673 return getStridedLoadVP(ISD::UNINDEXED, ExtType, VT, DL, Chain, Ptr, Undef,
10674 Stride, Mask, EVL, MemVT, MMO, IsExpanding);
10675}
10676
10678 SDValue Val, SDValue Ptr,
10679 SDValue Offset, SDValue Stride,
10680 SDValue Mask, SDValue EVL, EVT MemVT,
10681 MachineMemOperand *MMO,
10683 bool IsTruncating, bool IsCompressing) {
10684 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10685 bool Indexed = AM != ISD::UNINDEXED;
10686 assert((Indexed || Offset.isUndef()) && "Unindexed vp_store with an offset!");
10687 SDVTList VTs = Indexed ? getVTList(Ptr.getValueType(), MVT::Other)
10688 : getVTList(MVT::Other);
10689 SDValue Ops[] = {Chain, Val, Ptr, Offset, Stride, Mask, EVL};
10691 AddNodeIDNode(ID, ISD::EXPERIMENTAL_VP_STRIDED_STORE, VTs, Ops);
10692 ID.AddInteger(MemVT.getRawBits());
10693 ID.AddInteger(getSyntheticNodeSubclassData<VPStridedStoreSDNode>(
10694 DL.getIROrder(), VTs, AM, IsTruncating, IsCompressing, MemVT, MMO));
10695 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10696 void *IP = nullptr;
10697 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
10698 cast<VPStridedStoreSDNode>(E)->refineAlignment(MMO);
10699 return SDValue(E, 0);
10700 }
10701 auto *N = newSDNode<VPStridedStoreSDNode>(DL.getIROrder(), DL.getDebugLoc(),
10702 VTs, AM, IsTruncating,
10703 IsCompressing, MemVT, MMO);
10704 createOperands(N, Ops);
10705
10706 CSEMap.InsertNode(N, IP);
10707 InsertNode(N);
10708 SDValue V(N, 0);
10709 NewSDValueDbgMsg(V, "Creating new node: ", this);
10710 return V;
10711}
10712
10714 SDValue Val, SDValue Ptr,
10715 SDValue Stride, SDValue Mask,
10716 SDValue EVL, EVT SVT,
10717 MachineMemOperand *MMO,
10718 bool IsCompressing) {
10719 EVT VT = Val.getValueType();
10720
10721 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10722 if (VT == SVT)
10723 return getStridedStoreVP(Chain, DL, Val, Ptr, getUNDEF(Ptr.getValueType()),
10724 Stride, Mask, EVL, VT, MMO, ISD::UNINDEXED,
10725 /*IsTruncating*/ false, IsCompressing);
10726
10728 "Should only be a truncating store, not extending!");
10729 assert(VT.isInteger() == SVT.isInteger() && "Can't do FP-INT conversion!");
10730 assert(VT.isVector() == SVT.isVector() &&
10731 "Cannot use trunc store to convert to or from a vector!");
10732 assert((!VT.isVector() ||
10734 "Cannot use trunc store to change the number of vector elements!");
10735
10736 SDVTList VTs = getVTList(MVT::Other);
10738 SDValue Ops[] = {Chain, Val, Ptr, Undef, Stride, Mask, EVL};
10740 AddNodeIDNode(ID, ISD::EXPERIMENTAL_VP_STRIDED_STORE, VTs, Ops);
10741 ID.AddInteger(SVT.getRawBits());
10742 ID.AddInteger(getSyntheticNodeSubclassData<VPStridedStoreSDNode>(
10743 DL.getIROrder(), VTs, ISD::UNINDEXED, true, IsCompressing, SVT, MMO));
10744 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10745 void *IP = nullptr;
10746 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
10747 cast<VPStridedStoreSDNode>(E)->refineAlignment(MMO);
10748 return SDValue(E, 0);
10749 }
10750 auto *N = newSDNode<VPStridedStoreSDNode>(DL.getIROrder(), DL.getDebugLoc(),
10751 VTs, ISD::UNINDEXED, true,
10752 IsCompressing, SVT, MMO);
10753 createOperands(N, Ops);
10754
10755 CSEMap.InsertNode(N, IP);
10756 InsertNode(N);
10757 SDValue V(N, 0);
10758 NewSDValueDbgMsg(V, "Creating new node: ", this);
10759 return V;
10760}
10761
10764 ISD::MemIndexType IndexType) {
10765 assert(Ops.size() == 6 && "Incompatible number of operands");
10766
10768 AddNodeIDNode(ID, ISD::VP_GATHER, VTs, Ops);
10769 ID.AddInteger(VT.getRawBits());
10770 ID.AddInteger(getSyntheticNodeSubclassData<VPGatherSDNode>(
10771 dl.getIROrder(), VTs, VT, MMO, IndexType));
10772 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10773 ID.AddInteger(MMO->getFlags());
10774 void *IP = nullptr;
10775 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10776 cast<VPGatherSDNode>(E)->refineAlignment(MMO);
10777 return SDValue(E, 0);
10778 }
10779
10780 auto *N = newSDNode<VPGatherSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs,
10781 VT, MMO, IndexType);
10782 createOperands(N, Ops);
10783
10784 assert(N->getMask().getValueType().getVectorElementCount() ==
10785 N->getValueType(0).getVectorElementCount() &&
10786 "Vector width mismatch between mask and data");
10787 assert(N->getIndex().getValueType().getVectorElementCount().isScalable() ==
10788 N->getValueType(0).getVectorElementCount().isScalable() &&
10789 "Scalable flags of index and data do not match");
10791 N->getIndex().getValueType().getVectorElementCount(),
10792 N->getValueType(0).getVectorElementCount()) &&
10793 "Vector width mismatch between index and data");
10794 assert(isa<ConstantSDNode>(N->getScale()) &&
10795 N->getScale()->getAsAPIntVal().isPowerOf2() &&
10796 "Scale should be a constant power of 2");
10797
10798 CSEMap.InsertNode(N, IP);
10799 InsertNode(N);
10800 SDValue V(N, 0);
10801 NewSDValueDbgMsg(V, "Creating new node: ", this);
10802 return V;
10803}
10804
10807 MachineMemOperand *MMO,
10808 ISD::MemIndexType IndexType) {
10809 assert(Ops.size() == 7 && "Incompatible number of operands");
10810
10812 AddNodeIDNode(ID, ISD::VP_SCATTER, VTs, Ops);
10813 ID.AddInteger(VT.getRawBits());
10814 ID.AddInteger(getSyntheticNodeSubclassData<VPScatterSDNode>(
10815 dl.getIROrder(), VTs, VT, MMO, IndexType));
10816 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10817 ID.AddInteger(MMO->getFlags());
10818 void *IP = nullptr;
10819 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10820 cast<VPScatterSDNode>(E)->refineAlignment(MMO);
10821 return SDValue(E, 0);
10822 }
10823 auto *N = newSDNode<VPScatterSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs,
10824 VT, MMO, IndexType);
10825 createOperands(N, Ops);
10826
10827 assert(N->getMask().getValueType().getVectorElementCount() ==
10828 N->getValue().getValueType().getVectorElementCount() &&
10829 "Vector width mismatch between mask and data");
10830 assert(
10831 N->getIndex().getValueType().getVectorElementCount().isScalable() ==
10832 N->getValue().getValueType().getVectorElementCount().isScalable() &&
10833 "Scalable flags of index and data do not match");
10835 N->getIndex().getValueType().getVectorElementCount(),
10836 N->getValue().getValueType().getVectorElementCount()) &&
10837 "Vector width mismatch between index and data");
10838 assert(isa<ConstantSDNode>(N->getScale()) &&
10839 N->getScale()->getAsAPIntVal().isPowerOf2() &&
10840 "Scale should be a constant power of 2");
10841
10842 CSEMap.InsertNode(N, IP);
10843 InsertNode(N);
10844 SDValue V(N, 0);
10845 NewSDValueDbgMsg(V, "Creating new node: ", this);
10846 return V;
10847}
10848
10851 SDValue PassThru, EVT MemVT,
10852 MachineMemOperand *MMO,
10854 ISD::LoadExtType ExtTy, bool isExpanding) {
10855 bool Indexed = AM != ISD::UNINDEXED;
10856 assert((Indexed || Offset.isUndef()) &&
10857 "Unindexed masked load with an offset!");
10858 SDVTList VTs = Indexed ? getVTList(VT, Base.getValueType(), MVT::Other)
10859 : getVTList(VT, MVT::Other);
10860 SDValue Ops[] = {Chain, Base, Offset, Mask, PassThru};
10863 ID.AddInteger(MemVT.getRawBits());
10864 ID.AddInteger(getSyntheticNodeSubclassData<MaskedLoadSDNode>(
10865 dl.getIROrder(), VTs, AM, ExtTy, isExpanding, MemVT, MMO));
10866 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10867 ID.AddInteger(MMO->getFlags());
10868 void *IP = nullptr;
10869 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10870 cast<MaskedLoadSDNode>(E)->refineAlignment(MMO);
10871 return SDValue(E, 0);
10872 }
10873 auto *N = newSDNode<MaskedLoadSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs,
10874 AM, ExtTy, isExpanding, MemVT, MMO);
10875 createOperands(N, Ops);
10876
10877 CSEMap.InsertNode(N, IP);
10878 InsertNode(N);
10879 SDValue V(N, 0);
10880 NewSDValueDbgMsg(V, "Creating new node: ", this);
10881 return V;
10882}
10883
10888 assert(LD->getOffset().isUndef() && "Masked load is already a indexed load!");
10889 return getMaskedLoad(OrigLoad.getValueType(), dl, LD->getChain(), Base,
10890 Offset, LD->getMask(), LD->getPassThru(),
10891 LD->getMemoryVT(), LD->getMemOperand(), AM,
10892 LD->getExtensionType(), LD->isExpandingLoad());
10893}
10894
10897 SDValue Mask, EVT MemVT,
10898 MachineMemOperand *MMO,
10899 ISD::MemIndexedMode AM, bool IsTruncating,
10900 bool IsCompressing) {
10901 assert(Chain.getValueType() == MVT::Other &&
10902 "Invalid chain type");
10903 bool Indexed = AM != ISD::UNINDEXED;
10904 assert((Indexed || Offset.isUndef()) &&
10905 "Unindexed masked store with an offset!");
10906 SDVTList VTs = Indexed ? getVTList(Base.getValueType(), MVT::Other)
10907 : getVTList(MVT::Other);
10908 SDValue Ops[] = {Chain, Val, Base, Offset, Mask};
10911 ID.AddInteger(MemVT.getRawBits());
10912 ID.AddInteger(getSyntheticNodeSubclassData<MaskedStoreSDNode>(
10913 dl.getIROrder(), VTs, AM, IsTruncating, IsCompressing, MemVT, MMO));
10914 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10915 ID.AddInteger(MMO->getFlags());
10916 void *IP = nullptr;
10917 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10918 cast<MaskedStoreSDNode>(E)->refineAlignment(MMO);
10919 return SDValue(E, 0);
10920 }
10921 auto *N =
10922 newSDNode<MaskedStoreSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10923 IsTruncating, IsCompressing, MemVT, MMO);
10924 createOperands(N, Ops);
10925
10926 CSEMap.InsertNode(N, IP);
10927 InsertNode(N);
10928 SDValue V(N, 0);
10929 NewSDValueDbgMsg(V, "Creating new node: ", this);
10930 return V;
10931}
10932
10937 assert(ST->getOffset().isUndef() &&
10938 "Masked store is already a indexed store!");
10939 return getMaskedStore(ST->getChain(), dl, ST->getValue(), Base, Offset,
10940 ST->getMask(), ST->getMemoryVT(), ST->getMemOperand(),
10941 AM, ST->isTruncatingStore(), ST->isCompressingStore());
10942}
10943
10946 MachineMemOperand *MMO,
10947 ISD::MemIndexType IndexType,
10948 ISD::LoadExtType ExtTy) {
10949 assert(Ops.size() == 6 && "Incompatible number of operands");
10950
10953 ID.AddInteger(MemVT.getRawBits());
10954 ID.AddInteger(getSyntheticNodeSubclassData<MaskedGatherSDNode>(
10955 dl.getIROrder(), VTs, MemVT, MMO, IndexType, ExtTy));
10956 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10957 ID.AddInteger(MMO->getFlags());
10958 void *IP = nullptr;
10959 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10960 cast<MaskedGatherSDNode>(E)->refineAlignment(MMO);
10961 return SDValue(E, 0);
10962 }
10963
10964 auto *N = newSDNode<MaskedGatherSDNode>(dl.getIROrder(), dl.getDebugLoc(),
10965 VTs, MemVT, MMO, IndexType, ExtTy);
10966 createOperands(N, Ops);
10967
10968 assert(N->getPassThru().getValueType() == N->getValueType(0) &&
10969 "Incompatible type of the PassThru value in MaskedGatherSDNode");
10970 assert(N->getMask().getValueType().getVectorElementCount() ==
10971 N->getValueType(0).getVectorElementCount() &&
10972 "Vector width mismatch between mask and data");
10973 assert(N->getIndex().getValueType().getVectorElementCount().isScalable() ==
10974 N->getValueType(0).getVectorElementCount().isScalable() &&
10975 "Scalable flags of index and data do not match");
10977 N->getIndex().getValueType().getVectorElementCount(),
10978 N->getValueType(0).getVectorElementCount()) &&
10979 "Vector width mismatch between index and data");
10980 assert(isa<ConstantSDNode>(N->getScale()) &&
10981 N->getScale()->getAsAPIntVal().isPowerOf2() &&
10982 "Scale should be a constant power of 2");
10983
10984 CSEMap.InsertNode(N, IP);
10985 InsertNode(N);
10986 SDValue V(N, 0);
10987 NewSDValueDbgMsg(V, "Creating new node: ", this);
10988 return V;
10989}
10990
10993 MachineMemOperand *MMO,
10994 ISD::MemIndexType IndexType,
10995 bool IsTrunc) {
10996 assert(Ops.size() == 6 && "Incompatible number of operands");
10997
11000 ID.AddInteger(MemVT.getRawBits());
11001 ID.AddInteger(getSyntheticNodeSubclassData<MaskedScatterSDNode>(
11002 dl.getIROrder(), VTs, MemVT, MMO, IndexType, IsTrunc));
11003 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11004 ID.AddInteger(MMO->getFlags());
11005 void *IP = nullptr;
11006 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
11007 cast<MaskedScatterSDNode>(E)->refineAlignment(MMO);
11008 return SDValue(E, 0);
11009 }
11010
11011 auto *N = newSDNode<MaskedScatterSDNode>(dl.getIROrder(), dl.getDebugLoc(),
11012 VTs, MemVT, MMO, IndexType, IsTrunc);
11013 createOperands(N, Ops);
11014
11015 assert(N->getMask().getValueType().getVectorElementCount() ==
11016 N->getValue().getValueType().getVectorElementCount() &&
11017 "Vector width mismatch between mask and data");
11018 assert(
11019 N->getIndex().getValueType().getVectorElementCount().isScalable() ==
11020 N->getValue().getValueType().getVectorElementCount().isScalable() &&
11021 "Scalable flags of index and data do not match");
11023 N->getIndex().getValueType().getVectorElementCount(),
11024 N->getValue().getValueType().getVectorElementCount()) &&
11025 "Vector width mismatch between index and data");
11026 assert(isa<ConstantSDNode>(N->getScale()) &&
11027 N->getScale()->getAsAPIntVal().isPowerOf2() &&
11028 "Scale should be a constant power of 2");
11029
11030 CSEMap.InsertNode(N, IP);
11031 InsertNode(N);
11032 SDValue V(N, 0);
11033 NewSDValueDbgMsg(V, "Creating new node: ", this);
11034 return V;
11035}
11036
11038 const SDLoc &dl, ArrayRef<SDValue> Ops,
11039 MachineMemOperand *MMO,
11040 ISD::MemIndexType IndexType) {
11041 assert(Ops.size() == 7 && "Incompatible number of operands");
11042
11045 ID.AddInteger(MemVT.getRawBits());
11046 ID.AddInteger(getSyntheticNodeSubclassData<MaskedHistogramSDNode>(
11047 dl.getIROrder(), VTs, MemVT, MMO, IndexType));
11048 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11049 ID.AddInteger(MMO->getFlags());
11050 void *IP = nullptr;
11051 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
11052 cast<MaskedGatherSDNode>(E)->refineAlignment(MMO);
11053 return SDValue(E, 0);
11054 }
11055
11056 auto *N = newSDNode<MaskedHistogramSDNode>(dl.getIROrder(), dl.getDebugLoc(),
11057 VTs, MemVT, MMO, IndexType);
11058 createOperands(N, Ops);
11059
11060 assert(N->getMask().getValueType().getVectorElementCount() ==
11061 N->getIndex().getValueType().getVectorElementCount() &&
11062 "Vector width mismatch between mask and data");
11063 assert(isa<ConstantSDNode>(N->getScale()) &&
11064 N->getScale()->getAsAPIntVal().isPowerOf2() &&
11065 "Scale should be a constant power of 2");
11066 assert(N->getInc().getValueType().isInteger() && "Non integer update value");
11067
11068 CSEMap.InsertNode(N, IP);
11069 InsertNode(N);
11070 SDValue V(N, 0);
11071 NewSDValueDbgMsg(V, "Creating new node: ", this);
11072 return V;
11073}
11074
11076 SDValue Ptr, SDValue Mask, SDValue EVL,
11077 MachineMemOperand *MMO) {
11078 SDVTList VTs = getVTList(VT, EVL.getValueType(), MVT::Other);
11079 SDValue Ops[] = {Chain, Ptr, Mask, EVL};
11081 AddNodeIDNode(ID, ISD::VP_LOAD_FF, VTs, Ops);
11082 ID.AddInteger(VT.getRawBits());
11083 ID.AddInteger(getSyntheticNodeSubclassData<VPLoadFFSDNode>(DL.getIROrder(),
11084 VTs, VT, MMO));
11085 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11086 ID.AddInteger(MMO->getFlags());
11087 void *IP = nullptr;
11088 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
11089 cast<VPLoadFFSDNode>(E)->refineAlignment(MMO);
11090 return SDValue(E, 0);
11091 }
11092 auto *N = newSDNode<VPLoadFFSDNode>(DL.getIROrder(), DL.getDebugLoc(), VTs,
11093 VT, MMO);
11094 createOperands(N, Ops);
11095
11096 CSEMap.InsertNode(N, IP);
11097 InsertNode(N);
11098 SDValue V(N, 0);
11099 NewSDValueDbgMsg(V, "Creating new node: ", this);
11100 return V;
11101}
11102
11104 EVT MemVT, MachineMemOperand *MMO) {
11105 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
11106 SDVTList VTs = getVTList(MVT::Other);
11107 SDValue Ops[] = {Chain, Ptr};
11110 ID.AddInteger(MemVT.getRawBits());
11111 ID.AddInteger(getSyntheticNodeSubclassData<FPStateAccessSDNode>(
11112 ISD::GET_FPENV_MEM, dl.getIROrder(), VTs, MemVT, MMO));
11113 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11114 ID.AddInteger(MMO->getFlags());
11115 void *IP = nullptr;
11116 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
11117 return SDValue(E, 0);
11118
11119 auto *N = newSDNode<FPStateAccessSDNode>(ISD::GET_FPENV_MEM, dl.getIROrder(),
11120 dl.getDebugLoc(), VTs, MemVT, MMO);
11121 createOperands(N, Ops);
11122
11123 CSEMap.InsertNode(N, IP);
11124 InsertNode(N);
11125 SDValue V(N, 0);
11126 NewSDValueDbgMsg(V, "Creating new node: ", this);
11127 return V;
11128}
11129
11131 EVT MemVT, MachineMemOperand *MMO) {
11132 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
11133 SDVTList VTs = getVTList(MVT::Other);
11134 SDValue Ops[] = {Chain, Ptr};
11137 ID.AddInteger(MemVT.getRawBits());
11138 ID.AddInteger(getSyntheticNodeSubclassData<FPStateAccessSDNode>(
11139 ISD::SET_FPENV_MEM, dl.getIROrder(), VTs, MemVT, MMO));
11140 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11141 ID.AddInteger(MMO->getFlags());
11142 void *IP = nullptr;
11143 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
11144 return SDValue(E, 0);
11145
11146 auto *N = newSDNode<FPStateAccessSDNode>(ISD::SET_FPENV_MEM, dl.getIROrder(),
11147 dl.getDebugLoc(), VTs, MemVT, MMO);
11148 createOperands(N, Ops);
11149
11150 CSEMap.InsertNode(N, IP);
11151 InsertNode(N);
11152 SDValue V(N, 0);
11153 NewSDValueDbgMsg(V, "Creating new node: ", this);
11154 return V;
11155}
11156
11158 // select undef, T, F --> T (if T is a constant), otherwise F
11159 // select, ?, undef, F --> F
11160 // select, ?, T, undef --> T
11161 if (Cond.isUndef())
11162 return isConstantValueOfAnyType(T) ? T : F;
11163 if (T.isUndef())
11165 if (F.isUndef())
11167
11168 // select true, T, F --> T
11169 // select false, T, F --> F
11170 if (auto C = isBoolConstant(Cond))
11171 return *C ? T : F;
11172
11173 // select ?, T, T --> T
11174 if (T == F)
11175 return T;
11176
11177 return SDValue();
11178}
11179
11181 // shift undef, Y --> 0 (can always assume that the undef value is 0)
11182 if (X.isUndef())
11183 return getConstant(0, SDLoc(X.getNode()), X.getValueType());
11184 // shift X, undef --> undef (because it may shift by the bitwidth)
11185 if (Y.isUndef())
11186 return getUNDEF(X.getValueType());
11187
11188 // shift 0, Y --> 0
11189 // shift X, 0 --> X
11191 return X;
11192
11193 // shift X, C >= bitwidth(X) --> undef
11194 // All vector elements must be too big (or undef) to avoid partial undefs.
11195 auto isShiftTooBig = [X](ConstantSDNode *Val) {
11196 return !Val || Val->getAPIntValue().uge(X.getScalarValueSizeInBits());
11197 };
11198 if (ISD::matchUnaryPredicate(Y, isShiftTooBig, true))
11199 return getUNDEF(X.getValueType());
11200
11201 // shift i1/vXi1 X, Y --> X (any non-zero shift amount is undefined).
11202 if (X.getValueType().getScalarType() == MVT::i1)
11203 return X;
11204
11205 return SDValue();
11206}
11207
11209 SDNodeFlags Flags) {
11210 // If this operation has 'nnan' or 'ninf' and at least 1 disallowed operand
11211 // (an undef operand can be chosen to be Nan/Inf), then the result of this
11212 // operation is poison. That result can be relaxed to undef.
11213 ConstantFPSDNode *XC = isConstOrConstSplatFP(X, /* AllowUndefs */ true);
11214 ConstantFPSDNode *YC = isConstOrConstSplatFP(Y, /* AllowUndefs */ true);
11215 bool HasNan = (XC && XC->getValueAPF().isNaN()) ||
11216 (YC && YC->getValueAPF().isNaN());
11217 bool HasInf = (XC && XC->getValueAPF().isInfinity()) ||
11218 (YC && YC->getValueAPF().isInfinity());
11219
11220 if (Flags.hasNoNaNs() && (HasNan || X.isUndef() || Y.isUndef()))
11221 return getUNDEF(X.getValueType());
11222
11223 if (Flags.hasNoInfs() && (HasInf || X.isUndef() || Y.isUndef()))
11224 return getUNDEF(X.getValueType());
11225
11226 if (!YC)
11227 return SDValue();
11228
11229 // X + -0.0 --> X
11230 if (Opcode == ISD::FADD)
11231 if (YC->getValueAPF().isNegZero())
11232 return X;
11233
11234 // X - +0.0 --> X
11235 if (Opcode == ISD::FSUB)
11236 if (YC->getValueAPF().isPosZero())
11237 return X;
11238
11239 // X * 1.0 --> X
11240 // X / 1.0 --> X
11241 if (Opcode == ISD::FMUL || Opcode == ISD::FDIV)
11242 if (YC->getValueAPF().isExactlyValue(1.0))
11243 return X;
11244
11245 // X * 0.0 --> 0.0
11246 if (Opcode == ISD::FMUL && Flags.hasNoNaNs() && Flags.hasNoSignedZeros())
11247 if (YC->getValueAPF().isZero())
11248 return getConstantFP(0.0, SDLoc(Y), Y.getValueType());
11249
11250 return SDValue();
11251}
11252
11254 SDValue Ptr, SDValue SV, unsigned Align) {
11255 SDValue Ops[] = { Chain, Ptr, SV, getTargetConstant(Align, dl, MVT::i32) };
11256 return getNode(ISD::VAARG, dl, getVTList(VT, MVT::Other), Ops);
11257}
11258
11259SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
11261 switch (Ops.size()) {
11262 case 0: return getNode(Opcode, DL, VT);
11263 case 1: return getNode(Opcode, DL, VT, Ops[0].get());
11264 case 2: return getNode(Opcode, DL, VT, Ops[0], Ops[1]);
11265 case 3: return getNode(Opcode, DL, VT, Ops[0], Ops[1], Ops[2]);
11266 default: break;
11267 }
11268
11269 // Copy from an SDUse array into an SDValue array for use with
11270 // the regular getNode logic.
11272 return getNode(Opcode, DL, VT, NewOps);
11273}
11274
11275SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
11277 SDNodeFlags Flags;
11278 if (Inserter)
11279 Flags = Inserter->getFlags();
11280 return getNode(Opcode, DL, VT, Ops, Flags);
11281}
11282
11283SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
11284 ArrayRef<SDValue> Ops, const SDNodeFlags Flags) {
11285 unsigned NumOps = Ops.size();
11286 switch (NumOps) {
11287 case 0: return getNode(Opcode, DL, VT);
11288 case 1: return getNode(Opcode, DL, VT, Ops[0], Flags);
11289 case 2: return getNode(Opcode, DL, VT, Ops[0], Ops[1], Flags);
11290 case 3: return getNode(Opcode, DL, VT, Ops[0], Ops[1], Ops[2], Flags);
11291 default: break;
11292 }
11293
11294#ifndef NDEBUG
11295 for (const auto &Op : Ops)
11296 assert(Op.getOpcode() != ISD::DELETED_NODE &&
11297 "Operand is DELETED_NODE!");
11298#endif
11299
11300 switch (Opcode) {
11301 default: break;
11302 case ISD::BUILD_VECTOR:
11303 // Attempt to simplify BUILD_VECTOR.
11304 if (SDValue V = FoldBUILD_VECTOR(DL, VT, Ops, *this))
11305 return V;
11306 break;
11308 if (SDValue V = foldCONCAT_VECTORS(DL, VT, Ops, *this))
11309 return V;
11310 break;
11311 case ISD::SELECT_CC:
11312 assert(NumOps == 5 && "SELECT_CC takes 5 operands!");
11313 assert(Ops[0].getValueType() == Ops[1].getValueType() &&
11314 "LHS and RHS of condition must have same type!");
11315 assert(Ops[2].getValueType() == Ops[3].getValueType() &&
11316 "True and False arms of SelectCC must have same type!");
11317 assert(Ops[2].getValueType() == VT &&
11318 "select_cc node must be of same type as true and false value!");
11319 assert((!Ops[0].getValueType().isVector() ||
11320 Ops[0].getValueType().getVectorElementCount() ==
11321 VT.getVectorElementCount()) &&
11322 "Expected select_cc with vector result to have the same sized "
11323 "comparison type!");
11324 break;
11325 case ISD::BR_CC:
11326 assert(NumOps == 5 && "BR_CC takes 5 operands!");
11327 assert(Ops[2].getValueType() == Ops[3].getValueType() &&
11328 "LHS/RHS of comparison should match types!");
11329 break;
11330 case ISD::VP_ADD:
11331 case ISD::VP_SUB:
11332 // If it is VP_ADD/VP_SUB mask operation then turn it to VP_XOR
11333 if (VT.getScalarType() == MVT::i1)
11334 Opcode = ISD::VP_XOR;
11335 break;
11336 case ISD::VP_MUL:
11337 // If it is VP_MUL mask operation then turn it to VP_AND
11338 if (VT.getScalarType() == MVT::i1)
11339 Opcode = ISD::VP_AND;
11340 break;
11341 case ISD::VP_REDUCE_MUL:
11342 // If it is VP_REDUCE_MUL mask operation then turn it to VP_REDUCE_AND
11343 if (VT == MVT::i1)
11344 Opcode = ISD::VP_REDUCE_AND;
11345 break;
11346 case ISD::VP_REDUCE_ADD:
11347 // If it is VP_REDUCE_ADD mask operation then turn it to VP_REDUCE_XOR
11348 if (VT == MVT::i1)
11349 Opcode = ISD::VP_REDUCE_XOR;
11350 break;
11351 case ISD::VP_REDUCE_SMAX:
11352 case ISD::VP_REDUCE_UMIN:
11353 // If it is VP_REDUCE_SMAX/VP_REDUCE_UMIN mask operation then turn it to
11354 // VP_REDUCE_AND.
11355 if (VT == MVT::i1)
11356 Opcode = ISD::VP_REDUCE_AND;
11357 break;
11358 case ISD::VP_REDUCE_SMIN:
11359 case ISD::VP_REDUCE_UMAX:
11360 // If it is VP_REDUCE_SMIN/VP_REDUCE_UMAX mask operation then turn it to
11361 // VP_REDUCE_OR.
11362 if (VT == MVT::i1)
11363 Opcode = ISD::VP_REDUCE_OR;
11364 break;
11365 }
11366
11367 // Memoize nodes.
11368 SDNode *N;
11369 SDVTList VTs = getVTList(VT);
11370
11371 if (VT != MVT::Glue) {
11373 AddNodeIDNode(ID, Opcode, VTs, Ops);
11374 void *IP = nullptr;
11375
11376 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
11377 E->intersectFlagsWith(Flags);
11378 return SDValue(E, 0);
11379 }
11380
11381 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
11382 createOperands(N, Ops);
11383
11384 CSEMap.InsertNode(N, IP);
11385 } else {
11386 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
11387 createOperands(N, Ops);
11388 }
11389
11390 N->setFlags(Flags);
11391 InsertNode(N);
11392 SDValue V(N, 0);
11393 NewSDValueDbgMsg(V, "Creating new node: ", this);
11394 return V;
11395}
11396
11397SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL,
11398 ArrayRef<EVT> ResultTys, ArrayRef<SDValue> Ops) {
11399 SDNodeFlags Flags;
11400 if (Inserter)
11401 Flags = Inserter->getFlags();
11402 return getNode(Opcode, DL, getVTList(ResultTys), Ops, Flags);
11403}
11404
11405SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL,
11407 const SDNodeFlags Flags) {
11408 return getNode(Opcode, DL, getVTList(ResultTys), Ops, Flags);
11409}
11410
11411SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11413 SDNodeFlags Flags;
11414 if (Inserter)
11415 Flags = Inserter->getFlags();
11416 return getNode(Opcode, DL, VTList, Ops, Flags);
11417}
11418
11419SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11420 ArrayRef<SDValue> Ops, const SDNodeFlags Flags) {
11421 if (VTList.NumVTs == 1)
11422 return getNode(Opcode, DL, VTList.VTs[0], Ops, Flags);
11423
11424#ifndef NDEBUG
11425 for (const auto &Op : Ops)
11426 assert(Op.getOpcode() != ISD::DELETED_NODE &&
11427 "Operand is DELETED_NODE!");
11428#endif
11429
11430 switch (Opcode) {
11431 case ISD::SADDO:
11432 case ISD::UADDO:
11433 case ISD::SSUBO:
11434 case ISD::USUBO: {
11435 assert(VTList.NumVTs == 2 && Ops.size() == 2 &&
11436 "Invalid add/sub overflow op!");
11437 assert(VTList.VTs[0].isInteger() && VTList.VTs[1].isInteger() &&
11438 Ops[0].getValueType() == Ops[1].getValueType() &&
11439 Ops[0].getValueType() == VTList.VTs[0] &&
11440 "Binary operator types must match!");
11441 SDValue N1 = Ops[0], N2 = Ops[1];
11442 canonicalizeCommutativeBinop(Opcode, N1, N2);
11443
11444 // (X +- 0) -> X with zero-overflow.
11445 ConstantSDNode *N2CV = isConstOrConstSplat(N2, /*AllowUndefs*/ false,
11446 /*AllowTruncation*/ true);
11447 if (N2CV && N2CV->isZero()) {
11448 SDValue ZeroOverFlow = getConstant(0, DL, VTList.VTs[1]);
11449 return getNode(ISD::MERGE_VALUES, DL, VTList, {N1, ZeroOverFlow}, Flags);
11450 }
11451
11452 if (VTList.VTs[0].getScalarType() == MVT::i1 &&
11453 VTList.VTs[1].getScalarType() == MVT::i1) {
11454 SDValue F1 = getFreeze(N1);
11455 SDValue F2 = getFreeze(N2);
11456 // {vXi1,vXi1} (u/s)addo(vXi1 x, vXi1y) -> {xor(x,y),and(x,y)}
11457 if (Opcode == ISD::UADDO || Opcode == ISD::SADDO)
11458 return getNode(ISD::MERGE_VALUES, DL, VTList,
11459 {getNode(ISD::XOR, DL, VTList.VTs[0], F1, F2),
11460 getNode(ISD::AND, DL, VTList.VTs[1], F1, F2)},
11461 Flags);
11462 // {vXi1,vXi1} (u/s)subo(vXi1 x, vXi1y) -> {xor(x,y),and(~x,y)}
11463 if (Opcode == ISD::USUBO || Opcode == ISD::SSUBO) {
11464 SDValue NotF1 = getNOT(DL, F1, VTList.VTs[0]);
11465 return getNode(ISD::MERGE_VALUES, DL, VTList,
11466 {getNode(ISD::XOR, DL, VTList.VTs[0], F1, F2),
11467 getNode(ISD::AND, DL, VTList.VTs[1], NotF1, F2)},
11468 Flags);
11469 }
11470 }
11471 break;
11472 }
11473 case ISD::SADDO_CARRY:
11474 case ISD::UADDO_CARRY:
11475 case ISD::SSUBO_CARRY:
11476 case ISD::USUBO_CARRY:
11477 assert(VTList.NumVTs == 2 && Ops.size() == 3 &&
11478 "Invalid add/sub overflow op!");
11479 assert(VTList.VTs[0].isInteger() && VTList.VTs[1].isInteger() &&
11480 Ops[0].getValueType() == Ops[1].getValueType() &&
11481 Ops[0].getValueType() == VTList.VTs[0] &&
11482 Ops[2].getValueType() == VTList.VTs[1] &&
11483 "Binary operator types must match!");
11484 break;
11485 case ISD::SMUL_LOHI:
11486 case ISD::UMUL_LOHI: {
11487 assert(VTList.NumVTs == 2 && Ops.size() == 2 && "Invalid mul lo/hi op!");
11488 assert(VTList.VTs[0].isInteger() && VTList.VTs[0] == VTList.VTs[1] &&
11489 VTList.VTs[0] == Ops[0].getValueType() &&
11490 VTList.VTs[0] == Ops[1].getValueType() &&
11491 "Binary operator types must match!");
11492 // Constant fold.
11495 if (LHS && RHS) {
11496 unsigned Width = VTList.VTs[0].getScalarSizeInBits();
11497 unsigned OutWidth = Width * 2;
11498 APInt Val = LHS->getAPIntValue();
11499 APInt Mul = RHS->getAPIntValue();
11500 if (Opcode == ISD::SMUL_LOHI) {
11501 Val = Val.sext(OutWidth);
11502 Mul = Mul.sext(OutWidth);
11503 } else {
11504 Val = Val.zext(OutWidth);
11505 Mul = Mul.zext(OutWidth);
11506 }
11507 Val *= Mul;
11508
11509 SDValue Hi =
11510 getConstant(Val.extractBits(Width, Width), DL, VTList.VTs[0]);
11511 SDValue Lo = getConstant(Val.trunc(Width), DL, VTList.VTs[0]);
11512 return getNode(ISD::MERGE_VALUES, DL, VTList, {Lo, Hi}, Flags);
11513 }
11514 break;
11515 }
11516 case ISD::FFREXP: {
11517 assert(VTList.NumVTs == 2 && Ops.size() == 1 && "Invalid ffrexp op!");
11518 assert(VTList.VTs[0].isFloatingPoint() && VTList.VTs[1].isInteger() &&
11519 VTList.VTs[0] == Ops[0].getValueType() && "frexp type mismatch");
11520
11522 int FrexpExp;
11523 APFloat FrexpMant =
11524 frexp(C->getValueAPF(), FrexpExp, APFloat::rmNearestTiesToEven);
11525 SDValue Result0 = getConstantFP(FrexpMant, DL, VTList.VTs[0]);
11526 SDValue Result1 = getSignedConstant(FrexpMant.isFinite() ? FrexpExp : 0,
11527 DL, VTList.VTs[1]);
11528 return getNode(ISD::MERGE_VALUES, DL, VTList, {Result0, Result1}, Flags);
11529 }
11530
11531 break;
11532 }
11534 assert(VTList.NumVTs == 2 && Ops.size() == 2 &&
11535 "Invalid STRICT_FP_EXTEND!");
11536 assert(VTList.VTs[0].isFloatingPoint() &&
11537 Ops[1].getValueType().isFloatingPoint() && "Invalid FP cast!");
11538 assert(VTList.VTs[0].isVector() == Ops[1].getValueType().isVector() &&
11539 "STRICT_FP_EXTEND result type should be vector iff the operand "
11540 "type is vector!");
11541 assert((!VTList.VTs[0].isVector() ||
11542 VTList.VTs[0].getVectorElementCount() ==
11543 Ops[1].getValueType().getVectorElementCount()) &&
11544 "Vector element count mismatch!");
11545 assert(Ops[1].getValueType().bitsLT(VTList.VTs[0]) &&
11546 "Invalid fpext node, dst <= src!");
11547 break;
11549 assert(VTList.NumVTs == 2 && Ops.size() == 3 && "Invalid STRICT_FP_ROUND!");
11550 assert(VTList.VTs[0].isVector() == Ops[1].getValueType().isVector() &&
11551 "STRICT_FP_ROUND result type should be vector iff the operand "
11552 "type is vector!");
11553 assert((!VTList.VTs[0].isVector() ||
11554 VTList.VTs[0].getVectorElementCount() ==
11555 Ops[1].getValueType().getVectorElementCount()) &&
11556 "Vector element count mismatch!");
11557 assert(VTList.VTs[0].isFloatingPoint() &&
11558 Ops[1].getValueType().isFloatingPoint() &&
11559 VTList.VTs[0].bitsLT(Ops[1].getValueType()) &&
11560 Ops[2].getOpcode() == ISD::TargetConstant &&
11561 (Ops[2]->getAsZExtVal() == 0 || Ops[2]->getAsZExtVal() == 1) &&
11562 "Invalid STRICT_FP_ROUND!");
11563 break;
11564 }
11565
11566 // Memoize the node unless it returns a glue result.
11567 SDNode *N;
11568 if (VTList.VTs[VTList.NumVTs-1] != MVT::Glue) {
11570 AddNodeIDNode(ID, Opcode, VTList, Ops);
11571 void *IP = nullptr;
11572 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
11573 E->intersectFlagsWith(Flags);
11574 return SDValue(E, 0);
11575 }
11576
11577 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTList);
11578 createOperands(N, Ops);
11579 CSEMap.InsertNode(N, IP);
11580 } else {
11581 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTList);
11582 createOperands(N, Ops);
11583 }
11584
11585 N->setFlags(Flags);
11586 InsertNode(N);
11587 SDValue V(N, 0);
11588 NewSDValueDbgMsg(V, "Creating new node: ", this);
11589 return V;
11590}
11591
11592SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL,
11593 SDVTList VTList) {
11594 return getNode(Opcode, DL, VTList, ArrayRef<SDValue>());
11595}
11596
11597SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11598 SDValue N1) {
11599 SDValue Ops[] = { N1 };
11600 return getNode(Opcode, DL, VTList, Ops);
11601}
11602
11603SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11604 SDValue N1, SDValue N2) {
11605 SDValue Ops[] = { N1, N2 };
11606 return getNode(Opcode, DL, VTList, Ops);
11607}
11608
11609SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11610 SDValue N1, SDValue N2, SDValue N3) {
11611 SDValue Ops[] = { N1, N2, N3 };
11612 return getNode(Opcode, DL, VTList, Ops);
11613}
11614
11615SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11616 SDValue N1, SDValue N2, SDValue N3, SDValue N4) {
11617 SDValue Ops[] = { N1, N2, N3, N4 };
11618 return getNode(Opcode, DL, VTList, Ops);
11619}
11620
11621SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11622 SDValue N1, SDValue N2, SDValue N3, SDValue N4,
11623 SDValue N5) {
11624 SDValue Ops[] = { N1, N2, N3, N4, N5 };
11625 return getNode(Opcode, DL, VTList, Ops);
11626}
11627
11629 if (!VT.isExtended())
11630 return makeVTList(SDNode::getValueTypeList(VT.getSimpleVT()), 1);
11631
11632 return makeVTList(&(*EVTs.insert(VT).first), 1);
11633}
11634
11637 ID.AddInteger(2U);
11638 ID.AddInteger(VT1.getRawBits());
11639 ID.AddInteger(VT2.getRawBits());
11640
11641 void *IP = nullptr;
11642 SDVTListNode *Result = VTListMap.FindNodeOrInsertPos(ID, IP);
11643 if (!Result) {
11644 EVT *Array = Allocator.Allocate<EVT>(2);
11645 Array[0] = VT1;
11646 Array[1] = VT2;
11647 Result = new (Allocator) SDVTListNode(ID.Intern(Allocator), Array, 2);
11648 VTListMap.InsertNode(Result, IP);
11649 }
11650 return Result->getSDVTList();
11651}
11652
11655 ID.AddInteger(3U);
11656 ID.AddInteger(VT1.getRawBits());
11657 ID.AddInteger(VT2.getRawBits());
11658 ID.AddInteger(VT3.getRawBits());
11659
11660 void *IP = nullptr;
11661 SDVTListNode *Result = VTListMap.FindNodeOrInsertPos(ID, IP);
11662 if (!Result) {
11663 EVT *Array = Allocator.Allocate<EVT>(3);
11664 Array[0] = VT1;
11665 Array[1] = VT2;
11666 Array[2] = VT3;
11667 Result = new (Allocator) SDVTListNode(ID.Intern(Allocator), Array, 3);
11668 VTListMap.InsertNode(Result, IP);
11669 }
11670 return Result->getSDVTList();
11671}
11672
11675 ID.AddInteger(4U);
11676 ID.AddInteger(VT1.getRawBits());
11677 ID.AddInteger(VT2.getRawBits());
11678 ID.AddInteger(VT3.getRawBits());
11679 ID.AddInteger(VT4.getRawBits());
11680
11681 void *IP = nullptr;
11682 SDVTListNode *Result = VTListMap.FindNodeOrInsertPos(ID, IP);
11683 if (!Result) {
11684 EVT *Array = Allocator.Allocate<EVT>(4);
11685 Array[0] = VT1;
11686 Array[1] = VT2;
11687 Array[2] = VT3;
11688 Array[3] = VT4;
11689 Result = new (Allocator) SDVTListNode(ID.Intern(Allocator), Array, 4);
11690 VTListMap.InsertNode(Result, IP);
11691 }
11692 return Result->getSDVTList();
11693}
11694
11696 unsigned NumVTs = VTs.size();
11698 ID.AddInteger(NumVTs);
11699 for (unsigned index = 0; index < NumVTs; index++) {
11700 ID.AddInteger(VTs[index].getRawBits());
11701 }
11702
11703 void *IP = nullptr;
11704 SDVTListNode *Result = VTListMap.FindNodeOrInsertPos(ID, IP);
11705 if (!Result) {
11706 EVT *Array = Allocator.Allocate<EVT>(NumVTs);
11707 llvm::copy(VTs, Array);
11708 Result = new (Allocator) SDVTListNode(ID.Intern(Allocator), Array, NumVTs);
11709 VTListMap.InsertNode(Result, IP);
11710 }
11711 return Result->getSDVTList();
11712}
11713
11714
11715/// UpdateNodeOperands - *Mutate* the specified node in-place to have the
11716/// specified operands. If the resultant node already exists in the DAG,
11717/// this does not modify the specified node, instead it returns the node that
11718/// already exists. If the resultant node does not exist in the DAG, the
11719/// input node is returned. As a degenerate case, if you specify the same
11720/// input operands as the node already has, the input node is returned.
11722 assert(N->getNumOperands() == 1 && "Update with wrong number of operands");
11723
11724 // Check to see if there is no change.
11725 if (Op == N->getOperand(0)) return N;
11726
11727 // See if the modified node already exists.
11728 void *InsertPos = nullptr;
11729 if (SDNode *Existing = FindModifiedNodeSlot(N, Op, InsertPos))
11730 return Existing;
11731
11732 // Nope it doesn't. Remove the node from its current place in the maps.
11733 if (InsertPos)
11734 if (!RemoveNodeFromCSEMaps(N))
11735 InsertPos = nullptr;
11736
11737 // Now we update the operands.
11738 N->OperandList[0].set(Op);
11739
11741 // If this gets put into a CSE map, add it.
11742 if (InsertPos) CSEMap.InsertNode(N, InsertPos);
11743 return N;
11744}
11745
11747 assert(N->getNumOperands() == 2 && "Update with wrong number of operands");
11748
11749 // Check to see if there is no change.
11750 if (Op1 == N->getOperand(0) && Op2 == N->getOperand(1))
11751 return N; // No operands changed, just return the input node.
11752
11753 // See if the modified node already exists.
11754 void *InsertPos = nullptr;
11755 if (SDNode *Existing = FindModifiedNodeSlot(N, Op1, Op2, InsertPos))
11756 return Existing;
11757
11758 // Nope it doesn't. Remove the node from its current place in the maps.
11759 if (InsertPos)
11760 if (!RemoveNodeFromCSEMaps(N))
11761 InsertPos = nullptr;
11762
11763 // Now we update the operands.
11764 if (N->OperandList[0] != Op1)
11765 N->OperandList[0].set(Op1);
11766 if (N->OperandList[1] != Op2)
11767 N->OperandList[1].set(Op2);
11768
11770 // If this gets put into a CSE map, add it.
11771 if (InsertPos) CSEMap.InsertNode(N, InsertPos);
11772 return N;
11773}
11774
11777 SDValue Ops[] = { Op1, Op2, Op3 };
11778 return UpdateNodeOperands(N, Ops);
11779}
11780
11783 SDValue Op3, SDValue Op4) {
11784 SDValue Ops[] = { Op1, Op2, Op3, Op4 };
11785 return UpdateNodeOperands(N, Ops);
11786}
11787
11790 SDValue Op3, SDValue Op4, SDValue Op5) {
11791 SDValue Ops[] = { Op1, Op2, Op3, Op4, Op5 };
11792 return UpdateNodeOperands(N, Ops);
11793}
11794
11797 unsigned NumOps = Ops.size();
11798 assert(N->getNumOperands() == NumOps &&
11799 "Update with wrong number of operands");
11800
11801 // If no operands changed just return the input node.
11802 if (std::equal(Ops.begin(), Ops.end(), N->op_begin()))
11803 return N;
11804
11805 // See if the modified node already exists.
11806 void *InsertPos = nullptr;
11807 if (SDNode *Existing = FindModifiedNodeSlot(N, Ops, InsertPos))
11808 return Existing;
11809
11810 // Nope it doesn't. Remove the node from its current place in the maps.
11811 if (InsertPos)
11812 if (!RemoveNodeFromCSEMaps(N))
11813 InsertPos = nullptr;
11814
11815 // Now we update the operands.
11816 for (unsigned i = 0; i != NumOps; ++i)
11817 if (N->OperandList[i] != Ops[i])
11818 N->OperandList[i].set(Ops[i]);
11819
11821 // If this gets put into a CSE map, add it.
11822 if (InsertPos) CSEMap.InsertNode(N, InsertPos);
11823 return N;
11824}
11825
11826/// DropOperands - Release the operands and set this node to have
11827/// zero operands.
11829 // Unlike the code in MorphNodeTo that does this, we don't need to
11830 // watch for dead nodes here.
11831 for (op_iterator I = op_begin(), E = op_end(); I != E; ) {
11832 SDUse &Use = *I++;
11833 Use.set(SDValue());
11834 }
11835}
11836
11838 ArrayRef<MachineMemOperand *> NewMemRefs) {
11839 if (NewMemRefs.empty()) {
11840 N->clearMemRefs();
11841 return;
11842 }
11843
11844 // Check if we can avoid allocating by storing a single reference directly.
11845 if (NewMemRefs.size() == 1) {
11846 N->MemRefs = NewMemRefs[0];
11847 N->NumMemRefs = 1;
11848 return;
11849 }
11850
11851 MachineMemOperand **MemRefsBuffer =
11852 Allocator.template Allocate<MachineMemOperand *>(NewMemRefs.size());
11853 llvm::copy(NewMemRefs, MemRefsBuffer);
11854 N->MemRefs = MemRefsBuffer;
11855 N->NumMemRefs = static_cast<int>(NewMemRefs.size());
11856}
11857
11858/// SelectNodeTo - These are wrappers around MorphNodeTo that accept a
11859/// machine opcode.
11860///
11862 EVT VT) {
11863 SDVTList VTs = getVTList(VT);
11864 return SelectNodeTo(N, MachineOpc, VTs, {});
11865}
11866
11868 EVT VT, SDValue Op1) {
11869 SDVTList VTs = getVTList(VT);
11870 SDValue Ops[] = { Op1 };
11871 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11872}
11873
11875 EVT VT, SDValue Op1,
11876 SDValue Op2) {
11877 SDVTList VTs = getVTList(VT);
11878 SDValue Ops[] = { Op1, Op2 };
11879 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11880}
11881
11883 EVT VT, SDValue Op1,
11884 SDValue Op2, SDValue Op3) {
11885 SDVTList VTs = getVTList(VT);
11886 SDValue Ops[] = { Op1, Op2, Op3 };
11887 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11888}
11889
11892 SDVTList VTs = getVTList(VT);
11893 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11894}
11895
11897 EVT VT1, EVT VT2, ArrayRef<SDValue> Ops) {
11898 SDVTList VTs = getVTList(VT1, VT2);
11899 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11900}
11901
11903 EVT VT1, EVT VT2) {
11904 SDVTList VTs = getVTList(VT1, VT2);
11905 return SelectNodeTo(N, MachineOpc, VTs, {});
11906}
11907
11909 EVT VT1, EVT VT2, EVT VT3,
11911 SDVTList VTs = getVTList(VT1, VT2, VT3);
11912 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11913}
11914
11916 EVT VT1, EVT VT2,
11917 SDValue Op1, SDValue Op2) {
11918 SDVTList VTs = getVTList(VT1, VT2);
11919 SDValue Ops[] = { Op1, Op2 };
11920 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11921}
11922
11925 SDNode *New = MorphNodeTo(N, ~MachineOpc, VTs, Ops);
11926 // Reset the NodeID to -1.
11927 New->setNodeId(-1);
11928 if (New != N) {
11929 ReplaceAllUsesWith(N, New);
11931 }
11932 return New;
11933}
11934
11935/// UpdateSDLocOnMergeSDNode - If the opt level is -O0 then it throws away
11936/// the line number information on the merged node since it is not possible to
11937/// preserve the information that operation is associated with multiple lines.
11938/// This will make the debugger working better at -O0, were there is a higher
11939/// probability having other instructions associated with that line.
11940///
11941/// For IROrder, we keep the smaller of the two
11942SDNode *SelectionDAG::UpdateSDLocOnMergeSDNode(SDNode *N, const SDLoc &OLoc) {
11943 DebugLoc NLoc = N->getDebugLoc();
11944 if (NLoc && OptLevel == CodeGenOptLevel::None && OLoc.getDebugLoc() != NLoc) {
11945 N->setDebugLoc(DebugLoc());
11946 }
11947 unsigned Order = std::min(N->getIROrder(), OLoc.getIROrder());
11948 N->setIROrder(Order);
11949 return N;
11950}
11951
11952/// MorphNodeTo - This *mutates* the specified node to have the specified
11953/// return type, opcode, and operands.
11954///
11955/// Note that MorphNodeTo returns the resultant node. If there is already a
11956/// node of the specified opcode and operands, it returns that node instead of
11957/// the current one. Note that the SDLoc need not be the same.
11958///
11959/// Using MorphNodeTo is faster than creating a new node and swapping it in
11960/// with ReplaceAllUsesWith both because it often avoids allocating a new
11961/// node, and because it doesn't require CSE recalculation for any of
11962/// the node's users.
11963///
11964/// However, note that MorphNodeTo recursively deletes dead nodes from the DAG.
11965/// As a consequence it isn't appropriate to use from within the DAG combiner or
11966/// the legalizer which maintain worklists that would need to be updated when
11967/// deleting things.
11970 // If an identical node already exists, use it.
11971 void *IP = nullptr;
11972 if (VTs.VTs[VTs.NumVTs-1] != MVT::Glue) {
11974 AddNodeIDNode(ID, Opc, VTs, Ops);
11975 if (SDNode *ON = FindNodeOrInsertPos(ID, SDLoc(N), IP))
11976 return UpdateSDLocOnMergeSDNode(ON, SDLoc(N));
11977 }
11978
11979 if (!RemoveNodeFromCSEMaps(N))
11980 IP = nullptr;
11981
11982 // Start the morphing.
11983 N->NodeType = Opc;
11984 N->ValueList = VTs.VTs;
11985 N->NumValues = VTs.NumVTs;
11986
11987 // Clear the operands list, updating used nodes to remove this from their
11988 // use list. Keep track of any operands that become dead as a result.
11989 SmallPtrSet<SDNode*, 16> DeadNodeSet;
11990 for (SDNode::op_iterator I = N->op_begin(), E = N->op_end(); I != E; ) {
11991 SDUse &Use = *I++;
11992 SDNode *Used = Use.getNode();
11993 Use.set(SDValue());
11994 if (Used->use_empty())
11995 DeadNodeSet.insert(Used);
11996 }
11997
11998 // For MachineNode, initialize the memory references information.
12000 MN->clearMemRefs();
12001
12002 // Swap for an appropriately sized array from the recycler.
12003 removeOperands(N);
12004 createOperands(N, Ops);
12005
12006 // Delete any nodes that are still dead after adding the uses for the
12007 // new operands.
12008 if (!DeadNodeSet.empty()) {
12009 SmallVector<SDNode *, 16> DeadNodes;
12010 for (SDNode *N : DeadNodeSet)
12011 if (N->use_empty())
12012 DeadNodes.push_back(N);
12013 RemoveDeadNodes(DeadNodes);
12014 }
12015
12016 if (IP)
12017 CSEMap.InsertNode(N, IP); // Memoize the new node.
12018 return N;
12019}
12020
12022 unsigned OrigOpc = Node->getOpcode();
12023 unsigned NewOpc;
12024 switch (OrigOpc) {
12025 default:
12026 llvm_unreachable("mutateStrictFPToFP called with unexpected opcode!");
12027#define DAG_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
12028 case ISD::STRICT_##DAGN: NewOpc = ISD::DAGN; break;
12029#define CMP_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
12030 case ISD::STRICT_##DAGN: NewOpc = ISD::SETCC; break;
12031#include "llvm/IR/ConstrainedOps.def"
12032 }
12033
12034 assert(Node->getNumValues() == 2 && "Unexpected number of results!");
12035
12036 // We're taking this node out of the chain, so we need to re-link things.
12037 SDValue InputChain = Node->getOperand(0);
12038 SDValue OutputChain = SDValue(Node, 1);
12039 ReplaceAllUsesOfValueWith(OutputChain, InputChain);
12040
12042 for (unsigned i = 1, e = Node->getNumOperands(); i != e; ++i)
12043 Ops.push_back(Node->getOperand(i));
12044
12045 SDVTList VTs = getVTList(Node->getValueType(0));
12046 SDNode *Res = MorphNodeTo(Node, NewOpc, VTs, Ops);
12047
12048 // MorphNodeTo can operate in two ways: if an existing node with the
12049 // specified operands exists, it can just return it. Otherwise, it
12050 // updates the node in place to have the requested operands.
12051 if (Res == Node) {
12052 // If we updated the node in place, reset the node ID. To the isel,
12053 // this should be just like a newly allocated machine node.
12054 Res->setNodeId(-1);
12055 } else {
12058 }
12059
12060 return Res;
12061}
12062
12063/// getMachineNode - These are used for target selectors to create a new node
12064/// with specified return type(s), MachineInstr opcode, and operands.
12065///
12066/// Note that getMachineNode returns the resultant node. If there is already a
12067/// node of the specified opcode and operands, it returns that node instead of
12068/// the current one.
12070 EVT VT) {
12071 SDVTList VTs = getVTList(VT);
12072 return getMachineNode(Opcode, dl, VTs, {});
12073}
12074
12076 EVT VT, SDValue Op1) {
12077 SDVTList VTs = getVTList(VT);
12078 SDValue Ops[] = { Op1 };
12079 return getMachineNode(Opcode, dl, VTs, Ops);
12080}
12081
12083 EVT VT, SDValue Op1, SDValue Op2) {
12084 SDVTList VTs = getVTList(VT);
12085 SDValue Ops[] = { Op1, Op2 };
12086 return getMachineNode(Opcode, dl, VTs, Ops);
12087}
12088
12090 EVT VT, SDValue Op1, SDValue Op2,
12091 SDValue Op3) {
12092 SDVTList VTs = getVTList(VT);
12093 SDValue Ops[] = { Op1, Op2, Op3 };
12094 return getMachineNode(Opcode, dl, VTs, Ops);
12095}
12096
12099 SDVTList VTs = getVTList(VT);
12100 return getMachineNode(Opcode, dl, VTs, Ops);
12101}
12102
12104 EVT VT1, EVT VT2, SDValue Op1,
12105 SDValue Op2) {
12106 SDVTList VTs = getVTList(VT1, VT2);
12107 SDValue Ops[] = { Op1, Op2 };
12108 return getMachineNode(Opcode, dl, VTs, Ops);
12109}
12110
12112 EVT VT1, EVT VT2, SDValue Op1,
12113 SDValue Op2, SDValue Op3) {
12114 SDVTList VTs = getVTList(VT1, VT2);
12115 SDValue Ops[] = { Op1, Op2, Op3 };
12116 return getMachineNode(Opcode, dl, VTs, Ops);
12117}
12118
12120 EVT VT1, EVT VT2,
12122 SDVTList VTs = getVTList(VT1, VT2);
12123 return getMachineNode(Opcode, dl, VTs, Ops);
12124}
12125
12127 EVT VT1, EVT VT2, EVT VT3,
12128 SDValue Op1, SDValue Op2) {
12129 SDVTList VTs = getVTList(VT1, VT2, VT3);
12130 SDValue Ops[] = { Op1, Op2 };
12131 return getMachineNode(Opcode, dl, VTs, Ops);
12132}
12133
12135 EVT VT1, EVT VT2, EVT VT3,
12136 SDValue Op1, SDValue Op2,
12137 SDValue Op3) {
12138 SDVTList VTs = getVTList(VT1, VT2, VT3);
12139 SDValue Ops[] = { Op1, Op2, Op3 };
12140 return getMachineNode(Opcode, dl, VTs, Ops);
12141}
12142
12144 EVT VT1, EVT VT2, EVT VT3,
12146 SDVTList VTs = getVTList(VT1, VT2, VT3);
12147 return getMachineNode(Opcode, dl, VTs, Ops);
12148}
12149
12151 ArrayRef<EVT> ResultTys,
12153 SDVTList VTs = getVTList(ResultTys);
12154 return getMachineNode(Opcode, dl, VTs, Ops);
12155}
12156
12158 SDVTList VTs,
12160 bool DoCSE = VTs.VTs[VTs.NumVTs-1] != MVT::Glue;
12162 void *IP = nullptr;
12163
12164 if (DoCSE) {
12166 AddNodeIDNode(ID, ~Opcode, VTs, Ops);
12167 IP = nullptr;
12168 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
12169 return cast<MachineSDNode>(UpdateSDLocOnMergeSDNode(E, DL));
12170 }
12171 }
12172
12173 // Allocate a new MachineSDNode.
12174 N = newSDNode<MachineSDNode>(~Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
12175 createOperands(N, Ops);
12176
12177 if (DoCSE)
12178 CSEMap.InsertNode(N, IP);
12179
12180 InsertNode(N);
12181 NewSDValueDbgMsg(SDValue(N, 0), "Creating new machine node: ", this);
12182 return N;
12183}
12184
12185/// getTargetExtractSubreg - A convenience function for creating
12186/// TargetOpcode::EXTRACT_SUBREG nodes.
12188 SDValue Operand) {
12189 SDValue SRIdxVal = getTargetConstant(SRIdx, DL, MVT::i32);
12190 SDNode *Subreg = getMachineNode(TargetOpcode::EXTRACT_SUBREG, DL,
12191 VT, Operand, SRIdxVal);
12192 return SDValue(Subreg, 0);
12193}
12194
12195/// getTargetInsertSubreg - A convenience function for creating
12196/// TargetOpcode::INSERT_SUBREG nodes.
12198 SDValue Operand, SDValue Subreg) {
12199 SDValue SRIdxVal = getTargetConstant(SRIdx, DL, MVT::i32);
12200 SDNode *Result = getMachineNode(TargetOpcode::INSERT_SUBREG, DL,
12201 VT, Operand, Subreg, SRIdxVal);
12202 return SDValue(Result, 0);
12203}
12204
12205/// getNodeIfExists - Get the specified node if it's already available, or
12206/// else return NULL.
12209 bool AllowCommute) {
12210 SDNodeFlags Flags;
12211 if (Inserter)
12212 Flags = Inserter->getFlags();
12213 return getNodeIfExists(Opcode, VTList, Ops, Flags, AllowCommute);
12214}
12215
12218 const SDNodeFlags Flags,
12219 bool AllowCommute) {
12220 if (VTList.VTs[VTList.NumVTs - 1] == MVT::Glue)
12221 return nullptr;
12222
12223 auto Lookup = [&](ArrayRef<SDValue> LookupOps) -> SDNode * {
12225 AddNodeIDNode(ID, Opcode, VTList, LookupOps);
12226 void *IP = nullptr;
12227 if (SDNode *E = FindNodeOrInsertPos(ID, IP)) {
12228 E->intersectFlagsWith(Flags);
12229 return E;
12230 }
12231 return nullptr;
12232 };
12233
12234 if (SDNode *Existing = Lookup(Ops))
12235 return Existing;
12236
12237 if (AllowCommute && TLI->isCommutativeBinOp(Opcode))
12238 return Lookup({Ops[1], Ops[0]});
12239
12240 return nullptr;
12241}
12242
12243/// doesNodeExist - Check if a node exists without modifying its flags.
12244bool SelectionDAG::doesNodeExist(unsigned Opcode, SDVTList VTList,
12246 if (VTList.VTs[VTList.NumVTs - 1] != MVT::Glue) {
12248 AddNodeIDNode(ID, Opcode, VTList, Ops);
12249 void *IP = nullptr;
12250 if (FindNodeOrInsertPos(ID, SDLoc(), IP))
12251 return true;
12252 }
12253 return false;
12254}
12255
12256/// getDbgValue - Creates a SDDbgValue node.
12257///
12258/// SDNode
12260 SDNode *N, unsigned R, bool IsIndirect,
12261 const DebugLoc &DL, unsigned O) {
12262 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12263 "Expected inlined-at fields to agree");
12264 return new (DbgInfo->getAlloc())
12265 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, SDDbgOperand::fromNode(N, R),
12266 {}, IsIndirect, DL, O,
12267 /*IsVariadic=*/false);
12268}
12269
12270/// Constant
12272 DIExpression *Expr,
12273 const Value *C,
12274 const DebugLoc &DL, unsigned O) {
12275 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12276 "Expected inlined-at fields to agree");
12277 return new (DbgInfo->getAlloc())
12278 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, SDDbgOperand::fromConst(C), {},
12279 /*IsIndirect=*/false, DL, O,
12280 /*IsVariadic=*/false);
12281}
12282
12283/// FrameIndex
12285 DIExpression *Expr, unsigned FI,
12286 bool IsIndirect,
12287 const DebugLoc &DL,
12288 unsigned O) {
12289 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12290 "Expected inlined-at fields to agree");
12291 return getFrameIndexDbgValue(Var, Expr, FI, {}, IsIndirect, DL, O);
12292}
12293
12294/// FrameIndex with dependencies
12296 DIExpression *Expr, unsigned FI,
12297 ArrayRef<SDNode *> Dependencies,
12298 bool IsIndirect,
12299 const DebugLoc &DL,
12300 unsigned O) {
12301 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12302 "Expected inlined-at fields to agree");
12303 return new (DbgInfo->getAlloc())
12304 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, SDDbgOperand::fromFrameIdx(FI),
12305 Dependencies, IsIndirect, DL, O,
12306 /*IsVariadic=*/false);
12307}
12308
12309/// VReg
12311 Register VReg, bool IsIndirect,
12312 const DebugLoc &DL, unsigned O) {
12313 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12314 "Expected inlined-at fields to agree");
12315 return new (DbgInfo->getAlloc())
12316 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, SDDbgOperand::fromVReg(VReg),
12317 {}, IsIndirect, DL, O,
12318 /*IsVariadic=*/false);
12319}
12320
12323 ArrayRef<SDNode *> Dependencies,
12324 bool IsIndirect, const DebugLoc &DL,
12325 unsigned O, bool IsVariadic) {
12326 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12327 "Expected inlined-at fields to agree");
12328 return new (DbgInfo->getAlloc())
12329 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, Locs, Dependencies, IsIndirect,
12330 DL, O, IsVariadic);
12331}
12332
12334 unsigned OffsetInBits, unsigned SizeInBits,
12335 bool InvalidateDbg) {
12336 SDNode *FromNode = From.getNode();
12337 SDNode *ToNode = To.getNode();
12338 assert(FromNode && ToNode && "Can't modify dbg values");
12339
12340 // PR35338
12341 // TODO: assert(From != To && "Redundant dbg value transfer");
12342 // TODO: assert(FromNode != ToNode && "Intranode dbg value transfer");
12343 if (From == To || FromNode == ToNode)
12344 return;
12345
12346 if (!FromNode->getHasDebugValue())
12347 return;
12348
12349 SDDbgOperand FromLocOp =
12350 SDDbgOperand::fromNode(From.getNode(), From.getResNo());
12352
12354 for (SDDbgValue *Dbg : GetDbgValues(FromNode)) {
12355 if (Dbg->isInvalidated())
12356 continue;
12357
12358 // TODO: assert(!Dbg->isInvalidated() && "Transfer of invalid dbg value");
12359
12360 // Create a new location ops vector that is equal to the old vector, but
12361 // with each instance of FromLocOp replaced with ToLocOp.
12362 bool Changed = false;
12363 auto NewLocOps = Dbg->copyLocationOps();
12364 std::replace_if(
12365 NewLocOps.begin(), NewLocOps.end(),
12366 [&Changed, FromLocOp](const SDDbgOperand &Op) {
12367 bool Match = Op == FromLocOp;
12368 Changed |= Match;
12369 return Match;
12370 },
12371 ToLocOp);
12372 // Ignore this SDDbgValue if we didn't find a matching location.
12373 if (!Changed)
12374 continue;
12375
12376 DIVariable *Var = Dbg->getVariable();
12377 auto *Expr = Dbg->getExpression();
12378 // If a fragment is requested, update the expression.
12379 if (SizeInBits) {
12380 // When splitting a larger (e.g., sign-extended) value whose
12381 // lower bits are described with an SDDbgValue, do not attempt
12382 // to transfer the SDDbgValue to the upper bits.
12383 if (auto FI = Expr->getFragmentInfo())
12384 if (OffsetInBits + SizeInBits > FI->SizeInBits)
12385 continue;
12386 auto Fragment = DIExpression::createFragmentExpression(Expr, OffsetInBits,
12387 SizeInBits);
12388 if (!Fragment)
12389 continue;
12390 Expr = *Fragment;
12391 }
12392
12393 auto AdditionalDependencies = Dbg->getAdditionalDependencies();
12394 // Clone the SDDbgValue and move it to To.
12395 SDDbgValue *Clone = getDbgValueList(
12396 Var, Expr, NewLocOps, AdditionalDependencies, Dbg->isIndirect(),
12397 Dbg->getDebugLoc(), std::max(ToNode->getIROrder(), Dbg->getOrder()),
12398 Dbg->isVariadic());
12399 ClonedDVs.push_back(Clone);
12400
12401 if (InvalidateDbg) {
12402 // Invalidate value and indicate the SDDbgValue should not be emitted.
12403 Dbg->setIsInvalidated();
12404 Dbg->setIsEmitted();
12405 }
12406 }
12407
12408 for (SDDbgValue *Dbg : ClonedDVs) {
12409 assert(is_contained(Dbg->getSDNodes(), ToNode) &&
12410 "Transferred DbgValues should depend on the new SDNode");
12411 AddDbgValue(Dbg, false);
12412 }
12413}
12414
12416 if (!N.getHasDebugValue())
12417 return;
12418
12419 auto GetLocationOperand = [](SDNode *Node, unsigned ResNo) {
12420 if (auto *FISDN = dyn_cast<FrameIndexSDNode>(Node))
12421 return SDDbgOperand::fromFrameIdx(FISDN->getIndex());
12422 return SDDbgOperand::fromNode(Node, ResNo);
12423 };
12424
12426 for (auto *DV : GetDbgValues(&N)) {
12427 if (DV->isInvalidated())
12428 continue;
12429 switch (N.getOpcode()) {
12430 default:
12431 break;
12432 case ISD::ADD: {
12433 SDValue N0 = N.getOperand(0);
12434 SDValue N1 = N.getOperand(1);
12435 if (!isa<ConstantSDNode>(N0)) {
12436 bool RHSConstant = isa<ConstantSDNode>(N1);
12438 if (RHSConstant)
12439 Offset = N.getConstantOperandVal(1);
12440 // We are not allowed to turn indirect debug values variadic, so
12441 // don't salvage those.
12442 if (!RHSConstant && DV->isIndirect())
12443 continue;
12444
12445 // Rewrite an ADD constant node into a DIExpression. Since we are
12446 // performing arithmetic to compute the variable's *value* in the
12447 // DIExpression, we need to mark the expression with a
12448 // DW_OP_stack_value.
12449 auto *DIExpr = DV->getExpression();
12450 auto NewLocOps = DV->copyLocationOps();
12451 bool Changed = false;
12452 size_t OrigLocOpsSize = NewLocOps.size();
12453 for (size_t i = 0; i < OrigLocOpsSize; ++i) {
12454 // We're not given a ResNo to compare against because the whole
12455 // node is going away. We know that any ISD::ADD only has one
12456 // result, so we can assume any node match is using the result.
12457 if (NewLocOps[i].getKind() != SDDbgOperand::SDNODE ||
12458 NewLocOps[i].getSDNode() != &N)
12459 continue;
12460 NewLocOps[i] = GetLocationOperand(N0.getNode(), N0.getResNo());
12461 if (RHSConstant) {
12464 DIExpr = DIExpression::appendOpsToArg(DIExpr, ExprOps, i, true);
12465 } else {
12466 // Convert to a variadic expression (if not already).
12467 // convertToVariadicExpression() returns a const pointer, so we use
12468 // a temporary const variable here.
12469 const auto *TmpDIExpr =
12473 ExprOps.push_back(NewLocOps.size());
12474 ExprOps.push_back(dwarf::DW_OP_plus);
12475 SDDbgOperand RHS =
12477 NewLocOps.push_back(RHS);
12478 DIExpr = DIExpression::appendOpsToArg(TmpDIExpr, ExprOps, i, true);
12479 }
12480 Changed = true;
12481 }
12482 (void)Changed;
12483 assert(Changed && "Salvage target doesn't use N");
12484
12485 bool IsVariadic =
12486 DV->isVariadic() || OrigLocOpsSize != NewLocOps.size();
12487
12488 auto AdditionalDependencies = DV->getAdditionalDependencies();
12489 SDDbgValue *Clone = getDbgValueList(
12490 DV->getVariable(), DIExpr, NewLocOps, AdditionalDependencies,
12491 DV->isIndirect(), DV->getDebugLoc(), DV->getOrder(), IsVariadic);
12492 ClonedDVs.push_back(Clone);
12493 DV->setIsInvalidated();
12494 DV->setIsEmitted();
12495 LLVM_DEBUG(dbgs() << "SALVAGE: Rewriting";
12496 N0.getNode()->dumprFull(this);
12497 dbgs() << " into " << *DIExpr << '\n');
12498 }
12499 break;
12500 }
12501 case ISD::TRUNCATE: {
12502 SDValue N0 = N.getOperand(0);
12503 TypeSize FromSize = N0.getValueSizeInBits();
12504 TypeSize ToSize = N.getValueSizeInBits(0);
12505
12506 DIExpression *DbgExpression = DV->getExpression();
12507 auto ExtOps = DIExpression::getExtOps(FromSize, ToSize, false);
12508 auto NewLocOps = DV->copyLocationOps();
12509 bool Changed = false;
12510 for (size_t i = 0; i < NewLocOps.size(); ++i) {
12511 if (NewLocOps[i].getKind() != SDDbgOperand::SDNODE ||
12512 NewLocOps[i].getSDNode() != &N)
12513 continue;
12514
12515 NewLocOps[i] = GetLocationOperand(N0.getNode(), N0.getResNo());
12516 DbgExpression = DIExpression::appendOpsToArg(DbgExpression, ExtOps, i);
12517 Changed = true;
12518 }
12519 assert(Changed && "Salvage target doesn't use N");
12520 (void)Changed;
12521
12522 SDDbgValue *Clone =
12523 getDbgValueList(DV->getVariable(), DbgExpression, NewLocOps,
12524 DV->getAdditionalDependencies(), DV->isIndirect(),
12525 DV->getDebugLoc(), DV->getOrder(), DV->isVariadic());
12526
12527 ClonedDVs.push_back(Clone);
12528 DV->setIsInvalidated();
12529 DV->setIsEmitted();
12530 LLVM_DEBUG(dbgs() << "SALVAGE: Rewriting"; N0.getNode()->dumprFull(this);
12531 dbgs() << " into " << *DbgExpression << '\n');
12532 break;
12533 }
12534 }
12535 }
12536
12537 for (SDDbgValue *Dbg : ClonedDVs) {
12538 assert((!Dbg->getSDNodes().empty() ||
12539 llvm::any_of(Dbg->getLocationOps(),
12540 [&](const SDDbgOperand &Op) {
12541 return Op.getKind() == SDDbgOperand::FRAMEIX;
12542 })) &&
12543 "Salvaged DbgValue should depend on a new SDNode");
12544 AddDbgValue(Dbg, false);
12545 }
12546}
12547
12548/// Creates a SDDbgLabel node.
12550 const DebugLoc &DL, unsigned O) {
12551 assert(cast<DILabel>(Label)->isValidLocationForIntrinsic(DL) &&
12552 "Expected inlined-at fields to agree");
12553 return new (DbgInfo->getAlloc()) SDDbgLabel(Label, DL, O);
12554}
12555
12556namespace {
12557
12558/// RAUWUpdateListener - Helper for ReplaceAllUsesWith - When the node
12559/// pointed to by a use iterator is deleted, increment the use iterator
12560/// so that it doesn't dangle.
12561///
12562class RAUWUpdateListener : public SelectionDAG::DAGUpdateListener {
12565
12566 void NodeDeleted(SDNode *N, SDNode *E) override {
12567 // Increment the iterator as needed.
12568 while (UI != UE && N == UI->getUser())
12569 ++UI;
12570 }
12571
12572public:
12573 RAUWUpdateListener(SelectionDAG &d,
12576 : SelectionDAG::DAGUpdateListener(d), UI(ui), UE(ue) {}
12577};
12578
12579} // end anonymous namespace
12580
12581/// ReplaceAllUsesWith - Modify anything using 'From' to use 'To' instead.
12582/// This can cause recursive merging of nodes in the DAG.
12583///
12584/// This version assumes From has a single result value.
12585///
12587 SDNode *From = FromN.getNode();
12588 assert(From->getNumValues() == 1 && FromN.getResNo() == 0 &&
12589 "Cannot replace with this method!");
12590 assert(From != To.getNode() && "Cannot replace uses of with self");
12591
12592 // Preserve Debug Values
12593 transferDbgValues(FromN, To);
12594 // Preserve extra info.
12595 copyExtraInfo(From, To.getNode());
12596
12597 // Iterate over all the existing uses of From. New uses will be added
12598 // to the beginning of the use list, which we avoid visiting.
12599 // This specifically avoids visiting uses of From that arise while the
12600 // replacement is happening, because any such uses would be the result
12601 // of CSE: If an existing node looks like From after one of its operands
12602 // is replaced by To, we don't want to replace of all its users with To
12603 // too. See PR3018 for more info.
12604 SDNode::use_iterator UI = From->use_begin(), UE = From->use_end();
12605 RAUWUpdateListener Listener(*this, UI, UE);
12606 while (UI != UE) {
12607 SDNode *User = UI->getUser();
12608
12609 // This node is about to morph, remove its old self from the CSE maps.
12610 RemoveNodeFromCSEMaps(User);
12611
12612 // A user can appear in a use list multiple times, and when this
12613 // happens the uses are usually next to each other in the list.
12614 // To help reduce the number of CSE recomputations, process all
12615 // the uses of this user that we can find this way.
12616 do {
12617 SDUse &Use = *UI;
12618 ++UI;
12619 Use.set(To);
12620 if (To->isDivergent() != From->isDivergent())
12622 } while (UI != UE && UI->getUser() == User);
12623 // Now that we have modified User, add it back to the CSE maps. If it
12624 // already exists there, recursively merge the results together.
12625 AddModifiedNodeToCSEMaps(User);
12626 }
12627
12628 // If we just RAUW'd the root, take note.
12629 if (FromN == getRoot())
12630 setRoot(To);
12631}
12632
12633/// ReplaceAllUsesWith - Modify anything using 'From' to use 'To' instead.
12634/// This can cause recursive merging of nodes in the DAG.
12635///
12636/// This version assumes that for each value of From, there is a
12637/// corresponding value in To in the same position with the same type.
12638///
12640#ifndef NDEBUG
12641 for (unsigned i = 0, e = From->getNumValues(); i != e; ++i)
12642 assert((!From->hasAnyUseOfValue(i) ||
12643 From->getValueType(i) == To->getValueType(i)) &&
12644 "Cannot use this version of ReplaceAllUsesWith!");
12645#endif
12646
12647 // Handle the trivial case.
12648 if (From == To)
12649 return;
12650
12651 // Preserve Debug Info. Only do this if there's a use.
12652 for (unsigned i = 0, e = From->getNumValues(); i != e; ++i)
12653 if (From->hasAnyUseOfValue(i)) {
12654 assert((i < To->getNumValues()) && "Invalid To location");
12655 transferDbgValues(SDValue(From, i), SDValue(To, i));
12656 }
12657 // Preserve extra info.
12658 copyExtraInfo(From, To);
12659
12660 // Iterate over just the existing users of From. See the comments in
12661 // the ReplaceAllUsesWith above.
12662 SDNode::use_iterator UI = From->use_begin(), UE = From->use_end();
12663 RAUWUpdateListener Listener(*this, UI, UE);
12664 while (UI != UE) {
12665 SDNode *User = UI->getUser();
12666
12667 // This node is about to morph, remove its old self from the CSE maps.
12668 RemoveNodeFromCSEMaps(User);
12669
12670 // A user can appear in a use list multiple times, and when this
12671 // happens the uses are usually next to each other in the list.
12672 // To help reduce the number of CSE recomputations, process all
12673 // the uses of this user that we can find this way.
12674 do {
12675 SDUse &Use = *UI;
12676 ++UI;
12677 Use.setNode(To);
12678 if (To->isDivergent() != From->isDivergent())
12680 } while (UI != UE && UI->getUser() == User);
12681
12682 // Now that we have modified User, add it back to the CSE maps. If it
12683 // already exists there, recursively merge the results together.
12684 AddModifiedNodeToCSEMaps(User);
12685 }
12686
12687 // If we just RAUW'd the root, take note.
12688 if (From == getRoot().getNode())
12689 setRoot(SDValue(To, getRoot().getResNo()));
12690}
12691
12692/// ReplaceAllUsesWith - Modify anything using 'From' to use 'To' instead.
12693/// This can cause recursive merging of nodes in the DAG.
12694///
12695/// This version can replace From with any result values. To must match the
12696/// number and types of values returned by From.
12698 if (From->getNumValues() == 1) // Handle the simple case efficiently.
12699 return ReplaceAllUsesWith(SDValue(From, 0), To[0]);
12700
12701 for (unsigned i = 0, e = From->getNumValues(); i != e; ++i) {
12702 // Preserve Debug Info.
12703 transferDbgValues(SDValue(From, i), To[i]);
12704 // Preserve extra info.
12705 copyExtraInfo(From, To[i].getNode());
12706 }
12707
12708 // Iterate over just the existing users of From. See the comments in
12709 // the ReplaceAllUsesWith above.
12710 SDNode::use_iterator UI = From->use_begin(), UE = From->use_end();
12711 RAUWUpdateListener Listener(*this, UI, UE);
12712 while (UI != UE) {
12713 SDNode *User = UI->getUser();
12714
12715 // This node is about to morph, remove its old self from the CSE maps.
12716 RemoveNodeFromCSEMaps(User);
12717
12718 // A user can appear in a use list multiple times, and when this happens the
12719 // uses are usually next to each other in the list. To help reduce the
12720 // number of CSE and divergence recomputations, process all the uses of this
12721 // user that we can find this way.
12722 bool To_IsDivergent = false;
12723 do {
12724 SDUse &Use = *UI;
12725 const SDValue &ToOp = To[Use.getResNo()];
12726 ++UI;
12727 Use.set(ToOp);
12728 if (ToOp.getValueType() != MVT::Other)
12729 To_IsDivergent |= ToOp->isDivergent();
12730 } while (UI != UE && UI->getUser() == User);
12731
12732 if (To_IsDivergent != From->isDivergent())
12734
12735 // Now that we have modified User, add it back to the CSE maps. If it
12736 // already exists there, recursively merge the results together.
12737 AddModifiedNodeToCSEMaps(User);
12738 }
12739
12740 // If we just RAUW'd the root, take note.
12741 if (From == getRoot().getNode())
12742 setRoot(SDValue(To[getRoot().getResNo()]));
12743}
12744
12745/// ReplaceAllUsesOfValueWith - Replace any uses of From with To, leaving
12746/// uses of other values produced by From.getNode() alone. The Deleted
12747/// vector is handled the same way as for ReplaceAllUsesWith.
12749 // Handle the really simple, really trivial case efficiently.
12750 if (From == To) return;
12751
12752 // Handle the simple, trivial, case efficiently.
12753 if (From.getNode()->getNumValues() == 1) {
12754 ReplaceAllUsesWith(From, To);
12755 return;
12756 }
12757
12758 // Preserve Debug Info.
12759 transferDbgValues(From, To);
12760 copyExtraInfo(From.getNode(), To.getNode());
12761
12762 // Iterate over just the existing users of From. See the comments in
12763 // the ReplaceAllUsesWith above.
12764 SDNode::use_iterator UI = From.getNode()->use_begin(),
12765 UE = From.getNode()->use_end();
12766 RAUWUpdateListener Listener(*this, UI, UE);
12767 while (UI != UE) {
12768 SDNode *User = UI->getUser();
12769 bool UserRemovedFromCSEMaps = false;
12770
12771 // A user can appear in a use list multiple times, and when this
12772 // happens the uses are usually next to each other in the list.
12773 // To help reduce the number of CSE recomputations, process all
12774 // the uses of this user that we can find this way.
12775 do {
12776 SDUse &Use = *UI;
12777
12778 // Skip uses of different values from the same node.
12779 if (Use.getResNo() != From.getResNo()) {
12780 ++UI;
12781 continue;
12782 }
12783
12784 // If this node hasn't been modified yet, it's still in the CSE maps,
12785 // so remove its old self from the CSE maps.
12786 if (!UserRemovedFromCSEMaps) {
12787 RemoveNodeFromCSEMaps(User);
12788 UserRemovedFromCSEMaps = true;
12789 }
12790
12791 ++UI;
12792 Use.set(To);
12793 if (To->isDivergent() != From->isDivergent())
12795 } while (UI != UE && UI->getUser() == User);
12796 // We are iterating over all uses of the From node, so if a use
12797 // doesn't use the specific value, no changes are made.
12798 if (!UserRemovedFromCSEMaps)
12799 continue;
12800
12801 // Now that we have modified User, add it back to the CSE maps. If it
12802 // already exists there, recursively merge the results together.
12803 AddModifiedNodeToCSEMaps(User);
12804 }
12805
12806 // If we just RAUW'd the root, take note.
12807 if (From == getRoot())
12808 setRoot(To);
12809}
12810
12811namespace {
12812
12813/// UseMemo - This class is used by SelectionDAG::ReplaceAllUsesOfValuesWith
12814/// to record information about a use.
12815struct UseMemo {
12816 SDNode *User;
12817 unsigned Index;
12818 SDUse *Use;
12819};
12820
12821/// operator< - Sort Memos by User.
12822bool operator<(const UseMemo &L, const UseMemo &R) {
12823 return (intptr_t)L.User < (intptr_t)R.User;
12824}
12825
12826/// RAUOVWUpdateListener - Helper for ReplaceAllUsesOfValuesWith - When the node
12827/// pointed to by a UseMemo is deleted, set the User to nullptr to indicate that
12828/// the node already has been taken care of recursively.
12829class RAUOVWUpdateListener : public SelectionDAG::DAGUpdateListener {
12830 SmallVectorImpl<UseMemo> &Uses;
12831
12832 void NodeDeleted(SDNode *N, SDNode *E) override {
12833 for (UseMemo &Memo : Uses)
12834 if (Memo.User == N)
12835 Memo.User = nullptr;
12836 }
12837
12838public:
12839 RAUOVWUpdateListener(SelectionDAG &d, SmallVectorImpl<UseMemo> &uses)
12840 : SelectionDAG::DAGUpdateListener(d), Uses(uses) {}
12841};
12842
12843} // end anonymous namespace
12844
12845/// Return true if a glue output should propagate divergence information.
12847 switch (Node->getOpcode()) {
12848 case ISD::CopyFromReg:
12849 case ISD::CopyToReg:
12850 return false;
12851 default:
12852 return true;
12853 }
12854
12855 llvm_unreachable("covered opcode switch");
12856}
12857
12859 if (TLI->isSDNodeAlwaysUniform(N)) {
12860 assert(!TLI->isSDNodeSourceOfDivergence(N, FLI, UA) &&
12861 "Conflicting divergence information!");
12862 return false;
12863 }
12864 if (TLI->isSDNodeSourceOfDivergence(N, FLI, UA))
12865 return true;
12866 for (const auto &Op : N->ops()) {
12867 EVT VT = Op.getValueType();
12868
12869 // Skip Chain. It does not carry divergence.
12870 if (VT != MVT::Other && Op.getNode()->isDivergent() &&
12871 (VT != MVT::Glue || gluePropagatesDivergence(Op.getNode())))
12872 return true;
12873 }
12874 return false;
12875}
12876
12878 SmallVector<SDNode *, 16> Worklist(1, N);
12879 do {
12880 N = Worklist.pop_back_val();
12881 bool IsDivergent = calculateDivergence(N);
12882 if (N->SDNodeBits.IsDivergent != IsDivergent) {
12883 N->SDNodeBits.IsDivergent = IsDivergent;
12884 llvm::append_range(Worklist, N->users());
12885 }
12886 } while (!Worklist.empty());
12887}
12888
12889void SelectionDAG::CreateTopologicalOrder(std::vector<SDNode *> &Order) {
12891 Order.reserve(AllNodes.size());
12892 for (auto &N : allnodes()) {
12893 unsigned NOps = N.getNumOperands();
12894 Degree[&N] = NOps;
12895 if (0 == NOps)
12896 Order.push_back(&N);
12897 }
12898 for (size_t I = 0; I != Order.size(); ++I) {
12899 SDNode *N = Order[I];
12900 for (auto *U : N->users()) {
12901 unsigned &UnsortedOps = Degree[U];
12902 if (0 == --UnsortedOps)
12903 Order.push_back(U);
12904 }
12905 }
12906}
12907
12908#if !defined(NDEBUG) && LLVM_ENABLE_ABI_BREAKING_CHECKS
12909void SelectionDAG::VerifyDAGDivergence() {
12910 std::vector<SDNode *> TopoOrder;
12911 CreateTopologicalOrder(TopoOrder);
12912 for (auto *N : TopoOrder) {
12913 assert(calculateDivergence(N) == N->isDivergent() &&
12914 "Divergence bit inconsistency detected");
12915 }
12916}
12917#endif
12918
12919/// ReplaceAllUsesOfValuesWith - Replace any uses of From with To, leaving
12920/// uses of other values produced by From.getNode() alone. The same value
12921/// may appear in both the From and To list. The Deleted vector is
12922/// handled the same way as for ReplaceAllUsesWith.
12924 const SDValue *To,
12925 unsigned Num){
12926 // Handle the simple, trivial case efficiently.
12927 if (Num == 1)
12928 return ReplaceAllUsesOfValueWith(*From, *To);
12929
12930 transferDbgValues(*From, *To);
12931 copyExtraInfo(From->getNode(), To->getNode());
12932
12933 // Read up all the uses and make records of them. This helps
12934 // processing new uses that are introduced during the
12935 // replacement process.
12937 for (unsigned i = 0; i != Num; ++i) {
12938 unsigned FromResNo = From[i].getResNo();
12939 SDNode *FromNode = From[i].getNode();
12940 for (SDUse &Use : FromNode->uses()) {
12941 if (Use.getResNo() == FromResNo) {
12942 UseMemo Memo = {Use.getUser(), i, &Use};
12943 Uses.push_back(Memo);
12944 }
12945 }
12946 }
12947
12948 // Sort the uses, so that all the uses from a given User are together.
12950 RAUOVWUpdateListener Listener(*this, Uses);
12951
12952 for (unsigned UseIndex = 0, UseIndexEnd = Uses.size();
12953 UseIndex != UseIndexEnd; ) {
12954 // We know that this user uses some value of From. If it is the right
12955 // value, update it.
12956 SDNode *User = Uses[UseIndex].User;
12957 // If the node has been deleted by recursive CSE updates when updating
12958 // another node, then just skip this entry.
12959 if (User == nullptr) {
12960 ++UseIndex;
12961 continue;
12962 }
12963
12964 // This node is about to morph, remove its old self from the CSE maps.
12965 RemoveNodeFromCSEMaps(User);
12966
12967 // The Uses array is sorted, so all the uses for a given User
12968 // are next to each other in the list.
12969 // To help reduce the number of CSE recomputations, process all
12970 // the uses of this user that we can find this way.
12971 do {
12972 unsigned i = Uses[UseIndex].Index;
12973 SDUse &Use = *Uses[UseIndex].Use;
12974 ++UseIndex;
12975
12976 Use.set(To[i]);
12977 } while (UseIndex != UseIndexEnd && Uses[UseIndex].User == User);
12978
12979 // Now that we have modified User, add it back to the CSE maps. If it
12980 // already exists there, recursively merge the results together.
12981 AddModifiedNodeToCSEMaps(User);
12982 }
12983}
12984
12985/// AssignTopologicalOrder - Assign a unique node id for each node in the DAG
12986/// based on their topological order. It returns the maximum id and a vector
12987/// of the SDNodes* in assigned order by reference.
12989 unsigned DAGSize = 0;
12990
12991 // SortedPos tracks the progress of the algorithm. Nodes before it are
12992 // sorted, nodes after it are unsorted. When the algorithm completes
12993 // it is at the end of the list.
12994 allnodes_iterator SortedPos = allnodes_begin();
12995
12996 // Visit all the nodes. Move nodes with no operands to the front of
12997 // the list immediately. Annotate nodes that do have operands with their
12998 // operand count. Before we do this, the Node Id fields of the nodes
12999 // may contain arbitrary values. After, the Node Id fields for nodes
13000 // before SortedPos will contain the topological sort index, and the
13001 // Node Id fields for nodes At SortedPos and after will contain the
13002 // count of outstanding operands.
13004 checkForCycles(&N, this);
13005 unsigned Degree = N.getNumOperands();
13006 if (Degree == 0) {
13007 // A node with no uses, add it to the result array immediately.
13008 N.setNodeId(DAGSize++);
13009 allnodes_iterator Q(&N);
13010 if (Q != SortedPos)
13011 SortedPos = AllNodes.insert(SortedPos, AllNodes.remove(Q));
13012 assert(SortedPos != AllNodes.end() && "Overran node list");
13013 ++SortedPos;
13014 } else {
13015 // Temporarily use the Node Id as scratch space for the degree count.
13016 N.setNodeId(Degree);
13017 }
13018 }
13019
13020 // Visit all the nodes. As we iterate, move nodes into sorted order,
13021 // such that by the time the end is reached all nodes will be sorted.
13022 for (SDNode &Node : allnodes()) {
13023 SDNode *N = &Node;
13024 checkForCycles(N, this);
13025 // N is in sorted position, so all its uses have one less operand
13026 // that needs to be sorted.
13027 for (SDNode *P : N->users()) {
13028 unsigned Degree = P->getNodeId();
13029 assert(Degree != 0 && "Invalid node degree");
13030 --Degree;
13031 if (Degree == 0) {
13032 // All of P's operands are sorted, so P may sorted now.
13033 P->setNodeId(DAGSize++);
13034 if (P->getIterator() != SortedPos)
13035 SortedPos = AllNodes.insert(SortedPos, AllNodes.remove(P));
13036 assert(SortedPos != AllNodes.end() && "Overran node list");
13037 ++SortedPos;
13038 } else {
13039 // Update P's outstanding operand count.
13040 P->setNodeId(Degree);
13041 }
13042 }
13043 if (Node.getIterator() == SortedPos) {
13044#ifndef NDEBUG
13046 SDNode *S = &*++I;
13047 dbgs() << "Overran sorted position:\n";
13048 S->dumprFull(this); dbgs() << "\n";
13049 dbgs() << "Checking if this is due to cycles\n";
13050 checkForCycles(this, true);
13051#endif
13052 llvm_unreachable(nullptr);
13053 }
13054 }
13055
13056 assert(SortedPos == AllNodes.end() &&
13057 "Topological sort incomplete!");
13058 assert(AllNodes.front().getOpcode() == ISD::EntryToken &&
13059 "First node in topological sort is not the entry token!");
13060 assert(AllNodes.front().getNodeId() == 0 &&
13061 "First node in topological sort has non-zero id!");
13062 assert(AllNodes.front().getNumOperands() == 0 &&
13063 "First node in topological sort has operands!");
13064 assert(AllNodes.back().getNodeId() == (int)DAGSize-1 &&
13065 "Last node in topologic sort has unexpected id!");
13066 assert(AllNodes.back().use_empty() &&
13067 "Last node in topologic sort has users!");
13068 assert(DAGSize == allnodes_size() && "Node count mismatch!");
13069 return DAGSize;
13070}
13071
13073 SmallVectorImpl<const SDNode *> &SortedNodes) const {
13074 SortedNodes.clear();
13075 // Node -> remaining number of outstanding operands.
13076 DenseMap<const SDNode *, unsigned> RemainingOperands;
13077
13078 // Put nodes without any operands into SortedNodes first.
13079 for (const SDNode &N : allnodes()) {
13080 checkForCycles(&N, this);
13081 unsigned NumOperands = N.getNumOperands();
13082 if (NumOperands == 0)
13083 SortedNodes.push_back(&N);
13084 else
13085 // Record their total number of outstanding operands.
13086 RemainingOperands[&N] = NumOperands;
13087 }
13088
13089 // A node is pushed into SortedNodes when all of its operands (predecessors in
13090 // the graph) are also in SortedNodes.
13091 for (unsigned i = 0U; i < SortedNodes.size(); ++i) {
13092 const SDNode *N = SortedNodes[i];
13093 for (const SDNode *U : N->users()) {
13094 // HandleSDNode is never part of a DAG and therefore has no entry in
13095 // RemainingOperands.
13096 if (U->getOpcode() == ISD::HANDLENODE)
13097 continue;
13098 unsigned &NumRemOperands = RemainingOperands[U];
13099 assert(NumRemOperands && "Invalid number of remaining operands");
13100 --NumRemOperands;
13101 if (!NumRemOperands)
13102 SortedNodes.push_back(U);
13103 }
13104 }
13105
13106 assert(SortedNodes.size() == AllNodes.size() && "Node count mismatch");
13107 assert(SortedNodes.front()->getOpcode() == ISD::EntryToken &&
13108 "First node in topological sort is not the entry token");
13109 assert(SortedNodes.front()->getNumOperands() == 0 &&
13110 "First node in topological sort has operands");
13111}
13112
13113/// AddDbgValue - Add a dbg_value SDNode. If SD is non-null that means the
13114/// value is produced by SD.
13115void SelectionDAG::AddDbgValue(SDDbgValue *DB, bool isParameter) {
13116 for (SDNode *SD : DB->getSDNodes()) {
13117 if (!SD)
13118 continue;
13119 assert(DbgInfo->getSDDbgValues(SD).empty() || SD->getHasDebugValue());
13120 SD->setHasDebugValue(true);
13121 }
13122 DbgInfo->add(DB, isParameter);
13123}
13124
13125void SelectionDAG::AddDbgLabel(SDDbgLabel *DB) { DbgInfo->add(DB); }
13126
13128 SDValue NewMemOpChain) {
13129 assert(isa<MemSDNode>(NewMemOpChain) && "Expected a memop node");
13130 assert(NewMemOpChain.getValueType() == MVT::Other && "Expected a token VT");
13131 // The new memory operation must have the same position as the old load in
13132 // terms of memory dependency. Create a TokenFactor for the old load and new
13133 // memory operation and update uses of the old load's output chain to use that
13134 // TokenFactor.
13135 if (OldChain == NewMemOpChain || OldChain.use_empty())
13136 return NewMemOpChain;
13137
13138 SDValue TokenFactor = getNode(ISD::TokenFactor, SDLoc(OldChain), MVT::Other,
13139 OldChain, NewMemOpChain);
13140 ReplaceAllUsesOfValueWith(OldChain, TokenFactor);
13141 UpdateNodeOperands(TokenFactor.getNode(), OldChain, NewMemOpChain);
13142 return TokenFactor;
13143}
13144
13146 SDValue NewMemOp) {
13147 assert(isa<MemSDNode>(NewMemOp.getNode()) && "Expected a memop node");
13148 SDValue OldChain = SDValue(OldLoad, 1);
13149 SDValue NewMemOpChain = NewMemOp.getValue(1);
13150 return makeEquivalentMemoryOrdering(OldChain, NewMemOpChain);
13151}
13152
13154 Function **OutFunction) {
13155 assert(isa<ExternalSymbolSDNode>(Op) && "Node should be an ExternalSymbol");
13156
13157 auto *Symbol = cast<ExternalSymbolSDNode>(Op)->getSymbol();
13158 auto *Module = MF->getFunction().getParent();
13159 auto *Function = Module->getFunction(Symbol);
13160
13161 if (OutFunction != nullptr)
13162 *OutFunction = Function;
13163
13164 if (Function != nullptr) {
13165 auto PtrTy = TLI->getPointerTy(getDataLayout(), Function->getAddressSpace());
13166 return getGlobalAddress(Function, SDLoc(Op), PtrTy);
13167 }
13168
13169 std::string ErrorStr;
13170 raw_string_ostream ErrorFormatter(ErrorStr);
13171 ErrorFormatter << "Undefined external symbol ";
13172 ErrorFormatter << '"' << Symbol << '"';
13173 report_fatal_error(Twine(ErrorStr));
13174}
13175
13176//===----------------------------------------------------------------------===//
13177// SDNode Class
13178//===----------------------------------------------------------------------===//
13179
13182 return Const != nullptr && Const->isZero();
13183}
13184
13186 return V.isUndef() || isNullConstant(V);
13187}
13188
13191 return Const != nullptr && Const->isZero() && !Const->isNegative();
13192}
13193
13196 return Const != nullptr && Const->isAllOnes();
13197}
13198
13201 return Const != nullptr && Const->isOne();
13202}
13203
13206 return Const != nullptr && Const->isMinSignedValue();
13207}
13208
13209bool llvm::isNeutralConstant(unsigned Opcode, SDNodeFlags Flags, SDValue V,
13210 unsigned OperandNo) {
13211 // NOTE: The cases should match with IR's ConstantExpr::getBinOpIdentity().
13212 // TODO: Target-specific opcodes could be added.
13213 if (auto *ConstV = isConstOrConstSplat(V, /*AllowUndefs*/ false,
13214 /*AllowTruncation*/ true)) {
13215 APInt Const = ConstV->getAPIntValue().trunc(V.getScalarValueSizeInBits());
13216 switch (Opcode) {
13217 case ISD::ADD:
13218 case ISD::OR:
13219 case ISD::XOR:
13220 case ISD::UMAX:
13221 return Const.isZero();
13222 case ISD::MUL:
13223 return Const.isOne();
13224 case ISD::AND:
13225 case ISD::UMIN:
13226 return Const.isAllOnes();
13227 case ISD::SMAX:
13228 return Const.isMinSignedValue();
13229 case ISD::SMIN:
13230 return Const.isMaxSignedValue();
13231 case ISD::SUB:
13232 case ISD::SHL:
13233 case ISD::SRA:
13234 case ISD::SRL:
13235 return OperandNo == 1 && Const.isZero();
13236 case ISD::UDIV:
13237 case ISD::SDIV:
13238 return OperandNo == 1 && Const.isOne();
13239 }
13240 } else if (auto *ConstFP = isConstOrConstSplatFP(V)) {
13241 switch (Opcode) {
13242 case ISD::FADD:
13243 return ConstFP->isZero() &&
13244 (Flags.hasNoSignedZeros() || ConstFP->isNegative());
13245 case ISD::FSUB:
13246 return OperandNo == 1 && ConstFP->isZero() &&
13247 (Flags.hasNoSignedZeros() || !ConstFP->isNegative());
13248 case ISD::FMUL:
13249 return ConstFP->isExactlyValue(1.0);
13250 case ISD::FDIV:
13251 return OperandNo == 1 && ConstFP->isExactlyValue(1.0);
13252 case ISD::FMINNUM:
13253 case ISD::FMAXNUM: {
13254 // Neutral element for fminnum is NaN, Inf or FLT_MAX, depending on FMF.
13255 EVT VT = V.getValueType();
13256 const fltSemantics &Semantics = VT.getFltSemantics();
13257 APFloat NeutralAF = !Flags.hasNoNaNs()
13258 ? APFloat::getQNaN(Semantics)
13259 : !Flags.hasNoInfs()
13260 ? APFloat::getInf(Semantics)
13261 : APFloat::getLargest(Semantics);
13262 if (Opcode == ISD::FMAXNUM)
13263 NeutralAF.changeSign();
13264
13265 return ConstFP->isExactlyValue(NeutralAF);
13266 }
13267 }
13268 }
13269 return false;
13270}
13271
13273 while (V.getOpcode() == ISD::BITCAST)
13274 V = V.getOperand(0);
13275 return V;
13276}
13277
13279 while (V.getOpcode() == ISD::BITCAST && V.getOperand(0).hasOneUse())
13280 V = V.getOperand(0);
13281 return V;
13282}
13283
13285 while (V.getOpcode() == ISD::EXTRACT_SUBVECTOR)
13286 V = V.getOperand(0);
13287 return V;
13288}
13289
13291 while (V.getOpcode() == ISD::INSERT_VECTOR_ELT) {
13292 SDValue InVec = V.getOperand(0);
13293 SDValue EltNo = V.getOperand(2);
13294 EVT VT = InVec.getValueType();
13295 auto *IndexC = dyn_cast<ConstantSDNode>(EltNo);
13296 if (IndexC && VT.isFixedLengthVector() &&
13297 IndexC->getAPIntValue().ult(VT.getVectorNumElements()) &&
13298 !DemandedElts[IndexC->getZExtValue()]) {
13299 V = InVec;
13300 continue;
13301 }
13302 break;
13303 }
13304 return V;
13305}
13306
13308 while (V.getOpcode() == ISD::TRUNCATE)
13309 V = V.getOperand(0);
13310 return V;
13311}
13312
13313bool llvm::isBitwiseNot(SDValue V, bool AllowUndefs) {
13314 if (V.getOpcode() != ISD::XOR)
13315 return false;
13316 V = peekThroughBitcasts(V.getOperand(1));
13317 unsigned NumBits = V.getScalarValueSizeInBits();
13318 ConstantSDNode *C =
13319 isConstOrConstSplat(V, AllowUndefs, /*AllowTruncation*/ true);
13320 return C && (C->getAPIntValue().countr_one() >= NumBits);
13321}
13322
13324 bool AllowTruncation) {
13325 EVT VT = N.getValueType();
13326 APInt DemandedElts = VT.isFixedLengthVector()
13328 : APInt(1, 1);
13329 return isConstOrConstSplat(N, DemandedElts, AllowUndefs, AllowTruncation);
13330}
13331
13333 bool AllowUndefs,
13334 bool AllowTruncation) {
13336 return CN;
13337
13338 // SplatVectors can truncate their operands. Ignore that case here unless
13339 // AllowTruncation is set.
13340 if (N->getOpcode() == ISD::SPLAT_VECTOR) {
13341 EVT VecEltVT = N->getValueType(0).getVectorElementType();
13342 if (auto *CN = dyn_cast<ConstantSDNode>(N->getOperand(0))) {
13343 EVT CVT = CN->getValueType(0);
13344 assert(CVT.bitsGE(VecEltVT) && "Illegal splat_vector element extension");
13345 if (AllowTruncation || CVT == VecEltVT)
13346 return CN;
13347 }
13348 }
13349
13351 BitVector UndefElements;
13352 ConstantSDNode *CN = BV->getConstantSplatNode(DemandedElts, &UndefElements);
13353
13354 // BuildVectors can truncate their operands. Ignore that case here unless
13355 // AllowTruncation is set.
13356 // TODO: Look into whether we should allow UndefElements in non-DemandedElts
13357 if (CN && (UndefElements.none() || AllowUndefs)) {
13358 EVT CVT = CN->getValueType(0);
13359 EVT NSVT = N.getValueType().getScalarType();
13360 assert(CVT.bitsGE(NSVT) && "Illegal build vector element extension");
13361 if (AllowTruncation || (CVT == NSVT))
13362 return CN;
13363 }
13364 }
13365
13366 return nullptr;
13367}
13368
13370 EVT VT = N.getValueType();
13371 APInt DemandedElts = VT.isFixedLengthVector()
13373 : APInt(1, 1);
13374 return isConstOrConstSplatFP(N, DemandedElts, AllowUndefs);
13375}
13376
13378 const APInt &DemandedElts,
13379 bool AllowUndefs) {
13381 return CN;
13382
13384 BitVector UndefElements;
13385 ConstantFPSDNode *CN =
13386 BV->getConstantFPSplatNode(DemandedElts, &UndefElements);
13387 // TODO: Look into whether we should allow UndefElements in non-DemandedElts
13388 if (CN && (UndefElements.none() || AllowUndefs))
13389 return CN;
13390 }
13391
13392 if (N.getOpcode() == ISD::SPLAT_VECTOR)
13393 if (ConstantFPSDNode *CN = dyn_cast<ConstantFPSDNode>(N.getOperand(0)))
13394 return CN;
13395
13396 return nullptr;
13397}
13398
13399bool llvm::isNullOrNullSplat(SDValue N, bool AllowUndefs) {
13400 // TODO: may want to use peekThroughBitcast() here.
13401 ConstantSDNode *C =
13402 isConstOrConstSplat(N, AllowUndefs, /*AllowTruncation=*/true);
13403 return C && C->isZero();
13404}
13405
13406bool llvm::isOneOrOneSplat(SDValue N, bool AllowUndefs) {
13407 ConstantSDNode *C =
13408 isConstOrConstSplat(N, AllowUndefs, /*AllowTruncation*/ true);
13409 return C && C->isOne();
13410}
13411
13412bool llvm::isOneOrOneSplatFP(SDValue N, bool AllowUndefs) {
13413 ConstantFPSDNode *C = isConstOrConstSplatFP(N, AllowUndefs);
13414 return C && C->isExactlyValue(1.0);
13415}
13416
13417bool llvm::isAllOnesOrAllOnesSplat(SDValue N, bool AllowUndefs) {
13419 unsigned BitWidth = N.getScalarValueSizeInBits();
13420 ConstantSDNode *C = isConstOrConstSplat(N, AllowUndefs);
13421 return C && C->isAllOnes() && C->getValueSizeInBits(0) == BitWidth;
13422}
13423
13424bool llvm::isOnesOrOnesSplat(SDValue N, bool AllowUndefs) {
13425 ConstantSDNode *C = isConstOrConstSplat(N, AllowUndefs);
13426 return C && APInt::isSameValue(C->getAPIntValue(),
13427 APInt(C->getAPIntValue().getBitWidth(), 1));
13428}
13429
13430bool llvm::isZeroOrZeroSplat(SDValue N, bool AllowUndefs) {
13432 ConstantSDNode *C = isConstOrConstSplat(N, AllowUndefs, true);
13433 return C && C->isZero();
13434}
13435
13436bool llvm::isZeroOrZeroSplatFP(SDValue N, bool AllowUndefs) {
13437 ConstantFPSDNode *C = isConstOrConstSplatFP(N, AllowUndefs);
13438 return C && C->isZero();
13439}
13440
13444
13446 unsigned Opc, unsigned Order, const DebugLoc &dl, SDVTList VTs, EVT memvt,
13448 : SDNode(Opc, Order, dl, VTs), MemoryVT(memvt), MemRefs(memrefs) {
13449 bool IsVolatile = false;
13450 bool IsNonTemporal = false;
13451 bool IsDereferenceable = true;
13452 bool IsInvariant = true;
13453 for (const MachineMemOperand *MMO : memoperands()) {
13454 IsVolatile |= MMO->isVolatile();
13455 IsNonTemporal |= MMO->isNonTemporal();
13456 IsDereferenceable &= MMO->isDereferenceable();
13457 IsInvariant &= MMO->isInvariant();
13458 }
13459 MemSDNodeBits.IsVolatile = IsVolatile;
13460 MemSDNodeBits.IsNonTemporal = IsNonTemporal;
13461 MemSDNodeBits.IsDereferenceable = IsDereferenceable;
13462 MemSDNodeBits.IsInvariant = IsInvariant;
13463
13464 // For the single-MMO case, we check here that the size of the memory operand
13465 // fits within the size of the MMO. This is because the MMO might indicate
13466 // only a possible address range instead of specifying the affected memory
13467 // addresses precisely.
13470 getMemOperand()->getSize().getValue())) &&
13471 "Size mismatch!");
13472}
13473
13474/// Profile - Gather unique data for the node.
13475///
13477 AddNodeIDNode(ID, this);
13478}
13479
13480namespace {
13481
13482 struct EVTArray {
13483 std::vector<EVT> VTs;
13484
13485 EVTArray() {
13486 VTs.reserve(MVT::VALUETYPE_SIZE);
13487 for (unsigned i = 0; i < MVT::VALUETYPE_SIZE; ++i)
13488 VTs.push_back(MVT((MVT::SimpleValueType)i));
13489 }
13490 };
13491
13492} // end anonymous namespace
13493
13494/// getValueTypeList - Return a pointer to the specified value type.
13495///
13496const EVT *SDNode::getValueTypeList(MVT VT) {
13497 static EVTArray SimpleVTArray;
13498
13499 assert(VT < MVT::VALUETYPE_SIZE && "Value type out of range!");
13500 return &SimpleVTArray.VTs[VT.SimpleTy];
13501}
13502
13503/// hasAnyUseOfValue - Return true if there are any use of the indicated
13504/// value. This method ignores uses of other values defined by this operation.
13505bool SDNode::hasAnyUseOfValue(unsigned Value) const {
13506 assert(Value < getNumValues() && "Bad value!");
13507
13508 for (SDUse &U : uses())
13509 if (U.getResNo() == Value)
13510 return true;
13511
13512 return false;
13513}
13514
13515/// isOnlyUserOf - Return true if this node is the only use of N.
13516bool SDNode::isOnlyUserOf(const SDNode *N) const {
13517 bool Seen = false;
13518 for (const SDNode *User : N->users()) {
13519 if (User == this)
13520 Seen = true;
13521 else
13522 return false;
13523 }
13524
13525 return Seen;
13526}
13527
13528/// Return true if the only users of N are contained in Nodes.
13530 bool Seen = false;
13531 for (const SDNode *User : N->users()) {
13532 if (llvm::is_contained(Nodes, User))
13533 Seen = true;
13534 else
13535 return false;
13536 }
13537
13538 return Seen;
13539}
13540
13541/// Return true if the referenced return value is an operand of N.
13542bool SDValue::isOperandOf(const SDNode *N) const {
13543 return is_contained(N->op_values(), *this);
13544}
13545
13546bool SDNode::isOperandOf(const SDNode *N) const {
13547 return any_of(N->op_values(),
13548 [this](SDValue Op) { return this == Op.getNode(); });
13549}
13550
13551/// reachesChainWithoutSideEffects - Return true if this operand (which must
13552/// be a chain) reaches the specified operand without crossing any
13553/// side-effecting instructions on any chain path. In practice, this looks
13554/// through token factors and non-volatile loads. In order to remain efficient,
13555/// this only looks a couple of nodes in, it does not do an exhaustive search.
13556///
13557/// Note that we only need to examine chains when we're searching for
13558/// side-effects; SelectionDAG requires that all side-effects are represented
13559/// by chains, even if another operand would force a specific ordering. This
13560/// constraint is necessary to allow transformations like splitting loads.
13562 unsigned Depth) const {
13563 if (*this == Dest) return true;
13564
13565 // Don't search too deeply, we just want to be able to see through
13566 // TokenFactor's etc.
13567 if (Depth == 0) return false;
13568
13569 // If this is a token factor, all inputs to the TF happen in parallel.
13570 if (getOpcode() == ISD::TokenFactor) {
13571 // First, try a shallow search.
13572 if (is_contained((*this)->ops(), Dest)) {
13573 // We found the chain we want as an operand of this TokenFactor.
13574 // Essentially, we reach the chain without side-effects if we could
13575 // serialize the TokenFactor into a simple chain of operations with
13576 // Dest as the last operation. This is automatically true if the
13577 // chain has one use: there are no other ordering constraints.
13578 // If the chain has more than one use, we give up: some other
13579 // use of Dest might force a side-effect between Dest and the current
13580 // node.
13581 if (Dest.hasOneUse())
13582 return true;
13583 }
13584 // Next, try a deep search: check whether every operand of the TokenFactor
13585 // reaches Dest.
13586 return llvm::all_of((*this)->ops(), [=](SDValue Op) {
13587 return Op.reachesChainWithoutSideEffects(Dest, Depth - 1);
13588 });
13589 }
13590
13591 // Loads don't have side effects, look through them.
13592 if (LoadSDNode *Ld = dyn_cast<LoadSDNode>(*this)) {
13593 if (Ld->isUnordered())
13594 return Ld->getChain().reachesChainWithoutSideEffects(Dest, Depth-1);
13595 }
13596 return false;
13597}
13598
13599bool SDNode::hasPredecessor(const SDNode *N) const {
13602 Worklist.push_back(this);
13603 return hasPredecessorHelper(N, Visited, Worklist);
13604}
13605
13607 this->Flags &= Flags;
13608}
13609
13610SDValue
13612 ArrayRef<ISD::NodeType> CandidateBinOps,
13613 bool AllowPartials) {
13614 // The pattern must end in an extract from index 0.
13615 if (Extract->getOpcode() != ISD::EXTRACT_VECTOR_ELT ||
13616 !isNullConstant(Extract->getOperand(1)))
13617 return SDValue();
13618
13619 // Match against one of the candidate binary ops.
13620 SDValue Op = Extract->getOperand(0);
13621 if (llvm::none_of(CandidateBinOps, [Op](ISD::NodeType BinOp) {
13622 return Op.getOpcode() == unsigned(BinOp);
13623 }))
13624 return SDValue();
13625
13626 // Floating-point reductions may require relaxed constraints on the final step
13627 // of the reduction because they may reorder intermediate operations.
13628 unsigned CandidateBinOp = Op.getOpcode();
13629 if (Op.getValueType().isFloatingPoint()) {
13630 SDNodeFlags Flags = Op->getFlags();
13631 switch (CandidateBinOp) {
13632 case ISD::FADD:
13633 if (!Flags.hasNoSignedZeros() || !Flags.hasAllowReassociation())
13634 return SDValue();
13635 break;
13636 default:
13637 llvm_unreachable("Unhandled FP opcode for binop reduction");
13638 }
13639 }
13640
13641 // Matching failed - attempt to see if we did enough stages that a partial
13642 // reduction from a subvector is possible.
13643 auto PartialReduction = [&](SDValue Op, unsigned NumSubElts) {
13644 if (!AllowPartials || !Op)
13645 return SDValue();
13646 EVT OpVT = Op.getValueType();
13647 EVT OpSVT = OpVT.getScalarType();
13648 EVT SubVT = EVT::getVectorVT(*getContext(), OpSVT, NumSubElts);
13649 if (!TLI->isExtractSubvectorCheap(SubVT, OpVT, 0))
13650 return SDValue();
13651 BinOp = (ISD::NodeType)CandidateBinOp;
13652 return getExtractSubvector(SDLoc(Op), SubVT, Op, 0);
13653 };
13654
13655 // At each stage, we're looking for something that looks like:
13656 // %s = shufflevector <8 x i32> %op, <8 x i32> undef,
13657 // <8 x i32> <i32 2, i32 3, i32 undef, i32 undef,
13658 // i32 undef, i32 undef, i32 undef, i32 undef>
13659 // %a = binop <8 x i32> %op, %s
13660 // Where the mask changes according to the stage. E.g. for a 3-stage pyramid,
13661 // we expect something like:
13662 // <4,5,6,7,u,u,u,u>
13663 // <2,3,u,u,u,u,u,u>
13664 // <1,u,u,u,u,u,u,u>
13665 // While a partial reduction match would be:
13666 // <2,3,u,u,u,u,u,u>
13667 // <1,u,u,u,u,u,u,u>
13668 unsigned Stages = Log2_32(Op.getValueType().getVectorNumElements());
13669 SDValue PrevOp;
13670 for (unsigned i = 0; i < Stages; ++i) {
13671 unsigned MaskEnd = (1 << i);
13672
13673 if (Op.getOpcode() != CandidateBinOp)
13674 return PartialReduction(PrevOp, MaskEnd);
13675
13676 SDValue Op0 = Op.getOperand(0);
13677 SDValue Op1 = Op.getOperand(1);
13678
13680 if (Shuffle) {
13681 Op = Op1;
13682 } else {
13683 Shuffle = dyn_cast<ShuffleVectorSDNode>(Op1);
13684 Op = Op0;
13685 }
13686
13687 // The first operand of the shuffle should be the same as the other operand
13688 // of the binop.
13689 if (!Shuffle || Shuffle->getOperand(0) != Op)
13690 return PartialReduction(PrevOp, MaskEnd);
13691
13692 // Verify the shuffle has the expected (at this stage of the pyramid) mask.
13693 for (int Index = 0; Index < (int)MaskEnd; ++Index)
13694 if (Shuffle->getMaskElt(Index) != (int)(MaskEnd + Index))
13695 return PartialReduction(PrevOp, MaskEnd);
13696
13697 PrevOp = Op;
13698 }
13699
13700 // Handle subvector reductions, which tend to appear after the shuffle
13701 // reduction stages.
13702 while (Op.getOpcode() == CandidateBinOp) {
13703 unsigned NumElts = Op.getValueType().getVectorNumElements();
13704 SDValue Op0 = Op.getOperand(0);
13705 SDValue Op1 = Op.getOperand(1);
13706 if (Op0.getOpcode() != ISD::EXTRACT_SUBVECTOR ||
13708 Op0.getOperand(0) != Op1.getOperand(0))
13709 break;
13710 SDValue Src = Op0.getOperand(0);
13711 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
13712 if (NumSrcElts != (2 * NumElts))
13713 break;
13714 if (!(Op0.getConstantOperandAPInt(1) == 0 &&
13715 Op1.getConstantOperandAPInt(1) == NumElts) &&
13716 !(Op1.getConstantOperandAPInt(1) == 0 &&
13717 Op0.getConstantOperandAPInt(1) == NumElts))
13718 break;
13719 Op = Src;
13720 }
13721
13722 BinOp = (ISD::NodeType)CandidateBinOp;
13723 return Op;
13724}
13725
13727 EVT VT = N->getValueType(0);
13728 EVT EltVT = VT.getVectorElementType();
13729 unsigned NE = VT.getVectorNumElements();
13730
13731 SDLoc dl(N);
13732
13733 // If ResNE is 0, fully unroll the vector op.
13734 if (ResNE == 0)
13735 ResNE = NE;
13736 else if (NE > ResNE)
13737 NE = ResNE;
13738
13739 if (N->getNumValues() == 2) {
13740 SmallVector<SDValue, 8> Scalars0, Scalars1;
13741 SmallVector<SDValue, 4> Operands(N->getNumOperands());
13742 EVT VT1 = N->getValueType(1);
13743 EVT EltVT1 = VT1.getVectorElementType();
13744
13745 unsigned i;
13746 for (i = 0; i != NE; ++i) {
13747 for (unsigned j = 0, e = N->getNumOperands(); j != e; ++j) {
13748 SDValue Operand = N->getOperand(j);
13749 EVT OperandVT = Operand.getValueType();
13750
13751 // A vector operand; extract a single element.
13752 EVT OperandEltVT = OperandVT.getVectorElementType();
13753 Operands[j] = getExtractVectorElt(dl, OperandEltVT, Operand, i);
13754 }
13755
13756 SDValue EltOp = getNode(N->getOpcode(), dl, {EltVT, EltVT1}, Operands);
13757 Scalars0.push_back(EltOp);
13758 Scalars1.push_back(EltOp.getValue(1));
13759 }
13760
13761 for (; i < ResNE; ++i) {
13762 Scalars0.push_back(getUNDEF(EltVT));
13763 Scalars1.push_back(getUNDEF(EltVT1));
13764 }
13765
13766 EVT VecVT = EVT::getVectorVT(*getContext(), EltVT, ResNE);
13767 EVT VecVT1 = EVT::getVectorVT(*getContext(), EltVT1, ResNE);
13768 SDValue Vec0 = getBuildVector(VecVT, dl, Scalars0);
13769 SDValue Vec1 = getBuildVector(VecVT1, dl, Scalars1);
13770 return getMergeValues({Vec0, Vec1}, dl);
13771 }
13772
13773 assert(N->getNumValues() == 1 &&
13774 "Can't unroll a vector with multiple results!");
13775
13777 SmallVector<SDValue, 4> Operands(N->getNumOperands());
13778
13779 unsigned i;
13780 for (i= 0; i != NE; ++i) {
13781 for (unsigned j = 0, e = N->getNumOperands(); j != e; ++j) {
13782 SDValue Operand = N->getOperand(j);
13783 EVT OperandVT = Operand.getValueType();
13784 if (OperandVT.isVector()) {
13785 // A vector operand; extract a single element.
13786 EVT OperandEltVT = OperandVT.getVectorElementType();
13787 Operands[j] = getExtractVectorElt(dl, OperandEltVT, Operand, i);
13788 } else {
13789 // A scalar operand; just use it as is.
13790 Operands[j] = Operand;
13791 }
13792 }
13793
13794 switch (N->getOpcode()) {
13795 default: {
13796 Scalars.push_back(getNode(N->getOpcode(), dl, EltVT, Operands,
13797 N->getFlags()));
13798 break;
13799 }
13800 case ISD::VSELECT:
13801 Scalars.push_back(getNode(ISD::SELECT, dl, EltVT, Operands));
13802 break;
13803 case ISD::SHL:
13804 case ISD::SRA:
13805 case ISD::SRL:
13806 case ISD::ROTL:
13807 case ISD::ROTR:
13808 Scalars.push_back(getNode(N->getOpcode(), dl, EltVT, Operands[0],
13809 getShiftAmountOperand(Operands[0].getValueType(),
13810 Operands[1])));
13811 break;
13813 EVT ExtVT = cast<VTSDNode>(Operands[1])->getVT().getVectorElementType();
13814 Scalars.push_back(getNode(N->getOpcode(), dl, EltVT,
13815 Operands[0],
13816 getValueType(ExtVT)));
13817 break;
13818 }
13819 case ISD::ADDRSPACECAST: {
13820 const auto *ASC = cast<AddrSpaceCastSDNode>(N);
13821 Scalars.push_back(getAddrSpaceCast(dl, EltVT, Operands[0],
13822 ASC->getSrcAddressSpace(),
13823 ASC->getDestAddressSpace()));
13824 break;
13825 }
13826 }
13827 }
13828
13829 for (; i < ResNE; ++i)
13830 Scalars.push_back(getUNDEF(EltVT));
13831
13832 EVT VecVT = EVT::getVectorVT(*getContext(), EltVT, ResNE);
13833 return getBuildVector(VecVT, dl, Scalars);
13834}
13835
13836std::pair<SDValue, SDValue> SelectionDAG::UnrollVectorOverflowOp(
13837 SDNode *N, unsigned ResNE) {
13838 unsigned Opcode = N->getOpcode();
13839 assert((Opcode == ISD::UADDO || Opcode == ISD::SADDO ||
13840 Opcode == ISD::USUBO || Opcode == ISD::SSUBO ||
13841 Opcode == ISD::UMULO || Opcode == ISD::SMULO) &&
13842 "Expected an overflow opcode");
13843
13844 EVT ResVT = N->getValueType(0);
13845 EVT OvVT = N->getValueType(1);
13846 EVT ResEltVT = ResVT.getVectorElementType();
13847 EVT OvEltVT = OvVT.getVectorElementType();
13848 SDLoc dl(N);
13849
13850 // If ResNE is 0, fully unroll the vector op.
13851 unsigned NE = ResVT.getVectorNumElements();
13852 if (ResNE == 0)
13853 ResNE = NE;
13854 else if (NE > ResNE)
13855 NE = ResNE;
13856
13857 SmallVector<SDValue, 8> LHSScalars;
13858 SmallVector<SDValue, 8> RHSScalars;
13859 ExtractVectorElements(N->getOperand(0), LHSScalars, 0, NE);
13860 ExtractVectorElements(N->getOperand(1), RHSScalars, 0, NE);
13861
13862 EVT SVT = TLI->getSetCCResultType(getDataLayout(), *getContext(), ResEltVT);
13863 SDVTList VTs = getVTList(ResEltVT, SVT);
13864 SmallVector<SDValue, 8> ResScalars;
13865 SmallVector<SDValue, 8> OvScalars;
13866 for (unsigned i = 0; i < NE; ++i) {
13867 SDValue Res = getNode(Opcode, dl, VTs, LHSScalars[i], RHSScalars[i]);
13868 SDValue Ov =
13869 getSelect(dl, OvEltVT, Res.getValue(1),
13870 getBoolConstant(true, dl, OvEltVT, ResVT),
13871 getConstant(0, dl, OvEltVT));
13872
13873 ResScalars.push_back(Res);
13874 OvScalars.push_back(Ov);
13875 }
13876
13877 ResScalars.append(ResNE - NE, getUNDEF(ResEltVT));
13878 OvScalars.append(ResNE - NE, getUNDEF(OvEltVT));
13879
13880 EVT NewResVT = EVT::getVectorVT(*getContext(), ResEltVT, ResNE);
13881 EVT NewOvVT = EVT::getVectorVT(*getContext(), OvEltVT, ResNE);
13882 return std::make_pair(getBuildVector(NewResVT, dl, ResScalars),
13883 getBuildVector(NewOvVT, dl, OvScalars));
13884}
13885
13888 unsigned Bytes,
13889 int Dist) const {
13890 if (LD->isVolatile() || Base->isVolatile())
13891 return false;
13892 // TODO: probably too restrictive for atomics, revisit
13893 if (!LD->isSimple())
13894 return false;
13895 if (LD->isIndexed() || Base->isIndexed())
13896 return false;
13897 if (LD->getChain() != Base->getChain())
13898 return false;
13899 EVT VT = LD->getMemoryVT();
13900 if (VT.getSizeInBits() / 8 != Bytes)
13901 return false;
13902
13903 auto BaseLocDecomp = BaseIndexOffset::match(Base, *this);
13904 auto LocDecomp = BaseIndexOffset::match(LD, *this);
13905
13906 int64_t Offset = 0;
13907 if (BaseLocDecomp.equalBaseIndex(LocDecomp, *this, Offset))
13908 return (Dist * (int64_t)Bytes == Offset);
13909 return false;
13910}
13911
13912/// InferPtrAlignment - Infer alignment of a load / store address. Return
13913/// std::nullopt if it cannot be inferred.
13915 // If this is a GlobalAddress + cst, return the alignment.
13916 const GlobalValue *GV = nullptr;
13917 int64_t GVOffset = 0;
13918 if (TLI->isGAPlusOffset(Ptr.getNode(), GV, GVOffset)) {
13919 unsigned PtrWidth = getDataLayout().getPointerTypeSizeInBits(GV->getType());
13920 KnownBits Known(PtrWidth);
13922 unsigned AlignBits = Known.countMinTrailingZeros();
13923 if (AlignBits)
13924 return commonAlignment(Align(1ull << std::min(31U, AlignBits)), GVOffset);
13925 }
13926
13927 // If this is a direct reference to a stack slot, use information about the
13928 // stack slot's alignment.
13929 int FrameIdx = INT_MIN;
13930 int64_t FrameOffset = 0;
13932 FrameIdx = FI->getIndex();
13933 } else if (isBaseWithConstantOffset(Ptr) &&
13935 // Handle FI+Cst
13936 FrameIdx = cast<FrameIndexSDNode>(Ptr.getOperand(0))->getIndex();
13937 FrameOffset = Ptr.getConstantOperandVal(1);
13938 }
13939
13940 if (FrameIdx != INT_MIN) {
13942 return commonAlignment(MFI.getObjectAlign(FrameIdx), FrameOffset);
13943 }
13944
13945 return std::nullopt;
13946}
13947
13948/// Split the scalar node with EXTRACT_ELEMENT using the provided
13949/// VTs and return the low/high part.
13950std::pair<SDValue, SDValue> SelectionDAG::SplitScalar(const SDValue &N,
13951 const SDLoc &DL,
13952 const EVT &LoVT,
13953 const EVT &HiVT) {
13954 assert(!LoVT.isVector() && !HiVT.isVector() && !N.getValueType().isVector() &&
13955 "Split node must be a scalar type");
13956 SDValue Lo =
13958 SDValue Hi =
13960 return std::make_pair(Lo, Hi);
13961}
13962
13963/// GetSplitDestVTs - Compute the VTs needed for the low/hi parts of a type
13964/// which is split (or expanded) into two not necessarily identical pieces.
13965std::pair<EVT, EVT> SelectionDAG::GetSplitDestVTs(const EVT &VT) const {
13966 // Currently all types are split in half.
13967 EVT LoVT, HiVT;
13968 if (!VT.isVector())
13969 LoVT = HiVT = TLI->getTypeToTransformTo(*getContext(), VT);
13970 else
13971 LoVT = HiVT = VT.getHalfNumVectorElementsVT(*getContext());
13972
13973 return std::make_pair(LoVT, HiVT);
13974}
13975
13976/// GetDependentSplitDestVTs - Compute the VTs needed for the low/hi parts of a
13977/// type, dependent on an enveloping VT that has been split into two identical
13978/// pieces. Sets the HiIsEmpty flag when hi type has zero storage size.
13979std::pair<EVT, EVT>
13981 bool *HiIsEmpty) const {
13982 EVT EltTp = VT.getVectorElementType();
13983 // Examples:
13984 // custom VL=8 with enveloping VL=8/8 yields 8/0 (hi empty)
13985 // custom VL=9 with enveloping VL=8/8 yields 8/1
13986 // custom VL=10 with enveloping VL=8/8 yields 8/2
13987 // etc.
13988 ElementCount VTNumElts = VT.getVectorElementCount();
13989 ElementCount EnvNumElts = EnvVT.getVectorElementCount();
13990 assert(VTNumElts.isScalable() == EnvNumElts.isScalable() &&
13991 "Mixing fixed width and scalable vectors when enveloping a type");
13992 EVT LoVT, HiVT;
13993 if (VTNumElts.getKnownMinValue() > EnvNumElts.getKnownMinValue()) {
13994 LoVT = EVT::getVectorVT(*getContext(), EltTp, EnvNumElts);
13995 HiVT = EVT::getVectorVT(*getContext(), EltTp, VTNumElts - EnvNumElts);
13996 *HiIsEmpty = false;
13997 } else {
13998 // Flag that hi type has zero storage size, but return split envelop type
13999 // (this would be easier if vector types with zero elements were allowed).
14000 LoVT = EVT::getVectorVT(*getContext(), EltTp, VTNumElts);
14001 HiVT = EVT::getVectorVT(*getContext(), EltTp, EnvNumElts);
14002 *HiIsEmpty = true;
14003 }
14004 return std::make_pair(LoVT, HiVT);
14005}
14006
14007/// SplitVector - Split the vector with EXTRACT_SUBVECTOR and return the
14008/// low/high part.
14009std::pair<SDValue, SDValue>
14010SelectionDAG::SplitVector(const SDValue &N, const SDLoc &DL, const EVT &LoVT,
14011 const EVT &HiVT) {
14012 assert(LoVT.isScalableVector() == HiVT.isScalableVector() &&
14013 LoVT.isScalableVector() == N.getValueType().isScalableVector() &&
14014 "Splitting vector with an invalid mixture of fixed and scalable "
14015 "vector types");
14017 N.getValueType().getVectorMinNumElements() &&
14018 "More vector elements requested than available!");
14019 SDValue Lo, Hi;
14020 Lo = getExtractSubvector(DL, LoVT, N, 0);
14021 // For scalable vectors it is safe to use LoVT.getVectorMinNumElements()
14022 // (rather than having to use ElementCount), because EXTRACT_SUBVECTOR scales
14023 // IDX with the runtime scaling factor of the result vector type. For
14024 // fixed-width result vectors, that runtime scaling factor is 1.
14027 return std::make_pair(Lo, Hi);
14028}
14029
14030std::pair<SDValue, SDValue> SelectionDAG::SplitEVL(SDValue N, EVT VecVT,
14031 const SDLoc &DL) {
14032 // Split the vector length parameter.
14033 // %evl -> umin(%evl, %halfnumelts) and usubsat(%evl - %halfnumelts).
14034 EVT VT = N.getValueType();
14036 "Expecting the mask to be an evenly-sized vector");
14037 SDValue HalfNumElts = getElementCount(
14039 SDValue Lo = getNode(ISD::UMIN, DL, VT, N, HalfNumElts);
14040 SDValue Hi = getNode(ISD::USUBSAT, DL, VT, N, HalfNumElts);
14041 return std::make_pair(Lo, Hi);
14042}
14043
14044/// Widen the vector up to the next power of two using INSERT_SUBVECTOR.
14046 EVT VT = N.getValueType();
14049 return getInsertSubvector(DL, getUNDEF(WideVT), N, 0);
14050}
14051
14054 unsigned Start, unsigned Count,
14055 EVT EltVT) {
14056 EVT VT = Op.getValueType();
14057 if (Count == 0)
14059 if (EltVT == EVT())
14060 EltVT = VT.getVectorElementType();
14061 SDLoc SL(Op);
14062 for (unsigned i = Start, e = Start + Count; i != e; ++i) {
14063 Args.push_back(getExtractVectorElt(SL, EltVT, Op, i));
14064 }
14065}
14066
14067// getAddressSpace - Return the address space this GlobalAddress belongs to.
14069 return getGlobal()->getType()->getAddressSpace();
14070}
14071
14074 return Val.MachineCPVal->getType();
14075 return Val.ConstVal->getType();
14076}
14077
14078bool BuildVectorSDNode::isConstantSplat(APInt &SplatValue, APInt &SplatUndef,
14079 unsigned &SplatBitSize,
14080 bool &HasAnyUndefs,
14081 unsigned MinSplatBits,
14082 bool IsBigEndian) const {
14083 EVT VT = getValueType(0);
14084 assert(VT.isVector() && "Expected a vector type");
14085 unsigned VecWidth = VT.getSizeInBits();
14086 if (MinSplatBits > VecWidth)
14087 return false;
14088
14089 // FIXME: The widths are based on this node's type, but build vectors can
14090 // truncate their operands.
14091 SplatValue = APInt(VecWidth, 0);
14092 SplatUndef = APInt(VecWidth, 0);
14093
14094 // Get the bits. Bits with undefined values (when the corresponding element
14095 // of the vector is an ISD::UNDEF value) are set in SplatUndef and cleared
14096 // in SplatValue. If any of the values are not constant, give up and return
14097 // false.
14098 unsigned int NumOps = getNumOperands();
14099 assert(NumOps > 0 && "isConstantSplat has 0-size build vector");
14100 unsigned EltWidth = VT.getScalarSizeInBits();
14101
14102 for (unsigned j = 0; j < NumOps; ++j) {
14103 unsigned i = IsBigEndian ? NumOps - 1 - j : j;
14104 SDValue OpVal = getOperand(i);
14105 unsigned BitPos = j * EltWidth;
14106
14107 if (OpVal.isUndef())
14108 SplatUndef.setBits(BitPos, BitPos + EltWidth);
14109 else if (auto *CN = dyn_cast<ConstantSDNode>(OpVal))
14110 SplatValue.insertBits(CN->getAPIntValue().zextOrTrunc(EltWidth), BitPos);
14111 else if (auto *CN = dyn_cast<ConstantFPSDNode>(OpVal))
14112 SplatValue.insertBits(CN->getValueAPF().bitcastToAPInt(), BitPos);
14113 else
14114 return false;
14115 }
14116
14117 // The build_vector is all constants or undefs. Find the smallest element
14118 // size that splats the vector.
14119 HasAnyUndefs = (SplatUndef != 0);
14120
14121 // FIXME: This does not work for vectors with elements less than 8 bits.
14122 while (VecWidth > 8) {
14123 // If we can't split in half, stop here.
14124 if (VecWidth & 1)
14125 break;
14126
14127 unsigned HalfSize = VecWidth / 2;
14128 APInt HighValue = SplatValue.extractBits(HalfSize, HalfSize);
14129 APInt LowValue = SplatValue.extractBits(HalfSize, 0);
14130 APInt HighUndef = SplatUndef.extractBits(HalfSize, HalfSize);
14131 APInt LowUndef = SplatUndef.extractBits(HalfSize, 0);
14132
14133 // If the two halves do not match (ignoring undef bits), stop here.
14134 if ((HighValue & ~LowUndef) != (LowValue & ~HighUndef) ||
14135 MinSplatBits > HalfSize)
14136 break;
14137
14138 SplatValue = HighValue | LowValue;
14139 SplatUndef = HighUndef & LowUndef;
14140
14141 VecWidth = HalfSize;
14142 }
14143
14144 // FIXME: The loop above only tries to split in halves. But if the input
14145 // vector for example is <3 x i16> it wouldn't be able to detect a
14146 // SplatBitSize of 16. No idea if that is a design flaw currently limiting
14147 // optimizations. I guess that back in the days when this helper was created
14148 // vectors normally was power-of-2 sized.
14149
14150 SplatBitSize = VecWidth;
14151 return true;
14152}
14153
14155 BitVector *UndefElements) const {
14156 unsigned NumOps = getNumOperands();
14157 if (UndefElements) {
14158 UndefElements->clear();
14159 UndefElements->resize(NumOps);
14160 }
14161 assert(NumOps == DemandedElts.getBitWidth() && "Unexpected vector size");
14162 if (!DemandedElts)
14163 return SDValue();
14164 SDValue Splatted;
14165 for (unsigned i = 0; i != NumOps; ++i) {
14166 if (!DemandedElts[i])
14167 continue;
14168 SDValue Op = getOperand(i);
14169 if (Op.isUndef()) {
14170 if (UndefElements)
14171 (*UndefElements)[i] = true;
14172 } else if (!Splatted) {
14173 Splatted = Op;
14174 } else if (Splatted != Op) {
14175 return SDValue();
14176 }
14177 }
14178
14179 if (!Splatted) {
14180 unsigned FirstDemandedIdx = DemandedElts.countr_zero();
14181 assert(getOperand(FirstDemandedIdx).isUndef() &&
14182 "Can only have a splat without a constant for all undefs.");
14183 return getOperand(FirstDemandedIdx);
14184 }
14185
14186 return Splatted;
14187}
14188
14190 APInt DemandedElts = APInt::getAllOnes(getNumOperands());
14191 return getSplatValue(DemandedElts, UndefElements);
14192}
14193
14195 SmallVectorImpl<SDValue> &Sequence,
14196 BitVector *UndefElements) const {
14197 unsigned NumOps = getNumOperands();
14198 Sequence.clear();
14199 if (UndefElements) {
14200 UndefElements->clear();
14201 UndefElements->resize(NumOps);
14202 }
14203 assert(NumOps == DemandedElts.getBitWidth() && "Unexpected vector size");
14204 if (!DemandedElts || NumOps < 2 || !isPowerOf2_32(NumOps))
14205 return false;
14206
14207 // Set the undefs even if we don't find a sequence (like getSplatValue).
14208 if (UndefElements)
14209 for (unsigned I = 0; I != NumOps; ++I)
14210 if (DemandedElts[I] && getOperand(I).isUndef())
14211 (*UndefElements)[I] = true;
14212
14213 // Iteratively widen the sequence length looking for repetitions.
14214 for (unsigned SeqLen = 1; SeqLen < NumOps; SeqLen *= 2) {
14215 Sequence.append(SeqLen, SDValue());
14216 for (unsigned I = 0; I != NumOps; ++I) {
14217 if (!DemandedElts[I])
14218 continue;
14219 SDValue &SeqOp = Sequence[I % SeqLen];
14221 if (Op.isUndef()) {
14222 if (!SeqOp)
14223 SeqOp = Op;
14224 continue;
14225 }
14226 if (SeqOp && !SeqOp.isUndef() && SeqOp != Op) {
14227 Sequence.clear();
14228 break;
14229 }
14230 SeqOp = Op;
14231 }
14232 if (!Sequence.empty())
14233 return true;
14234 }
14235
14236 assert(Sequence.empty() && "Failed to empty non-repeating sequence pattern");
14237 return false;
14238}
14239
14241 BitVector *UndefElements) const {
14242 APInt DemandedElts = APInt::getAllOnes(getNumOperands());
14243 return getRepeatedSequence(DemandedElts, Sequence, UndefElements);
14244}
14245
14248 BitVector *UndefElements) const {
14250 getSplatValue(DemandedElts, UndefElements));
14251}
14252
14255 return dyn_cast_or_null<ConstantSDNode>(getSplatValue(UndefElements));
14256}
14257
14260 BitVector *UndefElements) const {
14262 getSplatValue(DemandedElts, UndefElements));
14263}
14264
14269
14270int32_t
14272 uint32_t BitWidth) const {
14273 if (ConstantFPSDNode *CN =
14275 bool IsExact;
14276 APSInt IntVal(BitWidth);
14277 const APFloat &APF = CN->getValueAPF();
14278 if (APF.convertToInteger(IntVal, APFloat::rmTowardZero, &IsExact) !=
14279 APFloat::opOK ||
14280 !IsExact)
14281 return -1;
14282
14283 return IntVal.exactLogBase2();
14284 }
14285 return -1;
14286}
14287
14289 bool IsLittleEndian, unsigned DstEltSizeInBits,
14290 SmallVectorImpl<APInt> &RawBitElements, BitVector &UndefElements) const {
14291 // Early-out if this contains anything but Undef/Constant/ConstantFP.
14292 if (!isConstant())
14293 return false;
14294
14295 unsigned NumSrcOps = getNumOperands();
14296 unsigned SrcEltSizeInBits = getValueType(0).getScalarSizeInBits();
14297 assert(((NumSrcOps * SrcEltSizeInBits) % DstEltSizeInBits) == 0 &&
14298 "Invalid bitcast scale");
14299
14300 // Extract raw src bits.
14301 SmallVector<APInt> SrcBitElements(NumSrcOps,
14302 APInt::getZero(SrcEltSizeInBits));
14303 BitVector SrcUndeElements(NumSrcOps, false);
14304
14305 for (unsigned I = 0; I != NumSrcOps; ++I) {
14307 if (Op.isUndef()) {
14308 SrcUndeElements.set(I);
14309 continue;
14310 }
14311 auto *CInt = dyn_cast<ConstantSDNode>(Op);
14312 auto *CFP = dyn_cast<ConstantFPSDNode>(Op);
14313 assert((CInt || CFP) && "Unknown constant");
14314 SrcBitElements[I] = CInt ? CInt->getAPIntValue().trunc(SrcEltSizeInBits)
14315 : CFP->getValueAPF().bitcastToAPInt();
14316 }
14317
14318 // Recast to dst width.
14319 recastRawBits(IsLittleEndian, DstEltSizeInBits, RawBitElements,
14320 SrcBitElements, UndefElements, SrcUndeElements);
14321 return true;
14322}
14323
14324void BuildVectorSDNode::recastRawBits(bool IsLittleEndian,
14325 unsigned DstEltSizeInBits,
14326 SmallVectorImpl<APInt> &DstBitElements,
14327 ArrayRef<APInt> SrcBitElements,
14328 BitVector &DstUndefElements,
14329 const BitVector &SrcUndefElements) {
14330 unsigned NumSrcOps = SrcBitElements.size();
14331 unsigned SrcEltSizeInBits = SrcBitElements[0].getBitWidth();
14332 assert(((NumSrcOps * SrcEltSizeInBits) % DstEltSizeInBits) == 0 &&
14333 "Invalid bitcast scale");
14334 assert(NumSrcOps == SrcUndefElements.size() &&
14335 "Vector size mismatch");
14336
14337 unsigned NumDstOps = (NumSrcOps * SrcEltSizeInBits) / DstEltSizeInBits;
14338 DstUndefElements.clear();
14339 DstUndefElements.resize(NumDstOps, false);
14340 DstBitElements.assign(NumDstOps, APInt::getZero(DstEltSizeInBits));
14341
14342 // Concatenate src elements constant bits together into dst element.
14343 if (SrcEltSizeInBits <= DstEltSizeInBits) {
14344 unsigned Scale = DstEltSizeInBits / SrcEltSizeInBits;
14345 for (unsigned I = 0; I != NumDstOps; ++I) {
14346 DstUndefElements.set(I);
14347 APInt &DstBits = DstBitElements[I];
14348 for (unsigned J = 0; J != Scale; ++J) {
14349 unsigned Idx = (I * Scale) + (IsLittleEndian ? J : (Scale - J - 1));
14350 if (SrcUndefElements[Idx])
14351 continue;
14352 DstUndefElements.reset(I);
14353 const APInt &SrcBits = SrcBitElements[Idx];
14354 assert(SrcBits.getBitWidth() == SrcEltSizeInBits &&
14355 "Illegal constant bitwidths");
14356 DstBits.insertBits(SrcBits, J * SrcEltSizeInBits);
14357 }
14358 }
14359 return;
14360 }
14361
14362 // Split src element constant bits into dst elements.
14363 unsigned Scale = SrcEltSizeInBits / DstEltSizeInBits;
14364 for (unsigned I = 0; I != NumSrcOps; ++I) {
14365 if (SrcUndefElements[I]) {
14366 DstUndefElements.set(I * Scale, (I + 1) * Scale);
14367 continue;
14368 }
14369 const APInt &SrcBits = SrcBitElements[I];
14370 for (unsigned J = 0; J != Scale; ++J) {
14371 unsigned Idx = (I * Scale) + (IsLittleEndian ? J : (Scale - J - 1));
14372 APInt &DstBits = DstBitElements[Idx];
14373 DstBits = SrcBits.extractBits(DstEltSizeInBits, J * DstEltSizeInBits);
14374 }
14375 }
14376}
14377
14379 for (const SDValue &Op : op_values()) {
14380 unsigned Opc = Op.getOpcode();
14381 if (!Op.isUndef() && Opc != ISD::Constant && Opc != ISD::ConstantFP)
14382 return false;
14383 }
14384 return true;
14385}
14386
14387std::optional<std::pair<APInt, APInt>>
14389 unsigned NumOps = getNumOperands();
14390 if (NumOps < 2)
14391 return std::nullopt;
14392
14393 unsigned EltSize = getValueType(0).getScalarSizeInBits();
14394 APInt Start, Stride;
14395 int FirstIdx = -1, SecondIdx = -1;
14396
14397 // Find the first two non-undef constant elements to determine Start and
14398 // Stride, then verify all remaining elements match the sequence.
14399 for (unsigned I = 0; I < NumOps; ++I) {
14401 if (Op->isUndef())
14402 continue;
14403 if (!isa<ConstantSDNode>(Op))
14404 return std::nullopt;
14405
14406 APInt Val = getConstantOperandAPInt(I).trunc(EltSize);
14407 if (FirstIdx < 0) {
14408 FirstIdx = I;
14409 Start = Val;
14410 } else if (SecondIdx < 0) {
14411 SecondIdx = I;
14412 // Compute stride using modular arithmetic. Simple division would handle
14413 // common strides (1, 2, -1, etc.), but modular inverse maximizes matches.
14414 // Example: <0, poison, poison, 0xFF> has stride 0x55 since 3*0x55 = 0xFF
14415 // Note that modular arithmetic is agnostic to signed/unsigned.
14416 unsigned IdxDiff = I - FirstIdx;
14417 APInt ValDiff = Val - Start;
14418
14419 // Step 1: Factor out common powers of 2 from IdxDiff and ValDiff.
14420 unsigned CommonPow2Bits = llvm::countr_zero(IdxDiff);
14421 if (ValDiff.countr_zero() < CommonPow2Bits)
14422 return std::nullopt; // ValDiff not divisible by 2^CommonPow2Bits
14423 IdxDiff >>= CommonPow2Bits;
14424 ValDiff.lshrInPlace(CommonPow2Bits);
14425
14426 // Step 2: IdxDiff is now odd, so its inverse mod 2^EltSize exists.
14427 // TODO: There are 2^CommonPow2Bits valid strides; currently we only try
14428 // one, but we could try all candidates to handle more cases.
14429 Stride = ValDiff * APInt(EltSize, IdxDiff).multiplicativeInverse();
14430 if (Stride.isZero())
14431 return std::nullopt;
14432
14433 // Step 3: Adjust Start based on the first defined element's index.
14434 Start -= Stride * FirstIdx;
14435 } else {
14436 // Verify this element matches the sequence.
14437 if (Val != Start + Stride * I)
14438 return std::nullopt;
14439 }
14440 }
14441
14442 // Need at least two defined elements.
14443 if (SecondIdx < 0)
14444 return std::nullopt;
14445
14446 return std::make_pair(Start, Stride);
14447}
14448
14450 // Find the first non-undef value in the shuffle mask.
14451 unsigned i, e;
14452 for (i = 0, e = Mask.size(); i != e && Mask[i] < 0; ++i)
14453 /* search */;
14454
14455 // If all elements are undefined, this shuffle can be considered a splat
14456 // (although it should eventually get simplified away completely).
14457 if (i == e)
14458 return true;
14459
14460 // Make sure all remaining elements are either undef or the same as the first
14461 // non-undef value.
14462 for (int Idx = Mask[i]; i != e; ++i)
14463 if (Mask[i] >= 0 && Mask[i] != Idx)
14464 return false;
14465 return true;
14466}
14467
14468// Returns true if it is a constant integer BuildVector or constant integer,
14469// possibly hidden by a bitcast.
14471 SDValue N, bool AllowOpaques) const {
14473
14474 if (auto *C = dyn_cast<ConstantSDNode>(N))
14475 return AllowOpaques || !C->isOpaque();
14476
14478 return true;
14479
14480 // Treat a GlobalAddress supporting constant offset folding as a
14481 // constant integer.
14482 if (auto *GA = dyn_cast<GlobalAddressSDNode>(N))
14483 if (GA->getOpcode() == ISD::GlobalAddress &&
14484 TLI->isOffsetFoldingLegal(GA))
14485 return true;
14486
14487 if ((N.getOpcode() == ISD::SPLAT_VECTOR) &&
14488 isa<ConstantSDNode>(N.getOperand(0)))
14489 return true;
14490 return false;
14491}
14492
14493// Returns true if it is a constant float BuildVector or constant float.
14496 return true;
14497
14499 return true;
14500
14501 if ((N.getOpcode() == ISD::SPLAT_VECTOR) &&
14502 isa<ConstantFPSDNode>(N.getOperand(0)))
14503 return true;
14504
14505 return false;
14506}
14507
14508std::optional<bool> SelectionDAG::isBoolConstant(SDValue N) const {
14509 ConstantSDNode *Const =
14510 isConstOrConstSplat(N, false, /*AllowTruncation=*/true);
14511 if (!Const)
14512 return std::nullopt;
14513
14514 EVT VT = N->getValueType(0);
14515 const APInt CVal = Const->getAPIntValue().trunc(VT.getScalarSizeInBits());
14516 switch (TLI->getBooleanContents(N.getValueType())) {
14518 if (CVal.isOne())
14519 return true;
14520 if (CVal.isZero())
14521 return false;
14522 return std::nullopt;
14524 if (CVal.isAllOnes())
14525 return true;
14526 if (CVal.isZero())
14527 return false;
14528 return std::nullopt;
14530 return CVal[0];
14531 }
14532 llvm_unreachable("Unknown BooleanContent enum");
14533}
14534
14535void SelectionDAG::createOperands(SDNode *Node, ArrayRef<SDValue> Vals) {
14536 assert(!Node->OperandList && "Node already has operands");
14538 "too many operands to fit into SDNode");
14539 SDUse *Ops = OperandRecycler.allocate(
14540 ArrayRecycler<SDUse>::Capacity::get(Vals.size()), OperandAllocator);
14541
14542 bool IsDivergent = false;
14543 for (unsigned I = 0; I != Vals.size(); ++I) {
14544 Ops[I].setUser(Node);
14545 Ops[I].setInitial(Vals[I]);
14546 EVT VT = Ops[I].getValueType();
14547
14548 // Skip Chain. It does not carry divergence.
14549 if (VT != MVT::Other &&
14550 (VT != MVT::Glue || gluePropagatesDivergence(Ops[I].getNode())) &&
14551 Ops[I].getNode()->isDivergent()) {
14552 IsDivergent = true;
14553 }
14554 }
14555 Node->NumOperands = Vals.size();
14556 Node->OperandList = Ops;
14557 if (!TLI->isSDNodeAlwaysUniform(Node)) {
14558 IsDivergent |= TLI->isSDNodeSourceOfDivergence(Node, FLI, UA);
14559 Node->SDNodeBits.IsDivergent = IsDivergent;
14560 }
14561 checkForCycles(Node);
14562}
14563
14566 size_t Limit = SDNode::getMaxNumOperands();
14567 while (Vals.size() > Limit) {
14568 unsigned SliceIdx = Vals.size() - Limit;
14569 auto ExtractedTFs = ArrayRef<SDValue>(Vals).slice(SliceIdx, Limit);
14570 SDValue NewTF = getNode(ISD::TokenFactor, DL, MVT::Other, ExtractedTFs);
14571 Vals.erase(Vals.begin() + SliceIdx, Vals.end());
14572 Vals.emplace_back(NewTF);
14573 }
14574 return getNode(ISD::TokenFactor, DL, MVT::Other, Vals);
14575}
14576
14578 EVT VT, SDNodeFlags Flags) {
14579 switch (Opcode) {
14580 default:
14581 return SDValue();
14582 case ISD::ADD:
14583 case ISD::OR:
14584 case ISD::XOR:
14585 case ISD::UMAX:
14586 return getConstant(0, DL, VT);
14587 case ISD::MUL:
14588 return getConstant(1, DL, VT);
14589 case ISD::AND:
14590 case ISD::UMIN:
14591 return getAllOnesConstant(DL, VT);
14592 case ISD::SMAX:
14594 case ISD::SMIN:
14596 case ISD::FADD:
14597 // If flags allow, prefer positive zero since it's generally cheaper
14598 // to materialize on most targets.
14599 return getConstantFP(Flags.hasNoSignedZeros() ? 0.0 : -0.0, DL, VT);
14600 case ISD::FMUL:
14601 return getConstantFP(1.0, DL, VT);
14602 case ISD::FMINNUM:
14603 case ISD::FMAXNUM: {
14604 // Neutral element for fminnum is NaN, Inf or FLT_MAX, depending on FMF.
14605 const fltSemantics &Semantics = VT.getFltSemantics();
14606 APFloat NeutralAF = !Flags.hasNoNaNs() ? APFloat::getQNaN(Semantics) :
14607 !Flags.hasNoInfs() ? APFloat::getInf(Semantics) :
14608 APFloat::getLargest(Semantics);
14609 if (Opcode == ISD::FMAXNUM)
14610 NeutralAF.changeSign();
14611
14612 return getConstantFP(NeutralAF, DL, VT);
14613 }
14614 case ISD::FMINIMUM:
14615 case ISD::FMAXIMUM: {
14616 // Neutral element for fminimum is Inf or FLT_MAX, depending on FMF.
14617 const fltSemantics &Semantics = VT.getFltSemantics();
14618 APFloat NeutralAF = !Flags.hasNoInfs() ? APFloat::getInf(Semantics)
14619 : APFloat::getLargest(Semantics);
14620 if (Opcode == ISD::FMAXIMUM)
14621 NeutralAF.changeSign();
14622
14623 return getConstantFP(NeutralAF, DL, VT);
14624 }
14625
14626 }
14627}
14628
14629/// Helper used to make a call to a library function that has one argument of
14630/// pointer type.
14631///
14632/// Such functions include 'fegetmode', 'fesetenv' and some others, which are
14633/// used to get or set floating-point state. They have one argument of pointer
14634/// type, which points to the memory region containing bits of the
14635/// floating-point state. The value returned by such function is ignored in the
14636/// created call.
14637///
14638/// \param LibFunc Reference to library function (value of RTLIB::Libcall).
14639/// \param Ptr Pointer used to save/load state.
14640/// \param InChain Ingoing token chain.
14641/// \returns Outgoing chain token.
14643 SDValue InChain,
14644 const SDLoc &DLoc) {
14645 assert(InChain.getValueType() == MVT::Other && "Expected token chain");
14647 Args.emplace_back(Ptr, Ptr.getValueType().getTypeForEVT(*getContext()));
14648 RTLIB::LibcallImpl LibcallImpl =
14649 Libcalls->getLibcallImpl(static_cast<RTLIB::Libcall>(LibFunc));
14650 if (LibcallImpl == RTLIB::Unsupported)
14651 reportFatalUsageError("emitting call to unsupported libcall");
14652
14653 SDValue Callee =
14654 getExternalSymbol(LibcallImpl, TLI->getPointerTy(getDataLayout()));
14656 CLI.setDebugLoc(DLoc).setChain(InChain).setLibCallee(
14657 Libcalls->getLibcallImplCallingConv(LibcallImpl),
14658 Type::getVoidTy(*getContext()), Callee, std::move(Args));
14659 return TLI->LowerCallTo(CLI).second;
14660}
14661
14663 assert(From && To && "Invalid SDNode; empty source SDValue?");
14664 auto I = SDEI.find(From);
14665 if (I == SDEI.end())
14666 return;
14667
14668 // Use of operator[] on the DenseMap may cause an insertion, which invalidates
14669 // the iterator, hence the need to make a copy to prevent a use-after-free.
14670 NodeExtraInfo NEI = I->second;
14671 if (LLVM_LIKELY(!NEI.PCSections)) {
14672 // No deep copy required for the types of extra info set.
14673 //
14674 // FIXME: Investigate if other types of extra info also need deep copy. This
14675 // depends on the types of nodes they can be attached to: if some extra info
14676 // is only ever attached to nodes where a replacement To node is always the
14677 // node where later use and propagation of the extra info has the intended
14678 // semantics, no deep copy is required.
14679 SDEI[To] = std::move(NEI);
14680 return;
14681 }
14682
14683 const SDNode *EntrySDN = getEntryNode().getNode();
14684
14685 // We need to copy NodeExtraInfo to all _new_ nodes that are being introduced
14686 // through the replacement of From with To. Otherwise, replacements of a node
14687 // (From) with more complex nodes (To and its operands) may result in lost
14688 // extra info where the root node (To) is insignificant in further propagating
14689 // and using extra info when further lowering to MIR.
14690 //
14691 // In the first step pre-populate the visited set with the nodes reachable
14692 // from the old From node. This avoids copying NodeExtraInfo to parts of the
14693 // DAG that is not new and should be left untouched.
14694 SmallVector<const SDNode *> Leafs{From}; // Leafs reachable with VisitFrom.
14695 DenseSet<const SDNode *> FromReach; // The set of nodes reachable from From.
14696 auto VisitFrom = [&](auto &&Self, const SDNode *N, int MaxDepth) {
14697 if (MaxDepth == 0) {
14698 // Remember this node in case we need to increase MaxDepth and continue
14699 // populating FromReach from this node.
14700 Leafs.emplace_back(N);
14701 return;
14702 }
14703 if (!FromReach.insert(N).second)
14704 return;
14705 for (const SDValue &Op : N->op_values())
14706 Self(Self, Op.getNode(), MaxDepth - 1);
14707 };
14708
14709 // Copy extra info to To and all its transitive operands (that are new).
14711 auto DeepCopyTo = [&](auto &&Self, const SDNode *N) {
14712 if (FromReach.contains(N))
14713 return true;
14714 if (!Visited.insert(N).second)
14715 return true;
14716 if (EntrySDN == N)
14717 return false;
14718 for (const SDValue &Op : N->op_values()) {
14719 if (N == To && Op.getNode() == EntrySDN) {
14720 // Special case: New node's operand is the entry node; just need to
14721 // copy extra info to new node.
14722 break;
14723 }
14724 if (!Self(Self, Op.getNode()))
14725 return false;
14726 }
14727 // Copy only if entry node was not reached.
14728 SDEI[N] = std::move(NEI);
14729 return true;
14730 };
14731
14732 // We first try with a lower MaxDepth, assuming that the path to common
14733 // operands between From and To is relatively short. This significantly
14734 // improves performance in the common case. The initial MaxDepth is big
14735 // enough to avoid retry in the common case; the last MaxDepth is large
14736 // enough to avoid having to use the fallback below (and protects from
14737 // potential stack exhaustion from recursion).
14738 for (int PrevDepth = 0, MaxDepth = 16; MaxDepth <= 1024;
14739 PrevDepth = MaxDepth, MaxDepth *= 2, Visited.clear()) {
14740 // StartFrom is the previous (or initial) set of leafs reachable at the
14741 // previous maximum depth.
14743 std::swap(StartFrom, Leafs);
14744 for (const SDNode *N : StartFrom)
14745 VisitFrom(VisitFrom, N, MaxDepth - PrevDepth);
14746 if (LLVM_LIKELY(DeepCopyTo(DeepCopyTo, To)))
14747 return;
14748 // This should happen very rarely (reached the entry node).
14749 LLVM_DEBUG(dbgs() << __func__ << ": MaxDepth=" << MaxDepth << " too low\n");
14750 assert(!Leafs.empty());
14751 }
14752
14753 // This should not happen - but if it did, that means the subgraph reachable
14754 // from From has depth greater or equal to maximum MaxDepth, and VisitFrom()
14755 // could not visit all reachable common operands. Consequently, we were able
14756 // to reach the entry node.
14757 errs() << "warning: incomplete propagation of SelectionDAG::NodeExtraInfo\n";
14758 assert(false && "From subgraph too complex - increase max. MaxDepth?");
14759 // Best-effort fallback if assertions disabled.
14760 SDEI[To] = std::move(NEI);
14761}
14762
14763#ifndef NDEBUG
14764static void checkForCyclesHelper(const SDNode *N,
14767 const llvm::SelectionDAG *DAG) {
14768 // If this node has already been checked, don't check it again.
14769 if (Checked.count(N))
14770 return;
14771
14772 // If a node has already been visited on this depth-first walk, reject it as
14773 // a cycle.
14774 if (!Visited.insert(N).second) {
14775 errs() << "Detected cycle in SelectionDAG\n";
14776 dbgs() << "Offending node:\n";
14777 N->dumprFull(DAG); dbgs() << "\n";
14778 abort();
14779 }
14780
14781 for (const SDValue &Op : N->op_values())
14782 checkForCyclesHelper(Op.getNode(), Visited, Checked, DAG);
14783
14784 Checked.insert(N);
14785 Visited.erase(N);
14786}
14787#endif
14788
14790 const llvm::SelectionDAG *DAG,
14791 bool force) {
14792#ifndef NDEBUG
14793 bool check = force;
14794#ifdef EXPENSIVE_CHECKS
14795 check = true;
14796#endif // EXPENSIVE_CHECKS
14797 if (check) {
14798 assert(N && "Checking nonexistent SDNode");
14801 checkForCyclesHelper(N, visited, checked, DAG);
14802 }
14803#endif // !NDEBUG
14804}
14805
14806void llvm::checkForCycles(const llvm::SelectionDAG *DAG, bool force) {
14807 checkForCycles(DAG->getRoot().getNode(), DAG, force);
14808}
return SDValue()
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static bool isConstant(const MachineInstr &MI)
constexpr LLT S1
This file declares a class to represent arbitrary precision floating point values and provide a varie...
This file implements a class to represent arbitrary precision integral constant values and operations...
This file implements the APSInt class, which is a simple class that represents an arbitrary sized int...
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
This file implements the BitVector class.
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static std::optional< bool > isBigEndian(const SmallDenseMap< int64_t, int64_t, 8 > &MemOffset2Idx, int64_t LowestIdx)
Given a map from byte offsets in memory to indices in a load/store, determine if that map corresponds...
#define __asan_unpoison_memory_region(p, size)
Definition Compiler.h:592
#define LLVM_LIKELY(EXPR)
Definition Compiler.h:335
This file contains the declarations for the subclasses of Constant, which represent the different fla...
This file defines the DenseSet and SmallDenseSet classes.
This file contains constants used for implementing Dwarf debug support.
This file defines a hash set that can be used to remove duplication of nodes in a graph.
iv users
Definition IVUsers.cpp:48
std::pair< Instruction::BinaryOps, Value * > OffsetOp
Find all possible pairs (BinOp, RHS) that BinOp V, RHS can be simplified.
const size_t AbstractManglingParser< Derived, Alloc >::NumOps
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
static Register getMemsetValue(Register Val, LLT Ty, MachineIRBuilder &MIB)
static bool shouldLowerMemFuncForSize(const MachineFunction &MF)
static bool isZero(Value *V, const DataLayout &DL, DominatorTree *DT, AssumptionCache *AC)
Definition Lint.cpp:539
static Align getPrefTypeAlign(EVT VT, SelectionDAG &DAG)
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
#define G(x, y, z)
Definition MD5.cpp:55
This file declares the MachineConstantPool class which is an abstract constant pool to keep track of ...
Register const TargetRegisterInfo * TRI
This file provides utility analysis objects describing memory locations.
This file contains the declarations for metadata subclasses.
#define T
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
#define P(N)
PowerPC Reduce CR logical Operation
const SmallVectorImpl< MachineOperand > & Cond
Remove Loads Into Fake Uses
static bool isValid(const char C)
Returns true if C is a valid mangled character: <0-9a-zA-Z_>.
Contains matchers for matching SelectionDAG nodes and values.
static Type * getValueType(Value *V)
Returns the type of the given value/instruction V.
This file contains some templates that are useful if you are working with the STL at all.
static uint64_t umul_ov(uint64_t i, uint64_t j, bool &Overflow)
static SDValue getMemcpyLoadsAndStores(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Dst, SDValue Src, uint64_t Size, Align Alignment, bool isVol, bool AlwaysInline, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo, BatchAAResults *BatchAA)
static SDValue getFixedOrScalableQuantity(SelectionDAG &DAG, const SDLoc &DL, EVT VT, Ty Quantity)
static std::pair< SDValue, SDValue > getRuntimeCallSDValueHelper(SDValue Chain, const SDLoc &dl, TargetLowering::ArgListTy &&Args, const CallInst *CI, RTLIB::Libcall Call, SelectionDAG *DAG, const TargetLowering *TLI)
static SDValue getMemsetStores(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Dst, SDValue Src, uint64_t Size, Align Alignment, bool isVol, bool AlwaysInline, MachinePointerInfo DstPtrInfo, const AAMDNodes &AAInfo)
Lower the call to 'memset' intrinsic function into a series of store operations.
static std::optional< APInt > FoldValueWithUndef(unsigned Opcode, const APInt &C1, bool IsUndef1, const APInt &C2, bool IsUndef2)
static SDValue FoldSTEP_VECTOR(const SDLoc &DL, EVT VT, SDValue Step, SelectionDAG &DAG)
static void AddNodeIDNode(FoldingSetNodeID &ID, unsigned OpC, SDVTList VTList, ArrayRef< SDValue > OpList)
static SDValue getMemsetStringVal(EVT VT, const SDLoc &dl, SelectionDAG &DAG, const TargetLowering &TLI, const ConstantDataArraySlice &Slice)
getMemsetStringVal - Similar to getMemsetValue.
static cl::opt< bool > EnableMemCpyDAGOpt("enable-memcpy-dag-opt", cl::Hidden, cl::init(true), cl::desc("Gang up loads and stores generated by inlining of memcpy"))
static bool haveNoCommonBitsSetCommutative(SDValue A, SDValue B)
static void AddNodeIDValueTypes(FoldingSetNodeID &ID, SDVTList VTList)
AddNodeIDValueTypes - Value type lists are intern'd so we can represent them solely with their pointe...
static void commuteShuffle(SDValue &N1, SDValue &N2, MutableArrayRef< int > M)
Swaps the values of N1 and N2.
static bool isMemSrcFromConstant(SDValue Src, ConstantDataArraySlice &Slice)
Returns true if memcpy source is constant data.
static SDValue getMemmoveLoadsAndStores(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Dst, SDValue Src, uint64_t Size, Align Alignment, bool isVol, bool AlwaysInline, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo)
static void AddNodeIDOpcode(FoldingSetNodeID &ID, unsigned OpC)
AddNodeIDOpcode - Add the node opcode to the NodeID data.
static ISD::CondCode getSetCCInverseImpl(ISD::CondCode Op, bool isIntegerLike)
static bool doNotCSE(SDNode *N)
doNotCSE - Return true if CSE should not be performed for this node.
static cl::opt< int > MaxLdStGlue("ldstmemcpy-glue-max", cl::desc("Number limit for gluing ld/st of memcpy."), cl::Hidden, cl::init(0))
static void AddNodeIDOperands(FoldingSetNodeID &ID, ArrayRef< SDValue > Ops)
AddNodeIDOperands - Various routines for adding operands to the NodeID data.
static SDValue foldCONCAT_VECTORS(const SDLoc &DL, EVT VT, ArrayRef< SDValue > Ops, SelectionDAG &DAG)
Try to simplify vector concatenation to an input value, undef, or build vector.
static MachinePointerInfo InferPointerInfo(const MachinePointerInfo &Info, SelectionDAG &DAG, SDValue Ptr, int64_t Offset=0)
InferPointerInfo - If the specified ptr/offset is a frame index, infer a MachinePointerInfo record fr...
static bool isInTailCallPositionWrapper(const CallInst *CI, const SelectionDAG *SelDAG, bool AllowReturnsFirstArg)
static void AddNodeIDCustom(FoldingSetNodeID &ID, const SDNode *N)
If this is an SDNode with special info, add this info to the NodeID data.
static bool gluePropagatesDivergence(const SDNode *Node)
Return true if a glue output should propagate divergence information.
static void NewSDValueDbgMsg(SDValue V, StringRef Msg, SelectionDAG *G)
static SDVTList makeVTList(const EVT *VTs, unsigned NumVTs)
makeVTList - Return an instance of the SDVTList struct initialized with the specified members.
static void checkForCyclesHelper(const SDNode *N, SmallPtrSetImpl< const SDNode * > &Visited, SmallPtrSetImpl< const SDNode * > &Checked, const llvm::SelectionDAG *DAG)
static void chainLoadsAndStoresForMemcpy(SelectionDAG &DAG, const SDLoc &dl, SmallVector< SDValue, 32 > &OutChains, unsigned From, unsigned To, SmallVector< SDValue, 16 > &OutLoadChains, SmallVector< SDValue, 16 > &OutStoreChains)
static int isSignedOp(ISD::CondCode Opcode)
For an integer comparison, return 1 if the comparison is a signed operation and 2 if the result is an...
static std::optional< APInt > FoldValue(unsigned Opcode, const APInt &C1, const APInt &C2)
static SDValue FoldBUILD_VECTOR(const SDLoc &DL, EVT VT, ArrayRef< SDValue > Ops, SelectionDAG &DAG)
static void checkAddrSpaceIsValidForLibcall(const TargetLowering *TLI, unsigned AS)
static cl::opt< unsigned > MaxSteps("has-predecessor-max-steps", cl::Hidden, cl::init(8192), cl::desc("DAG combiner limit number of steps when searching DAG " "for predecessor nodes"))
This file defines the SmallPtrSet class.
This file defines the SmallVector class.
#define LLVM_DEBUG(...)
Definition Debug.h:114
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static TableGen::Emitter::OptClass< SkeletonEmitter > X("gen-skeleton-class", "Generate example skeleton class")
static SymbolRef::Type getType(const Symbol *Sym)
Definition TapiFile.cpp:39
This file describes how to lower LLVM code to machine code.
static void removeOperands(MachineInstr &MI, unsigned i)
static std::optional< unsigned > getOpcode(ArrayRef< VPValue * > Values)
Returns the opcode of Values or ~0 if they do not all agree.
Definition VPlanSLP.cpp:247
static OverflowResult mapOverflowResult(ConstantRange::OverflowResult OR)
Convert ConstantRange OverflowResult into ValueTracking OverflowResult.
static int Lookup(ArrayRef< TableEntry > Table, unsigned Opcode)
static unsigned getSize(unsigned Kind)
static const fltSemantics & IEEEsingle()
Definition APFloat.h:296
cmpResult
IEEE-754R 5.11: Floating Point Comparison Relations.
Definition APFloat.h:334
static constexpr roundingMode rmTowardZero
Definition APFloat.h:348
static const fltSemantics & BFloat()
Definition APFloat.h:295
static const fltSemantics & IEEEquad()
Definition APFloat.h:298
static const fltSemantics & IEEEdouble()
Definition APFloat.h:297
static constexpr roundingMode rmTowardNegative
Definition APFloat.h:347
static constexpr roundingMode rmNearestTiesToEven
Definition APFloat.h:344
static constexpr roundingMode rmTowardPositive
Definition APFloat.h:346
static const fltSemantics & IEEEhalf()
Definition APFloat.h:294
opStatus
IEEE-754R 7: Default exception handling.
Definition APFloat.h:360
static APFloat getQNaN(const fltSemantics &Sem, bool Negative=false, const APInt *payload=nullptr)
Factory for QNaN values.
Definition APFloat.h:1175
opStatus divide(const APFloat &RHS, roundingMode RM)
Definition APFloat.h:1263
void copySign(const APFloat &RHS)
Definition APFloat.h:1357
LLVM_ABI opStatus convert(const fltSemantics &ToSemantics, roundingMode RM, bool *losesInfo)
Definition APFloat.cpp:5976
opStatus subtract(const APFloat &RHS, roundingMode RM)
Definition APFloat.h:1245
bool isExactlyValue(double V) const
We don't rely on operator== working on double values, as it returns true for things that are clearly ...
Definition APFloat.h:1499
opStatus add(const APFloat &RHS, roundingMode RM)
Definition APFloat.h:1236
bool isFinite() const
Definition APFloat.h:1521
opStatus convertFromAPInt(const APInt &Input, bool IsSigned, roundingMode RM)
Definition APFloat.h:1402
opStatus multiply(const APFloat &RHS, roundingMode RM)
Definition APFloat.h:1254
opStatus fusedMultiplyAdd(const APFloat &Multiplicand, const APFloat &Addend, roundingMode RM)
Definition APFloat.h:1290
bool isZero() const
Definition APFloat.h:1512
static APFloat getLargest(const fltSemantics &Sem, bool Negative=false)
Returns the largest finite number in the given semantics.
Definition APFloat.h:1193
opStatus convertToInteger(MutableArrayRef< integerPart > Input, unsigned int Width, bool IsSigned, roundingMode RM, bool *IsExact) const
Definition APFloat.h:1387
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
Definition APFloat.h:1153
opStatus mod(const APFloat &RHS)
Definition APFloat.h:1281
bool isPosZero() const
Definition APFloat.h:1527
bool isNegZero() const
Definition APFloat.h:1528
void changeSign()
Definition APFloat.h:1352
static APFloat getNaN(const fltSemantics &Sem, bool Negative=false, uint64_t payload=0)
Factory for NaN values.
Definition APFloat.h:1164
Class for arbitrary precision integers.
Definition APInt.h:78
LLVM_ABI APInt umul_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1982
LLVM_ABI APInt usub_sat(const APInt &RHS) const
Definition APInt.cpp:2066
LLVM_ABI APInt udiv(const APInt &RHS) const
Unsigned division operation.
Definition APInt.cpp:1584
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
Definition APInt.h:235
void clearBit(unsigned BitPosition)
Set a given bit to 0.
Definition APInt.h:1421
LLVM_ABI APInt zext(unsigned width) const
Zero extend to a new width.
Definition APInt.cpp:1023
static APInt getSignMask(unsigned BitWidth)
Get the SignMask for a specific bit width.
Definition APInt.h:230
uint64_t getZExtValue() const
Get zero extended value.
Definition APInt.h:1555
void setHighBits(unsigned hiBits)
Set the top hiBits bits.
Definition APInt.h:1406
unsigned popcount() const
Count the number of bits set.
Definition APInt.h:1685
void setBitsFrom(unsigned loBit)
Set the top bits starting from loBit.
Definition APInt.h:1400
LLVM_ABI APInt getHiBits(unsigned numBits) const
Compute an APInt containing numBits highbits from this APInt.
Definition APInt.cpp:639
LLVM_ABI APInt zextOrTrunc(unsigned width) const
Zero extend or truncate to width.
Definition APInt.cpp:1044
unsigned getActiveBits() const
Compute the number of active bits in the value.
Definition APInt.h:1527
LLVM_ABI APInt trunc(unsigned width) const
Truncate to new width.
Definition APInt.cpp:936
void setBit(unsigned BitPosition)
Set the given bit to 1 whose position is given as "bitPosition".
Definition APInt.h:1345
APInt abs() const
Get the absolute value.
Definition APInt.h:1810
LLVM_ABI APInt sadd_sat(const APInt &RHS) const
Definition APInt.cpp:2037
bool isAllOnes() const
Determine if all bits are set. This is true for zero-width values.
Definition APInt.h:372
bool ugt(const APInt &RHS) const
Unsigned greater than comparison.
Definition APInt.h:1189
static APInt getBitsSet(unsigned numBits, unsigned loBit, unsigned hiBit)
Get a value with a block of bits set.
Definition APInt.h:259
bool isZero() const
Determine if this value is zero, i.e. all bits are clear.
Definition APInt.h:381
LLVM_ABI APInt urem(const APInt &RHS) const
Unsigned remainder operation.
Definition APInt.cpp:1677
unsigned getBitWidth() const
Return the number of bits in the APInt.
Definition APInt.h:1503
bool ult(const APInt &RHS) const
Unsigned less than comparison.
Definition APInt.h:1118
static APInt getSignedMaxValue(unsigned numBits)
Gets maximum signed value of APInt for a specific bit width.
Definition APInt.h:210
bool isNegative() const
Determine sign of this APInt.
Definition APInt.h:330
LLVM_ABI APInt sdiv(const APInt &RHS) const
Signed division function for APInt.
Definition APInt.cpp:1655
void clearAllBits()
Set every bit to 0.
Definition APInt.h:1411
LLVM_ABI APInt rotr(unsigned rotateAmt) const
Rotate right by rotateAmt.
Definition APInt.cpp:1165
LLVM_ABI APInt reverseBits() const
Definition APInt.cpp:768
void ashrInPlace(unsigned ShiftAmt)
Arithmetic right-shift this APInt by ShiftAmt in place.
Definition APInt.h:841
bool sle(const APInt &RHS) const
Signed less or equal comparison.
Definition APInt.h:1173
unsigned countr_zero() const
Count the number of trailing zero bits.
Definition APInt.h:1654
unsigned getNumSignBits() const
Computes the number of leading bits of this APInt that are equal to its sign bit.
Definition APInt.h:1643
unsigned countl_zero() const
The APInt version of std::countl_zero.
Definition APInt.h:1613
static LLVM_ABI APInt getSplat(unsigned NewLen, const APInt &V)
Return a value containing V broadcasted over NewLen bits.
Definition APInt.cpp:651
static APInt getSignedMinValue(unsigned numBits)
Gets minimum signed value of APInt for a specific bit width.
Definition APInt.h:220
LLVM_ABI APInt sshl_sat(const APInt &RHS) const
Definition APInt.cpp:2097
LLVM_ABI APInt ushl_sat(const APInt &RHS) const
Definition APInt.cpp:2111
LLVM_ABI APInt sextOrTrunc(unsigned width) const
Sign extend or truncate to width.
Definition APInt.cpp:1052
static bool isSameValue(const APInt &I1, const APInt &I2, bool SignedCompare=false)
Determine if two APInts have the same value, after zero-extending or sign-extending (if SignedCompare...
Definition APInt.h:555
LLVM_ABI APInt rotl(unsigned rotateAmt) const
Rotate left by rotateAmt.
Definition APInt.cpp:1152
LLVM_ABI void insertBits(const APInt &SubBits, unsigned bitPosition)
Insert the bits from a smaller APInt starting at bitPosition.
Definition APInt.cpp:397
void clearLowBits(unsigned loBits)
Set bottom loBits bits to 0.
Definition APInt.h:1450
unsigned logBase2() const
Definition APInt.h:1776
LLVM_ABI APInt uadd_sat(const APInt &RHS) const
Definition APInt.cpp:2047
APInt ashr(unsigned ShiftAmt) const
Arithmetic right-shift function.
Definition APInt.h:834
LLVM_ABI APInt multiplicativeInverse() const
Definition APInt.cpp:1285
LLVM_ABI APInt srem(const APInt &RHS) const
Function for signed remainder operation.
Definition APInt.cpp:1747
bool isNonNegative() const
Determine if this APInt Value is non-negative (>= 0)
Definition APInt.h:335
bool ule(const APInt &RHS) const
Unsigned less or equal comparison.
Definition APInt.h:1157
LLVM_ABI APInt sext(unsigned width) const
Sign extend to a new width.
Definition APInt.cpp:996
void setBits(unsigned loBit, unsigned hiBit)
Set the bits from loBit (inclusive) to hiBit (exclusive) to 1.
Definition APInt.h:1382
APInt shl(unsigned shiftAmt) const
Left-shift function.
Definition APInt.h:880
LLVM_ABI APInt byteSwap() const
Definition APInt.cpp:746
bool isSubsetOf(const APInt &RHS) const
This operation checks that all bits set in this APInt are also set in RHS.
Definition APInt.h:1264
bool isPowerOf2() const
Check if this APInt's value is a power of two greater than zero.
Definition APInt.h:441
static APInt getLowBitsSet(unsigned numBits, unsigned loBitsSet)
Constructs an APInt value that has the bottom loBitsSet bits set.
Definition APInt.h:307
void clearBits(unsigned LoBit, unsigned HiBit)
Clear the bits from LoBit (inclusive) to HiBit (exclusive) to 0.
Definition APInt.h:1432
static APInt getZero(unsigned numBits)
Get the '0' value for the specified bit-width.
Definition APInt.h:201
void setLowBits(unsigned loBits)
Set the bottom loBits bits.
Definition APInt.h:1403
LLVM_ABI APInt extractBits(unsigned numBits, unsigned bitPosition) const
Return an APInt with the extracted bits [bitPosition,bitPosition+numBits).
Definition APInt.cpp:482
bool sge(const APInt &RHS) const
Signed greater or equal comparison.
Definition APInt.h:1244
bool isOne() const
Determine if this is a value of 1.
Definition APInt.h:390
static APInt getBitsSetFrom(unsigned numBits, unsigned loBit)
Constructs an APInt value that has a contiguous range of bits set.
Definition APInt.h:287
static APInt getOneBitSet(unsigned numBits, unsigned BitNo)
Return an APInt with exactly one bit set in the result.
Definition APInt.h:240
void lshrInPlace(unsigned ShiftAmt)
Logical right-shift this APInt by ShiftAmt in place.
Definition APInt.h:865
APInt lshr(unsigned shiftAmt) const
Logical right-shift function.
Definition APInt.h:858
bool uge(const APInt &RHS) const
Unsigned greater or equal comparison.
Definition APInt.h:1228
LLVM_ABI APInt ssub_sat(const APInt &RHS) const
Definition APInt.cpp:2056
An arbitrary precision integer that knows its signedness.
Definition APSInt.h:24
unsigned getSrcAddressSpace() const
unsigned getDestAddressSpace() const
static Capacity get(size_t N)
Get the capacity of an array that can hold at least N elements.
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
size_t size() const
size - Get the array size.
Definition ArrayRef.h:142
bool empty() const
empty - Check if the array is empty.
Definition ArrayRef.h:137
This is an SDNode representing atomic operations.
static LLVM_ABI BaseIndexOffset match(const SDNode *N, const SelectionDAG &DAG)
Parses tree in N for base, index, offset addresses.
This class is a wrapper over an AAResults, and it is intended to be used only when there are no IR ch...
bool pointsToConstantMemory(const MemoryLocation &Loc, bool OrLocal=false)
BitVector & reset()
Definition BitVector.h:411
void resize(unsigned N, bool t=false)
resize - Grow or shrink the bitvector.
Definition BitVector.h:360
void clear()
clear - Removes all bits from the bitvector.
Definition BitVector.h:354
BitVector & set()
Definition BitVector.h:370
bool none() const
none - Returns true if none of the bits are set.
Definition BitVector.h:207
size_type size() const
size - Returns the number of bits in this bitvector.
Definition BitVector.h:178
const BlockAddress * getBlockAddress() const
The address of a basic block.
Definition Constants.h:904
BlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate IR basic block frequen...
A "pseudo-class" with methods for operating on BUILD_VECTORs.
LLVM_ABI bool getConstantRawBits(bool IsLittleEndian, unsigned DstEltSizeInBits, SmallVectorImpl< APInt > &RawBitElements, BitVector &UndefElements) const
Extract the raw bit data from a build vector of Undef, Constant or ConstantFP node elements.
static LLVM_ABI void recastRawBits(bool IsLittleEndian, unsigned DstEltSizeInBits, SmallVectorImpl< APInt > &DstBitElements, ArrayRef< APInt > SrcBitElements, BitVector &DstUndefElements, const BitVector &SrcUndefElements)
Recast bit data SrcBitElements to DstEltSizeInBits wide elements.
LLVM_ABI bool getRepeatedSequence(const APInt &DemandedElts, SmallVectorImpl< SDValue > &Sequence, BitVector *UndefElements=nullptr) const
Find the shortest repeating sequence of values in the build vector.
LLVM_ABI ConstantFPSDNode * getConstantFPSplatNode(const APInt &DemandedElts, BitVector *UndefElements=nullptr) const
Returns the demanded splatted constant FP or null if this is not a constant FP splat.
LLVM_ABI SDValue getSplatValue(const APInt &DemandedElts, BitVector *UndefElements=nullptr) const
Returns the demanded splatted value or a null value if this is not a splat.
LLVM_ABI bool isConstantSplat(APInt &SplatValue, APInt &SplatUndef, unsigned &SplatBitSize, bool &HasAnyUndefs, unsigned MinSplatBits=0, bool isBigEndian=false) const
Check if this is a constant splat, and if so, find the smallest element size that splats the vector.
LLVM_ABI ConstantSDNode * getConstantSplatNode(const APInt &DemandedElts, BitVector *UndefElements=nullptr) const
Returns the demanded splatted constant or null if this is not a constant splat.
LLVM_ABI int32_t getConstantFPSplatPow2ToLog2Int(BitVector *UndefElements, uint32_t BitWidth) const
If this is a constant FP splat and the splatted constant FP is an exact power or 2,...
LLVM_ABI std::optional< std::pair< APInt, APInt > > isArithmeticSequence() const
If this BuildVector is constant and represents an arithmetic sequence "<a, a+n, a+2n,...
LLVM_ABI bool isConstant() const
This class represents a function call, abstracting a target machine's calling convention.
bool isTailCall() const
static LLVM_ABI bool isValueValidForType(EVT VT, const APFloat &Val)
const APFloat & getValueAPF() const
bool isExactlyValue(double V) const
We don't rely on operator== working on double values, as it returns true for things that are clearly ...
ConstantFP - Floating Point Values [float, double].
Definition Constants.h:282
const APFloat & getValue() const
Definition Constants.h:326
This is the shared class of boolean and integer constants.
Definition Constants.h:87
unsigned getBitWidth() const
getBitWidth - Return the scalar bitwidth of this constant.
Definition Constants.h:162
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition Constants.h:159
MachineConstantPoolValue * getMachineCPVal() const
const Constant * getConstVal() const
LLVM_ABI Type * getType() const
This class represents a range of values.
LLVM_ABI ConstantRange multiply(const ConstantRange &Other) const
Return a new range representing the possible values resulting from a multiplication of a value in thi...
const APInt * getSingleElement() const
If this set contains a single element, return it, otherwise return null.
static LLVM_ABI ConstantRange fromKnownBits(const KnownBits &Known, bool IsSigned)
Initialize a range based on a known bits constraint.
LLVM_ABI OverflowResult unsignedSubMayOverflow(const ConstantRange &Other) const
Return whether unsigned sub of the two ranges always/never overflows.
LLVM_ABI OverflowResult unsignedAddMayOverflow(const ConstantRange &Other) const
Return whether unsigned add of the two ranges always/never overflows.
LLVM_ABI KnownBits toKnownBits() const
Return known bits for values in this range.
LLVM_ABI ConstantRange zeroExtend(uint32_t BitWidth) const
Return a new range in the specified integer type, which must be strictly larger than the current type...
LLVM_ABI APInt getSignedMin() const
Return the smallest signed value contained in the ConstantRange.
LLVM_ABI OverflowResult unsignedMulMayOverflow(const ConstantRange &Other) const
Return whether unsigned mul of the two ranges always/never overflows.
LLVM_ABI ConstantRange signExtend(uint32_t BitWidth) const
Return a new range in the specified integer type, which must be strictly larger than the current type...
LLVM_ABI bool contains(const APInt &Val) const
Return true if the specified value is in the set.
LLVM_ABI APInt getUnsignedMax() const
Return the largest unsigned value contained in the ConstantRange.
LLVM_ABI APInt getSignedMax() const
Return the largest signed value contained in the ConstantRange.
OverflowResult
Represents whether an operation on the given constant range is known to always or never overflow.
@ AlwaysOverflowsHigh
Always overflows in the direction of signed/unsigned max value.
@ AlwaysOverflowsLow
Always overflows in the direction of signed/unsigned min value.
@ MayOverflow
May or may not overflow.
uint32_t getBitWidth() const
Get the bit width of this ConstantRange.
LLVM_ABI OverflowResult signedSubMayOverflow(const ConstantRange &Other) const
Return whether signed sub of the two ranges always/never overflows.
uint64_t getZExtValue() const
const APInt & getAPIntValue() const
This is an important base class in LLVM.
Definition Constant.h:43
LLVM_ABI Constant * getSplatValue(bool AllowPoison=false) const
If all elements of the vector constant have the same value, return that value.
LLVM_ABI Constant * getAggregateElement(unsigned Elt) const
For aggregates (struct/array/vector) return the constant that corresponds to the specified element if...
DWARF expression.
static LLVM_ABI ExtOps getExtOps(unsigned FromSize, unsigned ToSize, bool Signed)
Returns the ops for a zero- or sign-extension in a DIExpression.
static LLVM_ABI void appendOffset(SmallVectorImpl< uint64_t > &Ops, int64_t Offset)
Append Ops with operations to apply the Offset.
static LLVM_ABI DIExpression * appendOpsToArg(const DIExpression *Expr, ArrayRef< uint64_t > Ops, unsigned ArgNo, bool StackValue=false)
Create a copy of Expr by appending the given list of Ops to each instance of the operand DW_OP_LLVM_a...
static LLVM_ABI const DIExpression * convertToVariadicExpression(const DIExpression *Expr)
If Expr is a non-variadic expression (i.e.
static LLVM_ABI std::optional< DIExpression * > createFragmentExpression(const DIExpression *Expr, unsigned OffsetInBits, unsigned SizeInBits)
Create a DIExpression to describe one part of an aggregate variable that is fragmented across multipl...
Base class for variables.
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
bool isLittleEndian() const
Layout endianness...
Definition DataLayout.h:214
LLVM_ABI IntegerType * getIntPtrType(LLVMContext &C, unsigned AddressSpace=0) const
Returns an integer type with size at least as big as that of a pointer in the given address space.
LLVM_ABI Align getABITypeAlign(Type *Ty) const
Returns the minimum ABI-required alignment for the specified type.
LLVM_ABI unsigned getPointerTypeSizeInBits(Type *) const
The pointer representation size in bits for this type.
LLVM_ABI Align getPrefTypeAlign(Type *Ty) const
Returns the preferred stack/global alignment for the specified type.
A debug info location.
Definition DebugLoc.h:123
Implements a dense probed hash-table based set.
Definition DenseSet.h:279
const char * getSymbol() const
FoldingSetNodeID - This class is used to gather all the unique data bits of a node.
Definition FoldingSet.h:209
Data structure describing the variable locations in a function.
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Definition Function.h:711
AttributeList getAttributes() const
Return the attribute list for this Function.
Definition Function.h:354
LLVM_ABI unsigned getAddressSpace() const
const GlobalValue * getGlobal() const
bool isThreadLocal() const
If the value is "Thread Local", its value isn't shared by the threads.
unsigned getAddressSpace() const
Module * getParent()
Get the module that this global value is contained inside of...
PointerType * getType() const
Global values are always pointers.
This class is used to form a handle around another node that is persistent and is updated across invo...
const SDValue & getValue() const
static LLVM_ABI bool compare(const APInt &LHS, const APInt &RHS, ICmpInst::Predicate Pred)
Return result of LHS Pred RHS comparison.
This is an important class for using LLVM in a threaded context.
Definition LLVMContext.h:68
Tracks which library functions to use for a particular subtarget.
LLVM_ABI CallingConv::ID getLibcallImplCallingConv(RTLIB::LibcallImpl Call) const
Get the CallingConv that should be used for the specified libcall.
LLVM_ABI RTLIB::LibcallImpl getLibcallImpl(RTLIB::Libcall Call) const
Return the lowering's selection of implementation call for Call.
This SDNode is used for LIFETIME_START/LIFETIME_END values.
This class is used to represent ISD::LOAD nodes.
static LocationSize precise(uint64_t Value)
MCSymbol - Instances of this class represent a symbol name in the MC file, and MCSymbols are created ...
Definition MCSymbol.h:42
Metadata node.
Definition Metadata.h:1080
const MDOperand & getOperand(unsigned I) const
Definition Metadata.h:1444
Machine Value Type.
SimpleValueType SimpleTy
static MVT getIntegerVT(unsigned BitWidth)
Abstract base class for all machine specific constantpool value subclasses.
virtual void addSelectionDAGCSEId(FoldingSetNodeID &ID)=0
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
LLVM_ABI int CreateStackObject(uint64_t Size, Align Alignment, bool isSpillSlot, const AllocaInst *Alloca=nullptr, uint8_t ID=0)
Create a new statically sized stack object, returning a nonnegative identifier to represent it.
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
bool isFixedObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to a fixed stack object.
void setObjectAlignment(int ObjectIdx, Align Alignment)
setObjectAlignment - Change the alignment of the specified stack object.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
Function & getFunction()
Return the LLVM function that this machine code represents.
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
A description of a memory reference used in the backend.
const MDNode * getRanges() const
Return the range tag for the memory reference.
Flags
Flags values. These may be or'd together.
@ MOVolatile
The memory access is volatile.
@ MODereferenceable
The memory access is dereferenceable (i.e., doesn't trap).
@ MOLoad
The memory access reads data.
@ MOInvariant
The memory access always returns the same value (or traps).
@ MOStore
The memory access writes data.
const MachinePointerInfo & getPointerInfo() const
Flags getFlags() const
Return the raw flags of the source value,.
This class contains meta information specific to a module.
An SDNode that represents everything that will be needed to construct a MachineInstr.
This class is used to represent an MGATHER node.
This class is used to represent an MLOAD node.
This class is used to represent an MSCATTER node.
This class is used to represent an MSTORE node.
This SDNode is used for target intrinsics that touch memory and need an associated MachineMemOperand.
size_t getNumMemOperands() const
Return the number of memory operands.
LLVM_ABI MemSDNode(unsigned Opc, unsigned Order, const DebugLoc &dl, SDVTList VTs, EVT memvt, PointerUnion< MachineMemOperand *, MachineMemOperand ** > memrefs)
Constructor that supports single or multiple MMOs.
PointerUnion< MachineMemOperand *, MachineMemOperand ** > MemRefs
Memory reference information.
MachineMemOperand * getMemOperand() const
Return the unique MachineMemOperand object describing the memory reference performed by operation.
const MachinePointerInfo & getPointerInfo() const
ArrayRef< MachineMemOperand * > memoperands() const
Return the memory operands for this node.
unsigned getRawSubclassData() const
Return the SubclassData value, without HasDebugValue.
EVT getMemoryVT() const
Return the type of the in-memory value.
Representation for a specific memory location.
A Module instance is used to store all the information related to an LLVM module.
Definition Module.h:67
Function * getFunction(StringRef Name) const
Look up the specified function in the module symbol table.
Definition Module.cpp:235
MutableArrayRef - Represent a mutable reference to an array (0 or more elements consecutively in memo...
Definition ArrayRef.h:298
The optimization diagnostic interface.
Pass interface - Implemented by all 'passes'.
Definition Pass.h:99
Class to represent pointers.
static PointerType * getUnqual(Type *ElementType)
This constructs a pointer to an object of the specified type in the default address space (address sp...
unsigned getAddressSpace() const
Return the address space of the Pointer type.
static LLVM_ABI PointerType * get(Type *ElementType, unsigned AddressSpace)
This constructs a pointer to an object of the specified type in a numbered address space.
A discriminated union of two or more pointer types, with the discriminator in the low bit of the poin...
bool isNull() const
Test if the pointer held in the union is null, regardless of which type it is.
Analysis providing profile information.
void Deallocate(SubClass *E)
Deallocate - Release storage for the pointed-to object.
Wrapper class representing virtual and physical registers.
Definition Register.h:20
Keeps track of dbg_value information through SDISel.
LLVM_ABI void add(SDDbgValue *V, bool isParameter)
LLVM_ABI void erase(const SDNode *Node)
Invalidate all DbgValues attached to the node and remove it from the Node-to-DbgValues map.
Holds the information from a dbg_label node through SDISel.
Holds the information for a single machine location through SDISel; either an SDNode,...
static SDDbgOperand fromNode(SDNode *Node, unsigned ResNo)
static SDDbgOperand fromFrameIdx(unsigned FrameIdx)
static SDDbgOperand fromVReg(Register VReg)
static SDDbgOperand fromConst(const Value *Const)
@ SDNODE
Value is the result of an expression.
Holds the information from a dbg_value node through SDISel.
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
const DebugLoc & getDebugLoc() const
unsigned getIROrder() const
This class provides iterator support for SDUse operands that use a specific SDNode.
Represents one node in the SelectionDAG.
ArrayRef< SDUse > ops() const
const APInt & getAsAPIntVal() const
Helper method returns the APInt value of a ConstantSDNode.
LLVM_ABI void dumprFull(const SelectionDAG *G=nullptr) const
printrFull to dbgs().
unsigned getOpcode() const
Return the SelectionDAG opcode value for this node.
bool isDivergent() const
LLVM_ABI bool isOnlyUserOf(const SDNode *N) const
Return true if this node is the only use of N.
iterator_range< value_op_iterator > op_values() const
unsigned getIROrder() const
Return the node ordering.
static constexpr size_t getMaxNumOperands()
Return the maximum number of operands that a SDNode can hold.
iterator_range< use_iterator > uses()
MemSDNodeBitfields MemSDNodeBits
LLVM_ABI void Profile(FoldingSetNodeID &ID) const
Gather unique data for the node.
bool getHasDebugValue() const
SDNodeFlags getFlags() const
void setNodeId(int Id)
Set unique node id.
LLVM_ABI void intersectFlagsWith(const SDNodeFlags Flags)
Clear any flags in this node that aren't also set in Flags.
static bool hasPredecessorHelper(const SDNode *N, SmallPtrSetImpl< const SDNode * > &Visited, SmallVectorImpl< const SDNode * > &Worklist, unsigned int MaxSteps=0, bool TopologicalPrune=false)
Returns true if N is a predecessor of any node in Worklist.
uint64_t getAsZExtVal() const
Helper method returns the zero-extended integer value of a ConstantSDNode.
bool use_empty() const
Return true if there are no uses of this node.
unsigned getNumValues() const
Return the number of values defined/returned by this operator.
unsigned getNumOperands() const
Return the number of values used by this operation.
const SDValue & getOperand(unsigned Num) const
static LLVM_ABI bool areOnlyUsersOf(ArrayRef< const SDNode * > Nodes, const SDNode *N)
Return true if all the users of N are contained in Nodes.
use_iterator use_begin() const
Provide iteration support to walk over all uses of an SDNode.
LLVM_ABI bool isOperandOf(const SDNode *N) const
Return true if this node is an operand of N.
const APInt & getConstantOperandAPInt(unsigned Num) const
Helper method returns the APInt of a ConstantSDNode operand.
std::optional< APInt > bitcastToAPInt() const
LLVM_ABI bool hasPredecessor(const SDNode *N) const
Return true if N is a predecessor of this node.
LLVM_ABI bool hasAnyUseOfValue(unsigned Value) const
Return true if there are any use of the indicated value.
EVT getValueType(unsigned ResNo) const
Return the type of a specified result.
bool isUndef() const
Returns true if the node type is UNDEF or POISON.
op_iterator op_end() const
op_iterator op_begin() const
static use_iterator use_end()
LLVM_ABI void DropOperands()
Release the operands and set this node to have zero operands.
SDNode(unsigned Opc, unsigned Order, DebugLoc dl, SDVTList VTs)
Create an SDNode.
Represents a use of a SDNode.
SDNode * getUser()
This returns the SDNode that contains this Use.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
bool isUndef() const
SDNode * getNode() const
get the SDNode which holds the desired result
bool hasOneUse() const
Return true if there is exactly one node using value ResNo of Node.
LLVM_ABI bool isOperandOf(const SDNode *N) const
Return true if the referenced return value is an operand of N.
SDValue()=default
LLVM_ABI bool reachesChainWithoutSideEffects(SDValue Dest, unsigned Depth=2) const
Return true if this operand (which must be a chain) reaches the specified operand without crossing an...
SDValue getValue(unsigned R) const
EVT getValueType() const
Return the ValueType of the referenced return value.
TypeSize getValueSizeInBits() const
Returns the size of the value in bits.
const SDValue & getOperand(unsigned i) const
bool use_empty() const
Return true if there are no nodes using value ResNo of Node.
const APInt & getConstantOperandAPInt(unsigned i) const
uint64_t getScalarValueSizeInBits() const
unsigned getResNo() const
get the index which selects a specific result in the SDNode
uint64_t getConstantOperandVal(unsigned i) const
unsigned getOpcode() const
virtual void verifyTargetNode(const SelectionDAG &DAG, const SDNode *N) const
Checks that the given target-specific node is valid. Aborts if it is not.
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
LLVM_ABI SDValue getElementCount(const SDLoc &DL, EVT VT, ElementCount EC)
LLVM_ABI Align getReducedAlign(EVT VT, bool UseABI)
In most cases this function returns the ABI alignment for a given type, except for illegal vector typ...
LLVM_ABI SDValue getVPZeroExtendInReg(SDValue Op, SDValue Mask, SDValue EVL, const SDLoc &DL, EVT VT)
Return the expression required to zero extend the Op value assuming it was the smaller SrcTy value.
LLVM_ABI SDValue getShiftAmountOperand(EVT LHSTy, SDValue Op)
Return the specified value casted to the target's desired shift amount type.
LLVM_ABI SDValue getExtLoad(ISD::LoadExtType ExtType, const SDLoc &dl, EVT VT, SDValue Chain, SDValue Ptr, MachinePointerInfo PtrInfo, EVT MemVT, MaybeAlign Alignment=MaybeAlign(), MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
LLVM_ABI std::pair< SDValue, SDValue > getMemccpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue C, SDValue Size, const CallInst *CI)
Lower a memccpy operation into a target library call and return the resulting chain and call result a...
LLVM_ABI SDValue getExtLoadVP(ISD::LoadExtType ExtType, const SDLoc &dl, EVT VT, SDValue Chain, SDValue Ptr, SDValue Mask, SDValue EVL, MachinePointerInfo PtrInfo, EVT MemVT, MaybeAlign Alignment, MachineMemOperand::Flags MMOFlags, const AAMDNodes &AAInfo, bool IsExpanding=false)
SDValue getExtractVectorElt(const SDLoc &DL, EVT VT, SDValue Vec, unsigned Idx)
Extract element at Idx from Vec.
LLVM_ABI SDValue getSplatSourceVector(SDValue V, int &SplatIndex)
If V is a splatted value, return the source vector and its splat index.
LLVM_ABI SDValue getLabelNode(unsigned Opcode, const SDLoc &dl, SDValue Root, MCSymbol *Label)
LLVM_ABI OverflowKind computeOverflowForUnsignedSub(SDValue N0, SDValue N1) const
Determine if the result of the unsigned sub of 2 nodes can overflow.
LLVM_ABI unsigned ComputeMaxSignificantBits(SDValue Op, unsigned Depth=0) const
Get the upper bound on bit size for this Value Op as a signed integer.
const SDValue & getRoot() const
Return the root tag of the SelectionDAG.
LLVM_ABI std::pair< SDValue, SDValue > getStrlen(SDValue Chain, const SDLoc &dl, SDValue Src, const CallInst *CI)
Lower a strlen operation into a target library call and return the resulting chain and call result as...
LLVM_ABI SDValue getMaskedGather(SDVTList VTs, EVT MemVT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType, ISD::LoadExtType ExtTy)
LLVM_ABI SDValue getAddrSpaceCast(const SDLoc &dl, EVT VT, SDValue Ptr, unsigned SrcAS, unsigned DestAS)
Return an AddrSpaceCastSDNode.
LLVM_ABI SDValue FoldSetCC(EVT VT, SDValue N1, SDValue N2, ISD::CondCode Cond, const SDLoc &dl, SDNodeFlags Flags={})
Constant fold a setcc to true or false.
bool isKnownNeverSNaN(SDValue Op, const APInt &DemandedElts, unsigned Depth=0) const
LLVM_ABI std::optional< bool > isBoolConstant(SDValue N) const
Check if a value \op N is a constant using the target's BooleanContent for its type.
LLVM_ABI SDValue getStackArgumentTokenFactor(SDValue Chain)
Compute a TokenFactor to force all the incoming stack arguments to be loaded from the stack.
const TargetSubtargetInfo & getSubtarget() const
LLVM_ABI SDValue getMergeValues(ArrayRef< SDValue > Ops, const SDLoc &dl)
Create a MERGE_VALUES node from the given operands.
LLVM_ABI SDVTList getVTList(EVT VT)
Return an SDVTList that represents the list of values specified.
LLVM_ABI SDValue getShiftAmountConstant(uint64_t Val, EVT VT, const SDLoc &DL)
LLVM_ABI void updateDivergence(SDNode *N)
LLVM_ABI SDValue getSplatValue(SDValue V, bool LegalTypes=false)
If V is a splat vector, return its scalar source operand by extracting that element from the source v...
LLVM_ABI SDValue getAllOnesConstant(const SDLoc &DL, EVT VT, bool IsTarget=false, bool IsOpaque=false)
LLVM_ABI MachineSDNode * getMachineNode(unsigned Opcode, const SDLoc &dl, EVT VT)
These are used for target selectors to create a new node with specified return type(s),...
LLVM_ABI void ExtractVectorElements(SDValue Op, SmallVectorImpl< SDValue > &Args, unsigned Start=0, unsigned Count=0, EVT EltVT=EVT())
Append the extracted elements from Start to Count out of the vector Op in Args.
LLVM_ABI SDValue getNeutralElement(unsigned Opcode, const SDLoc &DL, EVT VT, SDNodeFlags Flags)
Get the (commutative) neutral element for the given opcode, if it exists.
LLVM_ABI SDValue getAtomicMemset(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Value, SDValue Size, Type *SizeTy, unsigned ElemSz, bool isTailCall, MachinePointerInfo DstPtrInfo)
LLVM_ABI SDValue getAtomicLoad(ISD::LoadExtType ExtType, const SDLoc &dl, EVT MemVT, EVT VT, SDValue Chain, SDValue Ptr, MachineMemOperand *MMO)
LLVM_ABI SDNode * getNodeIfExists(unsigned Opcode, SDVTList VTList, ArrayRef< SDValue > Ops, const SDNodeFlags Flags, bool AllowCommute=false)
Get the specified node if it's already available, or else return NULL.
LLVM_ABI SDValue getPseudoProbeNode(const SDLoc &Dl, SDValue Chain, uint64_t Guid, uint64_t Index, uint32_t Attr)
Creates a PseudoProbeSDNode with function GUID Guid and the index of the block Index it is probing,...
LLVM_ABI SDValue getFreeze(SDValue V)
Return a freeze using the SDLoc of the value operand.
LLVM_ABI SDNode * SelectNodeTo(SDNode *N, unsigned MachineOpc, EVT VT)
These are used for target selectors to mutate the specified node to have the specified return type,...
LLVM_ABI void init(MachineFunction &NewMF, OptimizationRemarkEmitter &NewORE, Pass *PassPtr, const TargetLibraryInfo *LibraryInfo, const LibcallLoweringInfo *LibcallsInfo, UniformityInfo *UA, ProfileSummaryInfo *PSIin, BlockFrequencyInfo *BFIin, MachineModuleInfo &MMI, FunctionVarLocs const *FnVarLocs)
Prepare this SelectionDAG to process code in the given MachineFunction.
LLVM_ABI SelectionDAG(const TargetMachine &TM, CodeGenOptLevel)
LLVM_ABI SDValue getMemset(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Align Alignment, bool isVol, bool AlwaysInline, const CallInst *CI, MachinePointerInfo DstPtrInfo, const AAMDNodes &AAInfo=AAMDNodes())
LLVM_ABI SDValue getBitcastedSExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by first bitcasting (from potentia...
LLVM_ABI SDValue getConstantPool(const Constant *C, EVT VT, MaybeAlign Align=std::nullopt, int Offs=0, bool isT=false, unsigned TargetFlags=0)
LLVM_ABI SDValue getStridedLoadVP(ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, EVT VT, const SDLoc &DL, SDValue Chain, SDValue Ptr, SDValue Offset, SDValue Stride, SDValue Mask, SDValue EVL, EVT MemVT, MachineMemOperand *MMO, bool IsExpanding=false)
LLVM_ABI SDValue getAtomicCmpSwap(unsigned Opcode, const SDLoc &dl, EVT MemVT, SDVTList VTs, SDValue Chain, SDValue Ptr, SDValue Cmp, SDValue Swp, MachineMemOperand *MMO)
Gets a node for an atomic cmpxchg op.
LLVM_ABI SDValue makeEquivalentMemoryOrdering(SDValue OldChain, SDValue NewMemOpChain)
If an existing load has uses of its chain, create a token factor node with that chain and the new mem...
LLVM_ABI bool isConstantIntBuildVectorOrConstantInt(SDValue N, bool AllowOpaques=true) const
Test whether the given value is a constant int or similar node.
LLVM_ABI void ReplaceAllUsesOfValuesWith(const SDValue *From, const SDValue *To, unsigned Num)
Like ReplaceAllUsesOfValueWith, but for multiple values at once.
LLVM_ABI SDValue getJumpTableDebugInfo(int JTI, SDValue Chain, const SDLoc &DL)
LLVM_ABI SDValue getSymbolFunctionGlobalAddress(SDValue Op, Function **TargetFunction=nullptr)
Return a GlobalAddress of the function from the current module with name matching the given ExternalS...
LLVM_ABI std::optional< unsigned > getValidMaximumShiftAmount(SDValue V, const APInt &DemandedElts, unsigned Depth=0) const
If a SHL/SRA/SRL node V has shift amounts that are all less than the element bit-width of the shift n...
LLVM_ABI SDValue UnrollVectorOp(SDNode *N, unsigned ResNE=0)
Utility function used by legalize and lowering to "unroll" a vector operation by splitting out the sc...
LLVM_ABI SDValue getVScale(const SDLoc &DL, EVT VT, APInt MulImm)
Return a node that represents the runtime scaling 'MulImm * RuntimeVL'.
LLVM_ABI SDValue getConstantFP(double Val, const SDLoc &DL, EVT VT, bool isTarget=false)
Create a ConstantFPSDNode wrapping a constant value.
OverflowKind
Used to represent the possible overflow behavior of an operation.
static LLVM_ABI unsigned getHasPredecessorMaxSteps()
LLVM_ABI bool haveNoCommonBitsSet(SDValue A, SDValue B) const
Return true if A and B have no common bits set.
SDValue getExtractSubvector(const SDLoc &DL, EVT VT, SDValue Vec, unsigned Idx)
Return the VT typed sub-vector of Vec at Idx.
LLVM_ABI bool cannotBeOrderedNegativeFP(SDValue Op) const
Test whether the given float value is known to be positive.
LLVM_ABI SDValue getRegister(Register Reg, EVT VT)
LLVM_ABI bool calculateDivergence(SDNode *N)
LLVM_ABI std::pair< SDValue, SDValue > getStrcmp(SDValue Chain, const SDLoc &dl, SDValue S0, SDValue S1, const CallInst *CI)
Lower a strcmp operation into a target library call and return the resulting chain and call result as...
LLVM_ABI SDValue getGetFPEnv(SDValue Chain, const SDLoc &dl, SDValue Ptr, EVT MemVT, MachineMemOperand *MMO)
LLVM_ABI SDValue getAssertAlign(const SDLoc &DL, SDValue V, Align A)
Return an AssertAlignSDNode.
LLVM_ABI SDNode * mutateStrictFPToFP(SDNode *Node)
Mutate the specified strict FP node to its non-strict equivalent, unlinking the node from its chain a...
LLVM_ABI SDValue getLoad(EVT VT, const SDLoc &dl, SDValue Chain, SDValue Ptr, MachinePointerInfo PtrInfo, MaybeAlign Alignment=MaybeAlign(), MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr)
Loads are not normal binary operators: their result type is not determined by their operands,...
LLVM_ABI bool canIgnoreSignBitOfZero(const SDUse &Use) const
Check if a use of a float value is insensitive to signed zeros.
LLVM_ABI bool SignBitIsZeroFP(SDValue Op, unsigned Depth=0) const
Return true if the sign bit of Op is known to be zero, for a floating-point value.
LLVM_ABI SDValue getMemIntrinsicNode(unsigned Opcode, const SDLoc &dl, SDVTList VTList, ArrayRef< SDValue > Ops, EVT MemVT, MachinePointerInfo PtrInfo, Align Alignment, MachineMemOperand::Flags Flags=MachineMemOperand::MOLoad|MachineMemOperand::MOStore, LocationSize Size=LocationSize::precise(0), const AAMDNodes &AAInfo=AAMDNodes())
Creates a MemIntrinsicNode that may produce a result and takes a list of operands.
SDValue getInsertSubvector(const SDLoc &DL, SDValue Vec, SDValue SubVec, unsigned Idx)
Insert SubVec at the Idx element of Vec.
LLVM_ABI SDValue getBitcastedZExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by first bitcasting (from potentia...
LLVM_ABI SDValue getStepVector(const SDLoc &DL, EVT ResVT, const APInt &StepVal)
Returns a vector of type ResVT whose elements contain the linear sequence <0, Step,...
SDValue getSetCC(const SDLoc &DL, EVT VT, SDValue LHS, SDValue RHS, ISD::CondCode Cond, SDValue Chain=SDValue(), bool IsSignaling=false, SDNodeFlags Flags={})
Helper function to make it easier to build SetCC's if you just have an ISD::CondCode instead of an SD...
LLVM_ABI SDValue getAtomic(unsigned Opcode, const SDLoc &dl, EVT MemVT, SDValue Chain, SDValue Ptr, SDValue Val, MachineMemOperand *MMO)
Gets a node for an atomic op, produces result (if relevant) and chain and takes 2 operands.
LLVM_ABI SDValue getMemcpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Align Alignment, bool isVol, bool AlwaysInline, const CallInst *CI, std::optional< bool > OverrideTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo=AAMDNodes(), BatchAAResults *BatchAA=nullptr)
LLVM_ABI Align getEVTAlign(EVT MemoryVT) const
Compute the default alignment value for the given type.
LLVM_ABI bool shouldOptForSize() const
LLVM_ABI SDValue getNOT(const SDLoc &DL, SDValue Val, EVT VT)
Create a bitwise NOT operation as (XOR Val, -1).
LLVM_ABI SDValue getVPZExtOrTrunc(const SDLoc &DL, EVT VT, SDValue Op, SDValue Mask, SDValue EVL)
Convert a vector-predicated Op, which must be an integer vector, to the vector-type VT,...
const TargetLowering & getTargetLoweringInfo() const
LLVM_ABI bool isEqualTo(SDValue A, SDValue B) const
Test whether two SDValues are known to compare equal.
static constexpr unsigned MaxRecursionDepth
LLVM_ABI SDValue getStridedStoreVP(SDValue Chain, const SDLoc &DL, SDValue Val, SDValue Ptr, SDValue Offset, SDValue Stride, SDValue Mask, SDValue EVL, EVT MemVT, MachineMemOperand *MMO, ISD::MemIndexedMode AM, bool IsTruncating=false, bool IsCompressing=false)
bool isGuaranteedNotToBePoison(SDValue Op, unsigned Depth=0) const
Return true if this function can prove that Op is never poison.
LLVM_ABI SDValue expandVACopy(SDNode *Node)
Expand the specified ISD::VACOPY node as the Legalize pass would.
LLVM_ABI SDValue getIndexedMaskedLoad(SDValue OrigLoad, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI void dump(bool Sorted=false) const
Dump the textual format of this DAG.
LLVM_ABI APInt computeVectorKnownZeroElements(SDValue Op, const APInt &DemandedElts, unsigned Depth=0) const
For each demanded element of a vector, see if it is known to be zero.
LLVM_ABI void AddDbgValue(SDDbgValue *DB, bool isParameter)
Add a dbg_value SDNode.
bool NewNodesMustHaveLegalTypes
When true, additional steps are taken to ensure that getConstant() and similar functions return DAG n...
LLVM_ABI std::pair< EVT, EVT > GetSplitDestVTs(const EVT &VT) const
Compute the VTs needed for the low/hi parts of a type which is split (or expanded) into two not neces...
LLVM_ABI void salvageDebugInfo(SDNode &N)
To be invoked on an SDNode that is slated to be erased.
LLVM_ABI SDNode * MorphNodeTo(SDNode *N, unsigned Opc, SDVTList VTs, ArrayRef< SDValue > Ops)
This mutates the specified node to have the specified return type, opcode, and operands.
LLVM_ABI std::pair< SDValue, SDValue > UnrollVectorOverflowOp(SDNode *N, unsigned ResNE=0)
Like UnrollVectorOp(), but for the [US](ADD|SUB|MUL)O family of opcodes.
allnodes_const_iterator allnodes_begin() const
SDValue getUNDEF(EVT VT)
Return an UNDEF node. UNDEF does not have a useful SDLoc.
LLVM_ABI SDValue getGatherVP(SDVTList VTs, EVT VT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType)
SDValue getBuildVector(EVT VT, const SDLoc &DL, ArrayRef< SDValue > Ops)
Return an ISD::BUILD_VECTOR node.
LLVM_ABI SDValue getBitcastedAnyExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by first bitcasting (from potentia...
LLVM_ABI bool isSplatValue(SDValue V, const APInt &DemandedElts, APInt &UndefElts, unsigned Depth=0) const
Test whether V has a splatted value for all the demanded elements.
LLVM_ABI void DeleteNode(SDNode *N)
Remove the specified node from the system.
LLVM_ABI SDValue getBitcast(EVT VT, SDValue V)
Return a bitcast using the SDLoc of the value operand, and casting to the provided type.
LLVM_ABI SDDbgValue * getDbgValueList(DIVariable *Var, DIExpression *Expr, ArrayRef< SDDbgOperand > Locs, ArrayRef< SDNode * > Dependencies, bool IsIndirect, const DebugLoc &DL, unsigned O, bool IsVariadic)
Creates a SDDbgValue node from a list of locations.
LLVM_ABI std::pair< SDValue, SDValue > getStrcpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, const CallInst *CI)
Lower a strcpy operation into a target library call and return the resulting chain and call result as...
SDValue getSelect(const SDLoc &DL, EVT VT, SDValue Cond, SDValue LHS, SDValue RHS, SDNodeFlags Flags=SDNodeFlags())
Helper function to make it easier to build Select's if you just have operands and don't want to check...
LLVM_ABI SDValue getNegative(SDValue Val, const SDLoc &DL, EVT VT)
Create negative operation as (SUB 0, Val).
LLVM_ABI std::optional< unsigned > getValidShiftAmount(SDValue V, const APInt &DemandedElts, unsigned Depth=0) const
If a SHL/SRA/SRL node V has a uniform shift amount that is less than the element bit-width of the shi...
LLVM_ABI void setNodeMemRefs(MachineSDNode *N, ArrayRef< MachineMemOperand * > NewMemRefs)
Mutate the specified machine node's memory references to the provided list.
LLVM_ABI SDValue simplifySelect(SDValue Cond, SDValue TVal, SDValue FVal)
Try to simplify a select/vselect into 1 of its operands or a constant.
LLVM_ABI SDValue getZeroExtendInReg(SDValue Op, const SDLoc &DL, EVT VT)
Return the expression required to zero extend the Op value assuming it was the smaller SrcTy value.
LLVM_ABI bool isConstantFPBuildVectorOrConstantFP(SDValue N) const
Test whether the given value is a constant FP or similar node.
const DataLayout & getDataLayout() const
LLVM_ABI SDValue expandVAArg(SDNode *Node)
Expand the specified ISD::VAARG node as the Legalize pass would.
LLVM_ABI SDValue getTokenFactor(const SDLoc &DL, SmallVectorImpl< SDValue > &Vals)
Creates a new TokenFactor containing Vals.
LLVM_ABI bool doesNodeExist(unsigned Opcode, SDVTList VTList, ArrayRef< SDValue > Ops)
Check if a node exists without modifying its flags.
const SelectionDAGTargetInfo & getSelectionDAGInfo() const
LLVM_ABI bool areNonVolatileConsecutiveLoads(LoadSDNode *LD, LoadSDNode *Base, unsigned Bytes, int Dist) const
Return true if loads are next to each other and can be merged.
LLVM_ABI SDValue getMaskedHistogram(SDVTList VTs, EVT MemVT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType)
LLVM_ABI SDDbgLabel * getDbgLabel(DILabel *Label, const DebugLoc &DL, unsigned O)
Creates a SDDbgLabel node.
LLVM_ABI SDValue getStoreVP(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, SDValue Offset, SDValue Mask, SDValue EVL, EVT MemVT, MachineMemOperand *MMO, ISD::MemIndexedMode AM, bool IsTruncating=false, bool IsCompressing=false)
LLVM_ABI OverflowKind computeOverflowForUnsignedMul(SDValue N0, SDValue N1) const
Determine if the result of the unsigned mul of 2 nodes can overflow.
LLVM_ABI void copyExtraInfo(SDNode *From, SDNode *To)
Copy extra info associated with one node to another.
LLVM_ABI SDValue getConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
Create a ConstantSDNode wrapping a constant value.
LLVM_ABI SDValue getMemBasePlusOffset(SDValue Base, TypeSize Offset, const SDLoc &DL, const SDNodeFlags Flags=SDNodeFlags())
Returns sum of the base pointer and offset.
LLVM_ABI SDValue getGlobalAddress(const GlobalValue *GV, const SDLoc &DL, EVT VT, int64_t offset=0, bool isTargetGA=false, unsigned TargetFlags=0)
LLVM_ABI SDValue getVAArg(EVT VT, const SDLoc &dl, SDValue Chain, SDValue Ptr, SDValue SV, unsigned Align)
VAArg produces a result and token chain, and takes a pointer and a source value as input.
LLVM_ABI SDValue getTruncStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, MachinePointerInfo PtrInfo, EVT SVT, Align Alignment, MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
LLVM_ABI SDValue getLoadFFVP(EVT VT, const SDLoc &DL, SDValue Chain, SDValue Ptr, SDValue Mask, SDValue EVL, MachineMemOperand *MMO)
LLVM_ABI SDValue getTypeSize(const SDLoc &DL, EVT VT, TypeSize TS)
LLVM_ABI SDValue getMDNode(const MDNode *MD)
Return an MDNodeSDNode which holds an MDNode.
LLVM_ABI void clear()
Clear state and free memory necessary to make this SelectionDAG ready to process a new block.
LLVM_ABI std::pair< SDValue, SDValue > getMemcmp(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, const CallInst *CI)
Lower a memcmp operation into a target library call and return the resulting chain and call result as...
LLVM_ABI void ReplaceAllUsesWith(SDValue From, SDValue To)
Modify anything using 'From' to use 'To' instead.
LLVM_ABI SDValue getCommutedVectorShuffle(const ShuffleVectorSDNode &SV)
Returns an ISD::VECTOR_SHUFFLE node semantically equivalent to the shuffle node in input but with swa...
LLVM_ABI std::pair< SDValue, SDValue > SplitVector(const SDValue &N, const SDLoc &DL, const EVT &LoVT, const EVT &HiVT)
Split the vector with EXTRACT_SUBVECTOR using the provided VTs and return the low/high part.
LLVM_ABI SDValue makeStateFunctionCall(unsigned LibFunc, SDValue Ptr, SDValue InChain, const SDLoc &DLoc)
Helper used to make a call to a library function that has one argument of pointer type.
LLVM_ABI bool isGuaranteedNotToBeUndefOrPoison(SDValue Op, bool PoisonOnly=false, unsigned Depth=0) const
Return true if this function can prove that Op is never poison and, if PoisonOnly is false,...
LLVM_ABI SDValue getStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, MachinePointerInfo PtrInfo, Align Alignment, MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
Helper function to build ISD::STORE nodes.
LLVM_ABI SDValue getSignedConstant(int64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
LLVM_ABI SDValue getIndexedLoadVP(SDValue OrigLoad, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI SDValue getSrcValue(const Value *v)
Construct a node to track a Value* through the backend.
SDValue getSplatVector(EVT VT, const SDLoc &DL, SDValue Op)
LLVM_ABI SDValue getAtomicMemcpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Type *SizeTy, unsigned ElemSz, bool isTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo)
LLVM_ABI OverflowKind computeOverflowForSignedMul(SDValue N0, SDValue N1) const
Determine if the result of the signed mul of 2 nodes can overflow.
LLVM_ABI MaybeAlign InferPtrAlign(SDValue Ptr) const
Infer alignment of a load / store address.
LLVM_ABI bool MaskedValueIsAllOnes(SDValue Op, const APInt &Mask, unsigned Depth=0) const
Return true if '(Op & Mask) == Mask'.
LLVM_ABI bool SignBitIsZero(SDValue Op, unsigned Depth=0) const
Return true if the sign bit of Op is known to be zero.
LLVM_ABI void RemoveDeadNodes()
This method deletes all unreachable nodes in the SelectionDAG.
LLVM_ABI void RemoveDeadNode(SDNode *N)
Remove the specified node from the system.
LLVM_ABI void AddDbgLabel(SDDbgLabel *DB)
Add a dbg_label SDNode.
bool isConstantValueOfAnyType(SDValue N) const
LLVM_ABI SDValue getTargetExtractSubreg(int SRIdx, const SDLoc &DL, EVT VT, SDValue Operand)
A convenience function for creating TargetInstrInfo::EXTRACT_SUBREG nodes.
LLVM_ABI SDValue getBasicBlock(MachineBasicBlock *MBB)
LLVM_ABI SDValue getSExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either sign-extending or trunca...
LLVM_ABI SDDbgValue * getVRegDbgValue(DIVariable *Var, DIExpression *Expr, Register VReg, bool IsIndirect, const DebugLoc &DL, unsigned O)
Creates a VReg SDDbgValue node.
LLVM_ABI SDValue getEHLabel(const SDLoc &dl, SDValue Root, MCSymbol *Label)
LLVM_ABI SDValue getIndexedStoreVP(SDValue OrigStore, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI bool isKnownNeverZero(SDValue Op, unsigned Depth=0) const
Test whether the given SDValue is known to contain non-zero value(s).
LLVM_ABI SDValue getIndexedStore(SDValue OrigStore, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI SDValue FoldConstantArithmetic(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDValue > Ops, SDNodeFlags Flags=SDNodeFlags())
LLVM_ABI std::optional< unsigned > getValidMinimumShiftAmount(SDValue V, const APInt &DemandedElts, unsigned Depth=0) const
If a SHL/SRA/SRL node V has shift amounts that are all less than the element bit-width of the shift n...
LLVM_ABI SDValue getSetFPEnv(SDValue Chain, const SDLoc &dl, SDValue Ptr, EVT MemVT, MachineMemOperand *MMO)
LLVM_ABI SDValue getBoolExtOrTrunc(SDValue Op, const SDLoc &SL, EVT VT, EVT OpVT)
Convert Op, which must be of integer type, to the integer type VT, by using an extension appropriate ...
LLVM_ABI SDValue getMaskedStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Base, SDValue Offset, SDValue Mask, EVT MemVT, MachineMemOperand *MMO, ISD::MemIndexedMode AM, bool IsTruncating=false, bool IsCompressing=false)
LLVM_ABI SDValue getExternalSymbol(const char *Sym, EVT VT)
const TargetMachine & getTarget() const
LLVM_ABI std::pair< SDValue, SDValue > getStrictFPExtendOrRound(SDValue Op, SDValue Chain, const SDLoc &DL, EVT VT)
Convert Op, which must be a STRICT operation of float type, to the float type VT, by either extending...
LLVM_ABI std::pair< SDValue, SDValue > SplitEVL(SDValue N, EVT VecVT, const SDLoc &DL)
Split the explicit vector length parameter of a VP operation.
LLVM_ABI SDValue getPtrExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either truncating it or perform...
LLVM_ABI SDValue getVPLogicalNOT(const SDLoc &DL, SDValue Val, SDValue Mask, SDValue EVL, EVT VT)
Create a vector-predicated logical NOT operation as (VP_XOR Val, BooleanOne, Mask,...
LLVM_ABI SDValue getMaskFromElementCount(const SDLoc &DL, EVT VT, ElementCount Len)
Return a vector with the first 'Len' lanes set to true and remaining lanes set to false.
LLVM_ABI SDValue getAnyExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either any-extending or truncat...
iterator_range< allnodes_iterator > allnodes()
LLVM_ABI SDValue getBlockAddress(const BlockAddress *BA, EVT VT, int64_t Offset=0, bool isTarget=false, unsigned TargetFlags=0)
LLVM_ABI SDValue WidenVector(const SDValue &N, const SDLoc &DL)
Widen the vector up to the next power of two using INSERT_SUBVECTOR.
LLVM_ABI bool isKnownNeverZeroFloat(SDValue Op) const
Test whether the given floating point SDValue is known to never be positive or negative zero.
const LibcallLoweringInfo & getLibcalls() const
LLVM_ABI SDValue getLoadVP(ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, EVT VT, const SDLoc &dl, SDValue Chain, SDValue Ptr, SDValue Offset, SDValue Mask, SDValue EVL, MachinePointerInfo PtrInfo, EVT MemVT, Align Alignment, MachineMemOperand::Flags MMOFlags, const AAMDNodes &AAInfo, const MDNode *Ranges=nullptr, bool IsExpanding=false)
LLVM_ABI SDValue getIntPtrConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
LLVM_ABI SDDbgValue * getConstantDbgValue(DIVariable *Var, DIExpression *Expr, const Value *C, const DebugLoc &DL, unsigned O)
Creates a constant SDDbgValue node.
LLVM_ABI SDValue getScatterVP(SDVTList VTs, EVT VT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType)
LLVM_ABI SDValue getValueType(EVT)
LLVM_ABI SDValue getLifetimeNode(bool IsStart, const SDLoc &dl, SDValue Chain, int FrameIndex)
Creates a LifetimeSDNode that starts (IsStart==true) or ends (IsStart==false) the lifetime of the Fra...
ArrayRef< SDDbgValue * > GetDbgValues(const SDNode *SD) const
Get the debug values which reference the given SDNode.
LLVM_ABI SDValue getNode(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDUse > Ops)
Gets or creates the specified node.
LLVM_ABI OverflowKind computeOverflowForSignedAdd(SDValue N0, SDValue N1) const
Determine if the result of the signed addition of 2 nodes can overflow.
LLVM_ABI SDValue getFPExtendOrRound(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of float type, to the float type VT, by either extending or rounding (by tr...
LLVM_ABI unsigned AssignTopologicalOrder()
Topological-sort the AllNodes list and a assign a unique node id for each node in the DAG based on th...
ilist< SDNode >::size_type allnodes_size() const
LLVM_ABI bool isKnownNeverNaN(SDValue Op, const APInt &DemandedElts, bool SNaN=false, unsigned Depth=0) const
Test whether the given SDValue (or all elements of it, if it is a vector) is known to never be NaN in...
LLVM_ABI SDValue FoldConstantBuildVector(BuildVectorSDNode *BV, const SDLoc &DL, EVT DstEltVT)
Fold BUILD_VECTOR of constants/undefs to the destination type BUILD_VECTOR of constants/undefs elemen...
LLVM_ABI SDValue getAtomicMemmove(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Type *SizeTy, unsigned ElemSz, bool isTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo)
LLVM_ABI SDValue getIndexedMaskedStore(SDValue OrigStore, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI SDValue getTruncStoreVP(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, SDValue Mask, SDValue EVL, MachinePointerInfo PtrInfo, EVT SVT, Align Alignment, MachineMemOperand::Flags MMOFlags, const AAMDNodes &AAInfo, bool IsCompressing=false)
SDValue getTargetConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isOpaque=false)
LLVM_ABI unsigned ComputeNumSignBits(SDValue Op, unsigned Depth=0) const
Return the number of times the sign bit of the register is replicated into the other bits.
LLVM_ABI bool MaskedVectorIsZero(SDValue Op, const APInt &DemandedElts, unsigned Depth=0) const
Return true if 'Op' is known to be zero in DemandedElts.
LLVM_ABI SDValue getBoolConstant(bool V, const SDLoc &DL, EVT VT, EVT OpVT)
Create a true or false constant of type VT using the target's BooleanContent for type OpVT.
LLVM_ABI SDDbgValue * getFrameIndexDbgValue(DIVariable *Var, DIExpression *Expr, unsigned FI, bool IsIndirect, const DebugLoc &DL, unsigned O)
Creates a FrameIndex SDDbgValue node.
LLVM_ABI SDValue getExtStridedLoadVP(ISD::LoadExtType ExtType, const SDLoc &DL, EVT VT, SDValue Chain, SDValue Ptr, SDValue Stride, SDValue Mask, SDValue EVL, EVT MemVT, MachineMemOperand *MMO, bool IsExpanding=false)
LLVM_ABI SDValue getMemmove(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Align Alignment, bool isVol, const CallInst *CI, std::optional< bool > OverrideTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo=AAMDNodes(), BatchAAResults *BatchAA=nullptr)
LLVM_ABI SDValue getJumpTable(int JTI, EVT VT, bool isTarget=false, unsigned TargetFlags=0)
LLVM_ABI bool isBaseWithConstantOffset(SDValue Op) const
Return true if the specified operand is an ISD::ADD with a ConstantSDNode on the right-hand side,...
LLVM_ABI SDValue getVPPtrExtOrTrunc(const SDLoc &DL, EVT VT, SDValue Op, SDValue Mask, SDValue EVL)
Convert a vector-predicated Op, which must be of integer type, to the vector-type integer type VT,...
LLVM_ABI SDValue getVectorIdxConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
LLVM_ABI void getTopologicallyOrderedNodes(SmallVectorImpl< const SDNode * > &SortedNodes) const
Get all the nodes in their topological order without modifying any states.
LLVM_ABI void ReplaceAllUsesOfValueWith(SDValue From, SDValue To)
Replace any uses of From with To, leaving uses of other values produced by From.getNode() alone.
MachineFunction & getMachineFunction() const
LLVM_ABI std::pair< SDValue, SDValue > getStrstr(SDValue Chain, const SDLoc &dl, SDValue S0, SDValue S1, const CallInst *CI)
Lower a strstr operation into a target library call and return the resulting chain and call result as...
LLVM_ABI SDValue getPtrExtendInReg(SDValue Op, const SDLoc &DL, EVT VT)
Return the expression required to extend the Op as a pointer value assuming it was the smaller SrcTy ...
LLVM_ABI bool canCreateUndefOrPoison(SDValue Op, const APInt &DemandedElts, bool PoisonOnly=false, bool ConsiderFlags=true, unsigned Depth=0) const
Return true if Op can create undef or poison from non-undef & non-poison operands.
LLVM_ABI OverflowKind computeOverflowForUnsignedAdd(SDValue N0, SDValue N1) const
Determine if the result of the unsigned addition of 2 nodes can overflow.
SDValue getPOISON(EVT VT)
Return a POISON node. POISON does not have a useful SDLoc.
SDValue getSplatBuildVector(EVT VT, const SDLoc &DL, SDValue Op)
Return a splat ISD::BUILD_VECTOR node, consisting of Op splatted to all elements.
LLVM_ABI SDValue getFrameIndex(int FI, EVT VT, bool isTarget=false)
LLVM_ABI SDValue getTruncStridedStoreVP(SDValue Chain, const SDLoc &DL, SDValue Val, SDValue Ptr, SDValue Stride, SDValue Mask, SDValue EVL, EVT SVT, MachineMemOperand *MMO, bool IsCompressing=false)
LLVM_ABI void canonicalizeCommutativeBinop(unsigned Opcode, SDValue &N1, SDValue &N2) const
Swap N1 and N2 if Opcode is a commutative binary opcode and the canonical form expects the opposite o...
LLVM_ABI KnownBits computeKnownBits(SDValue Op, unsigned Depth=0) const
Determine which bits of Op are known to be either zero or one and return them in Known.
LLVM_ABI SDValue getRegisterMask(const uint32_t *RegMask)
LLVM_ABI SDValue getZExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either zero-extending or trunca...
LLVM_ABI SDValue getCondCode(ISD::CondCode Cond)
LLVM_ABI bool MaskedValueIsZero(SDValue Op, const APInt &Mask, unsigned Depth=0) const
Return true if 'Op & Mask' is known to be zero.
LLVM_ABI bool isKnownToBeAPowerOfTwoFP(SDValue Val, unsigned Depth=0) const
Test if the given fp value is known to be an integer power-of-2, either positive or negative.
LLVM_ABI OverflowKind computeOverflowForSignedSub(SDValue N0, SDValue N1) const
Determine if the result of the signed sub of 2 nodes can overflow.
SDValue getObjectPtrOffset(const SDLoc &SL, SDValue Ptr, TypeSize Offset)
Create an add instruction with appropriate flags when used for addressing some offset of an object.
LLVMContext * getContext() const
LLVM_ABI SDValue simplifyFPBinop(unsigned Opcode, SDValue X, SDValue Y, SDNodeFlags Flags)
Try to simplify a floating-point binary operation into 1 of its operands or a constant.
const SDValue & setRoot(SDValue N)
Set the current root tag of the SelectionDAG.
LLVM_ABI bool isKnownToBeAPowerOfTwo(SDValue Val, bool OrZero=false, unsigned Depth=0) const
Test if the given value is known to have exactly one bit set.
LLVM_ABI SDValue getDeactivationSymbol(const GlobalValue *GV)
LLVM_ABI SDValue getTargetExternalSymbol(const char *Sym, EVT VT, unsigned TargetFlags=0)
LLVM_ABI SDValue getMCSymbol(MCSymbol *Sym, EVT VT)
LLVM_ABI bool isUndef(unsigned Opcode, ArrayRef< SDValue > Ops)
Return true if the result of this operation is always undefined.
LLVM_ABI SDValue CreateStackTemporary(TypeSize Bytes, Align Alignment)
Create a stack temporary based on the size in bytes and the alignment.
LLVM_ABI SDNode * UpdateNodeOperands(SDNode *N, SDValue Op)
Mutate the specified node in-place to have the specified operands.
LLVM_ABI std::pair< EVT, EVT > GetDependentSplitDestVTs(const EVT &VT, const EVT &EnvVT, bool *HiIsEmpty) const
Compute the VTs needed for the low/hi parts of a type, dependent on an enveloping VT that has been sp...
LLVM_ABI SDValue foldConstantFPMath(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDValue > Ops)
Fold floating-point operations when all operands are constants and/or undefined.
LLVM_ABI std::optional< ConstantRange > getValidShiftAmountRange(SDValue V, const APInt &DemandedElts, unsigned Depth) const
If a SHL/SRA/SRL node V has shift amounts that are all less than the element bit-width of the shift n...
LLVM_ABI SDValue FoldSymbolOffset(unsigned Opcode, EVT VT, const GlobalAddressSDNode *GA, const SDNode *N2)
LLVM_ABI SDValue getIndexedLoad(SDValue OrigLoad, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI SDValue getTargetInsertSubreg(int SRIdx, const SDLoc &DL, EVT VT, SDValue Operand, SDValue Subreg)
A convenience function for creating TargetInstrInfo::INSERT_SUBREG nodes.
SDValue getEntryNode() const
Return the token chain corresponding to the entry of the function.
LLVM_ABI SDDbgValue * getDbgValue(DIVariable *Var, DIExpression *Expr, SDNode *N, unsigned R, bool IsIndirect, const DebugLoc &DL, unsigned O)
Creates a SDDbgValue node.
LLVM_ABI SDValue getMaskedLoad(EVT VT, const SDLoc &dl, SDValue Chain, SDValue Base, SDValue Offset, SDValue Mask, SDValue Src0, EVT MemVT, MachineMemOperand *MMO, ISD::MemIndexedMode AM, ISD::LoadExtType, bool IsExpanding=false)
SDValue getSplat(EVT VT, const SDLoc &DL, SDValue Op)
Returns a node representing a splat of one value into all lanes of the provided vector type.
LLVM_ABI std::pair< SDValue, SDValue > SplitScalar(const SDValue &N, const SDLoc &DL, const EVT &LoVT, const EVT &HiVT)
Split the scalar node with EXTRACT_ELEMENT using the provided VTs and return the low/high part.
LLVM_ABI SDValue matchBinOpReduction(SDNode *Extract, ISD::NodeType &BinOp, ArrayRef< ISD::NodeType > CandidateBinOps, bool AllowPartials=false)
Match a binop + shuffle pyramid that represents a horizontal reduction over the elements of a vector ...
LLVM_ABI bool isADDLike(SDValue Op, bool NoWrap=false) const
Return true if the specified operand is an ISD::OR or ISD::XOR node that can be treated as an ISD::AD...
LLVM_ABI SDValue getVectorShuffle(EVT VT, const SDLoc &dl, SDValue N1, SDValue N2, ArrayRef< int > Mask)
Return an ISD::VECTOR_SHUFFLE node.
LLVM_ABI SDValue simplifyShift(SDValue X, SDValue Y)
Try to simplify a shift into 1 of its operands or a constant.
LLVM_ABI void transferDbgValues(SDValue From, SDValue To, unsigned OffsetInBits=0, unsigned SizeInBits=0, bool InvalidateDbg=true)
Transfer debug values from one node to another, while optionally generating fragment expressions for ...
LLVM_ABI SDValue getLogicalNOT(const SDLoc &DL, SDValue Val, EVT VT)
Create a logical NOT operation as (XOR Val, BooleanOne).
LLVM_ABI SDValue getMaskedScatter(SDVTList VTs, EVT MemVT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType, bool IsTruncating=false)
ilist< SDNode >::iterator allnodes_iterator
This SDNode is used to implement the code generator support for the llvm IR shufflevector instruction...
int getMaskElt(unsigned Idx) const
ArrayRef< int > getMask() const
static void commuteMask(MutableArrayRef< int > Mask)
Change values in a shuffle permute mask assuming the two vector operands have swapped position.
static LLVM_ABI bool isSplatMask(ArrayRef< int > Mask)
A templated base class for SmallPtrSet which provides the typesafe interface that is common across al...
bool erase(PtrType Ptr)
Remove pointer from the set.
size_type count(ConstPtrType Ptr) const
count - Return 1 if the specified pointer is in the set, 0 otherwise.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void assign(size_type NumElts, ValueParamT Elt)
reference emplace_back(ArgTypes &&... Args)
void reserve(size_type N)
iterator erase(const_iterator CI)
void append(ItTy in_start, ItTy in_end)
Add the specified range to the end of the SmallVector.
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
This class is used to represent ISD::STORE nodes.
StringRef - Represent a constant reference to a string, i.e.
Definition StringRef.h:55
constexpr const char * data() const
data - Get a pointer to the start of the string (which may not be null terminated).
Definition StringRef.h:137
Information about stack frame layout on the target.
virtual TargetStackID::Value getStackIDForScalableVectors() const
Returns the StackID that scalable vectors should be associated with.
Align getStackAlign() const
getStackAlignment - This method returns the number of bytes to which the stack pointer must be aligne...
Completely target-dependent object reference.
unsigned getTargetFlags() const
Provides information about what library functions are available for the current target.
virtual bool shouldConvertConstantLoadToIntImm(const APInt &Imm, Type *Ty) const
Return true if it is beneficial to convert a load of a constant to just the constant itself.
const TargetMachine & getTargetMachine() const
virtual bool isZExtFree(Type *FromTy, Type *ToTy) const
Return true if any actual instruction that defines a value of type FromTy implicitly zero-extends the...
unsigned getMaxStoresPerMemcpy(bool OptSize) const
Get maximum # of store operations permitted for llvm.memcpy.
unsigned getMaxStoresPerMemset(bool OptSize) const
Get maximum # of store operations permitted for llvm.memset.
virtual bool shallExtractConstSplatVectorElementToStore(Type *VectorTy, unsigned ElemSizeInBits, unsigned &Index) const
Return true if the target shall perform extract vector element and store given that the vector is kno...
virtual bool isTruncateFree(Type *FromTy, Type *ToTy) const
Return true if it's free to truncate a value of type FromTy to type ToTy.
virtual EVT getTypeToTransformTo(LLVMContext &Context, EVT VT) const
For types supported by the target, this is an identity function.
bool isTypeLegal(EVT VT) const
Return true if the target has native support for the specified value type.
virtual MVT getPointerTy(const DataLayout &DL, uint32_t AS=0) const
Return the pointer type for the given address space, defaults to the pointer type from the data layou...
BooleanContent
Enum that describes how the target represents true/false values.
virtual unsigned getMaxGluedStoresPerMemcpy() const
Get maximum # of store operations to be glued together.
std::vector< ArgListEntry > ArgListTy
unsigned getMaxStoresPerMemmove(bool OptSize) const
Get maximum # of store operations permitted for llvm.memmove.
virtual bool isLegalStoreImmediate(int64_t Value) const
Return true if the specified immediate is legal for the value input of a store instruction.
static ISD::NodeType getExtendForContent(BooleanContent Content)
This class defines information used to lower LLVM code to legal SelectionDAG operators that the targe...
virtual bool findOptimalMemOpLowering(LLVMContext &Context, std::vector< EVT > &MemOps, unsigned Limit, const MemOp &Op, unsigned DstAS, unsigned SrcAS, const AttributeList &FuncAttributes, EVT *LargestVT=nullptr) const
Determines the optimal series of memory ops to replace the memset / memcpy.
std::pair< SDValue, SDValue > LowerCallTo(CallLoweringInfo &CLI) const
This function lowers an abstract call to a function into an actual call.
Primary interface to the complete machine description for the target machine.
virtual bool isNoopAddrSpaceCast(unsigned SrcAS, unsigned DestAS) const
Returns true if a cast between SrcAS and DestAS is a noop.
const Triple & getTargetTriple() const
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
virtual const SelectionDAGTargetInfo * getSelectionDAGInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
virtual const TargetLowering * getTargetLowering() const
bool isOSDarwin() const
Is this a "Darwin" OS (macOS, iOS, tvOS, watchOS, DriverKit, XROS, or bridgeOS).
Definition Triple.h:639
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
Definition Twine.h:82
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Definition TypeSize.h:343
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:45
bool isVectorTy() const
True if this is an instance of VectorType.
Definition Type.h:273
static LLVM_ABI IntegerType * getInt32Ty(LLVMContext &C)
Definition Type.cpp:296
static LLVM_ABI Type * getVoidTy(LLVMContext &C)
Definition Type.cpp:280
static LLVM_ABI IntegerType * getInt8Ty(LLVMContext &C)
Definition Type.cpp:294
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
Definition Type.cpp:197
LLVM_ABI unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
Definition Type.cpp:230
A Use represents the edge between a Value definition and its users.
Definition Use.h:35
LLVM_ABI unsigned getOperandNo() const
Return the operand # of this use in its User.
Definition Use.cpp:35
LLVM_ABI void set(Value *Val)
Definition Value.h:907
User * getUser() const
Returns the User that contains this Use.
Definition Use.h:61
Value * getOperand(unsigned i) const
Definition User.h:207
This class is used to represent an VP_GATHER node.
This class is used to represent a VP_LOAD node.
This class is used to represent an VP_SCATTER node.
This class is used to represent a VP_STORE node.
This class is used to represent an EXPERIMENTAL_VP_STRIDED_LOAD node.
This class is used to represent an EXPERIMENTAL_VP_STRIDED_STORE node.
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:256
std::pair< iterator, bool > insert(const ValueT &V)
Definition DenseSet.h:202
bool contains(const_arg_type_t< ValueT > V) const
Check if the set contains the given element.
Definition DenseSet.h:175
constexpr bool hasKnownScalarFactor(const FixedOrScalableQuantity &RHS) const
Returns true if there exists a value X where RHS.multiplyCoefficientBy(X) will result in a value whos...
Definition TypeSize.h:269
constexpr ScalarTy getFixedValue() const
Definition TypeSize.h:200
static constexpr bool isKnownLE(const FixedOrScalableQuantity &LHS, const FixedOrScalableQuantity &RHS)
Definition TypeSize.h:230
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
Definition TypeSize.h:168
constexpr bool isKnownEven() const
A return value of true indicates we know at compile time that the number of elements (vscale * Min) i...
Definition TypeSize.h:176
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
Definition TypeSize.h:165
constexpr LeafTy divideCoefficientBy(ScalarTy RHS) const
We do not provide the '/' operator here because division for polynomial types does not work in the sa...
Definition TypeSize.h:252
static constexpr bool isKnownGE(const FixedOrScalableQuantity &LHS, const FixedOrScalableQuantity &RHS)
Definition TypeSize.h:237
A raw_ostream that writes to an std::string.
CallInst * Call
Changed
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
LLVM_ABI APInt clmulr(const APInt &LHS, const APInt &RHS)
Perform a reversed carry-less multiply.
Definition APInt.cpp:3212
LLVM_ABI APInt mulhu(const APInt &C1, const APInt &C2)
Performs (2*N)-bit multiplication on zero-extended operands.
Definition APInt.cpp:3142
LLVM_ABI APInt avgCeilU(const APInt &C1, const APInt &C2)
Compute the ceil of the unsigned average of C1 and C2.
Definition APInt.cpp:3129
LLVM_ABI APInt avgFloorU(const APInt &C1, const APInt &C2)
Compute the floor of the unsigned average of C1 and C2.
Definition APInt.cpp:3119
LLVM_ABI APInt fshr(const APInt &Hi, const APInt &Lo, const APInt &Shift)
Perform a funnel shift right.
Definition APInt.cpp:3193
LLVM_ABI APInt mulhs(const APInt &C1, const APInt &C2)
Performs (2*N)-bit multiplication on sign-extended operands.
Definition APInt.cpp:3134
LLVM_ABI APInt clmul(const APInt &LHS, const APInt &RHS)
Perform a carry-less multiply, also known as XOR multiplication, and return low-bits.
Definition APInt.cpp:3202
APInt abds(const APInt &A, const APInt &B)
Determine the absolute difference of two APInts considered to be signed.
Definition APInt.h:2283
LLVM_ABI APInt fshl(const APInt &Hi, const APInt &Lo, const APInt &Shift)
Perform a funnel shift left.
Definition APInt.cpp:3184
LLVM_ABI APInt ScaleBitMask(const APInt &A, unsigned NewBitWidth, bool MatchAllBits=false)
Splat/Merge neighboring bits to widen/narrow the bitmask represented by.
Definition APInt.cpp:3020
LLVM_ABI APInt clmulh(const APInt &LHS, const APInt &RHS)
Perform a carry-less multiply, and return high-bits.
Definition APInt.cpp:3217
APInt abdu(const APInt &A, const APInt &B)
Determine the absolute difference of two APInts considered to be unsigned.
Definition APInt.h:2288
LLVM_ABI APInt avgFloorS(const APInt &C1, const APInt &C2)
Compute the floor of the signed average of C1 and C2.
Definition APInt.cpp:3114
LLVM_ABI APInt avgCeilS(const APInt &C1, const APInt &C2)
Compute the ceil of the signed average of C1 and C2.
Definition APInt.cpp:3124
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
Definition CallingConv.h:24
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
LLVM_ABI CondCode getSetCCInverse(CondCode Operation, bool isIntegerLike)
Return the operation corresponding to !(X op Y), where 'op' is a valid SetCC operation.
ISD namespace - This namespace contains an enum which represents all of the SelectionDAG node types a...
Definition ISDOpcodes.h:24
LLVM_ABI CondCode getSetCCAndOperation(CondCode Op1, CondCode Op2, EVT Type)
Return the result of a logical AND between different comparisons of identical values: ((X op1 Y) & (X...
LLVM_ABI bool isConstantSplatVectorAllOnes(const SDNode *N, bool BuildVectorOnly=false)
Return true if the specified node is a BUILD_VECTOR or SPLAT_VECTOR where all of the elements are ~0 ...
bool isNON_EXTLoad(const SDNode *N)
Returns true if the specified node is a non-extending load.
NodeType
ISD::NodeType enum - This enum defines the target-independent operators for a SelectionDAG.
Definition ISDOpcodes.h:41
@ SETCC
SetCC operator - This evaluates to a true value iff the condition is true.
Definition ISDOpcodes.h:819
@ MERGE_VALUES
MERGE_VALUES - This node takes multiple discrete operands and returns them all as its individual resu...
Definition ISDOpcodes.h:261
@ CTLZ_ZERO_UNDEF
Definition ISDOpcodes.h:788
@ TargetConstantPool
Definition ISDOpcodes.h:189
@ MDNODE_SDNODE
MDNODE_SDNODE - This is a node that holdes an MDNode*, which is used to reference metadata in the IR.
@ STRICT_FSETCC
STRICT_FSETCC/STRICT_FSETCCS - Constrained versions of SETCC, used for floating-point operands only.
Definition ISDOpcodes.h:511
@ PTRADD
PTRADD represents pointer arithmetic semantics, for targets that opt in using shouldPreservePtrArith(...
@ DELETED_NODE
DELETED_NODE - This is an illegal value that is used to catch errors.
Definition ISDOpcodes.h:45
@ POISON
POISON - A poison node.
Definition ISDOpcodes.h:236
@ PARTIAL_REDUCE_SMLA
PARTIAL_REDUCE_[U|S]MLA(Accumulator, Input1, Input2) The partial reduction nodes sign or zero extend ...
@ VECREDUCE_SEQ_FADD
Generic reduction nodes.
@ MLOAD
Masked load and store - consecutive vector load and store operations with additional mask operand tha...
@ FGETSIGN
INT = FGETSIGN(FP) - Return the sign bit of the specified floating point value as an integer 0/1 valu...
Definition ISDOpcodes.h:538
@ SMUL_LOHI
SMUL_LOHI/UMUL_LOHI - Multiply two integers of type iN, producing a signed/unsigned value of type i[2...
Definition ISDOpcodes.h:275
@ INSERT_SUBVECTOR
INSERT_SUBVECTOR(VECTOR1, VECTOR2, IDX) - Returns a vector with VECTOR2 inserted into VECTOR1.
Definition ISDOpcodes.h:600
@ JUMP_TABLE_DEBUG_INFO
JUMP_TABLE_DEBUG_INFO - Jumptable debug info.
@ BSWAP
Byte Swap and Counting operators.
Definition ISDOpcodes.h:779
@ TargetBlockAddress
Definition ISDOpcodes.h:191
@ DEACTIVATION_SYMBOL
Untyped node storing deactivation symbol reference (DeactivationSymbolSDNode).
@ ATOMIC_STORE
OUTCHAIN = ATOMIC_STORE(INCHAIN, val, ptr) This corresponds to "store atomic" instruction.
@ ADDC
Carry-setting nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:294
@ FMAD
FMAD - Perform a * b + c, while getting the same result as the separately rounded operations.
Definition ISDOpcodes.h:522
@ ADD
Simple integer binary arithmetic operators.
Definition ISDOpcodes.h:264
@ LOAD
LOAD and STORE have token chains as their first operand, then the same operands as an LLVM load/store...
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
Definition ISDOpcodes.h:853
@ ATOMIC_LOAD_USUB_COND
@ FMA
FMA - Perform a * b + c with no intermediate rounding step.
Definition ISDOpcodes.h:518
@ FATAN2
FATAN2 - atan2, inspired by libm.
@ INTRINSIC_VOID
OUTCHAIN = INTRINSIC_VOID(INCHAIN, INTRINSICID, arg1, arg2, ...) This node represents a target intrin...
Definition ISDOpcodes.h:220
@ GlobalAddress
Definition ISDOpcodes.h:88
@ ATOMIC_CMP_SWAP_WITH_SUCCESS
Val, Success, OUTCHAIN = ATOMIC_CMP_SWAP_WITH_SUCCESS(INCHAIN, ptr, cmp, swap) N.b.
@ SINT_TO_FP
[SU]INT_TO_FP - These operators convert integers (whose interpreted sign depends on the first letter)...
Definition ISDOpcodes.h:880
@ CONCAT_VECTORS
CONCAT_VECTORS(VECTOR0, VECTOR1, ...) - Given a number of values of vector type with the same length ...
Definition ISDOpcodes.h:584
@ VECREDUCE_FMAX
FMIN/FMAX nodes can have flags, for NaN/NoNaN variants.
@ FADD
Simple binary floating point operators.
Definition ISDOpcodes.h:417
@ VECREDUCE_FMAXIMUM
FMINIMUM/FMAXIMUM nodes propatate NaNs and signed zeroes using the llvm.minimum and llvm....
@ ABS
ABS - Determine the unsigned absolute value of a signed integer value of the same bitwidth.
Definition ISDOpcodes.h:747
@ SIGN_EXTEND_VECTOR_INREG
SIGN_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register sign-extension of the low ...
Definition ISDOpcodes.h:910
@ FP16_TO_FP
FP16_TO_FP, FP_TO_FP16 - These operators are used to perform promotions and truncation for half-preci...
@ FMULADD
FMULADD - Performs a * b + c, with, or without, intermediate rounding.
Definition ISDOpcodes.h:528
@ BITCAST
BITCAST - This operator converts between integer, vector and FP values, as if the value was stored to...
Definition ISDOpcodes.h:993
@ BUILD_PAIR
BUILD_PAIR - This is the opposite of EXTRACT_ELEMENT in some ways.
Definition ISDOpcodes.h:254
@ CLMUL
Carry-less multiplication operations.
Definition ISDOpcodes.h:774
@ FLDEXP
FLDEXP - ldexp, inspired by libm (op0 * 2**op1).
@ BUILTIN_OP_END
BUILTIN_OP_END - This must be the last enum value in this list.
@ GlobalTLSAddress
Definition ISDOpcodes.h:89
@ SRCVALUE
SRCVALUE - This is a node type that holds a Value* that is used to make reference to a value in the L...
@ EH_LABEL
EH_LABEL - Represents a label in mid basic block used to track locations needed for debug and excepti...
@ ATOMIC_LOAD_USUB_SAT
@ PARTIAL_REDUCE_UMLA
@ SIGN_EXTEND
Conversion operators.
Definition ISDOpcodes.h:844
@ AVGCEILS
AVGCEILS/AVGCEILU - Rounding averaging add - Add two integers using an integer of type i[N+2],...
Definition ISDOpcodes.h:715
@ SCALAR_TO_VECTOR
SCALAR_TO_VECTOR(VAL) - This represents the operation of loading a scalar value into element 0 of the...
Definition ISDOpcodes.h:665
@ TargetExternalSymbol
Definition ISDOpcodes.h:190
@ VECREDUCE_FADD
These reductions have relaxed evaluation order semantics, and have a single vector operand.
@ CTTZ_ZERO_UNDEF
Bit counting operators with an undefined result for zero inputs.
Definition ISDOpcodes.h:787
@ TargetJumpTable
Definition ISDOpcodes.h:188
@ TargetIndex
TargetIndex - Like a constant pool entry, but with completely target-dependent semantics.
Definition ISDOpcodes.h:198
@ PARTIAL_REDUCE_FMLA
@ PREFETCH
PREFETCH - This corresponds to a prefetch intrinsic.
@ TRUNCATE_SSAT_U
Definition ISDOpcodes.h:873
@ SETCCCARRY
Like SetCC, ops #0 and #1 are the LHS and RHS operands to compare, but op #2 is a boolean indicating ...
Definition ISDOpcodes.h:827
@ FNEG
Perform various unary floating-point operations inspired by libm.
@ BR_CC
BR_CC - Conditional branch.
@ SSUBO
Same for subtraction.
Definition ISDOpcodes.h:352
@ STEP_VECTOR
STEP_VECTOR(IMM) - Returns a scalable vector whose lanes are comprised of a linear sequence of unsign...
Definition ISDOpcodes.h:691
@ FCANONICALIZE
Returns platform specific canonical encoding of a floating point number.
Definition ISDOpcodes.h:541
@ SSUBSAT
RESULT = [US]SUBSAT(LHS, RHS) - Perform saturation subtraction on 2 integers with the same bit width ...
Definition ISDOpcodes.h:374
@ SELECT
Select(COND, TRUEVAL, FALSEVAL).
Definition ISDOpcodes.h:796
@ ATOMIC_LOAD
Val, OUTCHAIN = ATOMIC_LOAD(INCHAIN, ptr) This corresponds to "load atomic" instruction.
@ UNDEF
UNDEF - An undefined node.
Definition ISDOpcodes.h:233
@ EXTRACT_ELEMENT
EXTRACT_ELEMENT - This is used to get the lower or upper (determined by a Constant,...
Definition ISDOpcodes.h:247
@ SPLAT_VECTOR
SPLAT_VECTOR(VAL) - Returns a vector with the scalar value VAL duplicated in all lanes.
Definition ISDOpcodes.h:672
@ AssertAlign
AssertAlign - These nodes record if a register contains a value that has a known alignment and the tr...
Definition ISDOpcodes.h:69
@ GET_ACTIVE_LANE_MASK
GET_ACTIVE_LANE_MASK - this corrosponds to the llvm.get.active.lane.mask intrinsic.
@ BasicBlock
Various leaf nodes.
Definition ISDOpcodes.h:81
@ CopyFromReg
CopyFromReg - This node indicates that the input value is a virtual or physical register that is defi...
Definition ISDOpcodes.h:230
@ SADDO
RESULT, BOOL = [SU]ADDO(LHS, RHS) - Overflow-aware nodes for addition.
Definition ISDOpcodes.h:348
@ TargetGlobalAddress
TargetGlobalAddress - Like GlobalAddress, but the DAG does no folding or anything else with this node...
Definition ISDOpcodes.h:185
@ ARITH_FENCE
ARITH_FENCE - This corresponds to a arithmetic fence intrinsic.
@ CTLS
Count leading redundant sign bits.
Definition ISDOpcodes.h:792
@ VECREDUCE_ADD
Integer reductions may have a result type larger than the vector element type.
@ MULHU
MULHU/MULHS - Multiply high - Multiply two integers of type iN, producing an unsigned/signed value of...
Definition ISDOpcodes.h:704
@ ATOMIC_LOAD_FMAXIMUM
@ SHL
Shift and rotation operations.
Definition ISDOpcodes.h:765
@ AssertNoFPClass
AssertNoFPClass - These nodes record if a register contains a float value that is known to be not som...
Definition ISDOpcodes.h:78
@ VECTOR_SHUFFLE
VECTOR_SHUFFLE(VEC1, VEC2) - Returns a vector, of the same type as VEC1/VEC2.
Definition ISDOpcodes.h:649
@ EXTRACT_SUBVECTOR
EXTRACT_SUBVECTOR(VECTOR, IDX) - Returns a subvector from VECTOR.
Definition ISDOpcodes.h:614
@ FMINNUM_IEEE
FMINNUM_IEEE/FMAXNUM_IEEE - Perform floating-point minimumNumber or maximumNumber on two values,...
@ EntryToken
EntryToken - This is the marker used to indicate the start of a region.
Definition ISDOpcodes.h:48
@ EXTRACT_VECTOR_ELT
EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR identified by the (potentially...
Definition ISDOpcodes.h:576
@ CopyToReg
CopyToReg - This node has three operands: a chain, a register number to set to this value,...
Definition ISDOpcodes.h:224
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
Definition ISDOpcodes.h:850
@ TargetConstantFP
Definition ISDOpcodes.h:180
@ SELECT_CC
Select with condition operator - This selects between a true value and a false value (ops #2 and #3) ...
Definition ISDOpcodes.h:811
@ VSCALE
VSCALE(IMM) - Returns the runtime scaling factor used to calculate the number of elements within a sc...
@ ATOMIC_CMP_SWAP
Val, OUTCHAIN = ATOMIC_CMP_SWAP(INCHAIN, ptr, cmp, swap) For double-word atomic operations: ValLo,...
@ FMINNUM
FMINNUM/FMAXNUM - Perform floating-point minimum maximum on two values, following IEEE-754 definition...
@ SSHLSAT
RESULT = [US]SHLSAT(LHS, RHS) - Perform saturation left shift.
Definition ISDOpcodes.h:386
@ SMULO
Same for multiplication.
Definition ISDOpcodes.h:356
@ ATOMIC_LOAD_FMINIMUM
@ TargetFrameIndex
Definition ISDOpcodes.h:187
@ VECTOR_SPLICE_LEFT
VECTOR_SPLICE_LEFT(VEC1, VEC2, OFFSET) - Shifts CONCAT_VECTORS(VEC1, VEC2) left by OFFSET elements an...
Definition ISDOpcodes.h:653
@ ANY_EXTEND_VECTOR_INREG
ANY_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register any-extension of the low la...
Definition ISDOpcodes.h:899
@ SIGN_EXTEND_INREG
SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to sign extend a small value in ...
Definition ISDOpcodes.h:888
@ SMIN
[US]{MIN/MAX} - Binary minimum or maximum of signed or unsigned integers.
Definition ISDOpcodes.h:727
@ LIFETIME_START
This corresponds to the llvm.lifetime.
@ FP_EXTEND
X = FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
Definition ISDOpcodes.h:978
@ VSELECT
Select with a vector condition (op #0) and two vector operands (ops #1 and #2), returning a vector re...
Definition ISDOpcodes.h:805
@ UADDO_CARRY
Carry-using nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:328
@ MGATHER
Masked gather and scatter - load and store operations for a vector of random addresses with additiona...
@ HANDLENODE
HANDLENODE node - Used as a handle for various purposes.
@ BF16_TO_FP
BF16_TO_FP, FP_TO_BF16 - These operators are used to perform promotions and truncation for bfloat16.
@ ATOMIC_LOAD_UDEC_WRAP
@ STRICT_FP_ROUND
X = STRICT_FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision ...
Definition ISDOpcodes.h:500
@ FMINIMUM
FMINIMUM/FMAXIMUM - NaN-propagating minimum/maximum that also treat -0.0 as less than 0....
@ FP_TO_SINT
FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
Definition ISDOpcodes.h:926
@ TargetConstant
TargetConstant* - Like Constant*, but the DAG does not do any folding, simplification,...
Definition ISDOpcodes.h:179
@ STRICT_FP_EXTEND
X = STRICT_FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
Definition ISDOpcodes.h:505
@ AND
Bitwise operators - logical and, logical or, logical xor.
Definition ISDOpcodes.h:739
@ INTRINSIC_WO_CHAIN
RESULT = INTRINSIC_WO_CHAIN(INTRINSICID, arg1, arg2, ...) This node represents a target intrinsic fun...
Definition ISDOpcodes.h:205
@ GET_FPENV_MEM
Gets the current floating-point environment.
@ PSEUDO_PROBE
Pseudo probe for AutoFDO, as a place holder in a basic block to improve the sample counts quality.
@ SCMP
[US]CMP - 3-way comparison of signed or unsigned integers.
Definition ISDOpcodes.h:735
@ AVGFLOORS
AVGFLOORS/AVGFLOORU - Averaging add - Add two integers using an integer of type i[N+1],...
Definition ISDOpcodes.h:710
@ VECTOR_SPLICE_RIGHT
VECTOR_SPLICE_RIGHT(VEC1, VEC2, OFFSET) - Shifts CONCAT_VECTORS(VEC1,VEC2) right by OFFSET elements a...
Definition ISDOpcodes.h:657
@ ADDE
Carry-using nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:304
@ SPLAT_VECTOR_PARTS
SPLAT_VECTOR_PARTS(SCALAR1, SCALAR2, ...) - Returns a vector with the scalar values joined together a...
Definition ISDOpcodes.h:681
@ FREEZE
FREEZE - FREEZE(VAL) returns an arbitrary value if VAL is UNDEF (or is evaluated to UNDEF),...
Definition ISDOpcodes.h:241
@ INSERT_VECTOR_ELT
INSERT_VECTOR_ELT(VECTOR, VAL, IDX) - Returns VECTOR with the element at IDX replaced with VAL.
Definition ISDOpcodes.h:565
@ TokenFactor
TokenFactor - This node takes multiple tokens as input and produces a single token result.
Definition ISDOpcodes.h:53
@ ATOMIC_SWAP
Val, OUTCHAIN = ATOMIC_SWAP(INCHAIN, ptr, amt) Val, OUTCHAIN = ATOMIC_LOAD_[OpName](INCHAIN,...
@ ExternalSymbol
Definition ISDOpcodes.h:93
@ FFREXP
FFREXP - frexp, extract fractional and exponent component of a floating-point value.
@ FP_ROUND
X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision of the ...
Definition ISDOpcodes.h:959
@ VECTOR_COMPRESS
VECTOR_COMPRESS(Vec, Mask, Passthru) consecutively place vector elements based on mask e....
Definition ISDOpcodes.h:699
@ ZERO_EXTEND_VECTOR_INREG
ZERO_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register zero-extension of the low ...
Definition ISDOpcodes.h:921
@ ADDRSPACECAST
ADDRSPACECAST - This operator converts between pointers of different address spaces.
Definition ISDOpcodes.h:997
@ EXPERIMENTAL_VECTOR_HISTOGRAM
Experimental vector histogram intrinsic Operands: Input Chain, Inc, Mask, Base, Index,...
@ FP_TO_SINT_SAT
FP_TO_[US]INT_SAT - Convert floating point value in operand 0 to a signed or unsigned scalar integer ...
Definition ISDOpcodes.h:945
@ VECREDUCE_FMINIMUM
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
Definition ISDOpcodes.h:856
@ VAARG
VAARG - VAARG has four operands: an input chain, a pointer, a SRCVALUE, and the alignment.
@ VECREDUCE_SEQ_FMUL
@ SHL_PARTS
SHL_PARTS/SRA_PARTS/SRL_PARTS - These operators are used for expanded integer shift operations.
Definition ISDOpcodes.h:833
@ AssertSext
AssertSext, AssertZext - These nodes record if a register contains a value that has already been zero...
Definition ISDOpcodes.h:62
@ ATOMIC_LOAD_UINC_WRAP
@ FCOPYSIGN
FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.
Definition ISDOpcodes.h:534
@ PARTIAL_REDUCE_SUMLA
@ SADDSAT
RESULT = [US]ADDSAT(LHS, RHS) - Perform saturation addition on 2 integers with the same bit width (W)...
Definition ISDOpcodes.h:365
@ SET_FPENV_MEM
Sets the current floating point environment.
@ FMINIMUMNUM
FMINIMUMNUM/FMAXIMUMNUM - minimumnum/maximumnum that is same with FMINNUM_IEEE and FMAXNUM_IEEE besid...
@ TRUNCATE_SSAT_S
TRUNCATE_[SU]SAT_[SU] - Truncate for saturated operand [SU] located in middle, prefix for SAT means i...
Definition ISDOpcodes.h:871
@ ABDS
ABDS/ABDU - Absolute difference - Return the absolute difference between two numbers interpreted as s...
Definition ISDOpcodes.h:722
@ TRUNCATE_USAT_U
Definition ISDOpcodes.h:875
@ SADDO_CARRY
Carry-using overflow-aware nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:338
@ INTRINSIC_W_CHAIN
RESULT,OUTCHAIN = INTRINSIC_W_CHAIN(INCHAIN, INTRINSICID, arg1, ...) This node represents a target in...
Definition ISDOpcodes.h:213
@ TargetGlobalTLSAddress
Definition ISDOpcodes.h:186
@ BUILD_VECTOR
BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a fixed-width vector with the specified,...
Definition ISDOpcodes.h:556
LLVM_ABI NodeType getOppositeSignednessMinMaxOpcode(unsigned MinMaxOpc)
Given a MinMaxOpc of ISD::(U|S)MIN or ISD::(U|S)MAX, returns the corresponding opcode with the opposi...
LLVM_ABI bool isBuildVectorOfConstantSDNodes(const SDNode *N)
Return true if the specified node is a BUILD_VECTOR node of all ConstantSDNode or undef.
LLVM_ABI NodeType getExtForLoadExtType(bool IsFP, LoadExtType)
bool isZEXTLoad(const SDNode *N)
Returns true if the specified node is a ZEXTLOAD.
bool matchUnaryFpPredicate(SDValue Op, std::function< bool(ConstantFPSDNode *)> Match, bool AllowUndefs=false)
Hook for matching ConstantFPSDNode predicate.
bool isExtOpcode(unsigned Opcode)
LLVM_ABI bool isConstantSplatVectorAllZeros(const SDNode *N, bool BuildVectorOnly=false)
Return true if the specified node is a BUILD_VECTOR or SPLAT_VECTOR where all of the elements are 0 o...
LLVM_ABI bool isVectorShrinkable(const SDNode *N, unsigned NewEltSize, bool Signed)
Returns true if the specified node is a vector where all elements can be truncated to the specified e...
LLVM_ABI bool isVPBinaryOp(unsigned Opcode)
Whether this is a vector-predicated binary operation opcode.
LLVM_ABI CondCode getSetCCInverse(CondCode Operation, EVT Type)
Return the operation corresponding to !(X op Y), where 'op' is a valid SetCC operation.
LLVM_ABI std::optional< unsigned > getBaseOpcodeForVP(unsigned Opcode, bool hasFPExcept)
Translate this VP Opcode to its corresponding non-VP Opcode.
bool isTrueWhenEqual(CondCode Cond)
Return true if the specified condition returns true if the two operands to the condition are equal.
LLVM_ABI std::optional< unsigned > getVPMaskIdx(unsigned Opcode)
The operand position of the vector mask.
unsigned getUnorderedFlavor(CondCode Cond)
This function returns 0 if the condition is always false if an operand is a NaN, 1 if the condition i...
LLVM_ABI std::optional< unsigned > getVPExplicitVectorLengthIdx(unsigned Opcode)
The operand position of the explicit vector length parameter.
bool isEXTLoad(const SDNode *N)
Returns true if the specified node is a EXTLOAD.
LLVM_ABI bool allOperandsUndef(const SDNode *N)
Return true if the node has at least one operand and all operands of the specified node are ISD::UNDE...
LLVM_ABI bool isFreezeUndef(const SDNode *N)
Return true if the specified node is FREEZE(UNDEF).
LLVM_ABI CondCode getSetCCSwappedOperands(CondCode Operation)
Return the operation corresponding to (Y op X) when given the operation for (X op Y).
LLVM_ABI std::optional< unsigned > getVPForBaseOpcode(unsigned Opcode)
Translate this non-VP Opcode to its corresponding VP Opcode.
MemIndexType
MemIndexType enum - This enum defines how to interpret MGATHER/SCATTER's index parameter when calcula...
LLVM_ABI bool isBuildVectorAllZeros(const SDNode *N)
Return true if the specified node is a BUILD_VECTOR where all of the elements are 0 or undef.
bool matchUnaryPredicateImpl(SDValue Op, std::function< bool(ConstNodeType *)> Match, bool AllowUndefs=false, bool AllowTruncation=false)
Attempt to match a unary predicate against a scalar/splat constant or every element of a constant BUI...
LLVM_ABI bool isConstantSplatVector(const SDNode *N, APInt &SplatValue)
Node predicates.
LLVM_ABI NodeType getInverseMinMaxOpcode(unsigned MinMaxOpc)
Given a MinMaxOpc of ISD::(U|S)MIN or ISD::(U|S)MAX, returns ISD::(U|S)MAX and ISD::(U|S)MIN,...
LLVM_ABI bool matchBinaryPredicate(SDValue LHS, SDValue RHS, std::function< bool(ConstantSDNode *, ConstantSDNode *)> Match, bool AllowUndefs=false, bool AllowTypeMismatch=false)
Attempt to match a binary predicate against a pair of scalar/splat constants or every element of a pa...
LLVM_ABI bool isVPReduction(unsigned Opcode)
Whether this is a vector-predicated reduction opcode.
bool matchUnaryPredicate(SDValue Op, std::function< bool(ConstantSDNode *)> Match, bool AllowUndefs=false, bool AllowTruncation=false)
Hook for matching ConstantSDNode predicate.
MemIndexedMode
MemIndexedMode enum - This enum defines the load / store indexed addressing modes.
LLVM_ABI bool isBuildVectorOfConstantFPSDNodes(const SDNode *N)
Return true if the specified node is a BUILD_VECTOR node of all ConstantFPSDNode or undef.
bool isSEXTLoad(const SDNode *N)
Returns true if the specified node is a SEXTLOAD.
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
LLVM_ABI bool isBuildVectorAllOnes(const SDNode *N)
Return true if the specified node is a BUILD_VECTOR where all of the elements are ~0 or undef.
LLVM_ABI NodeType getVecReduceBaseOpcode(unsigned VecReduceOpcode)
Get underlying scalar opcode for VECREDUCE opcode.
LoadExtType
LoadExtType enum - This enum defines the three variants of LOADEXT (load with extension).
LLVM_ABI bool isVPOpcode(unsigned Opcode)
Whether this is a vector-predicated Opcode.
LLVM_ABI CondCode getSetCCOrOperation(CondCode Op1, CondCode Op2, EVT Type)
Return the result of a logical OR between different comparisons of identical values: ((X op1 Y) | (X ...
BinaryOp_match< SpecificConstantMatch, SrcTy, TargetOpcode::G_SUB > m_Neg(const SrcTy &&Src)
Matches a register negated by a G_SUB.
BinaryOp_match< LHS, RHS, Instruction::And > m_And(const LHS &L, const RHS &R)
deferredval_ty< Value > m_Deferred(Value *const &V)
Like m_Specific(), but works if the specific value to match is determined as part of the same match()...
class_match< Value > m_Value()
Match an arbitrary value and ignore it.
LLVM_ABI Libcall getMEMCPY_ELEMENT_UNORDERED_ATOMIC(uint64_t ElementSize)
getMEMCPY_ELEMENT_UNORDERED_ATOMIC - Return MEMCPY_ELEMENT_UNORDERED_ATOMIC_* value for the given ele...
LLVM_ABI Libcall getMEMSET_ELEMENT_UNORDERED_ATOMIC(uint64_t ElementSize)
getMEMSET_ELEMENT_UNORDERED_ATOMIC - Return MEMSET_ELEMENT_UNORDERED_ATOMIC_* value for the given ele...
LLVM_ABI Libcall getMEMMOVE_ELEMENT_UNORDERED_ATOMIC(uint64_t ElementSize)
getMEMMOVE_ELEMENT_UNORDERED_ATOMIC - Return MEMMOVE_ELEMENT_UNORDERED_ATOMIC_* value for the given e...
bool sd_match(SDNode *N, const SelectionDAG *DAG, Pattern &&P)
initializer< Ty > init(const Ty &Val)
@ DW_OP_LLVM_arg
Only used in LLVM metadata.
Definition Dwarf.h:149
std::enable_if_t< detail::IsValidPointer< X, Y >::value, X * > extract(Y &&MD)
Extract a Value from Metadata.
Definition Metadata.h:668
NodeAddr< NodeBase * > Node
Definition RDFGraph.h:381
This is an optimization pass for GlobalISel generic memory operations.
Definition Types.h:26
GenericUniformityInfo< SSAContext > UniformityInfo
unsigned Log2_32_Ceil(uint32_t Value)
Return the ceil log base 2 of the specified value, 32 if the value is zero.
Definition MathExtras.h:344
@ Offset
Definition DWP.cpp:532
bool operator<(int64_t V1, const APSInt &V2)
Definition APSInt.h:360
ISD::CondCode getICmpCondCode(ICmpInst::Predicate Pred)
getICmpCondCode - Return the ISD condition code corresponding to the given LLVM IR integer condition ...
Definition Analysis.cpp:237
void fill(R &&Range, T &&Value)
Provide wrappers to std::fill which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1759
LLVM_ABI SDValue peekThroughExtractSubvectors(SDValue V)
Return the non-extracted vector source operand of V if it exists.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1739
MaybeAlign getAlign(const CallInst &I, unsigned Index)
LLVM_ABI bool isNullConstant(SDValue V)
Returns true if V is a constant integer zero.
LLVM_ABI bool isAllOnesOrAllOnesSplat(const MachineInstr &MI, const MachineRegisterInfo &MRI, bool AllowUndefs=false)
Return true if the value is a constant -1 integer or a splatted vector of a constant -1 integer (with...
Definition Utils.cpp:1608
LLVM_ABI SDValue getBitwiseNotOperand(SDValue V, SDValue Mask, bool AllowUndefs)
If V is a bitwise not, returns the inverted operand.
@ Undef
Value of the register doesn't matter.
LLVM_ABI SDValue peekThroughBitcasts(SDValue V)
Return the non-bitcasted source operand of V if it exists.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
Definition STLExtras.h:2554
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
int countr_one(T Value)
Count the number of ones from the least significant bit to the first zero bit.
Definition bit.h:293
bool isIntOrFPConstant(SDValue V)
Return true if V is either a integer or FP constant.
auto dyn_cast_if_present(const Y &Val)
dyn_cast_if_present<X> - Functionally identical to dyn_cast, except that a null (or none in the case ...
Definition Casting.h:732
LLVM_ABI bool getConstantDataArrayInfo(const Value *V, ConstantDataArraySlice &Slice, unsigned ElementSize, uint64_t Offset=0)
Returns true if the value V is a pointer into a ConstantDataArray.
LLVM_ABI bool isOneOrOneSplatFP(SDValue V, bool AllowUndefs=false)
Return true if the value is a constant floating-point value, or a splatted vector of a constant float...
int bit_width(T Value)
Returns the number of bits needed to represent Value if Value is nonzero.
Definition bit.h:303
LLVM_READONLY APFloat maximum(const APFloat &A, const APFloat &B)
Implements IEEE 754-2019 maximum semantics.
Definition APFloat.h:1710
void append_range(Container &C, Range &&R)
Wrapper function to append range R to container C.
Definition STLExtras.h:2208
constexpr bool isUIntN(unsigned N, uint64_t x)
Checks if an unsigned integer fits into the given (dynamic) bit width.
Definition MathExtras.h:243
LLVM_ABI bool shouldOptimizeForSize(const MachineFunction *MF, ProfileSummaryInfo *PSI, const MachineBlockFrequencyInfo *BFI, PGSOQueryType QueryType=PGSOQueryType::Other)
Returns true if machine function MF is suggested to be size-optimized based on the profile.
iterator_range< early_inc_iterator_impl< detail::IterOfRange< RangeT > > > make_early_inc_range(RangeT &&Range)
Make a range that does early increment to allow mutation of the underlying range without disrupting i...
Definition STLExtras.h:634
auto cast_or_null(const Y &Val)
Definition Casting.h:714
LLVM_ABI bool isNullOrNullSplat(const MachineInstr &MI, const MachineRegisterInfo &MRI, bool AllowUndefs=false)
Return true if the value is a constant 0 integer or a splatted vector of a constant 0 integer (with n...
Definition Utils.cpp:1590
LLVM_ABI bool isMinSignedConstant(SDValue V)
Returns true if V is a constant min signed integer value.
LLVM_ABI ConstantFPSDNode * isConstOrConstSplatFP(SDValue N, bool AllowUndefs=false)
Returns the SDNode if it is a constant splat BuildVector or constant float.
LLVM_ABI ConstantRange getConstantRangeFromMetadata(const MDNode &RangeMD)
Parse out a conservative ConstantRange from !range metadata.
APFloat frexp(const APFloat &X, int &Exp, APFloat::roundingMode RM)
Equivalent of C standard library function.
Definition APFloat.h:1622
int countr_zero(T Val)
Count number of 0's from the least significant bit to the most stopping at the first 1.
Definition bit.h:202
auto dyn_cast_or_null(const Y &Val)
Definition Casting.h:753
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1746
LLVM_ABI bool getShuffleDemandedElts(int SrcWidth, ArrayRef< int > Mask, const APInt &DemandedElts, APInt &DemandedLHS, APInt &DemandedRHS, bool AllowUndefElts=false)
Transform a shuffle mask's output demanded element mask into demanded element masks for the 2 operand...
LLVM_READONLY APFloat maxnum(const APFloat &A, const APFloat &B)
Implements IEEE-754 2008 maxNum semantics.
Definition APFloat.h:1665
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
Definition MathExtras.h:331
LLVM_ABI bool isBitwiseNot(SDValue V, bool AllowUndefs=false)
Returns true if V is a bitwise not operation.
LLVM_ABI SDValue peekThroughInsertVectorElt(SDValue V, const APInt &DemandedElts)
Recursively peek through INSERT_VECTOR_ELT nodes, returning the source vector operand of V,...
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition MathExtras.h:279
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
LLVM_ABI void checkForCycles(const SelectionDAG *DAG, bool force=false)
void sort(IteratorTy Start, IteratorTy End)
Definition STLExtras.h:1636
LLVM_READONLY APFloat minimumnum(const APFloat &A, const APFloat &B)
Implements IEEE 754-2019 minimumNumber semantics.
Definition APFloat.h:1696
FPClassTest
Floating-point class tests, supported by 'is_fpclass' intrinsic.
LLVM_ABI void computeKnownBits(const Value *V, KnownBits &Known, const DataLayout &DL, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Determine which bits of V are known to be either zero or one and return them in the KnownZero/KnownOn...
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition Debug.cpp:207
LLVM_ABI SDValue peekThroughTruncates(SDValue V)
Return the non-truncated source operand of V if it exists.
bool none_of(R &&Range, UnaryPredicate P)
Provide wrappers to std::none_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1753
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
Definition Error.cpp:163
constexpr std::underlying_type_t< Enum > to_underlying(Enum E)
Returns underlying integer value of an enum.
FunctionAddr VTableAddr Count
Definition InstrProf.h:139
LLVM_ABI ConstantRange getVScaleRange(const Function *F, unsigned BitWidth)
Determine the possible constant range of vscale with the given bit width, based on the vscale_range f...
LLVM_ABI SDValue peekThroughOneUseBitcasts(SDValue V)
Return the non-bitcasted and one-use source operand of V if it exists.
CodeGenOptLevel
Code generation optimization level.
Definition CodeGen.h:82
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
LLVM_ABI bool isOneOrOneSplat(SDValue V, bool AllowUndefs=false)
Return true if the value is a constant 1 integer or a splatted vector of a constant 1 integer (with n...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
@ Other
Any other memory.
Definition ModRef.h:68
LLVM_READONLY APFloat minnum(const APFloat &A, const APFloat &B)
Implements IEEE-754 2008 minNum semantics.
Definition APFloat.h:1646
@ Mul
Product of integers.
@ Sub
Subtraction of integers.
LLVM_ABI bool isNullConstantOrUndef(SDValue V)
Returns true if V is a constant integer zero or an UNDEF node.
bool isInTailCallPosition(const CallBase &Call, const TargetMachine &TM, bool ReturnsFirstArg=false)
Test if the given instruction is in a position to be optimized with a tail-call.
Definition Analysis.cpp:539
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
LLVM_ABI ConstantSDNode * isConstOrConstSplat(SDValue N, bool AllowUndefs=false, bool AllowTruncation=false)
Returns the SDNode if it is a constant splat BuildVector or constant int.
OutputIt copy(R &&Range, OutputIt Out)
Definition STLExtras.h:1885
constexpr unsigned BitWidth
bool funcReturnsFirstArgOfCall(const CallInst &CI)
Returns true if the parent of CI returns CI's first argument after calling CI.
Definition Analysis.cpp:719
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
LLVM_ABI bool isZeroOrZeroSplat(SDValue N, bool AllowUndefs=false)
Return true if the value is a constant 0 integer or a splatted vector of a constant 0 integer (with n...
LLVM_ABI bool isOneConstant(SDValue V)
Returns true if V is a constant integer one.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
Definition STLExtras.h:1947
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
Definition Alignment.h:201
LLVM_ABI bool isNullFPConstant(SDValue V)
Returns true if V is an FP constant with a value of positive zero.
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
Definition MathExtras.h:572
unsigned Log2(Align A)
Returns the log2 of the alignment.
Definition Alignment.h:197
LLVM_ABI bool isZeroOrZeroSplatFP(SDValue N, bool AllowUndefs=false)
Return true if the value is a constant (+/-)0.0 floating-point value or a splatted vector thereof (wi...
LLVM_ABI void computeKnownBitsFromRangeMetadata(const MDNode &Ranges, KnownBits &Known)
Compute known bits from the range metadata.
LLVM_READONLY APFloat minimum(const APFloat &A, const APFloat &B)
Implements IEEE 754-2019 minimum semantics.
Definition APFloat.h:1683
LLVM_READONLY APFloat maximumnum(const APFloat &A, const APFloat &B)
Implements IEEE 754-2019 maximumNumber semantics.
Definition APFloat.h:1723
LLVM_ABI bool isOnesOrOnesSplat(SDValue N, bool AllowUndefs=false)
Return true if the value is a constant 1 integer or a splatted vector of a constant 1 integer (with n...
LLVM_ABI bool isNeutralConstant(unsigned Opc, SDNodeFlags Flags, SDValue V, unsigned OperandNo)
Returns true if V is a neutral element of Opc with Flags.
LLVM_ABI bool isAllOnesConstant(SDValue V)
Returns true if V is an integer constant with all bits set.
constexpr uint64_t NextPowerOf2(uint64_t A)
Returns the next power of two (in 64-bits) that is strictly greater than A.
Definition MathExtras.h:373
LLVM_ABI void reportFatalUsageError(Error Err)
Report a fatal error that does not indicate a bug in LLVM.
Definition Error.cpp:177
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:872
#define N
A collection of metadata nodes that might be associated with a memory access used by the alias-analys...
Definition Metadata.h:763
MDNode * TBAAStruct
The tag for type-based alias analysis (tbaa struct).
Definition Metadata.h:783
MDNode * TBAA
The tag for type-based alias analysis.
Definition Metadata.h:780
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
constexpr uint64_t value() const
This is a hole in the type system and should not be abused.
Definition Alignment.h:77
Represents offset+length into a ConstantDataArray.
uint64_t Length
Length of the slice.
uint64_t Offset
Slice starts at this Offset.
void move(uint64_t Delta)
Moves the Offset and adjusts Length accordingly.
const ConstantDataArray * Array
ConstantDataArray pointer.
Extended Value Type.
Definition ValueTypes.h:35
TypeSize getStoreSize() const
Return the number of bytes overwritten by a store of the specified value type.
Definition ValueTypes.h:403
bool isSimple() const
Test if the given EVT is simple (as opposed to being extended).
Definition ValueTypes.h:145
intptr_t getRawBits() const
Definition ValueTypes.h:520
static EVT getVectorVT(LLVMContext &Context, EVT VT, unsigned NumElements, bool IsScalable=false)
Returns the EVT that represents a vector NumElements in length, where each element is of type VT.
Definition ValueTypes.h:70
EVT changeTypeToInteger() const
Return the type converted to an equivalently sized integer or vector with integer element type.
Definition ValueTypes.h:129
bool bitsGT(EVT VT) const
Return true if this has more bits than VT.
Definition ValueTypes.h:292
bool bitsLT(EVT VT) const
Return true if this has less bits than VT.
Definition ValueTypes.h:308
bool isFloatingPoint() const
Return true if this is a FP or a vector FP type.
Definition ValueTypes.h:155
ElementCount getVectorElementCount() const
Definition ValueTypes.h:358
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
Definition ValueTypes.h:381
unsigned getVectorMinNumElements() const
Given a vector type, return the minimum number of elements it contains.
Definition ValueTypes.h:367
uint64_t getScalarSizeInBits() const
Definition ValueTypes.h:393
MVT getSimpleVT() const
Return the SimpleValueType held in the specified simple EVT.
Definition ValueTypes.h:324
static EVT getIntegerVT(LLVMContext &Context, unsigned BitWidth)
Returns the EVT that represents an integer with the given number of bits.
Definition ValueTypes.h:61
bool isFixedLengthVector() const
Definition ValueTypes.h:189
bool isVector() const
Return true if this is a vector value type.
Definition ValueTypes.h:176
EVT getScalarType() const
If this is a vector type, return the element type, otherwise return this.
Definition ValueTypes.h:331
bool bitsGE(EVT VT) const
Return true if this has no less bits than VT.
Definition ValueTypes.h:300
bool bitsEq(EVT VT) const
Return true if this has the same number of bits as VT.
Definition ValueTypes.h:264
LLVM_ABI Type * getTypeForEVT(LLVMContext &Context) const
This method returns an LLVM type corresponding to the specified EVT.
bool isScalableVector() const
Return true if this is a vector type where the runtime length is machine dependent.
Definition ValueTypes.h:182
EVT getVectorElementType() const
Given a vector type, return the type of each element.
Definition ValueTypes.h:336
bool isExtended() const
Test if the given EVT is extended (as opposed to being simple).
Definition ValueTypes.h:150
LLVM_ABI const fltSemantics & getFltSemantics() const
Returns an APFloat semantics tag appropriate for the value type.
unsigned getVectorNumElements() const
Given a vector type, return the number of elements it contains.
Definition ValueTypes.h:344
bool bitsLE(EVT VT) const
Return true if this has no more bits than VT.
Definition ValueTypes.h:316
EVT getHalfNumVectorElementsVT(LLVMContext &Context) const
Definition ValueTypes.h:461
bool isInteger() const
Return true if this is an integer or a vector integer type.
Definition ValueTypes.h:160
static KnownBits makeConstant(const APInt &C)
Create known bits from a known constant.
Definition KnownBits.h:317
LLVM_ABI KnownBits sextInReg(unsigned SrcBitWidth) const
Return known bits for a in-register sign extension of the value we're tracking.
static LLVM_ABI KnownBits mulhu(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits from zero-extended multiply-hi.
unsigned countMinSignBits() const
Returns the number of times the sign bit is replicated into the other bits.
Definition KnownBits.h:271
static LLVM_ABI KnownBits smax(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for smax(LHS, RHS).
bool isNonNegative() const
Returns true if this value is known to be non-negative.
Definition KnownBits.h:108
bool isZero() const
Returns true if value is all zero.
Definition KnownBits.h:80
void makeNonNegative()
Make this value non-negative.
Definition KnownBits.h:127
static LLVM_ABI KnownBits usub_sat(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from llvm.usub.sat(LHS, RHS)
unsigned countMinTrailingZeros() const
Returns the minimum number of trailing zero bits.
Definition KnownBits.h:258
static LLVM_ABI KnownBits ashr(const KnownBits &LHS, const KnownBits &RHS, bool ShAmtNonZero=false, bool Exact=false)
Compute known bits for ashr(LHS, RHS).
static LLVM_ABI KnownBits urem(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for urem(LHS, RHS).
bool isUnknown() const
Returns true if we don't know any bits.
Definition KnownBits.h:66
unsigned countMaxTrailingZeros() const
Returns the maximum number of trailing zero bits possible.
Definition KnownBits.h:290
static LLVM_ABI std::optional< bool > ne(const KnownBits &LHS, const KnownBits &RHS)
Determine if these known bits always give the same ICMP_NE result.
void makeNegative()
Make this value negative.
Definition KnownBits.h:122
void setAllConflict()
Make all bits known to be both zero and one.
Definition KnownBits.h:99
KnownBits trunc(unsigned BitWidth) const
Return known bits for a truncation of the value we're tracking.
Definition KnownBits.h:167
KnownBits byteSwap() const
Definition KnownBits.h:538
unsigned countMaxPopulation() const
Returns the maximum number of bits that could be one.
Definition KnownBits.h:305
void setAllZero()
Make all bits known to be zero and discard any previous information.
Definition KnownBits.h:86
KnownBits reverseBits() const
Definition KnownBits.h:542
KnownBits concat(const KnownBits &Lo) const
Concatenate the bits from Lo onto the bottom of *this.
Definition KnownBits.h:249
unsigned getBitWidth() const
Get the bit width of this value.
Definition KnownBits.h:44
static LLVM_ABI KnownBits umax(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for umax(LHS, RHS).
KnownBits zext(unsigned BitWidth) const
Return known bits for a zero extension of the value we're tracking.
Definition KnownBits.h:178
void resetAll()
Resets the known state of all bits.
Definition KnownBits.h:74
KnownBits unionWith(const KnownBits &RHS) const
Returns KnownBits information that is known to be true for either this or RHS or both.
Definition KnownBits.h:337
static LLVM_ABI KnownBits lshr(const KnownBits &LHS, const KnownBits &RHS, bool ShAmtNonZero=false, bool Exact=false)
Compute known bits for lshr(LHS, RHS).
bool isNonZero() const
Returns true if this value is known to be non-zero.
Definition KnownBits.h:111
static LLVM_ABI KnownBits abdu(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for abdu(LHS, RHS).
KnownBits extractBits(unsigned NumBits, unsigned BitPosition) const
Return a subset of the known bits from [bitPosition,bitPosition+numBits).
Definition KnownBits.h:241
static LLVM_ABI KnownBits avgFloorU(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from APIntOps::avgFloorU.
KnownBits intersectWith(const KnownBits &RHS) const
Returns KnownBits information that is known to be true for both this and RHS.
Definition KnownBits.h:327
KnownBits sext(unsigned BitWidth) const
Return known bits for a sign extension of the value we're tracking.
Definition KnownBits.h:186
static LLVM_ABI KnownBits computeForSubBorrow(const KnownBits &LHS, KnownBits RHS, const KnownBits &Borrow)
Compute known bits results from subtracting RHS from LHS with 1-bit Borrow.
KnownBits zextOrTrunc(unsigned BitWidth) const
Return known bits for a zero extension or truncation of the value we're tracking.
Definition KnownBits.h:202
APInt getMaxValue() const
Return the maximal unsigned value possible given these KnownBits.
Definition KnownBits.h:148
static LLVM_ABI KnownBits abds(KnownBits LHS, KnownBits RHS)
Compute known bits for abds(LHS, RHS).
static LLVM_ABI KnownBits smin(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for smin(LHS, RHS).
static LLVM_ABI KnownBits mulhs(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits from sign-extended multiply-hi.
static LLVM_ABI KnownBits srem(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for srem(LHS, RHS).
static LLVM_ABI KnownBits udiv(const KnownBits &LHS, const KnownBits &RHS, bool Exact=false)
Compute known bits for udiv(LHS, RHS).
static LLVM_ABI KnownBits computeForAddSub(bool Add, bool NSW, bool NUW, const KnownBits &LHS, const KnownBits &RHS)
Compute known bits resulting from adding LHS and RHS.
Definition KnownBits.cpp:61
bool isStrictlyPositive() const
Returns true if this value is known to be positive.
Definition KnownBits.h:114
static LLVM_ABI KnownBits sdiv(const KnownBits &LHS, const KnownBits &RHS, bool Exact=false)
Compute known bits for sdiv(LHS, RHS).
static LLVM_ABI KnownBits avgFloorS(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from APIntOps::avgFloorS.
static bool haveNoCommonBitsSet(const KnownBits &LHS, const KnownBits &RHS)
Return true if LHS and RHS have no common bits set.
Definition KnownBits.h:342
bool isNegative() const
Returns true if this value is known to be negative.
Definition KnownBits.h:105
LLVM_ABI KnownBits truncSSat(unsigned BitWidth) const
Truncate with signed saturation (signed input -> signed output)
static LLVM_ABI KnownBits computeForAddCarry(const KnownBits &LHS, const KnownBits &RHS, const KnownBits &Carry)
Compute known bits resulting from adding LHS, RHS and a 1-bit Carry.
Definition KnownBits.cpp:54
unsigned countMaxLeadingZeros() const
Returns the maximum number of leading zero bits possible.
Definition KnownBits.h:296
void insertBits(const KnownBits &SubBits, unsigned BitPosition)
Insert the bits from a smaller known bits starting at bitPosition.
Definition KnownBits.h:235
static LLVM_ABI KnownBits avgCeilU(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from APIntOps::avgCeilU.
static LLVM_ABI KnownBits mul(const KnownBits &LHS, const KnownBits &RHS, bool NoUndefSelfMultiply=false)
Compute known bits resulting from multiplying LHS and RHS.
KnownBits anyext(unsigned BitWidth) const
Return known bits for an "any" extension of the value we're tracking, where we don't know anything ab...
Definition KnownBits.h:173
static LLVM_ABI KnownBits clmul(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for clmul(LHS, RHS).
LLVM_ABI KnownBits abs(bool IntMinIsPoison=false) const
Compute known bits for the absolute value.
LLVM_ABI KnownBits truncUSat(unsigned BitWidth) const
Truncate with unsigned saturation (unsigned input -> unsigned output)
static LLVM_ABI KnownBits shl(const KnownBits &LHS, const KnownBits &RHS, bool NUW=false, bool NSW=false, bool ShAmtNonZero=false)
Compute known bits for shl(LHS, RHS).
static LLVM_ABI KnownBits umin(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for umin(LHS, RHS).
LLVM_ABI KnownBits truncSSatU(unsigned BitWidth) const
Truncate with signed saturation to unsigned (signed input -> unsigned output)
static LLVM_ABI KnownBits avgCeilS(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from APIntOps::avgCeilS.
This class contains a discriminated union of information about pointers in memory operands,...
LLVM_ABI bool isDereferenceable(unsigned Size, LLVMContext &C, const DataLayout &DL) const
Return true if memory region [V, V+Offset+Size) is known to be dereferenceable.
LLVM_ABI unsigned getAddrSpace() const
Return the LLVM IR address space number that this pointer points into.
PointerUnion< const Value *, const PseudoSourceValue * > V
This is the IR pointer value for the access, or it is null if unknown.
MachinePointerInfo getWithOffset(int64_t O) const
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
This struct is a compact representation of a valid (power of two) or undefined (0) alignment.
Definition Alignment.h:106
static MemOp Set(uint64_t Size, bool DstAlignCanChange, Align DstAlign, bool IsZeroMemset, bool IsVolatile)
static MemOp Copy(uint64_t Size, bool DstAlignCanChange, Align DstAlign, Align SrcAlign, bool IsVolatile, bool MemcpyStrSrc=false)
static StringRef getLibcallImplName(RTLIB::LibcallImpl CallImpl)
Get the libcall routine name for the specified libcall implementation.
These are IR-level optimization flags that may be propagated to SDNodes.
This represents a list of ValueType's that has been intern'd by a SelectionDAG.
unsigned int NumVTs
Clients of various APIs that cause global effects on the DAG can optionally implement this interface.
virtual void NodeDeleted(SDNode *N, SDNode *E)
The node N that was deleted and, if E is not null, an equivalent node E that replaced it.
virtual void NodeInserted(SDNode *N)
The node N that was inserted.
virtual void NodeUpdated(SDNode *N)
The node N that was updated.
This structure contains all information that is necessary for lowering calls.
CallLoweringInfo & setLibCallee(CallingConv::ID CC, Type *ResultType, SDValue Target, ArgListTy &&ArgsList)
CallLoweringInfo & setDiscardResult(bool Value=true)
CallLoweringInfo & setDebugLoc(const SDLoc &dl)
CallLoweringInfo & setTailCall(bool Value=true)
CallLoweringInfo & setChain(SDValue InChain)