xref: /freebsd/contrib/llvm-project/llvm/lib/Target/NVPTX/NVPTXTargetTransformInfo.h (revision c66ec88fed842fbaad62c30d510644ceb7bd2d71)
1 //===-- NVPTXTargetTransformInfo.h - NVPTX specific TTI ---------*- C++ -*-===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 /// \file
9 /// This file a TargetTransformInfo::Concept conforming object specific to the
10 /// NVPTX target machine. It uses the target's detailed information to
11 /// provide more precise answers to certain TTI queries, while letting the
12 /// target independent and default TTI implementations handle the rest.
13 ///
14 //===----------------------------------------------------------------------===//
15 
16 #ifndef LLVM_LIB_TARGET_NVPTX_NVPTXTARGETTRANSFORMINFO_H
17 #define LLVM_LIB_TARGET_NVPTX_NVPTXTARGETTRANSFORMINFO_H
18 
19 #include "NVPTXTargetMachine.h"
20 #include "MCTargetDesc/NVPTXBaseInfo.h"
21 #include "llvm/Analysis/TargetTransformInfo.h"
22 #include "llvm/CodeGen/BasicTTIImpl.h"
23 #include "llvm/CodeGen/TargetLowering.h"
24 
25 namespace llvm {
26 
27 class NVPTXTTIImpl : public BasicTTIImplBase<NVPTXTTIImpl> {
28   typedef BasicTTIImplBase<NVPTXTTIImpl> BaseT;
29   typedef TargetTransformInfo TTI;
30   friend BaseT;
31 
32   const NVPTXSubtarget *ST;
33   const NVPTXTargetLowering *TLI;
34 
35   const NVPTXSubtarget *getST() const { return ST; };
36   const NVPTXTargetLowering *getTLI() const { return TLI; };
37 
38 public:
39   explicit NVPTXTTIImpl(const NVPTXTargetMachine *TM, const Function &F)
40       : BaseT(TM, F.getParent()->getDataLayout()), ST(TM->getSubtargetImpl()),
41         TLI(ST->getTargetLowering()) {}
42 
43   bool hasBranchDivergence() { return true; }
44 
45   bool isSourceOfDivergence(const Value *V);
46 
47   unsigned getFlatAddressSpace() const {
48     return AddressSpace::ADDRESS_SPACE_GENERIC;
49   }
50 
51   // Loads and stores can be vectorized if the alignment is at least as big as
52   // the load/store we want to vectorize.
53   bool isLegalToVectorizeLoadChain(unsigned ChainSizeInBytes, Align Alignment,
54                                    unsigned AddrSpace) const {
55     return Alignment >= ChainSizeInBytes;
56   }
57   bool isLegalToVectorizeStoreChain(unsigned ChainSizeInBytes, Align Alignment,
58                                     unsigned AddrSpace) const {
59     return isLegalToVectorizeLoadChain(ChainSizeInBytes, Alignment, AddrSpace);
60   }
61 
62   // NVPTX has infinite registers of all kinds, but the actual machine doesn't.
63   // We conservatively return 1 here which is just enough to enable the
64   // vectorizers but disables heuristics based on the number of registers.
65   // FIXME: Return a more reasonable number, while keeping an eye on
66   // LoopVectorizer's unrolling heuristics.
67   unsigned getNumberOfRegisters(bool Vector) const { return 1; }
68 
69   // Only <2 x half> should be vectorized, so always return 32 for the vector
70   // register size.
71   unsigned getRegisterBitWidth(bool Vector) const { return 32; }
72   unsigned getMinVectorRegisterBitWidth() const { return 32; }
73 
74   // We don't want to prevent inlining because of target-cpu and -features
75   // attributes that were added to newer versions of LLVM/Clang: There are
76   // no incompatible functions in PTX, ptxas will throw errors in such cases.
77   bool areInlineCompatible(const Function *Caller,
78                            const Function *Callee) const {
79     return true;
80   }
81 
82   // Increase the inlining cost threshold by a factor of 5, reflecting that
83   // calls are particularly expensive in NVPTX.
84   unsigned getInliningThresholdMultiplier() { return 5; }
85 
86   int getArithmeticInstrCost(
87       unsigned Opcode, Type *Ty,
88       TTI::TargetCostKind CostKind = TTI::TCK_RecipThroughput,
89       TTI::OperandValueKind Opd1Info = TTI::OK_AnyValue,
90       TTI::OperandValueKind Opd2Info = TTI::OK_AnyValue,
91       TTI::OperandValueProperties Opd1PropInfo = TTI::OP_None,
92       TTI::OperandValueProperties Opd2PropInfo = TTI::OP_None,
93       ArrayRef<const Value *> Args = ArrayRef<const Value *>(),
94       const Instruction *CxtI = nullptr);
95 
96   void getUnrollingPreferences(Loop *L, ScalarEvolution &SE,
97                                TTI::UnrollingPreferences &UP);
98 
99   void getPeelingPreferences(Loop *L, ScalarEvolution &SE,
100                              TTI::PeelingPreferences &PP);
101 
102   bool hasVolatileVariant(Instruction *I, unsigned AddrSpace) {
103     // Volatile loads/stores are only supported for shared and global address
104     // spaces, or for generic AS that maps to them.
105     if (!(AddrSpace == llvm::ADDRESS_SPACE_GENERIC ||
106           AddrSpace == llvm::ADDRESS_SPACE_GLOBAL ||
107           AddrSpace == llvm::ADDRESS_SPACE_SHARED))
108       return false;
109 
110     switch(I->getOpcode()){
111     default:
112       return false;
113     case Instruction::Load:
114     case Instruction::Store:
115       return true;
116     }
117   }
118 };
119 
120 } // end namespace llvm
121 
122 #endif
123