mirror of
https://github.com/capstone-engine/llvm-capstone.git
synced 2024-11-23 22:00:10 +00:00
a66f7769a3
Add applyStaticChunkedWorkshareLoop method implementing static schedule when chunk-size is specified. Unlike a static schedule without chunk-size (where chunk-size is chosen by the runtime such that each thread receives one chunk), we need two nested loops: one for looping over the iterations of a chunk, and a second for looping over all chunks assigned to the threads. This patch includes the following related changes: * Adapt applyWorkshareLoop to triage between the schedule types, now possible since all schedules have been implemented. The default schedule is assumed to be non-chunked static, as without OpenMPIRBuilder. * Remove the chunk parameter from applyStaticWorkshareLoop, it is ignored by the runtime. Change the value for the value passed to the init function to 0, as without OpenMPIRBuilder. * Refactor CanonicalLoopInfo::setTripCount and CanonicalLoopInfo::mapIndVar as used by both, applyStaticWorkshareLoop and applyStaticChunkedWorkshareLoop. * Enable Clang to use the OpenMPIRBuilder in the presence of the schedule clause. Differential Revision: https://reviews.llvm.org/D114413 |
||
---|---|---|
.. | ||
ABIInfo.h | ||
Address.h | ||
BackendUtil.cpp | ||
CGAtomic.cpp | ||
CGBlocks.cpp | ||
CGBlocks.h | ||
CGBuilder.h | ||
CGBuiltin.cpp | ||
CGCall.cpp | ||
CGCall.h | ||
CGClass.cpp | ||
CGCleanup.cpp | ||
CGCleanup.h | ||
CGCoroutine.cpp | ||
CGCUDANV.cpp | ||
CGCUDARuntime.cpp | ||
CGCUDARuntime.h | ||
CGCXX.cpp | ||
CGCXXABI.cpp | ||
CGCXXABI.h | ||
CGDebugInfo.cpp | ||
CGDebugInfo.h | ||
CGDecl.cpp | ||
CGDeclCXX.cpp | ||
CGException.cpp | ||
CGExpr.cpp | ||
CGExprAgg.cpp | ||
CGExprComplex.cpp | ||
CGExprConstant.cpp | ||
CGExprCXX.cpp | ||
CGExprScalar.cpp | ||
CGGPUBuiltin.cpp | ||
CGLoopInfo.cpp | ||
CGLoopInfo.h | ||
CGNonTrivialStruct.cpp | ||
CGObjC.cpp | ||
CGObjCGNU.cpp | ||
CGObjCMac.cpp | ||
CGObjCRuntime.cpp | ||
CGObjCRuntime.h | ||
CGOpenCLRuntime.cpp | ||
CGOpenCLRuntime.h | ||
CGOpenMPRuntime.cpp | ||
CGOpenMPRuntime.h | ||
CGOpenMPRuntimeGPU.cpp | ||
CGOpenMPRuntimeGPU.h | ||
CGRecordLayout.h | ||
CGRecordLayoutBuilder.cpp | ||
CGStmt.cpp | ||
CGStmtOpenMP.cpp | ||
CGValue.h | ||
CGVTables.cpp | ||
CGVTables.h | ||
CGVTT.cpp | ||
CMakeLists.txt | ||
CodeGenABITypes.cpp | ||
CodeGenAction.cpp | ||
CodeGenFunction.cpp | ||
CodeGenFunction.h | ||
CodeGenModule.cpp | ||
CodeGenModule.h | ||
CodeGenPGO.cpp | ||
CodeGenPGO.h | ||
CodeGenTBAA.cpp | ||
CodeGenTBAA.h | ||
CodeGenTypeCache.h | ||
CodeGenTypes.cpp | ||
CodeGenTypes.h | ||
ConstantEmitter.h | ||
ConstantInitBuilder.cpp | ||
CoverageMappingGen.cpp | ||
CoverageMappingGen.h | ||
EHScopeStack.h | ||
ItaniumCXXABI.cpp | ||
MacroPPCallbacks.cpp | ||
MacroPPCallbacks.h | ||
MicrosoftCXXABI.cpp | ||
ModuleBuilder.cpp | ||
ObjectFilePCHContainerOperations.cpp | ||
PatternInit.cpp | ||
PatternInit.h | ||
README.txt | ||
SanitizerMetadata.cpp | ||
SanitizerMetadata.h | ||
SwiftCallingConv.cpp | ||
TargetInfo.cpp | ||
TargetInfo.h | ||
VarBypassDetector.cpp | ||
VarBypassDetector.h |
IRgen optimization opportunities. //===---------------------------------------------------------------------===// The common pattern of -- short x; // or char, etc (x == 10) -- generates an zext/sext of x which can easily be avoided. //===---------------------------------------------------------------------===// Bitfields accesses can be shifted to simplify masking and sign extension. For example, if the bitfield width is 8 and it is appropriately aligned then is is a lot shorter to just load the char directly. //===---------------------------------------------------------------------===// It may be worth avoiding creation of alloca's for formal arguments for the common situation where the argument is never written to or has its address taken. The idea would be to begin generating code by using the argument directly and if its address is taken or it is stored to then generate the alloca and patch up the existing code. In theory, the same optimization could be a win for block local variables as long as the declaration dominates all statements in the block. NOTE: The main case we care about this for is for -O0 -g compile time performance, and in that scenario we will need to emit the alloca anyway currently to emit proper debug info. So this is blocked by being able to emit debug information which refers to an LLVM temporary, not an alloca. //===---------------------------------------------------------------------===// We should try and avoid generating basic blocks which only contain jumps. At -O0, this penalizes us all the way from IRgen (malloc & instruction overhead), all the way down through code generation and assembly time. On 176.gcc:expr.ll, it looks like over 12% of basic blocks are just direct branches! //===---------------------------------------------------------------------===//