[llvm] [LV] Teach the vectorizer to cost and vectorize llvm.sincos intrinsics (PR #123210)
Benjamin Maxwell via llvm-commits
llvm-commits at lists.llvm.org
Mon Feb 17 03:33:07 PST 2025
================
@@ -2005,10 +2049,13 @@ class BasicTTIImplBase : public TargetTransformInfoImplCRTPBase<T> {
InstructionCost ScalarizationCost = InstructionCost::getInvalid();
if (RetVF.isVector() && !RetVF.isScalable()) {
ScalarizationCost = 0;
- if (!RetTy->isVoidTy())
- ScalarizationCost += getScalarizationOverhead(
- cast<VectorType>(RetTy),
- /*Insert*/ true, /*Extract*/ false, CostKind);
+ if (!RetTy->isVoidTy()) {
+ for (Type *VectorTy : getContainedTypes(RetTy)) {
+ ScalarizationCost += getScalarizationOverhead(
+ cast<VectorType>(VectorTy),
+ /*Insert*/ true, /*Extract*/ false, CostKind);
+ }
+ }
----------------
MacDue wrote:
@arsenm this is what causes the changes in the AMDGPU `frexp.ll` costs. Previously, this would be skipped as `RetVF` was 1 for struct-of-vector returns. Now it costs the inserts for both returned vectors.
https://github.com/llvm/llvm-project/pull/123210
More information about the llvm-commits
mailing list