-
Notifications
You must be signed in to change notification settings - Fork 13.5k
[RISCV] Extract subregister if VLEN is known when lowering extract_subvector #65392
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -8622,17 +8622,20 @@ SDValue RISCVTargetLowering::lowerEXTRACT_SUBVECTOR(SDValue Op, | |
return DAG.getSetCC(DL, SubVecVT, Vec, SplatZero, ISD::SETNE); | ||
} | ||
} | ||
|
||
// With an index of 0 this is a cast-like subvector, which can be performed | ||
// with subregister operations. | ||
if (OrigIdx == 0) | ||
return Op; | ||
|
||
auto KnownVLen = Subtarget.getRealKnownVLen(); | ||
|
||
// If the subvector vector is a fixed-length type, we cannot use subregister | ||
// manipulation to simplify the codegen; we don't know which register of a | ||
// LMUL group contains the specific subvector as we only know the minimum | ||
// register size. Therefore we must slide the vector group down the full | ||
// amount. | ||
if (SubVecVT.isFixedLengthVector()) { | ||
// With an index of 0 this is a cast-like subvector, which can be performed | ||
// with subregister operations. | ||
if (OrigIdx == 0) | ||
return Op; | ||
// manipulation to simplify the codegen if we don't know VLEN; we don't know | ||
// which register of a LMUL group contains the specific subvector as we only | ||
// know the minimum register size. Therefore we must slide the vector group | ||
// down the full amount. | ||
if (SubVecVT.isFixedLengthVector() && !KnownVLen) { | ||
MVT ContainerVT = VecVT; | ||
if (VecVT.isFixedLengthVector()) { | ||
ContainerVT = getContainerForFixedLengthVector(VecVT); | ||
|
@@ -8653,36 +8656,68 @@ SDValue RISCVTargetLowering::lowerEXTRACT_SUBVECTOR(SDValue Op, | |
return DAG.getBitcast(Op.getValueType(), Slidedown); | ||
} | ||
|
||
if (VecVT.isFixedLengthVector()) { | ||
VecVT = getContainerForFixedLengthVector(VecVT); | ||
Vec = convertToScalableVector(VecVT, Vec, DAG, Subtarget); | ||
} | ||
|
||
// The semantics of extract_subvector are that if the extracted subvector is | ||
// scalable, then the index is scaled by vscale. So if we have a fixed length | ||
// subvector, we need to factor that in before we decompose it to | ||
// subregisters... | ||
MVT ContainerSubVecVT = SubVecVT; | ||
unsigned EffectiveIdx = OrigIdx; | ||
unsigned Vscale = *KnownVLen / RISCV::RVVBitsPerBlock; | ||
if (SubVecVT.isFixedLengthVector()) { | ||
assert(KnownVLen); | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. This assert isn't needed. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. IIUC There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I don't have any specific ideas, but I definitely wouldn't rely on UB. A library might put an assert in |
||
ContainerSubVecVT = getContainerForFixedLengthVector(SubVecVT); | ||
EffectiveIdx = OrigIdx / Vscale; | ||
} | ||
|
||
unsigned SubRegIdx, RemIdx; | ||
std::tie(SubRegIdx, RemIdx) = | ||
RISCVTargetLowering::decomposeSubvectorInsertExtractToSubRegs( | ||
VecVT, SubVecVT, OrigIdx, TRI); | ||
VecVT, ContainerSubVecVT, EffectiveIdx, TRI); | ||
|
||
// ... and scale the remainder back afterwards. | ||
if (SubVecVT.isFixedLengthVector()) | ||
RemIdx = (RemIdx * Vscale) + (OrigIdx % Vscale); | ||
|
||
// If the Idx has been completely eliminated then this is a subvector extract | ||
// which naturally aligns to a vector register. These can easily be handled | ||
// using subregister manipulation. | ||
if (RemIdx == 0) | ||
if (RemIdx == 0) { | ||
if (SubVecVT.isFixedLengthVector()) { | ||
Vec = DAG.getTargetExtractSubreg(SubRegIdx, DL, ContainerSubVecVT, Vec); | ||
return convertFromScalableVector(SubVecVT, Vec, DAG, Subtarget); | ||
} | ||
return Op; | ||
} | ||
|
||
// Else we must shift our vector register directly to extract the subvector. | ||
// Do this using VSLIDEDOWN. | ||
// Else SubVecVT is a fractional LMUL and needs to be slid down. | ||
assert(RISCVVType::decodeVLMUL(getLMUL(ContainerSubVecVT)).second); | ||
|
||
// If the vector type is an LMUL-group type, extract a subvector equal to the | ||
// nearest full vector register type. This should resolve to a EXTRACT_SUBREG | ||
// instruction. | ||
// nearest full vector register type. | ||
MVT InterSubVT = VecVT; | ||
if (VecVT.bitsGT(getLMUL1VT(VecVT))) { | ||
InterSubVT = getLMUL1VT(VecVT); | ||
Vec = DAG.getNode(ISD::EXTRACT_SUBVECTOR, DL, InterSubVT, Vec, | ||
DAG.getConstant(OrigIdx - RemIdx, DL, XLenVT)); | ||
Vec = DAG.getTargetExtractSubreg(SubRegIdx, DL, InterSubVT, Vec); | ||
} | ||
|
||
// Slide this vector register down by the desired number of elements in order | ||
// to place the desired subvector starting at element 0. | ||
SDValue SlidedownAmt = | ||
DAG.getVScale(DL, XLenVT, APInt(XLenVT.getSizeInBits(), RemIdx)); | ||
SDValue SlidedownAmt; | ||
if (SubVecVT.isFixedLengthVector()) | ||
SlidedownAmt = DAG.getConstant(RemIdx, DL, Subtarget.getXLenVT()); | ||
else | ||
SlidedownAmt = | ||
DAG.getVScale(DL, XLenVT, APInt(XLenVT.getSizeInBits(), RemIdx)); | ||
|
||
auto [Mask, VL] = getDefaultScalableVLOps(InterSubVT, DL, DAG, Subtarget); | ||
if (SubVecVT.isFixedLengthVector()) | ||
VL = getVLOp(SubVecVT.getVectorNumElements(), DL, DAG, Subtarget); | ||
|
||
SDValue Slidedown = | ||
getVSlidedown(DAG, Subtarget, DL, InterSubVT, DAG.getUNDEF(InterSubVT), | ||
Vec, SlidedownAmt, Mask, VL); | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -152,6 +152,11 @@ class RISCVSubtarget : public RISCVGenSubtargetInfo { | |
unsigned VLen = getMaxRVVVectorSizeInBits(); | ||
return VLen == 0 ? 65536 : VLen; | ||
} | ||
std::optional<unsigned> getRealKnownVLen() const { | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I'd suggest: getExactVLen(). |
||
if (getRealMinVLen() == getRealMaxVLen()) | ||
return getRealMinVLen(); | ||
return std::nullopt; | ||
} | ||
RISCVABI::ABI getTargetABI() const { return TargetABI; } | ||
bool isSoftFPABI() const { | ||
return TargetABI == RISCVABI::ABI_LP64 || | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This isn't fully general.
Consider: 8 x i64 on V (a lmul4 type), extract the second <2 x i64> quarter. We can still do know this is the first LMUL2 sub-register. This doesn't allow us to slide less, but it does allow us to reduce the lmul for the slide.