From d026f79bd33748a9ee591bc156ea39234dff9872 Mon Sep 17 00:00:00 2001 From: Stanislav Mekhanoshin Date: Thu, 13 Apr 2017 17:53:07 +0000 Subject: [PATCH] [AMDGPU] Combine DS operations with offsets bigger than byte In many cases ds operations can be combined even if offsets do not fit into 8 bit encoding. What it takes is to adjust base address. Differential Revision: https://reviews.llvm.org/D31993 llvm-svn: 300227 --- llvm/lib/Target/AMDGPU/SILoadStoreOptimizer.cpp | 316 +++++++++-------- .../test/CodeGen/AMDGPU/ds-combine-large-stride.ll | 385 +++++++++++++++++++++ 2 files changed, 551 insertions(+), 150 deletions(-) create mode 100644 llvm/test/CodeGen/AMDGPU/ds-combine-large-stride.ll diff --git a/llvm/lib/Target/AMDGPU/SILoadStoreOptimizer.cpp b/llvm/lib/Target/AMDGPU/SILoadStoreOptimizer.cpp index 6b0d18e..d583449 100644 --- a/llvm/lib/Target/AMDGPU/SILoadStoreOptimizer.cpp +++ b/llvm/lib/Target/AMDGPU/SILoadStoreOptimizer.cpp @@ -68,32 +68,31 @@ using namespace llvm; namespace { class SILoadStoreOptimizer : public MachineFunctionPass { + + typedef struct { + MachineBasicBlock::iterator I; + MachineBasicBlock::iterator Paired; + unsigned EltSize; + unsigned Offset0; + unsigned Offset1; + unsigned BaseOff; + bool UseST64; + SmallVector InstsToMove; + } CombineInfo; + private: const SIInstrInfo *TII = nullptr; const SIRegisterInfo *TRI = nullptr; MachineRegisterInfo *MRI = nullptr; AliasAnalysis *AA = nullptr; - static bool offsetsCanBeCombined(unsigned Offset0, - unsigned Offset1, - unsigned EltSize); + static bool offsetsCanBeCombined(CombineInfo &CI); - MachineBasicBlock::iterator findMatchingDSInst( - MachineBasicBlock::iterator I, - unsigned EltSize, - SmallVectorImpl &InstsToMove); + bool findMatchingDSInst(CombineInfo &CI); - MachineBasicBlock::iterator mergeRead2Pair( - MachineBasicBlock::iterator I, - MachineBasicBlock::iterator Paired, - unsigned EltSize, - ArrayRef InstsToMove); + MachineBasicBlock::iterator mergeRead2Pair(CombineInfo &CI); - MachineBasicBlock::iterator mergeWrite2Pair( - MachineBasicBlock::iterator I, - MachineBasicBlock::iterator Paired, - unsigned EltSize, - ArrayRef InstsToMove); + MachineBasicBlock::iterator mergeWrite2Pair(CombineInfo &CI); public: static char ID; @@ -199,46 +198,68 @@ canMoveInstsAcrossMemOp(MachineInstr &MemOp, return true; } -bool SILoadStoreOptimizer::offsetsCanBeCombined(unsigned Offset0, - unsigned Offset1, - unsigned Size) { +bool SILoadStoreOptimizer::offsetsCanBeCombined(CombineInfo &CI) { // XXX - Would the same offset be OK? Is there any reason this would happen or // be useful? - if (Offset0 == Offset1) + if (CI.Offset0 == CI.Offset1) return false; // This won't be valid if the offset isn't aligned. - if ((Offset0 % Size != 0) || (Offset1 % Size != 0)) + if ((CI.Offset0 % CI.EltSize != 0) || (CI.Offset1 % CI.EltSize != 0)) return false; - unsigned EltOffset0 = Offset0 / Size; - unsigned EltOffset1 = Offset1 / Size; + unsigned EltOffset0 = CI.Offset0 / CI.EltSize; + unsigned EltOffset1 = CI.Offset1 / CI.EltSize; + CI.UseST64 = false; + CI.BaseOff = 0; + + // If the offset in elements doesn't fit in 8-bits, we might be able to use + // the stride 64 versions. + if ((EltOffset0 % 64 == 0) && (EltOffset1 % 64) == 0 && + isUInt<8>(EltOffset0 / 64) && isUInt<8>(EltOffset1 / 64)) { + CI.Offset0 = EltOffset0 / 64; + CI.Offset1 = EltOffset1 / 64; + CI.UseST64 = true; + return true; + } // Check if the new offsets fit in the reduced 8-bit range. - if (isUInt<8>(EltOffset0) && isUInt<8>(EltOffset1)) + if (isUInt<8>(EltOffset0) && isUInt<8>(EltOffset1)) { + CI.Offset0 = EltOffset0; + CI.Offset1 = EltOffset1; + return true; + } + + // Try to shift base address to decrease offsets. + unsigned OffsetDiff = std::abs((int)EltOffset1 - (int)EltOffset0); + CI.BaseOff = std::min(CI.Offset0, CI.Offset1); + + if ((OffsetDiff % 64 == 0) && isUInt<8>(OffsetDiff / 64)) { + CI.Offset0 = (EltOffset0 - CI.BaseOff / CI.EltSize) / 64; + CI.Offset1 = (EltOffset1 - CI.BaseOff / CI.EltSize) / 64; + CI.UseST64 = true; return true; + } - // If the offset in elements doesn't fit in 8-bits, we might be able to use - // the stride 64 versions. - if ((EltOffset0 % 64 != 0) || (EltOffset1 % 64) != 0) - return false; + if (isUInt<8>(OffsetDiff)) { + CI.Offset0 = EltOffset0 - CI.BaseOff / CI.EltSize; + CI.Offset1 = EltOffset1 - CI.BaseOff / CI.EltSize; + return true; + } - return isUInt<8>(EltOffset0 / 64) && isUInt<8>(EltOffset1 / 64); + return false; } -MachineBasicBlock::iterator -SILoadStoreOptimizer::findMatchingDSInst(MachineBasicBlock::iterator I, - unsigned EltSize, - SmallVectorImpl &InstsToMove) { - MachineBasicBlock::iterator E = I->getParent()->end(); - MachineBasicBlock::iterator MBBI = I; +bool SILoadStoreOptimizer::findMatchingDSInst(CombineInfo &CI) { + MachineBasicBlock::iterator E = CI.I->getParent()->end(); + MachineBasicBlock::iterator MBBI = CI.I; ++MBBI; SmallVector DefsToMove; - addDefsToList(*I, DefsToMove); + addDefsToList(*CI.I, DefsToMove); for ( ; MBBI != E; ++MBBI) { - if (MBBI->getOpcode() != I->getOpcode()) { + if (MBBI->getOpcode() != CI.I->getOpcode()) { // This is not a matching DS instruction, but we can keep looking as // long as one of these conditions are met: @@ -249,14 +270,14 @@ SILoadStoreOptimizer::findMatchingDSInst(MachineBasicBlock::iterator I, if (MBBI->hasUnmodeledSideEffects()) // We can't re-order this instruction with respect to other memory // opeations, so we fail both conditions mentioned above. - return E; + return false; if (MBBI->mayLoadOrStore() && - !memAccessesCanBeReordered(*I, *MBBI, TII, AA)) { + !memAccessesCanBeReordered(*CI.I, *MBBI, TII, AA)) { // We fail condition #1, but we may still be able to satisfy condition // #2. Add this instruction to the move list and then we will check // if condition #2 holds once we have selected the matching instruction. - InstsToMove.push_back(&*MBBI); + CI.InstsToMove.push_back(&*MBBI); addDefsToList(*MBBI, DefsToMove); continue; } @@ -264,13 +285,13 @@ SILoadStoreOptimizer::findMatchingDSInst(MachineBasicBlock::iterator I, // When we match I with another DS instruction we will be moving I down // to the location of the matched instruction any uses of I will need to // be moved down as well. - addToListsIfDependent(*MBBI, DefsToMove, InstsToMove); + addToListsIfDependent(*MBBI, DefsToMove, CI.InstsToMove); continue; } // Don't merge volatiles. if (MBBI->hasOrderedMemoryRef()) - return E; + return false; // Handle a case like // DS_WRITE_B32 addr, v, idx0 @@ -278,77 +299,67 @@ SILoadStoreOptimizer::findMatchingDSInst(MachineBasicBlock::iterator I, // DS_WRITE_B32 addr, f(w), idx1 // where the DS_READ_B32 ends up in InstsToMove and therefore prevents // merging of the two writes. - if (addToListsIfDependent(*MBBI, DefsToMove, InstsToMove)) + if (addToListsIfDependent(*MBBI, DefsToMove, CI.InstsToMove)) continue; - int AddrIdx = AMDGPU::getNamedOperandIdx(I->getOpcode(), AMDGPU::OpName::addr); - const MachineOperand &AddrReg0 = I->getOperand(AddrIdx); + int AddrIdx = AMDGPU::getNamedOperandIdx(CI.I->getOpcode(), + AMDGPU::OpName::addr); + const MachineOperand &AddrReg0 = CI.I->getOperand(AddrIdx); const MachineOperand &AddrReg1 = MBBI->getOperand(AddrIdx); // Check same base pointer. Be careful of subregisters, which can occur with // vectors of pointers. if (AddrReg0.getReg() == AddrReg1.getReg() && AddrReg0.getSubReg() == AddrReg1.getSubReg()) { - int OffsetIdx = AMDGPU::getNamedOperandIdx(I->getOpcode(), + int OffsetIdx = AMDGPU::getNamedOperandIdx(CI.I->getOpcode(), AMDGPU::OpName::offset); - unsigned Offset0 = I->getOperand(OffsetIdx).getImm() & 0xffff; - unsigned Offset1 = MBBI->getOperand(OffsetIdx).getImm() & 0xffff; + CI.Offset0 = CI.I->getOperand(OffsetIdx).getImm() & 0xffff; + CI.Offset1 = MBBI->getOperand(OffsetIdx).getImm() & 0xffff; + CI.Paired = MBBI; // Check both offsets fit in the reduced range. // We also need to go through the list of instructions that we plan to // move and make sure they are all safe to move down past the merged // instruction. - if (offsetsCanBeCombined(Offset0, Offset1, EltSize) && - canMoveInstsAcrossMemOp(*MBBI, InstsToMove, TII, AA)) - return MBBI; + if (offsetsCanBeCombined(CI)) + if (canMoveInstsAcrossMemOp(*MBBI, CI.InstsToMove, TII, AA)) + return true; } // We've found a load/store that we couldn't merge for some reason. // We could potentially keep looking, but we'd need to make sure that // it was safe to move I and also all the instruction in InstsToMove // down past this instruction. - if (!memAccessesCanBeReordered(*I, *MBBI, TII, AA) || // check if we can move I across MBBI - !canMoveInstsAcrossMemOp(*MBBI, InstsToMove, TII, AA) // check if we can move all I's users - ) + // check if we can move I across MBBI and if we can move all I's users + if (!memAccessesCanBeReordered(*CI.I, *MBBI, TII, AA) || + !canMoveInstsAcrossMemOp(*MBBI, CI.InstsToMove, TII, AA)) break; } - return E; + return false; } MachineBasicBlock::iterator SILoadStoreOptimizer::mergeRead2Pair( - MachineBasicBlock::iterator I, - MachineBasicBlock::iterator Paired, - unsigned EltSize, - ArrayRef InstsToMove) { - MachineBasicBlock *MBB = I->getParent(); + CombineInfo &CI) { + MachineBasicBlock *MBB = CI.I->getParent(); // Be careful, since the addresses could be subregisters themselves in weird // cases, like vectors of pointers. - const MachineOperand *AddrReg = TII->getNamedOperand(*I, AMDGPU::OpName::addr); - - const MachineOperand *Dest0 = TII->getNamedOperand(*I, AMDGPU::OpName::vdst); - const MachineOperand *Dest1 = TII->getNamedOperand(*Paired, AMDGPU::OpName::vdst); - - unsigned Offset0 - = TII->getNamedOperand(*I, AMDGPU::OpName::offset)->getImm() & 0xffff; - unsigned Offset1 - = TII->getNamedOperand(*Paired, AMDGPU::OpName::offset)->getImm() & 0xffff; - - unsigned NewOffset0 = Offset0 / EltSize; - unsigned NewOffset1 = Offset1 / EltSize; - unsigned Opc = (EltSize == 4) ? AMDGPU::DS_READ2_B32 : AMDGPU::DS_READ2_B64; - - // Prefer the st64 form if we can use it, even if we can fit the offset in the - // non st64 version. I'm not sure if there's any real reason to do this. - bool UseST64 = (NewOffset0 % 64 == 0) && (NewOffset1 % 64 == 0); - if (UseST64) { - NewOffset0 /= 64; - NewOffset1 /= 64; - Opc = (EltSize == 4) ? AMDGPU::DS_READ2ST64_B32 : AMDGPU::DS_READ2ST64_B64; - } + const auto *AddrReg = TII->getNamedOperand(*CI.I, AMDGPU::OpName::addr); - unsigned SubRegIdx0 = (EltSize == 4) ? AMDGPU::sub0 : AMDGPU::sub0_sub1; - unsigned SubRegIdx1 = (EltSize == 4) ? AMDGPU::sub1 : AMDGPU::sub2_sub3; + const auto *Dest0 = TII->getNamedOperand(*CI.I, AMDGPU::OpName::vdst); + const auto *Dest1 = TII->getNamedOperand(*CI.Paired, AMDGPU::OpName::vdst); + + unsigned NewOffset0 = CI.Offset0; + unsigned NewOffset1 = CI.Offset1; + unsigned Opc = (CI.EltSize == 4) ? AMDGPU::DS_READ2_B32 + : AMDGPU::DS_READ2_B64; + + if (CI.UseST64) + Opc = (CI.EltSize == 4) ? AMDGPU::DS_READ2ST64_B32 + : AMDGPU::DS_READ2ST64_B64; + + unsigned SubRegIdx0 = (CI.EltSize == 4) ? AMDGPU::sub0 : AMDGPU::sub0_sub1; + unsigned SubRegIdx1 = (CI.EltSize == 4) ? AMDGPU::sub1 : AMDGPU::sub2_sub3; if (NewOffset0 > NewOffset1) { // Canonicalize the merged instruction so the smaller offset comes first. @@ -363,71 +374,69 @@ MachineBasicBlock::iterator SILoadStoreOptimizer::mergeRead2Pair( const MCInstrDesc &Read2Desc = TII->get(Opc); const TargetRegisterClass *SuperRC - = (EltSize == 4) ? &AMDGPU::VReg_64RegClass : &AMDGPU::VReg_128RegClass; + = (CI.EltSize == 4) ? &AMDGPU::VReg_64RegClass : &AMDGPU::VReg_128RegClass; unsigned DestReg = MRI->createVirtualRegister(SuperRC); - DebugLoc DL = I->getDebugLoc(); - MachineInstrBuilder Read2 = BuildMI(*MBB, Paired, DL, Read2Desc, DestReg) - .add(*AddrReg) // addr - .addImm(NewOffset0) // offset0 - .addImm(NewOffset1) // offset1 - .addImm(0) // gds - .addMemOperand(*I->memoperands_begin()) - .addMemOperand(*Paired->memoperands_begin()); + DebugLoc DL = CI.I->getDebugLoc(); + + unsigned BaseReg = AddrReg->getReg(); + unsigned BaseRegFlags = 0; + if (CI.BaseOff) { + BaseReg = MRI->createVirtualRegister(&AMDGPU::VGPR_32RegClass); + BaseRegFlags = RegState::Kill; + *BuildMI(*MBB, CI.Paired, DL, TII->get(AMDGPU::V_ADD_I32_e32), BaseReg) + .addImm(CI.BaseOff) + .addReg(AddrReg->getReg()); + } + + MachineInstrBuilder Read2 = BuildMI(*MBB, CI.Paired, DL, Read2Desc, DestReg) + .addReg(BaseReg, BaseRegFlags) // addr + .addImm(NewOffset0) // offset0 + .addImm(NewOffset1) // offset1 + .addImm(0) // gds + .addMemOperand(*CI.I->memoperands_begin()) + .addMemOperand(*CI.Paired->memoperands_begin()); (void)Read2; const MCInstrDesc &CopyDesc = TII->get(TargetOpcode::COPY); // Copy to the old destination registers. - BuildMI(*MBB, Paired, DL, CopyDesc) + BuildMI(*MBB, CI.Paired, DL, CopyDesc) .add(*Dest0) // Copy to same destination including flags and sub reg. .addReg(DestReg, 0, SubRegIdx0); - MachineInstr *Copy1 = BuildMI(*MBB, Paired, DL, CopyDesc) + MachineInstr *Copy1 = BuildMI(*MBB, CI.Paired, DL, CopyDesc) .add(*Dest1) .addReg(DestReg, RegState::Kill, SubRegIdx1); - moveInstsAfter(Copy1, InstsToMove); + moveInstsAfter(Copy1, CI.InstsToMove); - MachineBasicBlock::iterator Next = std::next(I); - I->eraseFromParent(); - Paired->eraseFromParent(); + MachineBasicBlock::iterator Next = std::next(CI.I); + CI.I->eraseFromParent(); + CI.Paired->eraseFromParent(); DEBUG(dbgs() << "Inserted read2: " << *Read2 << '\n'); return Next; } MachineBasicBlock::iterator SILoadStoreOptimizer::mergeWrite2Pair( - MachineBasicBlock::iterator I, - MachineBasicBlock::iterator Paired, - unsigned EltSize, - ArrayRef InstsToMove) { - MachineBasicBlock *MBB = I->getParent(); + CombineInfo &CI) { + MachineBasicBlock *MBB = CI.I->getParent(); // Be sure to use .addOperand(), and not .addReg() with these. We want to be // sure we preserve the subregister index and any register flags set on them. - const MachineOperand *Addr = TII->getNamedOperand(*I, AMDGPU::OpName::addr); - const MachineOperand *Data0 = TII->getNamedOperand(*I, AMDGPU::OpName::data0); + const MachineOperand *Addr = TII->getNamedOperand(*CI.I, AMDGPU::OpName::addr); + const MachineOperand *Data0 = TII->getNamedOperand(*CI.I, AMDGPU::OpName::data0); const MachineOperand *Data1 - = TII->getNamedOperand(*Paired, AMDGPU::OpName::data0); - + = TII->getNamedOperand(*CI.Paired, AMDGPU::OpName::data0); - unsigned Offset0 - = TII->getNamedOperand(*I, AMDGPU::OpName::offset)->getImm() & 0xffff; - unsigned Offset1 - = TII->getNamedOperand(*Paired, AMDGPU::OpName::offset)->getImm() & 0xffff; + unsigned NewOffset0 = CI.Offset0; + unsigned NewOffset1 = CI.Offset1; + unsigned Opc = (CI.EltSize == 4) ? AMDGPU::DS_WRITE2_B32 + : AMDGPU::DS_WRITE2_B64; - unsigned NewOffset0 = Offset0 / EltSize; - unsigned NewOffset1 = Offset1 / EltSize; - unsigned Opc = (EltSize == 4) ? AMDGPU::DS_WRITE2_B32 : AMDGPU::DS_WRITE2_B64; - - // Prefer the st64 form if we can use it, even if we can fit the offset in the - // non st64 version. I'm not sure if there's any real reason to do this. - bool UseST64 = (NewOffset0 % 64 == 0) && (NewOffset1 % 64 == 0); - if (UseST64) { - NewOffset0 /= 64; - NewOffset1 /= 64; - Opc = (EltSize == 4) ? AMDGPU::DS_WRITE2ST64_B32 : AMDGPU::DS_WRITE2ST64_B64; - } + if (CI.UseST64) + Opc = (CI.EltSize == 4) ? AMDGPU::DS_WRITE2ST64_B32 + : AMDGPU::DS_WRITE2ST64_B64; if (NewOffset0 > NewOffset1) { // Canonicalize the merged instruction so the smaller offset comes first. @@ -440,23 +449,33 @@ MachineBasicBlock::iterator SILoadStoreOptimizer::mergeWrite2Pair( "Computed offset doesn't fit"); const MCInstrDesc &Write2Desc = TII->get(Opc); - DebugLoc DL = I->getDebugLoc(); + DebugLoc DL = CI.I->getDebugLoc(); + + unsigned BaseReg = Addr->getReg(); + unsigned BaseRegFlags = 0; + if (CI.BaseOff) { + BaseReg = MRI->createVirtualRegister(&AMDGPU::VGPR_32RegClass); + BaseRegFlags = RegState::Kill; + *BuildMI(*MBB, CI.Paired, DL, TII->get(AMDGPU::V_ADD_I32_e32), BaseReg) + .addImm(CI.BaseOff) + .addReg(Addr->getReg()); + } - MachineInstrBuilder Write2 = BuildMI(*MBB, Paired, DL, Write2Desc) - .add(*Addr) // addr - .add(*Data0) // data0 - .add(*Data1) // data1 - .addImm(NewOffset0) // offset0 - .addImm(NewOffset1) // offset1 - .addImm(0) // gds - .addMemOperand(*I->memoperands_begin()) - .addMemOperand(*Paired->memoperands_begin()); + MachineInstrBuilder Write2 = BuildMI(*MBB, CI.Paired, DL, Write2Desc) + .addReg(BaseReg, BaseRegFlags) // addr + .add(*Data0) // data0 + .add(*Data1) // data1 + .addImm(NewOffset0) // offset0 + .addImm(NewOffset1) // offset1 + .addImm(0) // gds + .addMemOperand(*CI.I->memoperands_begin()) + .addMemOperand(*CI.Paired->memoperands_begin()); - moveInstsAfter(Write2, InstsToMove); + moveInstsAfter(Write2, CI.InstsToMove); - MachineBasicBlock::iterator Next = std::next(I); - I->eraseFromParent(); - Paired->eraseFromParent(); + MachineBasicBlock::iterator Next = std::next(CI.I); + CI.I->eraseFromParent(); + CI.Paired->eraseFromParent(); DEBUG(dbgs() << "Inserted write2 inst: " << *Write2 << '\n'); return Next; @@ -477,27 +496,24 @@ bool SILoadStoreOptimizer::optimizeBlock(MachineBasicBlock &MBB) { continue; } - SmallVector InstsToMove; + CombineInfo CI; + CI.I = I; unsigned Opc = MI.getOpcode(); if (Opc == AMDGPU::DS_READ_B32 || Opc == AMDGPU::DS_READ_B64) { - unsigned Size = (Opc == AMDGPU::DS_READ_B64) ? 8 : 4; - MachineBasicBlock::iterator Match = findMatchingDSInst(I, Size, - InstsToMove); - if (Match != E) { + CI.EltSize = (Opc == AMDGPU::DS_READ_B64) ? 8 : 4; + if (findMatchingDSInst(CI)) { Modified = true; - I = mergeRead2Pair(I, Match, Size, InstsToMove); + I = mergeRead2Pair(CI); } else { ++I; } continue; } else if (Opc == AMDGPU::DS_WRITE_B32 || Opc == AMDGPU::DS_WRITE_B64) { - unsigned Size = (Opc == AMDGPU::DS_WRITE_B64) ? 8 : 4; - MachineBasicBlock::iterator Match = findMatchingDSInst(I, Size, - InstsToMove); - if (Match != E) { + CI.EltSize = (Opc == AMDGPU::DS_WRITE_B64) ? 8 : 4; + if (findMatchingDSInst(CI)) { Modified = true; - I = mergeWrite2Pair(I, Match, Size, InstsToMove); + I = mergeWrite2Pair(CI); } else { ++I; } diff --git a/llvm/test/CodeGen/AMDGPU/ds-combine-large-stride.ll b/llvm/test/CodeGen/AMDGPU/ds-combine-large-stride.ll new file mode 100644 index 0000000..582b87b --- /dev/null +++ b/llvm/test/CodeGen/AMDGPU/ds-combine-large-stride.ll @@ -0,0 +1,385 @@ +; RUN: llc -march=amdgcn -mtriple=amdgcn--amdhsa -verify-machineinstrs < %s | FileCheck %s + +; CHECK-LABEL: ds_read32_combine_stride_400: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 0x320, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B2:v[0-9]+]], vcc, 0x640, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B3:v[0-9]+]], vcc, 0x960, [[BASE]] +; CHECK-DAG: ds_read2_b32 v[{{[0-9]+:[0-9]+}}], [[BASE]] offset1:100 +; CHECK-DAG: ds_read2_b32 v[{{[0-9]+:[0-9]+}}], [[B1]] offset1:100 +; CHECK-DAG: ds_read2_b32 v[{{[0-9]+:[0-9]+}}], [[B2]] offset1:100 +; CHECK-DAG: ds_read2_b32 v[{{[0-9]+:[0-9]+}}], [[B3]] offset1:100 +define void @ds_read32_combine_stride_400(float addrspace(3)* nocapture readonly %arg, float *nocapture %arg1) { +bb: + %tmp = load float, float addrspace(3)* %arg, align 4 + %tmp2 = fadd float %tmp, 0.000000e+00 + %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 100 + %tmp4 = load float, float addrspace(3)* %tmp3, align 4 + %tmp5 = fadd float %tmp2, %tmp4 + %tmp6 = getelementptr inbounds float, float addrspace(3)* %arg, i32 200 + %tmp7 = load float, float addrspace(3)* %tmp6, align 4 + %tmp8 = fadd float %tmp5, %tmp7 + %tmp9 = getelementptr inbounds float, float addrspace(3)* %arg, i32 300 + %tmp10 = load float, float addrspace(3)* %tmp9, align 4 + %tmp11 = fadd float %tmp8, %tmp10 + %tmp12 = getelementptr inbounds float, float addrspace(3)* %arg, i32 400 + %tmp13 = load float, float addrspace(3)* %tmp12, align 4 + %tmp14 = fadd float %tmp11, %tmp13 + %tmp15 = getelementptr inbounds float, float addrspace(3)* %arg, i32 500 + %tmp16 = load float, float addrspace(3)* %tmp15, align 4 + %tmp17 = fadd float %tmp14, %tmp16 + %tmp18 = getelementptr inbounds float, float addrspace(3)* %arg, i32 600 + %tmp19 = load float, float addrspace(3)* %tmp18, align 4 + %tmp20 = fadd float %tmp17, %tmp19 + %tmp21 = getelementptr inbounds float, float addrspace(3)* %arg, i32 700 + %tmp22 = load float, float addrspace(3)* %tmp21, align 4 + %tmp23 = fadd float %tmp20, %tmp22 + store float %tmp23, float *%arg1, align 4 + ret void +} + +; CHECK-LABEL: ds_read32_combine_stride_400_back: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 0x320, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B2:v[0-9]+]], vcc, 0x640, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B3:v[0-9]+]], vcc, 0x960, [[BASE]] +; CHECK-DAG: ds_read2_b32 v[{{[0-9]+:[0-9]+}}], [[BASE]] offset1:100 +; CHECK-DAG: ds_read2_b32 v[{{[0-9]+:[0-9]+}}], [[B1]] offset1:100 +; CHECK-DAG: ds_read2_b32 v[{{[0-9]+:[0-9]+}}], [[B2]] offset1:100 +; CHECK-DAG: ds_read2_b32 v[{{[0-9]+:[0-9]+}}], [[B3]] offset1:100 +define void @ds_read32_combine_stride_400_back(float addrspace(3)* nocapture readonly %arg, float *nocapture %arg1) { +bb: + %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 700 + %tmp2 = load float, float addrspace(3)* %tmp, align 4 + %tmp3 = fadd float %tmp2, 0.000000e+00 + %tmp4 = getelementptr inbounds float, float addrspace(3)* %arg, i32 600 + %tmp5 = load float, float addrspace(3)* %tmp4, align 4 + %tmp6 = fadd float %tmp3, %tmp5 + %tmp7 = getelementptr inbounds float, float addrspace(3)* %arg, i32 500 + %tmp8 = load float, float addrspace(3)* %tmp7, align 4 + %tmp9 = fadd float %tmp6, %tmp8 + %tmp10 = getelementptr inbounds float, float addrspace(3)* %arg, i32 400 + %tmp11 = load float, float addrspace(3)* %tmp10, align 4 + %tmp12 = fadd float %tmp9, %tmp11 + %tmp13 = getelementptr inbounds float, float addrspace(3)* %arg, i32 300 + %tmp14 = load float, float addrspace(3)* %tmp13, align 4 + %tmp15 = fadd float %tmp12, %tmp14 + %tmp16 = getelementptr inbounds float, float addrspace(3)* %arg, i32 200 + %tmp17 = load float, float addrspace(3)* %tmp16, align 4 + %tmp18 = fadd float %tmp15, %tmp17 + %tmp19 = getelementptr inbounds float, float addrspace(3)* %arg, i32 100 + %tmp20 = load float, float addrspace(3)* %tmp19, align 4 + %tmp21 = fadd float %tmp18, %tmp20 + %tmp22 = load float, float addrspace(3)* %arg, align 4 + %tmp23 = fadd float %tmp21, %tmp22 + store float %tmp23, float *%arg1, align 4 + ret void +} + +; CHECK-LABEL: ds_read32_combine_stride_8192: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: ds_read2st64_b32 v[{{[0-9]+:[0-9]+}}], [[BASE]] offset1:32 +; CHECK-DAG: ds_read2st64_b32 v[{{[0-9]+:[0-9]+}}], [[BASE]] offset0:64 offset1:96 +; CHECK-DAG: ds_read2st64_b32 v[{{[0-9]+:[0-9]+}}], [[BASE]] offset0:128 offset1:160 +; CHECK-DAG: ds_read2st64_b32 v[{{[0-9]+:[0-9]+}}], [[BASE]] offset0:192 offset1:224 +define void @ds_read32_combine_stride_8192(float addrspace(3)* nocapture readonly %arg, float *nocapture %arg1) { +bb: + %tmp = load float, float addrspace(3)* %arg, align 4 + %tmp2 = fadd float %tmp, 0.000000e+00 + %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2048 + %tmp4 = load float, float addrspace(3)* %tmp3, align 4 + %tmp5 = fadd float %tmp2, %tmp4 + %tmp6 = getelementptr inbounds float, float addrspace(3)* %arg, i32 4096 + %tmp7 = load float, float addrspace(3)* %tmp6, align 4 + %tmp8 = fadd float %tmp5, %tmp7 + %tmp9 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6144 + %tmp10 = load float, float addrspace(3)* %tmp9, align 4 + %tmp11 = fadd float %tmp8, %tmp10 + %tmp12 = getelementptr inbounds float, float addrspace(3)* %arg, i32 8192 + %tmp13 = load float, float addrspace(3)* %tmp12, align 4 + %tmp14 = fadd float %tmp11, %tmp13 + %tmp15 = getelementptr inbounds float, float addrspace(3)* %arg, i32 10240 + %tmp16 = load float, float addrspace(3)* %tmp15, align 4 + %tmp17 = fadd float %tmp14, %tmp16 + %tmp18 = getelementptr inbounds float, float addrspace(3)* %arg, i32 12288 + %tmp19 = load float, float addrspace(3)* %tmp18, align 4 + %tmp20 = fadd float %tmp17, %tmp19 + %tmp21 = getelementptr inbounds float, float addrspace(3)* %arg, i32 14336 + %tmp22 = load float, float addrspace(3)* %tmp21, align 4 + %tmp23 = fadd float %tmp20, %tmp22 + store float %tmp23, float *%arg1, align 4 + ret void +} + +; CHECK-LABEL: ds_read32_combine_stride_8192_shifted: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 8, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B2:v[0-9]+]], vcc, 0x4008, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B3:v[0-9]+]], vcc, 0x8008, [[BASE]] +; CHECK-DAG: ds_read2st64_b32 v[{{[0-9]+:[0-9]+}}], [[B1]] offset1:32 +; CHECK-DAG: ds_read2st64_b32 v[{{[0-9]+:[0-9]+}}], [[B2]] offset1:32 +; CHECK-DAG: ds_read2st64_b32 v[{{[0-9]+:[0-9]+}}], [[B3]] offset1:32 +define void @ds_read32_combine_stride_8192_shifted(float addrspace(3)* nocapture readonly %arg, float *nocapture %arg1) { +bb: + %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 2 + %tmp2 = load float, float addrspace(3)* %tmp, align 4 + %tmp3 = fadd float %tmp2, 0.000000e+00 + %tmp4 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2050 + %tmp5 = load float, float addrspace(3)* %tmp4, align 4 + %tmp6 = fadd float %tmp3, %tmp5 + %tmp7 = getelementptr inbounds float, float addrspace(3)* %arg, i32 4098 + %tmp8 = load float, float addrspace(3)* %tmp7, align 4 + %tmp9 = fadd float %tmp6, %tmp8 + %tmp10 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6146 + %tmp11 = load float, float addrspace(3)* %tmp10, align 4 + %tmp12 = fadd float %tmp9, %tmp11 + %tmp13 = getelementptr inbounds float, float addrspace(3)* %arg, i32 8194 + %tmp14 = load float, float addrspace(3)* %tmp13, align 4 + %tmp15 = fadd float %tmp12, %tmp14 + %tmp16 = getelementptr inbounds float, float addrspace(3)* %arg, i32 10242 + %tmp17 = load float, float addrspace(3)* %tmp16, align 4 + %tmp18 = fadd float %tmp15, %tmp17 + store float %tmp18, float *%arg1, align 4 + ret void +} + +; CHECK-LABEL: ds_read64_combine_stride_400: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 0x960, [[BASE]] +; CHECK-DAG: ds_read2_b64 v[{{[0-9]+:[0-9]+}}], [[BASE]] offset1:50 +; CHECK-DAG: ds_read2_b64 v[{{[0-9]+:[0-9]+}}], [[BASE]] offset0:100 offset1:150 +; CHECK-DAG: ds_read2_b64 v[{{[0-9]+:[0-9]+}}], [[BASE]] offset0:200 offset1:250 +; CHECK-DAG: ds_read2_b64 v[{{[0-9]+:[0-9]+}}], [[B1]] offset1:50 +define void @ds_read64_combine_stride_400(double addrspace(3)* nocapture readonly %arg, double *nocapture %arg1) { +bb: + %tmp = load double, double addrspace(3)* %arg, align 8 + %tmp2 = fadd double %tmp, 0.000000e+00 + %tmp3 = getelementptr inbounds double, double addrspace(3)* %arg, i32 50 + %tmp4 = load double, double addrspace(3)* %tmp3, align 8 + %tmp5 = fadd double %tmp2, %tmp4 + %tmp6 = getelementptr inbounds double, double addrspace(3)* %arg, i32 100 + %tmp7 = load double, double addrspace(3)* %tmp6, align 8 + %tmp8 = fadd double %tmp5, %tmp7 + %tmp9 = getelementptr inbounds double, double addrspace(3)* %arg, i32 150 + %tmp10 = load double, double addrspace(3)* %tmp9, align 8 + %tmp11 = fadd double %tmp8, %tmp10 + %tmp12 = getelementptr inbounds double, double addrspace(3)* %arg, i32 200 + %tmp13 = load double, double addrspace(3)* %tmp12, align 8 + %tmp14 = fadd double %tmp11, %tmp13 + %tmp15 = getelementptr inbounds double, double addrspace(3)* %arg, i32 250 + %tmp16 = load double, double addrspace(3)* %tmp15, align 8 + %tmp17 = fadd double %tmp14, %tmp16 + %tmp18 = getelementptr inbounds double, double addrspace(3)* %arg, i32 300 + %tmp19 = load double, double addrspace(3)* %tmp18, align 8 + %tmp20 = fadd double %tmp17, %tmp19 + %tmp21 = getelementptr inbounds double, double addrspace(3)* %arg, i32 350 + %tmp22 = load double, double addrspace(3)* %tmp21, align 8 + %tmp23 = fadd double %tmp20, %tmp22 + store double %tmp23, double *%arg1, align 8 + ret void +} + +; CHECK-LABEL: ds_read64_combine_stride_8192_shifted: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 8, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B2:v[0-9]+]], vcc, 0x4008, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B3:v[0-9]+]], vcc, 0x8008, [[BASE]] +; CHECK-DAG: ds_read2st64_b64 v[{{[0-9]+:[0-9]+}}], [[B1]] offset1:16 +; CHECK-DAG: ds_read2st64_b64 v[{{[0-9]+:[0-9]+}}], [[B2]] offset1:16 +; CHECK-DAG: ds_read2st64_b64 v[{{[0-9]+:[0-9]+}}], [[B3]] offset1:16 +define void @ds_read64_combine_stride_8192_shifted(double addrspace(3)* nocapture readonly %arg, double *nocapture %arg1) { +bb: + %tmp = getelementptr inbounds double, double addrspace(3)* %arg, i32 1 + %tmp2 = load double, double addrspace(3)* %tmp, align 8 + %tmp3 = fadd double %tmp2, 0.000000e+00 + %tmp4 = getelementptr inbounds double, double addrspace(3)* %arg, i32 1025 + %tmp5 = load double, double addrspace(3)* %tmp4, align 8 + %tmp6 = fadd double %tmp3, %tmp5 + %tmp7 = getelementptr inbounds double, double addrspace(3)* %arg, i32 2049 + %tmp8 = load double, double addrspace(3)* %tmp7, align 8 + %tmp9 = fadd double %tmp6, %tmp8 + %tmp10 = getelementptr inbounds double, double addrspace(3)* %arg, i32 3073 + %tmp11 = load double, double addrspace(3)* %tmp10, align 8 + %tmp12 = fadd double %tmp9, %tmp11 + %tmp13 = getelementptr inbounds double, double addrspace(3)* %arg, i32 4097 + %tmp14 = load double, double addrspace(3)* %tmp13, align 8 + %tmp15 = fadd double %tmp12, %tmp14 + %tmp16 = getelementptr inbounds double, double addrspace(3)* %arg, i32 5121 + %tmp17 = load double, double addrspace(3)* %tmp16, align 8 + %tmp18 = fadd double %tmp15, %tmp17 + store double %tmp18, double *%arg1, align 8 + ret void +} + +; CHECK-LABEL: ds_write32_combine_stride_400: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 0x320, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B2:v[0-9]+]], vcc, 0x640, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B3:v[0-9]+]], vcc, 0x960, [[BASE]] +; CHECK-DAG: ds_write2_b32 [[BASE]], v{{[0-9]+}}, v{{[0-9]+}} offset1:100 +; CHECK-DAG: ds_write2_b32 [[B1]], v{{[0-9]+}}, v{{[0-9]+}} offset1:100 +; CHECK-DAG: ds_write2_b32 [[B2]], v{{[0-9]+}}, v{{[0-9]+}} offset1:100 +; CHECK-DAG: ds_write2_b32 [[B3]], v{{[0-9]+}}, v{{[0-9]+}} offset1:100 +define void @ds_write32_combine_stride_400(float addrspace(3)* nocapture %arg) { +bb: + store float 1.000000e+00, float addrspace(3)* %arg, align 4 + %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 100 + store float 1.000000e+00, float addrspace(3)* %tmp, align 4 + %tmp1 = getelementptr inbounds float, float addrspace(3)* %arg, i32 200 + store float 1.000000e+00, float addrspace(3)* %tmp1, align 4 + %tmp2 = getelementptr inbounds float, float addrspace(3)* %arg, i32 300 + store float 1.000000e+00, float addrspace(3)* %tmp2, align 4 + %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 400 + store float 1.000000e+00, float addrspace(3)* %tmp3, align 4 + %tmp4 = getelementptr inbounds float, float addrspace(3)* %arg, i32 500 + store float 1.000000e+00, float addrspace(3)* %tmp4, align 4 + %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 600 + store float 1.000000e+00, float addrspace(3)* %tmp5, align 4 + %tmp6 = getelementptr inbounds float, float addrspace(3)* %arg, i32 700 + store float 1.000000e+00, float addrspace(3)* %tmp6, align 4 + ret void +} + +; CHECK-LABEL: ds_write32_combine_stride_400_back: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 0x320, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B2:v[0-9]+]], vcc, 0x640, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B3:v[0-9]+]], vcc, 0x960, [[BASE]] +; CHECK-DAG: ds_write2_b32 [[BASE]], v{{[0-9]+}}, v{{[0-9]+}} offset1:100 +; CHECK-DAG: ds_write2_b32 [[B1]], v{{[0-9]+}}, v{{[0-9]+}} offset1:100 +; CHECK-DAG: ds_write2_b32 [[B2]], v{{[0-9]+}}, v{{[0-9]+}} offset1:100 +; CHECK-DAG: ds_write2_b32 [[B3]], v{{[0-9]+}}, v{{[0-9]+}} offset1:100 +define void @ds_write32_combine_stride_400_back(float addrspace(3)* nocapture %arg) { +bb: + %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 700 + store float 1.000000e+00, float addrspace(3)* %tmp, align 4 + %tmp1 = getelementptr inbounds float, float addrspace(3)* %arg, i32 600 + store float 1.000000e+00, float addrspace(3)* %tmp1, align 4 + %tmp2 = getelementptr inbounds float, float addrspace(3)* %arg, i32 500 + store float 1.000000e+00, float addrspace(3)* %tmp2, align 4 + %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 400 + store float 1.000000e+00, float addrspace(3)* %tmp3, align 4 + %tmp4 = getelementptr inbounds float, float addrspace(3)* %arg, i32 300 + store float 1.000000e+00, float addrspace(3)* %tmp4, align 4 + %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 200 + store float 1.000000e+00, float addrspace(3)* %tmp5, align 4 + %tmp6 = getelementptr inbounds float, float addrspace(3)* %arg, i32 100 + store float 1.000000e+00, float addrspace(3)* %tmp6, align 4 + store float 1.000000e+00, float addrspace(3)* %arg, align 4 + ret void +} + +; CHECK-LABEL: ds_write32_combine_stride_8192: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: ds_write2st64_b32 [[BASE]], v{{[0-9]+}}, v{{[0-9]+}} offset1:32 +; CHECK-DAG: ds_write2st64_b32 [[BASE]], v{{[0-9]+}}, v{{[0-9]+}} offset0:64 offset1:96 +; CHECK-DAG: ds_write2st64_b32 [[BASE]], v{{[0-9]+}}, v{{[0-9]+}} offset0:128 offset1:160 +; CHECK-DAG: ds_write2st64_b32 [[BASE]], v{{[0-9]+}}, v{{[0-9]+}} offset0:192 offset1:224 +define void @ds_write32_combine_stride_8192(float addrspace(3)* nocapture %arg) { +bb: + store float 1.000000e+00, float addrspace(3)* %arg, align 4 + %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 2048 + store float 1.000000e+00, float addrspace(3)* %tmp, align 4 + %tmp1 = getelementptr inbounds float, float addrspace(3)* %arg, i32 4096 + store float 1.000000e+00, float addrspace(3)* %tmp1, align 4 + %tmp2 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6144 + store float 1.000000e+00, float addrspace(3)* %tmp2, align 4 + %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 8192 + store float 1.000000e+00, float addrspace(3)* %tmp3, align 4 + %tmp4 = getelementptr inbounds float, float addrspace(3)* %arg, i32 10240 + store float 1.000000e+00, float addrspace(3)* %tmp4, align 4 + %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 12288 + store float 1.000000e+00, float addrspace(3)* %tmp5, align 4 + %tmp6 = getelementptr inbounds float, float addrspace(3)* %arg, i32 14336 + store float 1.000000e+00, float addrspace(3)* %tmp6, align 4 + ret void +} + +; CHECK-LABEL: ds_write32_combine_stride_8192_shifted: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 4, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B2:v[0-9]+]], vcc, 0x4004, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B3:v[0-9]+]], vcc, 0x8004, [[BASE]] +; CHECK-DAG: ds_write2st64_b32 [[B1]], v{{[0-9]+}}, v{{[0-9]+}} offset1:32 +; CHECK-DAG: ds_write2st64_b32 [[B2]], v{{[0-9]+}}, v{{[0-9]+}} offset1:32 +; CHECK-DAG: ds_write2st64_b32 [[B3]], v{{[0-9]+}}, v{{[0-9]+}} offset1:32 +define void @ds_write32_combine_stride_8192_shifted(float addrspace(3)* nocapture %arg) { +bb: + %tmp = getelementptr inbounds float, float addrspace(3)* %arg, i32 1 + store float 1.000000e+00, float addrspace(3)* %tmp, align 4 + %tmp1 = getelementptr inbounds float, float addrspace(3)* %arg, i32 2049 + store float 1.000000e+00, float addrspace(3)* %tmp1, align 4 + %tmp2 = getelementptr inbounds float, float addrspace(3)* %arg, i32 4097 + store float 1.000000e+00, float addrspace(3)* %tmp2, align 4 + %tmp3 = getelementptr inbounds float, float addrspace(3)* %arg, i32 6145 + store float 1.000000e+00, float addrspace(3)* %tmp3, align 4 + %tmp4 = getelementptr inbounds float, float addrspace(3)* %arg, i32 8193 + store float 1.000000e+00, float addrspace(3)* %tmp4, align 4 + %tmp5 = getelementptr inbounds float, float addrspace(3)* %arg, i32 10241 + store float 1.000000e+00, float addrspace(3)* %tmp5, align 4 + ret void +} + +; CHECK-LABEL: ds_write64_combine_stride_400: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 0x960, [[BASE]] +; CHECK-DAG: ds_write2_b64 [[BASE]], v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}] offset1:50 +; CHECK-DAG: ds_write2_b64 [[BASE]], v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}] offset0:100 offset1:150 +; CHECK-DAG: ds_write2_b64 [[BASE]], v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}] offset0:200 offset1:250 +; CHECK-DAG: ds_write2_b64 [[B1]], v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}] offset1:50 +define void @ds_write64_combine_stride_400(double addrspace(3)* nocapture %arg) { +bb: + store double 1.000000e+00, double addrspace(3)* %arg, align 8 + %tmp = getelementptr inbounds double, double addrspace(3)* %arg, i32 50 + store double 1.000000e+00, double addrspace(3)* %tmp, align 8 + %tmp1 = getelementptr inbounds double, double addrspace(3)* %arg, i32 100 + store double 1.000000e+00, double addrspace(3)* %tmp1, align 8 + %tmp2 = getelementptr inbounds double, double addrspace(3)* %arg, i32 150 + store double 1.000000e+00, double addrspace(3)* %tmp2, align 8 + %tmp3 = getelementptr inbounds double, double addrspace(3)* %arg, i32 200 + store double 1.000000e+00, double addrspace(3)* %tmp3, align 8 + %tmp4 = getelementptr inbounds double, double addrspace(3)* %arg, i32 250 + store double 1.000000e+00, double addrspace(3)* %tmp4, align 8 + %tmp5 = getelementptr inbounds double, double addrspace(3)* %arg, i32 300 + store double 1.000000e+00, double addrspace(3)* %tmp5, align 8 + %tmp6 = getelementptr inbounds double, double addrspace(3)* %arg, i32 350 + store double 1.000000e+00, double addrspace(3)* %tmp6, align 8 + ret void +} + +; CHECK-LABEL: ds_write64_combine_stride_8192_shifted: +; CHECK: s_load_dword [[ARG:s[0-9]+]], s[4:5], 0x0 +; CHECK: v_mov_b32_e32 [[BASE:v[0-9]+]], [[ARG]] +; CHECK-DAG: v_add_i32_e32 [[B1:v[0-9]+]], vcc, 8, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B2:v[0-9]+]], vcc, 0x4008, [[BASE]] +; CHECK-DAG: v_add_i32_e32 [[B3:v[0-9]+]], vcc, 0x8008, [[BASE]] +; CHECK-DAG: ds_write2st64_b64 [[B1]], v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}] offset1:16 +; CHECK-DAG: ds_write2st64_b64 [[B2]], v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}] offset1:16 +; CHECK-DAG: ds_write2st64_b64 [[B3]], v[{{[0-9]+:[0-9]+}}], v[{{[0-9]+:[0-9]+}}] offset1:16 +define void @ds_write64_combine_stride_8192_shifted(double addrspace(3)* nocapture %arg) { +bb: + %tmp = getelementptr inbounds double, double addrspace(3)* %arg, i32 1 + store double 1.000000e+00, double addrspace(3)* %tmp, align 8 + %tmp1 = getelementptr inbounds double, double addrspace(3)* %arg, i32 1025 + store double 1.000000e+00, double addrspace(3)* %tmp1, align 8 + %tmp2 = getelementptr inbounds double, double addrspace(3)* %arg, i32 2049 + store double 1.000000e+00, double addrspace(3)* %tmp2, align 8 + %tmp3 = getelementptr inbounds double, double addrspace(3)* %arg, i32 3073 + store double 1.000000e+00, double addrspace(3)* %tmp3, align 8 + %tmp4 = getelementptr inbounds double, double addrspace(3)* %arg, i32 4097 + store double 1.000000e+00, double addrspace(3)* %tmp4, align 8 + %tmp5 = getelementptr inbounds double, double addrspace(3)* %arg, i32 5121 + store double 1.000000e+00, double addrspace(3)* %tmp5, align 8 + ret void +} -- 2.7.4