/gem5/src/gpu-compute/ |
H A D | global_memory_pipeline.cc | 176 GlobalMemPipeline::completeRequest(GPUDynInstPtr gpuDynInst) argument 178 if (gpuDynInst->isLoad() || gpuDynInst->isAtomic()) { 181 } else if (gpuDynInst->isStore()) { 187 if (gpuDynInst->isLoad() || gpuDynInst->isAtomic()) { 190 } else if (gpuDynInst->isStore()) { 197 assert(gmOrderedRespBuffer.begin()->first == gpuDynInst->seqNum()); 198 assert(gmOrderedRespBuffer.begin()->second.first == gpuDynInst); 202 gmOrderedRespBuffer.erase(gpuDynInst 207 issueRequest(GPUDynInstPtr gpuDynInst) argument 213 handleResponse(GPUDynInstPtr gpuDynInst) argument [all...] |
H A D | compute_unit.cc | 323 GPUDynInstPtr gpuDynInst = local 327 gpuDynInst->useContinuation = false; 328 injectGlobalMemFence(gpuDynInst, true); 628 GPUDynInstPtr gpuDynInst = sender_state->_gpuDynInst; local 633 computeUnit->wfList[gpuDynInst->simdId][gpuDynInst->wfSlotId]; 648 computeUnit->cu_id, gpuDynInst->simdId, 649 gpuDynInst->wfSlotId, w->barrierCnt); 651 if (gpuDynInst->useContinuation) { 652 assert(!gpuDynInst 743 sendRequest(GPUDynInstPtr gpuDynInst, int index, PacketPtr pkt) argument 923 sendSyncRequest(GPUDynInstPtr gpuDynInst, int index, PacketPtr pkt) argument 941 injectGlobalMemFence(GPUDynInstPtr gpuDynInst, bool kernelLaunch, RequestPtr req) argument 981 GPUDynInstPtr gpuDynInst = sender_state->_gpuDynInst; local 1113 GPUDynInstPtr gpuDynInst = sender_state->_gpuDynInst; local [all...] |
H A D | global_memory_pipeline.hh | 83 void completeRequest(GPUDynInstPtr gpuDynInst); 89 void issueRequest(GPUDynInstPtr gpuDynInst); 98 void handleResponse(GPUDynInstPtr gpuDynInst);
|
H A D | gpu_static_inst.hh | 76 virtual void execute(GPUDynInstPtr gpuDynInst) = 0; 88 GPUDynInstPtr gpuDynInst) = 0; 203 initiateAcc(GPUDynInstPtr gpuDynInst) argument 210 completeAcc(GPUDynInstPtr gpuDynInst) argument 225 execLdAcq(GPUDynInstPtr gpuDynInst) argument 231 execSt(GPUDynInstPtr gpuDynInst) argument 237 execAtomic(GPUDynInstPtr gpuDynInst) argument 243 execAtomicAcq(GPUDynInstPtr gpuDynInst) argument 274 execute(GPUDynInstPtr gpuDynInst) override 294 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) overrid [all...] |
H A D | lds_state.cc | 109 GPUDynInstPtr gpuDynInst = senderState->getMemInst(); local 111 return countBankConflicts(gpuDynInst, bankAccesses); 116 LdsState::countBankConflicts(GPUDynInstPtr gpuDynInst, argument 138 if (gpuDynInst->exec_mask[(i*numBanks)+j]) { 139 addr_array[j] = gpuDynInst->addr[(i*numBanks)+j]; 145 if (gpuDynInst->isLoad() || gpuDynInst->isStore()) { 289 GPUDynInstPtr gpuDynInst = ss->getMemInst(); local 291 gpuDynInst->initiateAcc(gpuDynInst); [all...] |
H A D | gpu_dyn_inst.cc | 70 GPUDynInst::execute(GPUDynInstPtr gpuDynInst) argument 72 _staticInst->execute(gpuDynInst); 112 GPUDynInst::getRegisterIndex(int operandIdx, GPUDynInstPtr gpuDynInst) argument 114 return _staticInst->getRegisterIndex(operandIdx, gpuDynInst); 155 GPUDynInst::initiateAcc(GPUDynInstPtr gpuDynInst) argument 160 _staticInst->initiateAcc(gpuDynInst); 165 GPUDynInst::completeAcc(GPUDynInstPtr gpuDynInst) argument 167 _staticInst->completeAcc(gpuDynInst);
|
H A D | compute_unit.hh | 282 template<typename c0, typename c1> void doSmReturn(GPUDynInstPtr gpuDynInst); 284 void sendRequest(GPUDynInstPtr gpuDynInst, int index, PacketPtr pkt); 285 void sendSyncRequest(GPUDynInstPtr gpuDynInst, int index, PacketPtr pkt); 286 void injectGlobalMemFence(GPUDynInstPtr gpuDynInst, 328 void updateInstStats(GPUDynInstPtr gpuDynInst); 397 sendToLds(GPUDynInstPtr gpuDynInst) __attribute__((warn_unused_result)); 437 SenderState(GPUDynInstPtr gpuDynInst, PortID _port_index, argument 439 : _gpuDynInst(gpuDynInst), 543 SenderState(GPUDynInstPtr gpuDynInst, PortID port_index) argument 544 : _gpuDynInst(gpuDynInst), portInde 631 SenderState(GPUDynInstPtr gpuDynInst) argument [all...] |
H A D | gpu_dyn_inst.hh | 203 void execute(GPUDynInstPtr gpuDynInst); 210 int getRegisterIndex(int operandIdx, GPUDynInstPtr gpuDynInst); 273 void initiateAcc(GPUDynInstPtr gpuDynInst); 277 void completeAcc(GPUDynInstPtr gpuDynInst);
|
H A D | fetch_unit.cc | 299 GPUDynInstPtr gpuDynInst = local 303 wavefront->instructionBuffer.push_back(gpuDynInst);
|
H A D | lds_state.hh | 248 countBankConflicts(GPUDynInstPtr gpuDynInst,
|
/gem5/src/arch/hsail/insts/ |
H A D | mem.hh | 150 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) override 180 void execute(GPUDynInstPtr gpuDynInst); 382 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) override 436 initiateAcc(GPUDynInstPtr gpuDynInst) override 440 gpuDynInst->statusBitVector = gpuDynInst->exec_mask; 443 for (int i = 0; i < gpuDynInst->computeUnit()->wfSize(); ++i) 444 if (gpuDynInst->exec_mask[i]) 445 gpuDynInst->statusVector.push_back(num_dest_operands); 447 gpuDynInst 1029 execSt(gpuDynInst); variable 1443 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) argument 1503 execAtomic(gpuDynInst); variable [all...] |
H A D | branch.hh | 99 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) override 109 void execute(GPUDynInstPtr gpuDynInst) override; 128 BrnInstBase<TargetType>::execute(GPUDynInstPtr gpuDynInst) argument 130 Wavefront *w = gpuDynInst->wavefront(); 188 void execute(GPUDynInstPtr gpuDynInst) override; 229 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) override 260 CbrInstBase<TargetType>::execute(GPUDynInstPtr gpuDynInst) argument 262 Wavefront *w = gpuDynInst->wavefront(); 355 void execute(GPUDynInstPtr gpuDynInst) override; 378 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) overrid 402 execute(GPUDynInstPtr gpuDynInst) argument [all...] |
H A D | main.cc | 130 Ret::execute(GPUDynInstPtr gpuDynInst) argument 132 Wavefront *w = gpuDynInst->wavefront(); 188 GPUDynInstPtr local_mempacket = gpuDynInst; 201 Barrier::execute(GPUDynInstPtr gpuDynInst) argument 203 Wavefront *w = gpuDynInst->wavefront();
|
H A D | decl.hh | 182 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) argument 320 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) argument 487 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) argument 770 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) override 797 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) override 847 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) argument 922 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) argument 959 void execute(GPUDynInstPtr gpuDynInst); 976 void execute(GPUDynInstPtr gpuDynInst); 1111 initiateAcc(GPUDynInstPtr gpuDynInst) argument 1118 execute(GPUDynInstPtr gpuDynInst) argument 1250 getRegisterIndex(int operandIndex, GPUDynInstPtr gpuDynInst) argument 1256 execute(GPUDynInstPtr gpuDynInst) argument [all...] |
H A D | mem_impl.hh | 56 LdaInst<DestDataType, AddrRegOperandType>::execute(GPUDynInstPtr gpuDynInst) argument 58 Wavefront *w = gpuDynInst->wavefront(); 165 AddrRegOperandType>::execute(GPUDynInstPtr gpuDynInst) argument 167 Wavefront *w = gpuDynInst->wavefront(); 213 GPUDynInstPtr m = gpuDynInst; 351 AddrRegOperandType>::execute(GPUDynInstPtr gpuDynInst) argument 353 Wavefront *w = gpuDynInst->wavefront(); 377 GPUDynInstPtr m = gpuDynInst; 545 HasDst>::execute(GPUDynInstPtr gpuDynInst) argument 549 Wavefront *w = gpuDynInst [all...] |
H A D | pseudo_inst.cc | 80 Call::execPseudoInst(Wavefront *w, GPUDynInstPtr gpuDynInst) argument 145 MagicAtomicNRAddGlobalU32Reg(w, gpuDynInst); 149 MagicAtomicNRAddGroupU32Reg(w, gpuDynInst); 153 MagicLoadGlobalU32Reg(w, gpuDynInst); 620 Call::MagicAtomicNRAddGlobalU32Reg(Wavefront *w, GPUDynInstPtr gpuDynInst) argument 622 GPUDynInstPtr m = gpuDynInst; 661 Call::MagicAtomicNRAddGroupU32Reg(Wavefront *w, GPUDynInstPtr gpuDynInst) argument 663 GPUDynInstPtr m = gpuDynInst; 701 Call::MagicLoadGlobalU32Reg(Wavefront *w, GPUDynInstPtr gpuDynInst) argument 703 GPUDynInstPtr m = gpuDynInst; [all...] |
/gem5/src/arch/hsail/ |
H A D | gpu_isa.hh | 70 advancePC(uint32_t old_pc, GPUDynInstPtr gpuDynInst) argument
|