/gem5/src/mem/ruby/network/garnet2.0/ |
H A D | flit.cc | 53 m_type = HEAD_TAIL_; 57 m_type = HEAD_; 59 m_type = TAIL_; 61 m_type = BODY_; 70 out << "Type=" << m_type << " "; local
|
H A D | NetworkLink.hh | 57 void setType(link_type type) { m_type = type; } 58 link_type getType() { return m_type; } 77 link_type m_type; member in class:NetworkLink
|
H A D | flit.hh | 61 flit_type get_type() { return m_type; } 109 flit_type m_type; member in class:flit
|
H A D | NetworkLink.cc | 40 m_type(NUM_LINK_TYPES_),
|
/gem5/src/mem/ruby/system/ |
H A D | GPUCoalescer.cc | 350 pkt->req->getPaddr(), i->second->m_type); 366 pkt->req->getPaddr(), i->second->m_type); 396 if ((srequest->m_type == RubyRequestType_ST) || 397 (srequest->m_type == RubyRequestType_RMW_Read) || 398 (srequest->m_type == RubyRequestType_RMW_Write) || 399 (srequest->m_type == RubyRequestType_Load_Linked) || 400 (srequest->m_type == RubyRequestType_Store_Conditional) || 401 (srequest->m_type == RubyRequestType_Locked_RMW_Read) || 402 (srequest->m_type == RubyRequestType_Locked_RMW_Write)) { 420 if (request->m_type [all...] |
H A D | CacheRecorder.cc | 43 << m_type << ", Time: " << m_time << "]"; 118 if (traceRecord->m_type == RubyRequestType_LD) { 123 } else if (traceRecord->m_type == RubyRequestType_IFETCH) { 161 rec->m_type = type;
|
H A D | Sequencer.cc | 275 if (request->m_type == RubyRequestType_Store_Conditional) { 294 } else if (request->m_type == RubyRequestType_Load_Linked) { 373 assert((request->m_type == RubyRequestType_ST) || 374 (request->m_type == RubyRequestType_ATOMIC) || 375 (request->m_type == RubyRequestType_RMW_Read) || 376 (request->m_type == RubyRequestType_RMW_Write) || 377 (request->m_type == RubyRequestType_Load_Linked) || 378 (request->m_type == RubyRequestType_Store_Conditional) || 379 (request->m_type == RubyRequestType_Locked_RMW_Read) || 380 (request->m_type [all...] |
H A D | CacheRecorder.hh | 60 RubyRequestType m_type; member in class:TraceRecord
|
H A D | Sequencer.hh | 46 RubyRequestType m_type; member in struct:SequencerRequest 51 : pkt(_pkt), m_type(_m_type), issue_time(_issue_time)
|
H A D | GPUCoalescer.hh | 67 RubyRequestType m_type; member in struct:GPUCoalescerRequest 72 : pkt(_pkt), m_type(_m_type), issue_time(_issue_time)
|
/gem5/src/mem/ruby/structures/ |
H A D | Prefetcher.cc | 257 m_controller->enqueuePrefetch(line_addr, stream->m_type); 299 mystream->m_type = type; 321 m_controller->enqueuePrefetch(line_addr, m_array[index].m_type);
|
H A D | Prefetcher.hh | 73 RubyRequestType m_type; member in class:PrefetchEntry
|
/gem5/src/arch/hsail/insts/ |
H A D | mem.hh | 53 MemInst(Enums::MemType m_type) argument 55 if (m_type == Enums::M_U64 || 56 m_type == Enums::M_S64 || 57 m_type == Enums::M_F64) { 59 } else if (m_type == Enums::M_U32 || 60 m_type == Enums::M_S32 || 61 m_type == Enums::M_F32) { 63 } else if (m_type == Enums::M_U16 || 64 m_type == Enums::M_S16 || 65 m_type [all...] |
H A D | pseudo_inst.cc | 636 m->m_type = U32::memType; 676 m->m_type = U32::memType; 712 m->m_type = U32::memType; //MemDataType::memType;
|
H A D | mem_impl.hh | 217 m->m_type = MemDataType::memType; 401 m->m_type = OperationType::memType; 570 m->m_type = DataType::memType;
|
/gem5/src/gpu-compute/ |
H A D | gpu_dyn_inst.hh | 235 Enums::MemType m_type; member in class:GPUDynInst
|
/gem5/ext/pybind11/include/pybind11/ |
H A D | pytypes.h | 327 PyErr_Fetch(&m_type.ptr(), &m_value.ptr(), &m_trace.ptr()); 338 void restore() { PyErr_Restore(m_type.release().ptr(), m_value.release().ptr(), m_trace.release().ptr()); } 347 bool matches(handle exc) const { return PyErr_GivenExceptionMatches(m_type.ptr(), exc.ptr()); } 349 const object& type() const { return m_type; } 354 object m_type, m_value, m_trace; member in class:error_already_set
|
H A D | pybind11.h | 2015 if (m_type) { 2018 m_type.release().dec_ref();
|