/gem5/ext/systemc/src/sysc/utils/ |
H A D | sc_list.cpp | 55 static void* operator new(std::size_t sz) { return sc_mempool::allocate(sz); }
|
H A D | sc_hash.cpp | 63 { return sc_mempool::allocate(sz); }
|
/gem5/src/systemc/tests/tlm/update_original/ |
H A D | update_original.cpp | 89 trans = m_mm.allocate();
|
/gem5/src/mem/cache/ |
H A D | mshr.hh | 156 //!< target list allocate in the cache? 170 /** Set when the response should allocate on fill */ 186 * @param alloc_on_fill Whether the pkt would allocate on a fill 274 * @param alloc_on_fill Whether it should allocate on a fill 418 * @param alloc_on_fill Should the cache allocate a block on fill 420 void allocate(Addr blk_addr, unsigned blk_size, PacketPtr pkt,
|
H A D | noncoherent_cache.cc | 90 // Writeback and WriteClean can allocate and fill even if the 166 pkt->allocate();
|
H A D | base.cc | 281 // specifies the latency to allocate an internal 473 const bool allocate = (writeAllocator && mshr->wasWholeLineWrite) ? local 474 writeAllocator->allocate() : mshr->allocOnFill(); 475 blk = handleFill(pkt, blk, writebacks, allocate); 784 // allocate an MSHR and return it, note 1202 // WriteClean handling is a special case. We can allocate a 1215 // allocate if the block is not present 1218 // a writeback that misses needs to allocate a new block 1336 bool allocate) 1355 blk = allocate 1335 handleFill(PacketPtr pkt, CacheBlk *blk, PacketList &writebacks, bool allocate) argument [all...] |
/gem5/src/systemc/ext/dt/int/ |
H A D | sc_signed.hh | 1101 sc_signed_bitref *result_p = sc_signed_bitref::m_pool.allocate(); 1110 sc_signed_bitref *result_p = sc_signed_bitref::m_pool.allocate(); 1119 sc_signed_bitref *result_p = sc_signed_bitref::m_pool.allocate(); 1128 sc_signed_bitref *result_p = sc_signed_bitref::m_pool.allocate(); 1162 sc_signed_subref *result_p = sc_signed_subref::m_pool.allocate(); 1171 sc_signed_subref *result_p = sc_signed_subref::m_pool.allocate(); 1180 sc_signed_subref *result_p = sc_signed_subref::m_pool.allocate(); 1189 sc_signed_subref *result_p = sc_signed_subref::m_pool.allocate();
|
H A D | sc_unsigned.hh | 1012 sc_unsigned_bitref *result_p = sc_unsigned_bitref::m_pool.allocate(); 1021 sc_unsigned_bitref *result_p = sc_unsigned_bitref::m_pool.allocate(); 1030 sc_unsigned_bitref *result_p = sc_unsigned_bitref::m_pool.allocate(); 1039 sc_unsigned_bitref *result_p = sc_unsigned_bitref::m_pool.allocate(); 1070 sc_unsigned_subref *result_p = sc_unsigned_subref::m_pool.allocate(); 1079 sc_unsigned_subref *result_p = sc_unsigned_subref::m_pool.allocate(); 1088 sc_unsigned_subref *result_p = sc_unsigned_subref::m_pool.allocate(); 1097 sc_unsigned_subref *result_p = sc_unsigned_subref::m_pool.allocate();
|
/gem5/src/mem/ |
H A D | packet.hh | 611 * to allocate the block downstream either as: 840 * packet should allocate its own data. 858 // should we allocate space for data, or not, the express 862 // even if asked to allocate data, if the original packet 870 allocate(); 1070 * will allocate their own data. When a packet reaches the final 1251 allocate() function 1254 // payload, actually allocate space
|
/gem5/src/arch/x86/ |
H A D | pagetable_walker.cc | 526 read->allocate(); 596 read->allocate();
|
/gem5/ext/dsent/ |
H A D | DSENT.cc | 280 Log::allocate("/tmp/dsent.log");
|
/gem5/src/gpu-compute/ |
H A D | dispatcher.cc | 133 pkt->allocate();
|
/gem5/src/systemc/tlm_bridge/ |
H A D | gem5_to_tlm.cc | 88 tlm::tlm_generic_payload *trans = mm.allocate();
|
/gem5/src/learning_gem5/part2/ |
H A D | simple_cache.cc | 325 new_pkt->allocate();
|
/gem5/ext/systemc/src/sysc/kernel/ |
H A D | sc_event.cpp | 498 sc_event_timed::allocate() function in class:sc_core::sc_event_timed
|
/gem5/src/mem/cache/prefetch/ |
H A D | queued.cc | 64 pkt->allocate();
|
/gem5/ext/systemc/src/sysc/datatypes/int/ |
H A D | sc_signed.h | 1200 sc_signed_bitref::m_pool.allocate(); 1209 sc_signed_bitref::m_pool.allocate(); 1218 sc_signed_bitref::m_pool.allocate(); 1227 sc_signed_bitref::m_pool.allocate(); 1260 sc_signed_subref::m_pool.allocate(); 1269 sc_signed_subref::m_pool.allocate(); 1278 sc_signed_subref::m_pool.allocate(); 1287 sc_signed_subref::m_pool.allocate();
|
H A D | sc_unsigned.h | 1102 sc_unsigned_bitref::m_pool.allocate(); 1111 sc_unsigned_bitref::m_pool.allocate(); 1120 sc_unsigned_bitref::m_pool.allocate(); 1129 sc_unsigned_bitref::m_pool.allocate(); 1162 sc_unsigned_subref::m_pool.allocate(); 1171 sc_unsigned_subref::m_pool.allocate(); 1180 sc_unsigned_subref::m_pool.allocate(); 1189 sc_unsigned_subref::m_pool.allocate();
|
/gem5/src/cpu/minor/ |
H A D | fetch1.cc | 232 packet->allocate();
|
H A D | lsq.cc | 1689 ret->allocate(); 1692 // stores otherwise need to allocate for data.
|
/gem5/src/mem/ruby/structures/ |
H A D | CacheMemory.cc | 255 CacheMemory::allocate(Addr address, AbstractCacheEntry *entry, bool touch) function in class:CacheMemory
|
/gem5/src/systemc/tests/include/specialized_signals/ |
H A D | scx_signal_int.h | 411 sc_int_sigref* result_p = sc_int_sigref::m_pool.allocate(); 539 result_p = sc_int_sigref::m_pool.allocate();
|
H A D | scx_signal_uint.h | 570 result_p = sc_uint_sigref::m_pool.allocate(); 775 sc_uint_sigref* result_p = sc_uint_sigref::m_pool.allocate();
|
H A D | scx_signal_signed.h | 678 result_p = sc_signed_sigref::m_pool.allocate(); 877 sc_signed_sigref* result_p = sc_signed_sigref::m_pool.allocate();
|
H A D | scx_signal_unsigned.h | 668 result_p = sc_unsigned_sigref::m_pool.allocate(); 882 sc_unsigned_sigref* result_p = sc_unsigned_sigref::m_pool.allocate();
|