/gem5/src/base/ |
H A D | callback.cc | 35 queue::iterator i = callbacks.begin(); 36 queue::iterator end = callbacks.end();
|
H A D | callback.hh | 39 * function that gets called when the callback queue is processed. 56 * queue is executed. 91 typedef std::list<Callback *> queue; typedef in class:CallbackQueue 96 queue callbacks; 103 * Add a callback to the end of the queue 104 * @param callback the callback to be added to the queue 127 * Find out if there are any callbacks in the queue 137 queue::iterator i = callbacks.begin(); 138 queue::iterator end = callbacks.end(); 147 * clear the callback queue [all...] |
H A D | pollevent.cc | 62 : queue(NULL), enabled(true) 71 if (queue) 72 queue->remove(this); 81 if (queue) 82 queue->copy(); 91 if (queue) 92 queue->copy(); 150 event->queue = NULL; 163 if (event->queue) 166 event->queue [all...] |
H A D | pollevent.hh | 50 PollQueue *queue; member in class:PollEvent 61 bool queued() { return queue != 0; }
|
/gem5/src/gpu-compute/ |
H A D | local_memory_pipeline.hh | 39 #include <queue> 64 std::queue<GPUDynInstPtr> &getLMReqFIFO() { return lmIssuedRequests; } 65 std::queue<GPUDynInstPtr> &getLMRespFIFO() { return lmReturnedRequests; } 95 std::queue<GPUDynInstPtr> lmIssuedRequests; 99 std::queue<GPUDynInstPtr> lmReturnedRequests;
|
H A D | global_memory_pipeline.hh | 40 #include <queue> 66 std::queue<GPUDynInstPtr> &getGMStRespFIFO() { return gmReturnedStores; } 67 std::queue<GPUDynInstPtr> &getGMLdRespFIFO() { return gmReturnedLoads; } 152 * is finished, and if so it will be removed from the queue. 164 std::queue<GPUDynInstPtr> gmIssuedRequests; 168 std::queue<GPUDynInstPtr> gmReturnedStores; 172 std::queue<GPUDynInstPtr> gmReturnedLoads;
|
H A D | dispatcher.hh | 41 #include <queue> 76 std::queue<int> execIds; 78 std::queue<int> doneIds;
|
/gem5/src/mem/qos/ |
H A D | q_policy.hh | 57 * within the same QoS priority queue 65 * This factory method is used for generating the queue policy. It takes 68 * the QoSMemCtrlParams, the method will default to a LIFO queue policy. 77 * packet. It is a way for the queue policy to hook into the packet 88 * to be serviced from the packet queue passed as an argument. 90 * @param queue Packet queue 91 * @return Iterator pointing to the packet in the queue to be 94 virtual PacketQueue::iterator selectPacket(PacketQueue* queue) = 0; 124 * @param queue Th [all...] |
H A D | mem_sink.cc | 55 // Resize read and write queue to allocate space 151 "%s Read queue full, not accepting\n", __func__); 158 // QoS priority queue 165 "%s Write queue full, not accepting\n", __func__); 172 // priority queue 237 for (auto queue = (*queue_ptr).rbegin(); 238 queue != (*queue_ptr).rend(); ++queue) { 243 "%s checking %s queue [%d] priority [%d packets]\n", 245 curr_prio, queue [all...] |
/gem5/src/arch/arm/tracers/ |
H A D | tarmac_record.hh | 193 virtual void addInstEntry(std::vector<InstPtr>& queue, 197 virtual void addMemEntry(std::vector<MemPtr>& queue, 201 virtual void addRegEntry(std::vector<RegPtr>& queue, 218 mergeCCEntry(std::vector<RegPtr>& queue, const TarmacContext& tarmCtx) argument 220 // Find all CC Entries and move them at the end of the queue 222 queue.begin(), queue.end(), 226 if (it != queue.end()) { 228 queue.erase(it, queue [all...] |
H A D | tarmac_record.cc | 282 TarmacTracerRecord::addInstEntry(std::vector<InstPtr>& queue, argument 287 queue.push_back( 293 TarmacTracerRecord::addMemEntry(std::vector<MemPtr>& queue, argument 300 queue.push_back( 309 TarmacTracerRecord::addRegEntry(std::vector<RegPtr>& queue, argument 323 queue.push_back( 331 mergeCCEntry<TraceRegEntry>(queue, tarmCtx); 383 TarmacTracerRecord::flushQueues(Queue& queue) argument 387 for (const auto &single_entry : queue) { 391 queue 396 flushQueues(Queue& queue, Args & ... args) argument [all...] |
H A D | tarmac_record_v8.cc | 129 TarmacTracerRecordV8::addInstEntry(std::vector<InstPtr>& queue, argument 134 queue.push_back( 140 TarmacTracerRecordV8::addMemEntry(std::vector<MemPtr>& queue, argument 147 queue.push_back( 156 TarmacTracerRecordV8::addRegEntry(std::vector<RegPtr>& queue, argument 170 queue.push_back( 178 mergeCCEntry<TraceRegEntryV8>(queue, tarmCtx);
|
H A D | tarmac_record_v8.hh | 142 void addInstEntry(std::vector<InstPtr>& queue, const TarmacContext& ptr); 145 void addMemEntry(std::vector<MemPtr>& queue, const TarmacContext& ptr); 148 void addRegEntry(std::vector<RegPtr>& queue, const TarmacContext& ptr);
|
/gem5/src/cpu/minor/ |
H A D | buffers.hh | 43 * Classes for buffer, queue and FIFO behaviour. 50 #include <queue> 388 /** Wrapper for a queue type to act as a pipeline stage input queue. 402 std::deque<ElemType> queue; member in class:Minor::Queue 425 * just discarded. It is assummed that any push into a queue with 432 queue.push_back(data); 434 if (queue.size() > capacity) { 435 warn("%s: No space to push data into queue of capacity" 452 /** Reserve space in the queue fo 571 mutable Queue<ElemType, ReportTraits, BubbleTraits> queue; member in class:Minor::InputBuffer [all...] |
/gem5/src/systemc/tests/systemc/misc/sim_tests/biquad/biquad2/ |
H A D | op_queue.cpp | 43 // Queue management. Assuming a circular queue. 58 queue[tail] = in.read(); 64 cout << "Warning: Data is being lost because queue is full" << endl; 70 out.write(queue[head]); 76 cout << "Warning: No data in queue to be popped" << endl;
|
H A D | op_queue.h | 54 float *queue; //internal variable local 70 queue = new float[queue_size];
|
/gem5/src/cpu/ |
H A D | pc_event.hh | 48 PCEventQueue *queue; member in class:PCEvent 54 virtual ~PCEvent() { if (queue) remove(); } 120 : description(desc), queue(q), evpc(pc) 122 queue->schedule(this); 128 if (!queue) 131 return queue->remove(this);
|
/gem5/src/mem/ |
H A D | dramsim2.hh | 47 #include <queue> 117 std::unordered_map<Addr, std::queue<PacketPtr> > outstandingReads; 118 std::unordered_map<Addr, std::queue<PacketPtr> > outstandingWrites; 123 * the sending queue we need to buffer the response packets.
|
H A D | noncoherent_xbar.hh | 94 /** A normal packet queue used to store responses. */ 95 RespPacketQueue queue; member in class:NoncoherentXBar::NoncoherentXBarSlavePort 101 : QueuedSlavePort(_name, &_xbar, queue, _id), xbar(_xbar), 102 queue(_xbar, *this)
|
/gem5/src/dev/arm/ |
H A D | smmu_v3_proc.hh | 44 #include <queue> 87 std::queue<SMMUProcess *> queue; member in struct:SMMUSemaphore
|
H A D | smmu_v3.hh | 45 #include <queue> 152 std::queue<SMMUAction> packetsToRetry; 153 std::queue<SMMUAction> packetsTableWalkToRetry;
|
H A D | smmu_v3_proc.cc | 149 sem.queue.push(this); 161 if (!sem.queue.empty()) { 162 SMMUProcess *next_proc = sem.queue.front(); 163 sem.queue.pop();
|
/gem5/src/mem/cache/prefetch/ |
H A D | queued.cc | 185 "inserting into prefetch queue.\n", new_pfi.getAddr()); 204 // If the queue is empty, attempt first to fill it with requests 205 // from the queue of missing translations 237 .desc("number of redundant prefetches already in prefetch queue"); 245 .desc("number of prefetches dropped due to prefetch queue size"); 306 QueuedPrefetcher::alreadyInQueue(std::list<DeferredPacket> &queue, argument 311 for (it = queue.begin(); it != queue.end() && !found; it++) { 315 /* If the address is already in the queue, update priority and leave */ 316 if (it != queue 450 addToQueue(std::list<DeferredPacket> &queue, DeferredPacket &dpp) argument [all...] |
/gem5/src/systemc/ext/channel/ |
H A D | sc_event_queue.hh | 33 #include <queue>
|
/gem5/src/cpu/o3/ |
H A D | decode.hh | 46 #include <queue> 230 /** Decode instruction queue. */ 236 /** Fetch instruction queue interface. */ 239 /** Wire to get fetch's output from fetch queue. */ 243 std::queue<DynInstPtr> insts[Impl::MaxThreads]; 246 std::queue<DynInstPtr> skidBuffer[Impl::MaxThreads];
|