/gem5/src/mem/ruby/network/simple/ |
H A D | Throttle.cc | 75 for (int vnet = 0; vnet < in_vec.size(); ++vnet) { 76 MessageBuffer *in_ptr = in_vec[vnet]; 77 MessageBuffer *out_ptr = out_vec[vnet]; 92 Throttle::operateVnet(int vnet, int &bw_remaining, bool &schedule_wakeup, argument 99 assert(m_units_remaining[vnet] >= 0); 103 m_units_remaining[vnet] > 0) && 107 if (m_units_remaining[vnet] == 0 && in->isReady(current_time)) { 111 m_units_remaining[vnet] [all...] |
H A D | PerfectSwitch.cc | 105 PerfectSwitch::operateVnet(int vnet) argument 114 if (m_pending_message_count[vnet] > 0) { 124 if (m_in[incoming].size() <= vnet) { 128 MessageBuffer *buffer = m_in[incoming][vnet]; 133 operateMessageBuffer(buffer, incoming, vnet); 140 int vnet) 170 if (m_network_ptr->isVNetOrdered(vnet)) { 225 if (!m_out[outgoing][vnet]->areNSlotsAvailable(1, current_time)) 229 "outgoing: %d, vnet: %d, enough: %d\n", 230 outgoing, vnet, enoug 139 operateMessageBuffer(MessageBuffer *buffer, int incoming, int vnet) argument [all...] |
H A D | SimpleNetwork.hh | 59 bool isVNetOrdered(int vnet) const { return m_ordered[vnet]; }
|
H A D | PerfectSwitch.hh | 88 void operateVnet(int vnet); 89 void operateMessageBuffer(MessageBuffer *b, int incoming, int vnet);
|
H A D | Throttle.hh | 86 void operateVnet(int vnet, int &bw_remainin, bool &schedule_wakeup,
|
/gem5/src/mem/ruby/network/garnet2.0/ |
H A D | GarnetNetwork.hh | 79 bool isVNetOrdered(int vnet) const { return m_ordered[vnet]; } 83 int vnet = vc/getVCsPerVnet(); local 84 return m_vnet_type[vnet]; 110 void increment_injected_packets(int vnet) { m_packets_injected[vnet]++; } argument 111 void increment_received_packets(int vnet) { m_packets_received[vnet]++; } argument 114 increment_packet_network_latency(Cycles latency, int vnet) argument 116 m_packet_network_latency[vnet] 120 increment_packet_queueing_latency(Cycles latency, int vnet) argument 125 increment_injected_flits(int vnet) argument 126 increment_received_flits(int vnet) argument 129 increment_flit_network_latency(Cycles latency, int vnet) argument 135 increment_flit_queueing_latency(Cycles latency, int vnet) argument [all...] |
H A D | NetworkInterface.cc | 69 m_vc_allocator.resize(m_virtual_networks); // 1 allocator per vnet 145 int vnet = t_flit->get_vnet(); local 148 m_net_ptr->increment_received_flits(vnet); 155 m_net_ptr->increment_flit_network_latency(network_delay, vnet); 156 m_net_ptr->increment_flit_queueing_latency(queueing_delay, vnet); 159 m_net_ptr->increment_received_packets(vnet); 160 m_net_ptr->increment_packet_network_latency(network_delay, vnet); 161 m_net_ptr->increment_packet_queueing_latency(queueing_delay, vnet); 189 for (int vnet = 0; vnet < inNode_pt 213 int vnet = t_flit->get_vnet(); local 290 int vnet = stallFlit->get_vnet(); local 327 flitisizeMessage(MsgPtr msg_ptr, int vnet) argument 406 calculateVC(int vnet) argument [all...] |
H A D | InputUnit.hh | 143 double get_buf_read_activity(unsigned int vnet) const 144 { return m_num_buffer_reads[vnet]; } 145 double get_buf_write_activity(unsigned int vnet) const 146 { return m_num_buffer_writes[vnet]; }
|
H A D | OutputUnit.cc | 98 OutputUnit::has_free_vc(int vnet) argument 100 int vc_base = vnet*m_vc_per_vnet; 111 OutputUnit::select_free_vc(int vnet) argument 113 int vc_base = vnet*m_vc_per_vnet;
|
H A D | CommonTypes.hh | 52 int vnet; member in struct:RouteInfo
|
H A D | flit.cc | 37 flit::flit(int id, int vc, int vnet, RouteInfo route, int size, argument 46 m_vnet = vnet;
|
H A D | NetworkInterface.hh | 111 bool flitisizeMessage(MsgPtr msg_ptr, int vnet); 112 int calculateVC(int vnet);
|
H A D | OutputUnit.hh | 61 bool has_free_vc(int vnet); 62 int select_free_vc(int vnet);
|
H A D | RoutingUnit.cc | 69 RoutingUnit::lookupRoutingTable(int vnet, NetDest msg_destination) argument 72 // For ordered vnet, just choose the first 74 // For unordered vnet, randomly choose any of the links 111 if (!(m_router->get_net_ptr())->isVNetOrdered(vnet)) 150 outport = lookupRoutingTable(route.vnet, route.net_dest); 161 lookupRoutingTable(route.vnet, route.net_dest); break; 168 lookupRoutingTable(route.vnet, route.net_dest); break;
|
H A D | SwitchAllocator.cc | 279 * within an ordered vnet 288 // Check if ordering violated (in ordered vnet) 290 int vnet = get_vnet(invc); local 299 if (m_output_unit[outport]->has_free_vc(vnet)) { 317 if ((m_router->get_net_ptr())->isVNetOrdered(vnet)) { 324 int vc_base = vnet*m_vc_per_vnet; 373 int vnet = invc/m_vc_per_vnet; local 374 assert(vnet < m_router->get_num_vnets()); 375 return vnet;
|
H A D | InputUnit.cc | 117 int vnet = vc/m_vc_per_vnet; local 120 m_num_buffer_writes[vnet]++; 121 m_num_buffer_reads[vnet]++;
|
H A D | RoutingUnit.hh | 59 int lookupRoutingTable(int vnet, NetDest net_dest);
|
H A D | flit.hh | 48 flit(int id, int vc, int vnet, RouteInfo route, int size,
|
/gem5/src/mem/ruby/slicc_interface/ |
H A D | Message.hh | 103 int getVnet() const { return vnet; } 104 void setVnet(int net) { vnet = net; } 114 int vnet; member in class:Message
|
/gem5/src/mem/slicc/symbols/ |
H A D | StateMachine.py | 571 # set for maintaining the vnet, direction pairs already seen for this 573 # multiple message buffers being mapped to the same vnet. 586 vnet = var["virtual_network"] 589 assert (vnet, network) not in vnet_dir_set 590 vnet_dir_set.add((vnet,network)) 593 m_net_ptr->set${network}NetQueue(m_version + base, $vid->getOrdered(), $vnet,
|