Throttle.cc revision 6154
1 2/* 3 * Copyright (c) 1999-2008 Mark D. Hill and David A. Wood 4 * All rights reserved. 5 * 6 * Redistribution and use in source and binary forms, with or without 7 * modification, are permitted provided that the following conditions are 8 * met: redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer; 10 * redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution; 13 * neither the name of the copyright holders nor the names of its 14 * contributors may be used to endorse or promote products derived from 15 * this software without specific prior written permission. 16 * 17 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 18 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 19 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 20 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 21 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 22 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT 23 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 27 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 28 */ 29 30/* 31 * $Id$ 32 * 33 * Description: see Throttle.h 34 * 35 */ 36 37#include "mem/ruby/network/simple/Throttle.hh" 38#include "mem/ruby/buffers/MessageBuffer.hh" 39#include "mem/ruby/network/Network.hh" 40#include "mem/ruby/system/System.hh" 41#include "mem/ruby/slicc_interface/NetworkMessage.hh" 42#include "mem/protocol/Protocol.hh" 43 44const int HIGH_RANGE = 256; 45const int ADJUST_INTERVAL = 50000; 46const int MESSAGE_SIZE_MULTIPLIER = 1000; 47//const int BROADCAST_SCALING = 4; // Have a 16p system act like a 64p systems 48const int BROADCAST_SCALING = 1; 49const int PRIORITY_SWITCH_LIMIT = 128; 50 51static int network_message_to_size(NetworkMessage* net_msg_ptr); 52 53extern std::ostream * debug_cout_ptr; 54 55Throttle::Throttle(int sID, NodeID node, int link_latency, int link_bandwidth_multiplier) 56{ 57 init(node, link_latency, link_bandwidth_multiplier); 58 m_sID = sID; 59} 60 61Throttle::Throttle(NodeID node, int link_latency, int link_bandwidth_multiplier) 62{ 63 init(node, link_latency, link_bandwidth_multiplier); 64 m_sID = 0; 65} 66 67void Throttle::init(NodeID node, int link_latency, int link_bandwidth_multiplier) 68{ 69 m_node = node; 70 m_vnets = 0; 71 72 ASSERT(link_bandwidth_multiplier > 0); 73 m_link_bandwidth_multiplier = link_bandwidth_multiplier; 74 m_link_latency = link_latency; 75 76 m_bash_counter = HIGH_RANGE; 77 m_bandwidth_since_sample = 0; 78 m_last_bandwidth_sample = 0; 79 m_wakeups_wo_switch = 0; 80 clearStats(); 81} 82 83void Throttle::clear() 84{ 85 for (int counter = 0; counter < m_vnets; counter++) { 86 m_in[counter]->clear(); 87 m_out[counter]->clear(); 88 } 89} 90 91void Throttle::addLinks(const Vector<MessageBuffer*>& in_vec, const Vector<MessageBuffer*>& out_vec) 92{ 93 assert(in_vec.size() == out_vec.size()); 94 for (int i=0; i<in_vec.size(); i++) { 95 addVirtualNetwork(in_vec[i], out_vec[i]); 96 } 97 98 m_message_counters.setSize(MessageSizeType_NUM); 99 for (int i=0; i<MessageSizeType_NUM; i++) { 100 m_message_counters[i].setSize(in_vec.size()); 101 for (int j=0; j<m_message_counters[i].size(); j++) { 102 m_message_counters[i][j] = 0; 103 } 104 } 105 106 if (g_PRINT_TOPOLOGY) { 107 m_out_link_vec.insertAtBottom(out_vec); 108 } 109} 110 111void Throttle::addVirtualNetwork(MessageBuffer* in_ptr, MessageBuffer* out_ptr) 112{ 113 m_units_remaining.insertAtBottom(0); 114 m_in.insertAtBottom(in_ptr); 115 m_out.insertAtBottom(out_ptr); 116 117 // Set consumer and description 118 m_in[m_vnets]->setConsumer(this); 119 string desc = "[Queue to Throttle " + NodeIDToString(m_sID) + " " + NodeIDToString(m_node) + "]"; 120 m_in[m_vnets]->setDescription(desc); 121 m_vnets++; 122} 123 124void Throttle::wakeup() 125{ 126 // Limits the number of message sent to a limited number of bytes/cycle. 127 assert(getLinkBandwidth() > 0); 128 int bw_remaining = getLinkBandwidth(); 129 130 // Give the highest numbered link priority most of the time 131 m_wakeups_wo_switch++; 132 int highest_prio_vnet = m_vnets-1; 133 int lowest_prio_vnet = 0; 134 int counter = 1; 135 bool schedule_wakeup = false; 136 137 // invert priorities to avoid starvation seen in the component network 138 if (m_wakeups_wo_switch > PRIORITY_SWITCH_LIMIT) { 139 m_wakeups_wo_switch = 0; 140 highest_prio_vnet = 0; 141 lowest_prio_vnet = m_vnets-1; 142 counter = -1; 143 } 144 145 for (int vnet = highest_prio_vnet; (vnet*counter) >= (counter*lowest_prio_vnet); vnet -= counter) { 146 147 assert(m_out[vnet] != NULL); 148 assert(m_in[vnet] != NULL); 149 assert(m_units_remaining[vnet] >= 0); 150 151 while ((bw_remaining > 0) && ((m_in[vnet]->isReady()) || (m_units_remaining[vnet] > 0)) && m_out[vnet]->areNSlotsAvailable(1)) { 152 153 // See if we are done transferring the previous message on this virtual network 154 if (m_units_remaining[vnet] == 0 && m_in[vnet]->isReady()) { 155 156 // Find the size of the message we are moving 157 MsgPtr msg_ptr = m_in[vnet]->peekMsgPtr(); 158 NetworkMessage* net_msg_ptr = dynamic_cast<NetworkMessage*>(msg_ptr.ref()); 159 m_units_remaining[vnet] += network_message_to_size(net_msg_ptr); 160 161 DEBUG_NEWLINE(NETWORK_COMP,HighPrio); 162 DEBUG_MSG(NETWORK_COMP,HighPrio,"throttle: " + int_to_string(m_node) 163 + " my bw " + int_to_string(getLinkBandwidth()) 164 + " bw spent enqueueing net msg " + int_to_string(m_units_remaining[vnet]) 165 + " time: " + int_to_string(g_eventQueue_ptr->getTime()) + "."); 166 167 // Move the message 168 m_out[vnet]->enqueue(m_in[vnet]->peekMsgPtr(), m_link_latency); 169 m_in[vnet]->pop(); 170 171 // Count the message 172 m_message_counters[net_msg_ptr->getMessageSize()][vnet]++; 173 174 DEBUG_MSG(NETWORK_COMP,LowPrio,*m_out[vnet]); 175 DEBUG_NEWLINE(NETWORK_COMP,HighPrio); 176 } 177 178 // Calculate the amount of bandwidth we spent on this message 179 int diff = m_units_remaining[vnet] - bw_remaining; 180 m_units_remaining[vnet] = max(0, diff); 181 bw_remaining = max(0, -diff); 182 } 183 184 if ((bw_remaining > 0) && ((m_in[vnet]->isReady()) || (m_units_remaining[vnet] > 0)) && !m_out[vnet]->areNSlotsAvailable(1)) { 185 DEBUG_MSG(NETWORK_COMP,LowPrio,vnet); 186 schedule_wakeup = true; // schedule me to wakeup again because I'm waiting for my output queue to become available 187 } 188 } 189 190 // We should only wake up when we use the bandwidth 191 // assert(bw_remaining != getLinkBandwidth()); // This is only mostly true 192 193 // Record that we used some or all of the link bandwidth this cycle 194 double ratio = 1.0-(double(bw_remaining)/double(getLinkBandwidth())); 195 // If ratio = 0, we used no bandwidth, if ratio = 1, we used all 196 linkUtilized(ratio); 197 198 // Sample the link bandwidth utilization over a number of cycles 199 int bw_used = getLinkBandwidth()-bw_remaining; 200 m_bandwidth_since_sample += bw_used; 201 202 // FIXME - comment out the bash specific code for faster performance 203 // Start Bash code 204 // Update the predictor 205 Time current_time = g_eventQueue_ptr->getTime(); 206 while ((current_time - m_last_bandwidth_sample) > ADJUST_INTERVAL) { 207 double utilization = m_bandwidth_since_sample/double(ADJUST_INTERVAL * getLinkBandwidth()); 208 209 if (utilization > g_bash_bandwidth_adaptive_threshold) { 210 // Used more bandwidth 211 m_bash_counter++; 212 } else { 213 // Used less bandwidth 214 m_bash_counter--; 215 } 216 217 // Make sure we don't overflow 218 m_bash_counter = min(HIGH_RANGE, m_bash_counter); 219 m_bash_counter = max(0, m_bash_counter); 220 221 // Reset samples 222 m_last_bandwidth_sample += ADJUST_INTERVAL; 223 m_bandwidth_since_sample = 0; 224 } 225 // End Bash code 226 227 if ((bw_remaining > 0) && !schedule_wakeup) { 228 // We have extra bandwidth and our output buffer was available, so we must not have anything else to do until another message arrives. 229 DEBUG_MSG(NETWORK_COMP,LowPrio,*this); 230 DEBUG_MSG(NETWORK_COMP,LowPrio,"not scheduled again"); 231 } else { 232 DEBUG_MSG(NETWORK_COMP,LowPrio,*this); 233 DEBUG_MSG(NETWORK_COMP,LowPrio,"scheduled again"); 234 // We are out of bandwidth for this cycle, so wakeup next cycle and continue 235 g_eventQueue_ptr->scheduleEvent(this, 1); 236 } 237} 238 239bool Throttle::broadcastBandwidthAvailable(int rand) const 240{ 241 bool result = !(m_bash_counter > ((HIGH_RANGE/4) + (rand % (HIGH_RANGE/2)))); 242 return result; 243} 244 245void Throttle::printStats(ostream& out) const 246{ 247 out << "utilized_percent: " << getUtilization() << endl; 248} 249 250void Throttle::clearStats() 251{ 252 m_ruby_start = g_eventQueue_ptr->getTime(); 253 m_links_utilized = 0.0; 254 255 for (int i=0; i<m_message_counters.size(); i++) { 256 for (int j=0; j<m_message_counters[i].size(); j++) { 257 m_message_counters[i][j] = 0; 258 } 259 } 260} 261 262void Throttle::printConfig(ostream& out) const 263{ 264 265} 266 267double Throttle::getUtilization() const 268{ 269 return (100.0 * double(m_links_utilized)) / (double(g_eventQueue_ptr->getTime()-m_ruby_start)); 270} 271 272void Throttle::print(ostream& out) const 273{ 274 out << "[Throttle: " << m_sID << " " << m_node << " bw: " << getLinkBandwidth() << "]"; 275} 276 277// Helper function 278 279static 280int network_message_to_size(NetworkMessage* net_msg_ptr) 281{ 282 assert(net_msg_ptr != NULL); 283 284 // Artificially increase the size of broadcast messages 285 if (BROADCAST_SCALING > 1) { 286 if (net_msg_ptr->getDestination().isBroadcast()) { 287 return (MessageSizeType_to_int(net_msg_ptr->getMessageSize()) * MESSAGE_SIZE_MULTIPLIER * BROADCAST_SCALING); 288 } 289 } 290 return (MessageSizeType_to_int(net_msg_ptr->getMessageSize()) * MESSAGE_SIZE_MULTIPLIER); 291} 292