GPU_VIPER-TCC.sm revision 14184
1/* 2 * Copyright (c) 2010-2015 Advanced Micro Devices, Inc. 3 * All rights reserved. 4 * 5 * For use for simulation and test purposes only 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions are met: 9 * 10 * 1. Redistributions of source code must retain the above copyright notice, 11 * this list of conditions and the following disclaimer. 12 * 13 * 2. Redistributions in binary form must reproduce the above copyright notice, 14 * this list of conditions and the following disclaimer in the documentation 15 * and/or other materials provided with the distribution. 16 * 17 * 3. Neither the name of the copyright holder nor the names of its 18 * contributors may be used to endorse or promote products derived from this 19 * software without specific prior written permission. 20 * 21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" 22 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 23 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 24 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE 25 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 26 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 27 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 28 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 29 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 30 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 31 * POSSIBILITY OF SUCH DAMAGE. 32 * 33 * Author: Blake Hechtman 34 */ 35 36machine(MachineType:TCC, "TCC Cache") 37 : CacheMemory * L2cache; 38 bool WB; /*is this cache Writeback?*/ 39 Cycles l2_request_latency := 50; 40 Cycles l2_response_latency := 20; 41 42 // From the TCPs or SQCs 43 MessageBuffer * requestFromTCP, network="From", virtual_network="1", vnet_type="request"; 44 // To the Cores. TCC deals only with TCPs/SQCs. 45 MessageBuffer * responseToCore, network="To", virtual_network="3", vnet_type="response"; 46 // From the NB 47 MessageBuffer * probeFromNB, network="From", virtual_network="0", vnet_type="request"; 48 MessageBuffer * responseFromNB, network="From", virtual_network="2", vnet_type="response"; 49 // To the NB 50 MessageBuffer * requestToNB, network="To", virtual_network="0", vnet_type="request"; 51 MessageBuffer * responseToNB, network="To", virtual_network="2", vnet_type="response"; 52 MessageBuffer * unblockToNB, network="To", virtual_network="4", vnet_type="unblock"; 53 54 MessageBuffer * triggerQueue; 55 56{ 57 // EVENTS 58 enumeration(Event, desc="TCC Events") { 59 // Requests coming from the Cores 60 RdBlk, desc="RdBlk event"; 61 WrVicBlk, desc="L1 Write Through"; 62 WrVicBlkBack, desc="L1 Write Through(dirty cache)"; 63 Atomic, desc="Atomic Op"; 64 AtomicDone, desc="AtomicOps Complete"; 65 AtomicNotDone, desc="AtomicOps not Complete"; 66 Data, desc="data messgae"; 67 // Coming from this TCC 68 L2_Repl, desc="L2 Replacement"; 69 // Probes 70 PrbInv, desc="Invalidating probe"; 71 // Coming from Memory Controller 72 WBAck, desc="writethrough ack from memory"; 73 } 74 75 // STATES 76 state_declaration(State, desc="TCC State", default="TCC_State_I") { 77 M, AccessPermission:Read_Write, desc="Modified(dirty cache only)"; 78 W, AccessPermission:Read_Write, desc="Written(dirty cache only)"; 79 V, AccessPermission:Read_Only, desc="Valid"; 80 I, AccessPermission:Invalid, desc="Invalid"; 81 IV, AccessPermission:Busy, desc="Waiting for Data"; 82 WI, AccessPermission:Busy, desc="Waiting on Writethrough Ack"; 83 A, AccessPermission:Busy, desc="Invalid waiting on atomici Data"; 84 } 85 86 enumeration(RequestType, desc="To communicate stats from transitions to recordStats") { 87 DataArrayRead, desc="Read the data array"; 88 DataArrayWrite, desc="Write the data array"; 89 TagArrayRead, desc="Read the data array"; 90 TagArrayWrite, desc="Write the data array"; 91 } 92 93 94 // STRUCTURES 95 96 structure(Entry, desc="...", interface="AbstractCacheEntry") { 97 State CacheState, desc="cache state"; 98 bool Dirty, desc="Is the data dirty (diff from memory?)"; 99 DataBlock DataBlk, desc="Data for the block"; 100 WriteMask writeMask, desc="Dirty byte mask"; 101 } 102 103 structure(TBE, desc="...") { 104 State TBEState, desc="Transient state"; 105 DataBlock DataBlk, desc="data for the block"; 106 bool Dirty, desc="Is the data dirty?"; 107 bool Shared, desc="Victim hit by shared probe"; 108 MachineID From, desc="Waiting for writeback from..."; 109 NetDest Destination, desc="Data destination"; 110 int numAtomics, desc="number remaining atomics"; 111 } 112 113 structure(TBETable, external="yes") { 114 TBE lookup(Addr); 115 void allocate(Addr); 116 void deallocate(Addr); 117 bool isPresent(Addr); 118 } 119 120 TBETable TBEs, template="<TCC_TBE>", constructor="m_number_of_TBEs"; 121 122 void set_cache_entry(AbstractCacheEntry b); 123 void unset_cache_entry(); 124 void set_tbe(TBE b); 125 void unset_tbe(); 126 void wakeUpAllBuffers(); 127 void wakeUpBuffers(Addr a); 128 129 MachineID mapAddressToMachine(Addr addr, MachineType mtype); 130 131 // FUNCTION DEFINITIONS 132 Tick clockEdge(); 133 134 Entry getCacheEntry(Addr addr), return_by_pointer="yes" { 135 return static_cast(Entry, "pointer", L2cache.lookup(addr)); 136 } 137 138 DataBlock getDataBlock(Addr addr), return_by_ref="yes" { 139 return getCacheEntry(addr).DataBlk; 140 } 141 142 bool presentOrAvail(Addr addr) { 143 return L2cache.isTagPresent(addr) || L2cache.cacheAvail(addr); 144 } 145 146 State getState(TBE tbe, Entry cache_entry, Addr addr) { 147 if (is_valid(tbe)) { 148 return tbe.TBEState; 149 } else if (is_valid(cache_entry)) { 150 return cache_entry.CacheState; 151 } 152 return State:I; 153 } 154 155 void setState(TBE tbe, Entry cache_entry, Addr addr, State state) { 156 if (is_valid(tbe)) { 157 tbe.TBEState := state; 158 } 159 160 if (is_valid(cache_entry)) { 161 cache_entry.CacheState := state; 162 } 163 } 164 165 void functionalRead(Addr addr, Packet *pkt) { 166 TBE tbe := TBEs.lookup(addr); 167 if(is_valid(tbe)) { 168 testAndRead(addr, tbe.DataBlk, pkt); 169 } else { 170 functionalMemoryRead(pkt); 171 } 172 } 173 174 int functionalWrite(Addr addr, Packet *pkt) { 175 int num_functional_writes := 0; 176 177 TBE tbe := TBEs.lookup(addr); 178 if(is_valid(tbe)) { 179 num_functional_writes := num_functional_writes + 180 testAndWrite(addr, tbe.DataBlk, pkt); 181 } 182 183 num_functional_writes := num_functional_writes + 184 functionalMemoryWrite(pkt); 185 return num_functional_writes; 186 } 187 188 AccessPermission getAccessPermission(Addr addr) { 189 TBE tbe := TBEs.lookup(addr); 190 if(is_valid(tbe)) { 191 return TCC_State_to_permission(tbe.TBEState); 192 } 193 194 Entry cache_entry := getCacheEntry(addr); 195 if(is_valid(cache_entry)) { 196 return TCC_State_to_permission(cache_entry.CacheState); 197 } 198 199 return AccessPermission:NotPresent; 200 } 201 202 void setAccessPermission(Entry cache_entry, Addr addr, State state) { 203 if (is_valid(cache_entry)) { 204 cache_entry.changePermission(TCC_State_to_permission(state)); 205 } 206 } 207 208 void recordRequestType(RequestType request_type, Addr addr) { 209 if (request_type == RequestType:DataArrayRead) { 210 L2cache.recordRequestType(CacheRequestType:DataArrayRead, addr); 211 } else if (request_type == RequestType:DataArrayWrite) { 212 L2cache.recordRequestType(CacheRequestType:DataArrayWrite, addr); 213 } else if (request_type == RequestType:TagArrayRead) { 214 L2cache.recordRequestType(CacheRequestType:TagArrayRead, addr); 215 } else if (request_type == RequestType:TagArrayWrite) { 216 L2cache.recordRequestType(CacheRequestType:TagArrayWrite, addr); 217 } 218 } 219 220 bool checkResourceAvailable(RequestType request_type, Addr addr) { 221 if (request_type == RequestType:DataArrayRead) { 222 return L2cache.checkResourceAvailable(CacheResourceType:DataArray, addr); 223 } else if (request_type == RequestType:DataArrayWrite) { 224 return L2cache.checkResourceAvailable(CacheResourceType:DataArray, addr); 225 } else if (request_type == RequestType:TagArrayRead) { 226 return L2cache.checkResourceAvailable(CacheResourceType:TagArray, addr); 227 } else if (request_type == RequestType:TagArrayWrite) { 228 return L2cache.checkResourceAvailable(CacheResourceType:TagArray, addr); 229 } else { 230 error("Invalid RequestType type in checkResourceAvailable"); 231 return true; 232 } 233 } 234 235 236 // ** OUT_PORTS ** 237 238 // Three classes of ports 239 // Class 1: downward facing network links to NB 240 out_port(requestToNB_out, CPURequestMsg, requestToNB); 241 out_port(responseToNB_out, ResponseMsg, responseToNB); 242 out_port(unblockToNB_out, UnblockMsg, unblockToNB); 243 244 // Class 2: upward facing ports to GPU cores 245 out_port(responseToCore_out, ResponseMsg, responseToCore); 246 247 out_port(triggerQueue_out, TriggerMsg, triggerQueue); 248 // 249 // request queue going to NB 250 // 251 252 253// ** IN_PORTS ** 254 in_port(triggerQueue_in, TiggerMsg, triggerQueue) { 255 if (triggerQueue_in.isReady(clockEdge())) { 256 peek(triggerQueue_in, TriggerMsg) { 257 TBE tbe := TBEs.lookup(in_msg.addr); 258 Entry cache_entry := getCacheEntry(in_msg.addr); 259 if (tbe.numAtomics == 0) { 260 trigger(Event:AtomicDone, in_msg.addr, cache_entry, tbe); 261 } else { 262 trigger(Event:AtomicNotDone, in_msg.addr, cache_entry, tbe); 263 } 264 } 265 } 266 } 267 268 269 270 in_port(responseFromNB_in, ResponseMsg, responseFromNB) { 271 if (responseFromNB_in.isReady(clockEdge())) { 272 peek(responseFromNB_in, ResponseMsg, block_on="addr") { 273 TBE tbe := TBEs.lookup(in_msg.addr); 274 Entry cache_entry := getCacheEntry(in_msg.addr); 275 if (in_msg.Type == CoherenceResponseType:NBSysResp) { 276 if(presentOrAvail(in_msg.addr)) { 277 trigger(Event:Data, in_msg.addr, cache_entry, tbe); 278 } else { 279 Addr victim := L2cache.cacheProbe(in_msg.addr); 280 trigger(Event:L2_Repl, victim, getCacheEntry(victim), TBEs.lookup(victim)); 281 } 282 } else if (in_msg.Type == CoherenceResponseType:NBSysWBAck) { 283 trigger(Event:WBAck, in_msg.addr, cache_entry, tbe); 284 } else { 285 error("Unexpected Response Message to Core"); 286 } 287 } 288 } 289 } 290 291 // Finally handling incoming requests (from TCP) and probes (from NB). 292 in_port(probeNetwork_in, NBProbeRequestMsg, probeFromNB) { 293 if (probeNetwork_in.isReady(clockEdge())) { 294 peek(probeNetwork_in, NBProbeRequestMsg) { 295 DPRINTF(RubySlicc, "%s\n", in_msg); 296 Entry cache_entry := getCacheEntry(in_msg.addr); 297 TBE tbe := TBEs.lookup(in_msg.addr); 298 trigger(Event:PrbInv, in_msg.addr, cache_entry, tbe); 299 } 300 } 301 } 302 303 in_port(coreRequestNetwork_in, CPURequestMsg, requestFromTCP, rank=0) { 304 if (coreRequestNetwork_in.isReady(clockEdge())) { 305 peek(coreRequestNetwork_in, CPURequestMsg) { 306 TBE tbe := TBEs.lookup(in_msg.addr); 307 Entry cache_entry := getCacheEntry(in_msg.addr); 308 if (in_msg.Type == CoherenceRequestType:WriteThrough) { 309 if(WB) { 310 if(presentOrAvail(in_msg.addr)) { 311 trigger(Event:WrVicBlkBack, in_msg.addr, cache_entry, tbe); 312 } else { 313 Addr victim := L2cache.cacheProbe(in_msg.addr); 314 trigger(Event:L2_Repl, victim, getCacheEntry(victim), TBEs.lookup(victim)); 315 } 316 } else { 317 trigger(Event:WrVicBlk, in_msg.addr, cache_entry, tbe); 318 } 319 } else if (in_msg.Type == CoherenceRequestType:Atomic) { 320 trigger(Event:Atomic, in_msg.addr, cache_entry, tbe); 321 } else if (in_msg.Type == CoherenceRequestType:RdBlk) { 322 trigger(Event:RdBlk, in_msg.addr, cache_entry, tbe); 323 } else { 324 DPRINTF(RubySlicc, "%s\n", in_msg); 325 error("Unexpected Response Message to Core"); 326 } 327 } 328 } 329 } 330 // BEGIN ACTIONS 331 332 action(i_invL2, "i", desc="invalidate TCC cache block") { 333 if (is_valid(cache_entry)) { 334 L2cache.deallocate(address); 335 } 336 unset_cache_entry(); 337 } 338 339 action(sd_sendData, "sd", desc="send Shared response") { 340 peek(coreRequestNetwork_in, CPURequestMsg) { 341 enqueue(responseToCore_out, ResponseMsg, l2_response_latency) { 342 out_msg.addr := address; 343 out_msg.Type := CoherenceResponseType:TDSysResp; 344 out_msg.Sender := machineID; 345 out_msg.Destination.add(in_msg.Requestor); 346 out_msg.DataBlk := cache_entry.DataBlk; 347 out_msg.MessageSize := MessageSizeType:Response_Data; 348 out_msg.Dirty := false; 349 out_msg.State := CoherenceState:Shared; 350 DPRINTF(RubySlicc, "%s\n", out_msg); 351 } 352 } 353 } 354 355 356 action(sdr_sendDataResponse, "sdr", desc="send Shared response") { 357 enqueue(responseToCore_out, ResponseMsg, l2_response_latency) { 358 out_msg.addr := address; 359 out_msg.Type := CoherenceResponseType:TDSysResp; 360 out_msg.Sender := machineID; 361 out_msg.Destination := tbe.Destination; 362 out_msg.DataBlk := cache_entry.DataBlk; 363 out_msg.MessageSize := MessageSizeType:Response_Data; 364 out_msg.Dirty := false; 365 out_msg.State := CoherenceState:Shared; 366 DPRINTF(RubySlicc, "%s\n", out_msg); 367 } 368 enqueue(unblockToNB_out, UnblockMsg, 1) { 369 out_msg.addr := address; 370 out_msg.Destination.add(mapAddressToMachine(address, MachineType:Directory)); 371 out_msg.MessageSize := MessageSizeType:Unblock_Control; 372 DPRINTF(RubySlicc, "%s\n", out_msg); 373 } 374 } 375 376 377 action(rd_requestData, "r", desc="Miss in L2, pass on") { 378 if(tbe.Destination.count()==1){ 379 peek(coreRequestNetwork_in, CPURequestMsg) { 380 enqueue(requestToNB_out, CPURequestMsg, l2_request_latency) { 381 out_msg.addr := address; 382 out_msg.Type := in_msg.Type; 383 out_msg.Requestor := machineID; 384 out_msg.Destination.add(mapAddressToMachine(address, MachineType:Directory)); 385 out_msg.Shared := false; // unneeded for this request 386 out_msg.MessageSize := in_msg.MessageSize; 387 DPRINTF(RubySlicc, "%s\n", out_msg); 388 } 389 } 390 } 391 } 392 393 action(w_sendResponseWBAck, "w", desc="send WB Ack") { 394 peek(responseFromNB_in, ResponseMsg) { 395 enqueue(responseToCore_out, ResponseMsg, l2_response_latency) { 396 out_msg.addr := address; 397 out_msg.Type := CoherenceResponseType:TDSysWBAck; 398 out_msg.Destination.clear(); 399 out_msg.Destination.add(in_msg.WTRequestor); 400 out_msg.Sender := machineID; 401 out_msg.MessageSize := MessageSizeType:Writeback_Control; 402 } 403 } 404 } 405 406 action(swb_sendWBAck, "swb", desc="send WB Ack") { 407 peek(coreRequestNetwork_in, CPURequestMsg) { 408 enqueue(responseToCore_out, ResponseMsg, l2_response_latency) { 409 out_msg.addr := address; 410 out_msg.Type := CoherenceResponseType:TDSysWBAck; 411 out_msg.Destination.clear(); 412 out_msg.Destination.add(in_msg.Requestor); 413 out_msg.Sender := machineID; 414 out_msg.MessageSize := MessageSizeType:Writeback_Control; 415 } 416 } 417 } 418 419 action(ar_sendAtomicResponse, "ar", desc="send Atomic Ack") { 420 peek(responseFromNB_in, ResponseMsg) { 421 enqueue(responseToCore_out, ResponseMsg, l2_response_latency) { 422 out_msg.addr := address; 423 out_msg.Type := CoherenceResponseType:TDSysResp; 424 out_msg.Destination.add(in_msg.WTRequestor); 425 out_msg.Sender := machineID; 426 out_msg.MessageSize := in_msg.MessageSize; 427 out_msg.DataBlk := in_msg.DataBlk; 428 } 429 } 430 } 431 432 action(a_allocateBlock, "a", desc="allocate TCC block") { 433 if (is_invalid(cache_entry)) { 434 set_cache_entry(L2cache.allocate(address, new Entry)); 435 cache_entry.writeMask.clear(); 436 } 437 } 438 439 action(t_allocateTBE, "t", desc="allocate TBE Entry") { 440 if (is_invalid(tbe)) { 441 check_allocate(TBEs); 442 TBEs.allocate(address); 443 set_tbe(TBEs.lookup(address)); 444 tbe.Destination.clear(); 445 tbe.numAtomics := 0; 446 } 447 if (coreRequestNetwork_in.isReady(clockEdge())) { 448 peek(coreRequestNetwork_in, CPURequestMsg) { 449 if(in_msg.Type == CoherenceRequestType:RdBlk || in_msg.Type == CoherenceRequestType:Atomic){ 450 tbe.Destination.add(in_msg.Requestor); 451 } 452 } 453 } 454 } 455 456 action(dt_deallocateTBE, "dt", desc="Deallocate TBE entry") { 457 tbe.Destination.clear(); 458 TBEs.deallocate(address); 459 unset_tbe(); 460 } 461 462 action(wcb_writeCacheBlock, "wcb", desc="write data to TCC") { 463 peek(responseFromNB_in, ResponseMsg) { 464 cache_entry.DataBlk := in_msg.DataBlk; 465 DPRINTF(RubySlicc, "Writing to TCC: %s\n", in_msg); 466 } 467 } 468 469 action(wdb_writeDirtyBytes, "wdb", desc="write data to TCC") { 470 peek(coreRequestNetwork_in, CPURequestMsg) { 471 cache_entry.DataBlk.copyPartial(in_msg.DataBlk,in_msg.writeMask); 472 cache_entry.writeMask.orMask(in_msg.writeMask); 473 DPRINTF(RubySlicc, "Writing to TCC: %s\n", in_msg); 474 } 475 } 476 477 action(wt_writeThrough, "wt", desc="write back data") { 478 peek(coreRequestNetwork_in, CPURequestMsg) { 479 enqueue(requestToNB_out, CPURequestMsg, l2_request_latency) { 480 out_msg.addr := address; 481 out_msg.Requestor := machineID; 482 out_msg.WTRequestor := in_msg.Requestor; 483 out_msg.Destination.add(mapAddressToMachine(address, MachineType:Directory)); 484 out_msg.MessageSize := MessageSizeType:Data; 485 out_msg.Type := CoherenceRequestType:WriteThrough; 486 out_msg.Dirty := true; 487 out_msg.DataBlk := in_msg.DataBlk; 488 out_msg.writeMask.orMask(in_msg.writeMask); 489 } 490 } 491 } 492 493 action(wb_writeBack, "wb", desc="write back data") { 494 enqueue(requestToNB_out, CPURequestMsg, l2_request_latency) { 495 out_msg.addr := address; 496 out_msg.Requestor := machineID; 497 out_msg.WTRequestor := machineID; 498 out_msg.Destination.add(mapAddressToMachine(address, MachineType:Directory)); 499 out_msg.MessageSize := MessageSizeType:Data; 500 out_msg.Type := CoherenceRequestType:WriteThrough; 501 out_msg.Dirty := true; 502 out_msg.DataBlk := cache_entry.DataBlk; 503 out_msg.writeMask.orMask(cache_entry.writeMask); 504 } 505 } 506 507 action(at_atomicThrough, "at", desc="write back data") { 508 peek(coreRequestNetwork_in, CPURequestMsg) { 509 enqueue(requestToNB_out, CPURequestMsg, l2_request_latency) { 510 out_msg.addr := address; 511 out_msg.Requestor := machineID; 512 out_msg.WTRequestor := in_msg.Requestor; 513 out_msg.Destination.add(mapAddressToMachine(address, MachineType:Directory)); 514 out_msg.MessageSize := MessageSizeType:Data; 515 out_msg.Type := CoherenceRequestType:Atomic; 516 out_msg.Dirty := true; 517 out_msg.writeMask.orMask(in_msg.writeMask); 518 } 519 } 520 } 521 522 action(pi_sendProbeResponseInv, "pi", desc="send probe ack inv, no data") { 523 enqueue(responseToNB_out, ResponseMsg, 1) { 524 out_msg.addr := address; 525 out_msg.Type := CoherenceResponseType:CPUPrbResp; // TCC, L3 respond in same way to probes 526 out_msg.Sender := machineID; 527 out_msg.Destination.add(mapAddressToMachine(address, MachineType:Directory)); 528 out_msg.Dirty := false; 529 out_msg.Hit := false; 530 out_msg.Ntsl := true; 531 out_msg.State := CoherenceState:NA; 532 out_msg.MessageSize := MessageSizeType:Response_Control; 533 } 534 } 535 action(ut_updateTag, "ut", desc="update Tag (i.e. set MRU)") { 536 L2cache.setMRU(address); 537 } 538 539 action(p_popRequestQueue, "p", desc="pop request queue") { 540 coreRequestNetwork_in.dequeue(clockEdge()); 541 } 542 543 action(pr_popResponseQueue, "pr", desc="pop response queue") { 544 responseFromNB_in.dequeue(clockEdge()); 545 } 546 547 action(pp_popProbeQueue, "pp", desc="pop probe queue") { 548 probeNetwork_in.dequeue(clockEdge()); 549 } 550 551 action(z_stall, "z", desc="stall") { 552 // built-in 553 } 554 555 556 action(ina_incrementNumAtomics, "ina", desc="inc num atomics") { 557 tbe.numAtomics := tbe.numAtomics + 1; 558 } 559 560 561 action(dna_decrementNumAtomics, "dna", desc="inc num atomics") { 562 tbe.numAtomics := tbe.numAtomics - 1; 563 if (tbe.numAtomics==0) { 564 enqueue(triggerQueue_out, TriggerMsg, 1) { 565 out_msg.addr := address; 566 out_msg.Type := TriggerType:AtomicDone; 567 } 568 } 569 } 570 571 action(ptr_popTriggerQueue, "ptr", desc="pop Trigger") { 572 triggerQueue_in.dequeue(clockEdge()); 573 } 574 575 // END ACTIONS 576 577 // BEGIN TRANSITIONS 578 // transitions from base 579 // Assumptions for ArrayRead/Write 580 // TBE checked before tags 581 // Data Read/Write requires Tag Read 582 583 // Stalling transitions do NOT check the tag array...and if they do, 584 // they can cause a resource stall deadlock! 585 586 transition(WI, {RdBlk, WrVicBlk, Atomic, WrVicBlkBack}) { //TagArrayRead} { 587 z_stall; 588 } 589 transition(A, {RdBlk, WrVicBlk, WrVicBlkBack}) { //TagArrayRead} { 590 z_stall; 591 } 592 transition(IV, {WrVicBlk, Atomic, WrVicBlkBack}) { //TagArrayRead} { 593 z_stall; 594 } 595 transition({M, V}, RdBlk) {TagArrayRead, DataArrayRead} { 596 sd_sendData; 597 ut_updateTag; 598 p_popRequestQueue; 599 } 600 transition(W, RdBlk, WI) {TagArrayRead, DataArrayRead} { 601 t_allocateTBE; 602 wb_writeBack; 603 } 604 605 transition(I, RdBlk, IV) {TagArrayRead} { 606 t_allocateTBE; 607 rd_requestData; 608 p_popRequestQueue; 609 } 610 611 transition(IV, RdBlk) { 612 t_allocateTBE; 613 rd_requestData; 614 p_popRequestQueue; 615 } 616 617 transition({V, I},Atomic, A) {TagArrayRead} { 618 i_invL2; 619 t_allocateTBE; 620 at_atomicThrough; 621 ina_incrementNumAtomics; 622 p_popRequestQueue; 623 } 624 625 transition(A, Atomic) { 626 at_atomicThrough; 627 ina_incrementNumAtomics; 628 p_popRequestQueue; 629 } 630 631 transition({M, W}, Atomic, WI) {TagArrayRead} { 632 t_allocateTBE; 633 wb_writeBack; 634 } 635 636 transition(I, WrVicBlk) {TagArrayRead} { 637 wt_writeThrough; 638 p_popRequestQueue; 639 } 640 641 transition(V, WrVicBlk) {TagArrayRead, DataArrayWrite} { 642 ut_updateTag; 643 wdb_writeDirtyBytes; 644 wt_writeThrough; 645 p_popRequestQueue; 646 } 647 648 transition({V, M}, WrVicBlkBack, M) {TagArrayRead, TagArrayWrite, DataArrayWrite} { 649 ut_updateTag; 650 swb_sendWBAck; 651 wdb_writeDirtyBytes; 652 p_popRequestQueue; 653 } 654 655 transition(W, WrVicBlkBack) {TagArrayRead, TagArrayWrite, DataArrayWrite} { 656 ut_updateTag; 657 swb_sendWBAck; 658 wdb_writeDirtyBytes; 659 p_popRequestQueue; 660 } 661 662 transition(I, WrVicBlkBack, W) {TagArrayRead, TagArrayWrite, DataArrayWrite} { 663 a_allocateBlock; 664 ut_updateTag; 665 swb_sendWBAck; 666 wdb_writeDirtyBytes; 667 p_popRequestQueue; 668 } 669 670 transition({W, M}, L2_Repl, WI) {TagArrayRead, DataArrayRead} { 671 t_allocateTBE; 672 wb_writeBack; 673 i_invL2; 674 } 675 676 transition({I, V}, L2_Repl, I) {TagArrayRead, TagArrayWrite} { 677 i_invL2; 678 } 679 680 transition({A, IV, WI}, L2_Repl) { 681 i_invL2; 682 } 683 684 transition({I, V}, PrbInv, I) {TagArrayRead, TagArrayWrite} { 685 pi_sendProbeResponseInv; 686 pp_popProbeQueue; 687 } 688 689 transition(M, PrbInv, W) {TagArrayRead, TagArrayWrite} { 690 pi_sendProbeResponseInv; 691 pp_popProbeQueue; 692 } 693 694 transition(W, PrbInv) {TagArrayRead} { 695 pi_sendProbeResponseInv; 696 pp_popProbeQueue; 697 } 698 699 transition({A, IV, WI}, PrbInv) { 700 pi_sendProbeResponseInv; 701 pp_popProbeQueue; 702 } 703 704 transition(IV, Data, V) {TagArrayRead, TagArrayWrite, DataArrayWrite} { 705 a_allocateBlock; 706 ut_updateTag; 707 wcb_writeCacheBlock; 708 sdr_sendDataResponse; 709 pr_popResponseQueue; 710 dt_deallocateTBE; 711 } 712 713 transition(A, Data) {TagArrayRead, TagArrayWrite, DataArrayWrite} { 714 a_allocateBlock; 715 ar_sendAtomicResponse; 716 dna_decrementNumAtomics; 717 pr_popResponseQueue; 718 } 719 720 transition(A, AtomicDone, I) {TagArrayRead, TagArrayWrite} { 721 dt_deallocateTBE; 722 ptr_popTriggerQueue; 723 } 724 725 transition(A, AtomicNotDone) {TagArrayRead} { 726 ptr_popTriggerQueue; 727 } 728 729 //M,W should not see WBAck as the cache is in WB mode 730 //WBAcks do not need to check tags 731 transition({I, V, IV, A}, WBAck) { 732 w_sendResponseWBAck; 733 pr_popResponseQueue; 734 } 735 736 transition(WI, WBAck,I) { 737 dt_deallocateTBE; 738 pr_popResponseQueue; 739 } 740} 741