elastic_trace.cc revision 12255:9ef9176e4bb2
1/*
2 * Copyright (c) 2013 - 2015 ARM Limited
3 * All rights reserved
4 *
5 * The license below extends only to copyright in the software and shall
6 * not be construed as granting a license to any other intellectual
7 * property including but not limited to intellectual property relating
8 * to a hardware implementation of the functionality of the software
9 * licensed hereunder.  You may use the software subject to the license
10 * terms below provided that you ensure that this notice is replicated
11 * unmodified and in its entirety in all distributions of the software,
12 * modified or unmodified, in source code or in binary form.
13 *
14 * Redistribution and use in source and binary forms, with or without
15 * modification, are permitted provided that the following conditions are
16 * met: redistributions of source code must retain the above copyright
17 * notice, this list of conditions and the following disclaimer;
18 * redistributions in binary form must reproduce the above copyright
19 * notice, this list of conditions and the following disclaimer in the
20 * documentation and/or other materials provided with the distribution;
21 * neither the name of the copyright holders nor the names of its
22 * contributors may be used to endorse or promote products derived from
23 * this software without specific prior written permission.
24 *
25 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
26 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
27 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
28 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
29 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
30 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
31 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
32 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
33 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
34 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
35 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
36 *
37 * Authors: Radhika Jagtap
38 *          Andreas Hansson
39 *          Thomas Grass
40 */
41
42#include "cpu/o3/probe/elastic_trace.hh"
43
44#include "base/callback.hh"
45#include "base/output.hh"
46#include "base/trace.hh"
47#include "cpu/reg_class.hh"
48#include "debug/ElasticTrace.hh"
49#include "mem/packet.hh"
50
51ElasticTrace::ElasticTrace(const ElasticTraceParams* params)
52    :  ProbeListenerObject(params),
53       regEtraceListenersEvent([this]{ regEtraceListeners(); }, name()),
54       firstWin(true),
55       lastClearedSeqNum(0),
56       depWindowSize(params->depWindowSize),
57       dataTraceStream(nullptr),
58       instTraceStream(nullptr),
59       startTraceInst(params->startTraceInst),
60       allProbesReg(false),
61       traceVirtAddr(params->traceVirtAddr)
62{
63    cpu = dynamic_cast<FullO3CPU<O3CPUImpl>*>(params->manager);
64    fatal_if(!cpu, "Manager of %s is not of type O3CPU and thus does not "\
65                "support dependency tracing.\n", name());
66
67    fatal_if(depWindowSize == 0, "depWindowSize parameter must be non-zero. "\
68                "Recommended size is 3x ROB size in the O3CPU.\n");
69
70    fatal_if(cpu->numThreads > 1, "numThreads = %i, %s supports tracing for"\
71                "single-threaded workload only", cpu->numThreads, name());
72    // Initialize the protobuf output stream
73    fatal_if(params->instFetchTraceFile == "", "Assign instruction fetch "\
74                "trace file path to instFetchTraceFile");
75    fatal_if(params->dataDepTraceFile == "", "Assign data dependency "\
76                "trace file path to dataDepTraceFile");
77    std::string filename = simout.resolve(name() + "." +
78                                            params->instFetchTraceFile);
79    instTraceStream = new ProtoOutputStream(filename);
80    filename = simout.resolve(name() + "." + params->dataDepTraceFile);
81    dataTraceStream = new ProtoOutputStream(filename);
82    // Create a protobuf message for the header and write it to the stream
83    ProtoMessage::PacketHeader inst_pkt_header;
84    inst_pkt_header.set_obj_id(name());
85    inst_pkt_header.set_tick_freq(SimClock::Frequency);
86    instTraceStream->write(inst_pkt_header);
87    // Create a protobuf message for the header and write it to
88    // the stream
89    ProtoMessage::InstDepRecordHeader data_rec_header;
90    data_rec_header.set_obj_id(name());
91    data_rec_header.set_tick_freq(SimClock::Frequency);
92    data_rec_header.set_window_size(depWindowSize);
93    dataTraceStream->write(data_rec_header);
94    // Register a callback to flush trace records and close the output streams.
95    Callback* cb = new MakeCallback<ElasticTrace,
96        &ElasticTrace::flushTraces>(this);
97    registerExitCallback(cb);
98}
99
100void
101ElasticTrace::regProbeListeners()
102{
103    inform("@%llu: regProbeListeners() called, startTraceInst = %llu",
104        curTick(), startTraceInst);
105    if (startTraceInst == 0) {
106        // If we want to start tracing from the start of the simulation,
107        // register all elastic trace probes now.
108        regEtraceListeners();
109    } else {
110        // Schedule an event to register all elastic trace probes when
111        // specified no. of instructions are committed.
112        cpu->comInstEventQueue[(ThreadID)0]->schedule(&regEtraceListenersEvent,
113                                                      startTraceInst);
114    }
115}
116
117void
118ElasticTrace::regEtraceListeners()
119{
120    assert(!allProbesReg);
121    inform("@%llu: No. of instructions committed = %llu, registering elastic"
122        " probe listeners", curTick(), cpu->numSimulatedInsts());
123    // Create new listeners: provide method to be called upon a notify() for
124    // each probe point.
125    listeners.push_back(new ProbeListenerArg<ElasticTrace, RequestPtr>(this,
126                        "FetchRequest", &ElasticTrace::fetchReqTrace));
127    listeners.push_back(new ProbeListenerArg<ElasticTrace, DynInstPtr>(this,
128                        "Execute", &ElasticTrace::recordExecTick));
129    listeners.push_back(new ProbeListenerArg<ElasticTrace, DynInstPtr>(this,
130                        "ToCommit", &ElasticTrace::recordToCommTick));
131    listeners.push_back(new ProbeListenerArg<ElasticTrace, DynInstPtr>(this,
132                        "Rename", &ElasticTrace::updateRegDep));
133    listeners.push_back(new ProbeListenerArg<ElasticTrace, SeqNumRegPair>(this,
134                        "SquashInRename", &ElasticTrace::removeRegDepMapEntry));
135    listeners.push_back(new ProbeListenerArg<ElasticTrace, DynInstPtr>(this,
136                        "Squash", &ElasticTrace::addSquashedInst));
137    listeners.push_back(new ProbeListenerArg<ElasticTrace, DynInstPtr>(this,
138                        "Commit", &ElasticTrace::addCommittedInst));
139    allProbesReg = true;
140}
141
142void
143ElasticTrace::fetchReqTrace(const RequestPtr &req)
144{
145
146    DPRINTFR(ElasticTrace, "Fetch Req %i,(%lli,%lli,%lli),%i,%i,%lli\n",
147             (MemCmd::ReadReq),
148             req->getPC(), req->getVaddr(), req->getPaddr(),
149             req->getFlags(), req->getSize(), curTick());
150
151    // Create a protobuf message including the request fields necessary to
152    // recreate the request in the TraceCPU.
153    ProtoMessage::Packet inst_fetch_pkt;
154    inst_fetch_pkt.set_tick(curTick());
155    inst_fetch_pkt.set_cmd(MemCmd::ReadReq);
156    inst_fetch_pkt.set_pc(req->getPC());
157    inst_fetch_pkt.set_flags(req->getFlags());
158    inst_fetch_pkt.set_addr(req->getPaddr());
159    inst_fetch_pkt.set_size(req->getSize());
160    // Write the message to the stream.
161    instTraceStream->write(inst_fetch_pkt);
162}
163
164void
165ElasticTrace::recordExecTick(const DynInstPtr &dyn_inst)
166{
167
168    // In a corner case, a retired instruction is propagated backward to the
169    // IEW instruction queue to handle some side-channel information. But we
170    // must not process an instruction again. So we test the sequence number
171    // against the lastClearedSeqNum and skip adding the instruction for such
172    // corner cases.
173    if (dyn_inst->seqNum <= lastClearedSeqNum) {
174        DPRINTFR(ElasticTrace, "[sn:%lli] Ignoring in execute as instruction \
175        has already retired (mostly squashed)", dyn_inst->seqNum);
176        // Do nothing as program has proceeded and this inst has been
177        // propagated backwards to handle something.
178        return;
179    }
180
181    DPRINTFR(ElasticTrace, "[sn:%lli] Execute Tick = %i\n", dyn_inst->seqNum,
182                curTick());
183    // Either the execution info object will already exist if this
184    // instruction had a register dependency recorded in the rename probe
185    // listener before entering execute stage or it will not exist and will
186    // need to be created here.
187    InstExecInfo* exec_info_ptr;
188    auto itr_exec_info = tempStore.find(dyn_inst->seqNum);
189    if (itr_exec_info != tempStore.end()) {
190        exec_info_ptr = itr_exec_info->second;
191    } else {
192        exec_info_ptr = new InstExecInfo;
193        tempStore[dyn_inst->seqNum] = exec_info_ptr;
194    }
195
196    exec_info_ptr->executeTick = curTick();
197    maxTempStoreSize = std::max(tempStore.size(),
198                                (std::size_t)maxTempStoreSize.value());
199}
200
201void
202ElasticTrace::recordToCommTick(const DynInstPtr &dyn_inst)
203{
204    // If tracing has just been enabled then the instruction at this stage of
205    // execution is far enough that we cannot gather info about its past like
206    // the tick it started execution. Simply return until we see an instruction
207    // that is found in the tempStore.
208    auto itr_exec_info = tempStore.find(dyn_inst->seqNum);
209    if (itr_exec_info == tempStore.end()) {
210        DPRINTFR(ElasticTrace, "recordToCommTick: [sn:%lli] Not in temp store,"
211                    " skipping.\n", dyn_inst->seqNum);
212        return;
213    }
214
215    DPRINTFR(ElasticTrace, "[sn:%lli] To Commit Tick = %i\n", dyn_inst->seqNum,
216                curTick());
217    InstExecInfo* exec_info_ptr = itr_exec_info->second;
218    exec_info_ptr->toCommitTick = curTick();
219
220}
221
222void
223ElasticTrace::updateRegDep(const DynInstPtr &dyn_inst)
224{
225    // Get the sequence number of the instruction
226    InstSeqNum seq_num = dyn_inst->seqNum;
227
228    assert(dyn_inst->seqNum > lastClearedSeqNum);
229
230    // Since this is the first probe activated in the pipeline, create
231    // a new execution info object to track this instruction as it
232    // progresses through the pipeline.
233    InstExecInfo* exec_info_ptr = new InstExecInfo;
234    tempStore[seq_num] = exec_info_ptr;
235
236    // Loop through the source registers and look up the dependency map. If
237    // the source register entry is found in the dependency map, add a
238    // dependency on the last writer.
239    int8_t max_regs = dyn_inst->numSrcRegs();
240    for (int src_idx = 0; src_idx < max_regs; src_idx++) {
241
242        const RegId& src_reg = dyn_inst->srcRegIdx(src_idx);
243        if (!src_reg.isMiscReg() &&
244            !src_reg.isZeroReg()) {
245            // Get the physical register index of the i'th source register.
246            PhysRegIdPtr phys_src_reg = dyn_inst->renamedSrcRegIdx(src_idx);
247            DPRINTFR(ElasticTrace, "[sn:%lli] Check map for src reg"
248                     " %i (%s)\n", seq_num,
249                     phys_src_reg->flatIndex(), phys_src_reg->className());
250            auto itr_writer = physRegDepMap.find(phys_src_reg->flatIndex());
251            if (itr_writer != physRegDepMap.end()) {
252                InstSeqNum last_writer = itr_writer->second;
253                // Additionally the dependency distance is kept less than the
254                // window size parameter to limit the memory allocation to
255                // nodes in the graph. If the window were tending to infinite
256                // we would have to load a large number of node objects during
257                // replay.
258                if (seq_num - last_writer < depWindowSize) {
259                    // Record a physical register dependency.
260                    exec_info_ptr->physRegDepSet.insert(last_writer);
261                }
262            }
263
264        }
265
266    }
267
268    // Loop through the destination registers of this instruction and update
269    // the physical register dependency map for last writers to registers.
270    max_regs = dyn_inst->numDestRegs();
271    for (int dest_idx = 0; dest_idx < max_regs; dest_idx++) {
272        // For data dependency tracking the register must be an int, float or
273        // CC register and not a Misc register.
274        const RegId& dest_reg = dyn_inst->destRegIdx(dest_idx);
275        if (!dest_reg.isMiscReg() &&
276            !dest_reg.isZeroReg()) {
277            // Get the physical register index of the i'th destination
278            // register.
279            PhysRegIdPtr phys_dest_reg = dyn_inst->renamedDestRegIdx(dest_idx);
280            DPRINTFR(ElasticTrace, "[sn:%lli] Update map for dest reg"
281                     " %i (%s)\n", seq_num, phys_dest_reg->flatIndex(),
282                     dest_reg.className());
283            physRegDepMap[phys_dest_reg->flatIndex()] = seq_num;
284        }
285    }
286    maxPhysRegDepMapSize = std::max(physRegDepMap.size(),
287                                    (std::size_t)maxPhysRegDepMapSize.value());
288}
289
290void
291ElasticTrace::removeRegDepMapEntry(const SeqNumRegPair &inst_reg_pair)
292{
293    DPRINTFR(ElasticTrace, "Remove Map entry for Reg %i\n",
294            inst_reg_pair.second);
295    auto itr_regdep_map = physRegDepMap.find(inst_reg_pair.second);
296    if (itr_regdep_map != physRegDepMap.end())
297        physRegDepMap.erase(itr_regdep_map);
298}
299
300void
301ElasticTrace::addSquashedInst(const DynInstPtr &head_inst)
302{
303    // If the squashed instruction was squashed before being processed by
304    // execute stage then it will not be in the temporary store. In this case
305    // do nothing and return.
306    auto itr_exec_info = tempStore.find(head_inst->seqNum);
307    if (itr_exec_info == tempStore.end())
308        return;
309
310    // If there is a squashed load for which a read request was
311    // sent before it got squashed then add it to the trace.
312    DPRINTFR(ElasticTrace, "Attempt to add squashed inst [sn:%lli]\n",
313                head_inst->seqNum);
314    // Get pointer to the execution info object corresponding to the inst.
315    InstExecInfo* exec_info_ptr = itr_exec_info->second;
316    if (head_inst->isLoad() && exec_info_ptr->executeTick != MaxTick &&
317        exec_info_ptr->toCommitTick != MaxTick &&
318        head_inst->hasRequest() &&
319        head_inst->getFault() == NoFault) {
320        // Add record to depTrace with commit parameter as false.
321        addDepTraceRecord(head_inst, exec_info_ptr, false);
322    }
323    // As the information contained is no longer needed, remove the execution
324    // info object from the temporary store.
325    clearTempStoreUntil(head_inst);
326}
327
328void
329ElasticTrace::addCommittedInst(const DynInstPtr &head_inst)
330{
331    DPRINTFR(ElasticTrace, "Attempt to add committed inst [sn:%lli]\n",
332                head_inst->seqNum);
333
334    // Add the instruction to the depTrace.
335    if (!head_inst->isNop()) {
336
337        // If tracing has just been enabled then the instruction at this stage
338        // of execution is far enough that we cannot gather info about its past
339        // like the tick it started execution. Simply return until we see an
340        // instruction that is found in the tempStore.
341        auto itr_temp_store = tempStore.find(head_inst->seqNum);
342        if (itr_temp_store == tempStore.end()) {
343            DPRINTFR(ElasticTrace, "addCommittedInst: [sn:%lli] Not in temp "
344                "store, skipping.\n", head_inst->seqNum);
345            return;
346        }
347
348        // Get pointer to the execution info object corresponding to the inst.
349        InstExecInfo* exec_info_ptr = itr_temp_store->second;
350        assert(exec_info_ptr->executeTick != MaxTick);
351        assert(exec_info_ptr->toCommitTick != MaxTick);
352
353        // Check if the instruction had a fault, if it predicated false and
354        // thus previous register values were restored or if it was a
355        // load/store that did not have a request (e.g. when the size of the
356        // request is zero). In all these cases the instruction is set as
357        // executed and is picked up by the commit probe listener. But a
358        // request is not issued and registers are not written. So practically,
359        // skipping these should not hurt as execution would not stall on them.
360        // Alternatively, these could be included merely as a compute node in
361        // the graph. Removing these for now. If correlation accuracy needs to
362        // be improved in future these can be turned into comp nodes at the
363        // cost of bigger traces.
364        if (head_inst->getFault() != NoFault) {
365            DPRINTF(ElasticTrace, "%s [sn:%lli] has faulted so "
366                    "skip adding it to the trace\n",
367                    (head_inst->isMemRef() ? "Load/store" : "Comp inst."),
368                    head_inst->seqNum);
369        } else if (head_inst->isMemRef() && !head_inst->hasRequest()) {
370            DPRINTF(ElasticTrace, "Load/store [sn:%lli]  has no request so "
371                    "skip adding it to the trace\n", head_inst->seqNum);
372        } else if (!head_inst->readPredicate()) {
373            DPRINTF(ElasticTrace, "%s [sn:%lli] is predicated false so "
374                    "skip adding it to the trace\n",
375                    (head_inst->isMemRef() ? "Load/store" : "Comp inst."),
376                    head_inst->seqNum);
377        } else {
378            // Add record to depTrace with commit parameter as true.
379            addDepTraceRecord(head_inst, exec_info_ptr, true);
380        }
381    }
382    // As the information contained is no longer needed, remove the execution
383    // info object from the temporary store.
384    clearTempStoreUntil(head_inst);
385}
386
387void
388ElasticTrace::addDepTraceRecord(const DynInstPtr &head_inst,
389                                InstExecInfo* exec_info_ptr, bool commit)
390{
391    // Create a record to assign dynamic intruction related fields.
392    TraceInfo* new_record = new TraceInfo;
393    // Add to map for sequence number look up to retrieve the TraceInfo pointer
394    traceInfoMap[head_inst->seqNum] = new_record;
395
396    // Assign fields from the instruction
397    new_record->instNum = head_inst->seqNum;
398    new_record->commit = commit;
399    new_record->type = head_inst->isLoad() ? Record::LOAD :
400                        (head_inst->isStore() ? Record::STORE :
401                        Record::COMP);
402
403    // Assign fields for creating a request in case of a load/store
404    new_record->reqFlags = head_inst->memReqFlags;
405    new_record->virtAddr = head_inst->effAddr;
406    new_record->asid = head_inst->asid;
407    new_record->physAddr = head_inst->physEffAddrLow;
408    // Currently the tracing does not support split requests.
409    new_record->size = head_inst->effSize;
410    new_record->pc = head_inst->instAddr();
411
412    // Assign the timing information stored in the execution info object
413    new_record->executeTick = exec_info_ptr->executeTick;
414    new_record->toCommitTick = exec_info_ptr->toCommitTick;
415    new_record->commitTick = curTick();
416
417    // Assign initial values for number of dependents and computational delay
418    new_record->numDepts = 0;
419    new_record->compDelay = -1;
420
421    // The physical register dependency set of the first instruction is
422    // empty. Since there are no records in the depTrace at this point, the
423    // case of adding an ROB dependency by using a reverse iterator is not
424    // applicable. Thus, populate the fields of the record corresponding to the
425    // first instruction and return.
426    if (depTrace.empty()) {
427        // Store the record in depTrace.
428        depTrace.push_back(new_record);
429        DPRINTF(ElasticTrace, "Added first inst record %lli to DepTrace.\n",
430                new_record->instNum);
431        return;
432    }
433
434    // Clear register dependencies for squashed loads as they may be dependent
435    // on squashed instructions and we do not add those to the trace.
436    if (head_inst->isLoad() && !commit) {
437         (exec_info_ptr->physRegDepSet).clear();
438    }
439
440    // Assign the register dependencies stored in the execution info object
441    std::set<InstSeqNum>::const_iterator dep_set_it;
442    for (dep_set_it = (exec_info_ptr->physRegDepSet).begin();
443         dep_set_it != (exec_info_ptr->physRegDepSet).end();
444         ++dep_set_it) {
445        auto trace_info_itr = traceInfoMap.find(*dep_set_it);
446        if (trace_info_itr != traceInfoMap.end()) {
447            // The register dependency is valid. Assign it and calculate
448            // computational delay
449            new_record->physRegDepList.push_back(*dep_set_it);
450            DPRINTF(ElasticTrace, "Inst %lli has register dependency on "
451                    "%lli\n", new_record->instNum, *dep_set_it);
452            TraceInfo* reg_dep = trace_info_itr->second;
453            reg_dep->numDepts++;
454            compDelayPhysRegDep(reg_dep, new_record);
455            ++numRegDep;
456        } else {
457            // The instruction that this has a register dependency on was
458            // not added to the trace because of one of the following
459            // 1. it was an instruction that had a fault
460            // 2. it was an instruction that was predicated false and
461            // previous register values were restored
462            // 3. it was load/store that did not have a request (e.g. when
463            // the size of the request is zero but this may not be a fault)
464            // In all these cases the instruction is set as executed and is
465            // picked up by the commit probe listener. But a request is not
466            // issued and registers are not written to in these cases.
467            DPRINTF(ElasticTrace, "Inst %lli has register dependency on "
468                    "%lli is skipped\n",new_record->instNum, *dep_set_it);
469        }
470    }
471
472    // Check for and assign an ROB dependency in addition to register
473    // dependency before adding the record to the trace.
474    // As stores have to commit in order a store is dependent on the last
475    // committed load/store. This is recorded in the ROB dependency.
476    if (head_inst->isStore()) {
477        // Look up store-after-store order dependency
478        updateCommitOrderDep(new_record, false);
479        // Look up store-after-load order dependency
480        updateCommitOrderDep(new_record, true);
481    }
482
483    // In case a node is dependency-free or its dependency got discarded
484    // because it was outside the window, it is marked ready in the ROB at the
485    // time of issue. A request is sent as soon as possible. To model this, a
486    // node is assigned an issue order dependency on a committed instruction
487    // that completed earlier than it. This is done to avoid the problem of
488    // determining the issue times of such dependency-free nodes during replay
489    // which could lead to too much parallelism, thinking conservatively.
490    if (new_record->robDepList.empty() && new_record->physRegDepList.empty()) {
491        updateIssueOrderDep(new_record);
492    }
493
494    // Store the record in depTrace.
495    depTrace.push_back(new_record);
496    DPRINTF(ElasticTrace, "Added %s inst %lli to DepTrace.\n",
497            (commit ? "committed" : "squashed"), new_record->instNum);
498
499    // To process the number of records specified by depWindowSize in the
500    // forward direction, the depTrace must have twice as many records
501    // to check for dependencies.
502    if (depTrace.size() == 2 * depWindowSize) {
503
504        DPRINTF(ElasticTrace, "Writing out trace...\n");
505
506        // Write out the records which have been processed to the trace
507        // and remove them from the depTrace.
508        writeDepTrace(depWindowSize);
509
510        // After the first window, writeDepTrace() must check for valid
511        // compDelay.
512        firstWin = false;
513    }
514}
515
516void
517ElasticTrace::updateCommitOrderDep(TraceInfo* new_record,
518                                    bool find_load_not_store)
519{
520    assert(new_record->isStore());
521    // Iterate in reverse direction to search for the last committed
522    // load/store that completed earlier than the new record
523    depTraceRevItr from_itr(depTrace.end());
524    depTraceRevItr until_itr(depTrace.begin());
525    TraceInfo* past_record = *from_itr;
526    uint32_t num_go_back = 0;
527
528    // The execution time of this store is when it is sent, that is committed
529    Tick execute_tick = curTick();
530    // Search for store-after-load or store-after-store order dependency
531    while (num_go_back < depWindowSize && from_itr != until_itr) {
532        if (find_load_not_store) {
533            // Check if previous inst is a load completed earlier by comparing
534            // with execute tick
535            if (hasLoadCompleted(past_record, execute_tick)) {
536                // Assign rob dependency and calculate the computational delay
537                assignRobDep(past_record, new_record);
538                ++numOrderDepStores;
539                return;
540            }
541        } else {
542            // Check if previous inst is a store sent earlier by comparing with
543            // execute tick
544            if (hasStoreCommitted(past_record, execute_tick)) {
545                // Assign rob dependency and calculate the computational delay
546                assignRobDep(past_record, new_record);
547                ++numOrderDepStores;
548                return;
549            }
550        }
551        ++from_itr;
552        past_record = *from_itr;
553        ++num_go_back;
554    }
555}
556
557void
558ElasticTrace::updateIssueOrderDep(TraceInfo* new_record)
559{
560    // Interate in reverse direction to search for the last committed
561    // record that completed earlier than the new record
562    depTraceRevItr from_itr(depTrace.end());
563    depTraceRevItr until_itr(depTrace.begin());
564    TraceInfo* past_record = *from_itr;
565
566    uint32_t num_go_back = 0;
567    Tick execute_tick = 0;
568
569    if (new_record->isLoad()) {
570        // The execution time of a load is when a request is sent
571        execute_tick = new_record->executeTick;
572        ++numIssueOrderDepLoads;
573    } else if (new_record->isStore()) {
574        // The execution time of a store is when it is sent, i.e. committed
575        execute_tick = curTick();
576        ++numIssueOrderDepStores;
577    } else {
578        // The execution time of a non load/store is when it completes
579        execute_tick = new_record->toCommitTick;
580        ++numIssueOrderDepOther;
581    }
582
583    // We search if this record has an issue order dependency on a past record.
584    // Once we find it, we update both the new record and the record it depends
585    // on and return.
586    while (num_go_back < depWindowSize && from_itr != until_itr) {
587        // Check if a previous inst is a load sent earlier, or a store sent
588        // earlier, or a comp inst completed earlier by comparing with execute
589        // tick
590        if (hasLoadBeenSent(past_record, execute_tick) ||
591            hasStoreCommitted(past_record, execute_tick) ||
592            hasCompCompleted(past_record, execute_tick)) {
593            // Assign rob dependency and calculate the computational delay
594            assignRobDep(past_record, new_record);
595            return;
596        }
597        ++from_itr;
598        past_record = *from_itr;
599        ++num_go_back;
600    }
601}
602
603void
604ElasticTrace::assignRobDep(TraceInfo* past_record, TraceInfo* new_record) {
605    DPRINTF(ElasticTrace, "%s %lli has ROB dependency on %lli\n",
606            new_record->typeToStr(), new_record->instNum,
607            past_record->instNum);
608    // Add dependency on past record
609    new_record->robDepList.push_back(past_record->instNum);
610    // Update new_record's compute delay with respect to the past record
611    compDelayRob(past_record, new_record);
612    // Increment number of dependents of the past record
613    ++(past_record->numDepts);
614    // Update stat to log max number of dependents
615    maxNumDependents = std::max(past_record->numDepts,
616                                (uint32_t)maxNumDependents.value());
617}
618
619bool
620ElasticTrace::hasStoreCommitted(TraceInfo* past_record,
621                                    Tick execute_tick) const
622{
623    return (past_record->isStore() && past_record->commitTick <= execute_tick);
624}
625
626bool
627ElasticTrace::hasLoadCompleted(TraceInfo* past_record,
628                                    Tick execute_tick) const
629{
630    return(past_record->isLoad() && past_record->commit &&
631                past_record->toCommitTick <= execute_tick);
632}
633
634bool
635ElasticTrace::hasLoadBeenSent(TraceInfo* past_record,
636                                Tick execute_tick) const
637{
638    // Check if previous inst is a load sent earlier than this
639    return (past_record->isLoad() && past_record->commit &&
640        past_record->executeTick <= execute_tick);
641}
642
643bool
644ElasticTrace::hasCompCompleted(TraceInfo* past_record,
645                                    Tick execute_tick) const
646{
647    return(past_record->isComp() && past_record->toCommitTick <= execute_tick);
648}
649
650void
651ElasticTrace::clearTempStoreUntil(const DynInstPtr head_inst)
652{
653    // Clear from temp store starting with the execution info object
654    // corresponding the head_inst and continue clearing by decrementing the
655    // sequence number until the last cleared sequence number.
656    InstSeqNum temp_sn = (head_inst->seqNum);
657    while (temp_sn > lastClearedSeqNum) {
658        auto itr_exec_info = tempStore.find(temp_sn);
659        if (itr_exec_info != tempStore.end()) {
660            InstExecInfo* exec_info_ptr = itr_exec_info->second;
661            // Free allocated memory for the info object
662            delete exec_info_ptr;
663            // Remove entry from temporary store
664            tempStore.erase(itr_exec_info);
665        }
666        temp_sn--;
667    }
668    // Update the last cleared sequence number to that of the head_inst
669    lastClearedSeqNum = head_inst->seqNum;
670}
671
672void
673ElasticTrace::compDelayRob(TraceInfo* past_record, TraceInfo* new_record)
674{
675    // The computation delay is the delay between the completion tick of the
676    // inst. pointed to by past_record and the execution tick of its dependent
677    // inst. pointed to by new_record.
678    int64_t comp_delay = -1;
679    Tick execution_tick = 0, completion_tick = 0;
680
681    DPRINTF(ElasticTrace, "Seq num %lli has ROB dependency on seq num %lli.\n",
682            new_record->instNum, past_record->instNum);
683
684    // Get the tick when the node is executed as per the modelling of
685    // computation delay
686    execution_tick = new_record->getExecuteTick();
687
688    if (past_record->isLoad()) {
689        if (new_record->isStore()) {
690            completion_tick = past_record->toCommitTick;
691        } else {
692            completion_tick = past_record->executeTick;
693        }
694    } else if (past_record->isStore()) {
695        completion_tick = past_record->commitTick;
696    } else if (past_record->isComp()){
697        completion_tick = past_record->toCommitTick;
698    }
699    assert(execution_tick >= completion_tick);
700    comp_delay = execution_tick - completion_tick;
701
702    DPRINTF(ElasticTrace, "Computational delay is %lli - %lli = %lli\n",
703            execution_tick, completion_tick, comp_delay);
704
705    // Assign the computational delay with respect to the dependency which
706    // completes the latest.
707    if (new_record->compDelay == -1)
708        new_record->compDelay = comp_delay;
709    else
710        new_record->compDelay = std::min(comp_delay, new_record->compDelay);
711    DPRINTF(ElasticTrace, "Final computational delay = %lli.\n",
712            new_record->compDelay);
713}
714
715void
716ElasticTrace::compDelayPhysRegDep(TraceInfo* past_record,
717                                    TraceInfo* new_record)
718{
719    // The computation delay is the delay between the completion tick of the
720    // inst. pointed to by past_record and the execution tick of its dependent
721    // inst. pointed to by new_record.
722    int64_t comp_delay = -1;
723    Tick execution_tick = 0, completion_tick = 0;
724
725    DPRINTF(ElasticTrace, "Seq. num %lli has register dependency on seq. num"
726            " %lli.\n", new_record->instNum, past_record->instNum);
727
728    // Get the tick when the node is executed as per the modelling of
729    // computation delay
730    execution_tick = new_record->getExecuteTick();
731
732    // When there is a physical register dependency on an instruction, the
733    // completion tick of that instruction is when it wrote to the register,
734    // that is toCommitTick. In case, of a store updating a destination
735    // register, this is approximated to commitTick instead
736    if (past_record->isStore()) {
737        completion_tick = past_record->commitTick;
738    } else {
739        completion_tick = past_record->toCommitTick;
740    }
741    assert(execution_tick >= completion_tick);
742    comp_delay = execution_tick - completion_tick;
743    DPRINTF(ElasticTrace, "Computational delay is %lli - %lli = %lli\n",
744            execution_tick, completion_tick, comp_delay);
745
746    // Assign the computational delay with respect to the dependency which
747    // completes the latest.
748    if (new_record->compDelay == -1)
749        new_record->compDelay = comp_delay;
750    else
751        new_record->compDelay = std::min(comp_delay, new_record->compDelay);
752    DPRINTF(ElasticTrace, "Final computational delay = %lli.\n",
753            new_record->compDelay);
754}
755
756Tick
757ElasticTrace::TraceInfo::getExecuteTick() const
758{
759    if (isLoad()) {
760        // Execution tick for a load instruction is when the request was sent,
761        // that is executeTick.
762        return executeTick;
763    } else if (isStore()) {
764        // Execution tick for a store instruction is when the request was sent,
765        // that is commitTick.
766        return commitTick;
767    } else {
768        // Execution tick for a non load/store instruction is when the register
769        // value was written to, that is commitTick.
770        return toCommitTick;
771    }
772}
773
774void
775ElasticTrace::writeDepTrace(uint32_t num_to_write)
776{
777    // Write the trace with fields as follows:
778    // Instruction sequence number
779    // If instruction was a load
780    // If instruction was a store
781    // If instruction has addr
782    // If instruction has size
783    // If instruction has flags
784    // List of order dependencies - optional, repeated
785    // Computational delay with respect to last completed dependency
786    // List of physical register RAW dependencies - optional, repeated
787    // Weight of a node equal to no. of filtered nodes before it - optional
788    uint16_t num_filtered_nodes = 0;
789    depTraceItr dep_trace_itr(depTrace.begin());
790    depTraceItr dep_trace_itr_start = dep_trace_itr;
791    while (num_to_write > 0) {
792        TraceInfo* temp_ptr = *dep_trace_itr;
793        assert(temp_ptr->type != Record::INVALID);
794        // If no node dependends on a comp node then there is no reason to
795        // track the comp node in the dependency graph. We filter out such
796        // nodes but count them and add a weight field to the subsequent node
797        // that we do include in the trace.
798        if (!temp_ptr->isComp() || temp_ptr->numDepts != 0) {
799            DPRINTFR(ElasticTrace, "Instruction with seq. num %lli "
800                     "is as follows:\n", temp_ptr->instNum);
801            if (temp_ptr->isLoad() || temp_ptr->isStore()) {
802                DPRINTFR(ElasticTrace, "\tis a %s\n", temp_ptr->typeToStr());
803                DPRINTFR(ElasticTrace, "\thas a request with phys addr %i, "
804                         "size %i, flags %i\n", temp_ptr->physAddr,
805                         temp_ptr->size, temp_ptr->reqFlags);
806            } else {
807                 DPRINTFR(ElasticTrace, "\tis a %s\n", temp_ptr->typeToStr());
808            }
809            if (firstWin && temp_ptr->compDelay == -1) {
810                if (temp_ptr->isLoad()) {
811                    temp_ptr->compDelay = temp_ptr->executeTick;
812                } else if (temp_ptr->isStore()) {
813                    temp_ptr->compDelay = temp_ptr->commitTick;
814                } else {
815                    temp_ptr->compDelay = temp_ptr->toCommitTick;
816                }
817            }
818            assert(temp_ptr->compDelay != -1);
819            DPRINTFR(ElasticTrace, "\thas computational delay %lli\n",
820                     temp_ptr->compDelay);
821
822            // Create a protobuf message for the dependency record
823            ProtoMessage::InstDepRecord dep_pkt;
824            dep_pkt.set_seq_num(temp_ptr->instNum);
825            dep_pkt.set_type(temp_ptr->type);
826            dep_pkt.set_pc(temp_ptr->pc);
827            if (temp_ptr->isLoad() || temp_ptr->isStore()) {
828                dep_pkt.set_flags(temp_ptr->reqFlags);
829                dep_pkt.set_p_addr(temp_ptr->physAddr);
830                // If tracing of virtual addresses is enabled, set the optional
831                // field for it
832                if (traceVirtAddr) {
833                    dep_pkt.set_v_addr(temp_ptr->virtAddr);
834                    dep_pkt.set_asid(temp_ptr->asid);
835                }
836                dep_pkt.set_size(temp_ptr->size);
837            }
838            dep_pkt.set_comp_delay(temp_ptr->compDelay);
839            if (temp_ptr->robDepList.empty()) {
840                DPRINTFR(ElasticTrace, "\thas no order (rob) dependencies\n");
841            }
842            while (!temp_ptr->robDepList.empty()) {
843                DPRINTFR(ElasticTrace, "\thas order (rob) dependency on %lli\n",
844                         temp_ptr->robDepList.front());
845                dep_pkt.add_rob_dep(temp_ptr->robDepList.front());
846                temp_ptr->robDepList.pop_front();
847            }
848            if (temp_ptr->physRegDepList.empty()) {
849                DPRINTFR(ElasticTrace, "\thas no register dependencies\n");
850            }
851            while (!temp_ptr->physRegDepList.empty()) {
852                DPRINTFR(ElasticTrace, "\thas register dependency on %lli\n",
853                         temp_ptr->physRegDepList.front());
854                dep_pkt.add_reg_dep(temp_ptr->physRegDepList.front());
855                temp_ptr->physRegDepList.pop_front();
856            }
857            if (num_filtered_nodes != 0) {
858                // Set the weight of this node as the no. of filtered nodes
859                // between this node and the last node that we wrote to output
860                // stream. The weight will be used during replay to model ROB
861                // occupancy of filtered nodes.
862                dep_pkt.set_weight(num_filtered_nodes);
863                num_filtered_nodes = 0;
864            }
865            // Write the message to the protobuf output stream
866            dataTraceStream->write(dep_pkt);
867        } else {
868            // Don't write the node to the trace but note that we have filtered
869            // out a node.
870            ++numFilteredNodes;
871            ++num_filtered_nodes;
872        }
873        dep_trace_itr++;
874        traceInfoMap.erase(temp_ptr->instNum);
875        delete temp_ptr;
876        num_to_write--;
877    }
878    depTrace.erase(dep_trace_itr_start, dep_trace_itr);
879}
880
881void
882ElasticTrace::regStats() {
883    ProbeListenerObject::regStats();
884
885    using namespace Stats;
886    numRegDep
887        .name(name() + ".numRegDep")
888        .desc("Number of register dependencies recorded during tracing")
889        ;
890
891    numOrderDepStores
892        .name(name() + ".numOrderDepStores")
893        .desc("Number of commit order (rob) dependencies for a store recorded"
894              " on a past load/store during tracing")
895        ;
896
897    numIssueOrderDepLoads
898        .name(name() + ".numIssueOrderDepLoads")
899        .desc("Number of loads that got assigned issue order dependency"
900              " because they were dependency-free")
901        ;
902
903    numIssueOrderDepStores
904        .name(name() + ".numIssueOrderDepStores")
905        .desc("Number of stores that got assigned issue order dependency"
906              " because they were dependency-free")
907        ;
908
909    numIssueOrderDepOther
910        .name(name() + ".numIssueOrderDepOther")
911        .desc("Number of non load/store insts that got assigned issue order"
912              " dependency because they were dependency-free")
913        ;
914
915    numFilteredNodes
916        .name(name() + ".numFilteredNodes")
917        .desc("No. of nodes filtered out before writing the output trace")
918        ;
919
920    maxNumDependents
921        .name(name() + ".maxNumDependents")
922        .desc("Maximum number or dependents on any instruction")
923        ;
924
925    maxTempStoreSize
926        .name(name() + ".maxTempStoreSize")
927        .desc("Maximum size of the temporary store during the run")
928        ;
929
930    maxPhysRegDepMapSize
931        .name(name() + ".maxPhysRegDepMapSize")
932        .desc("Maximum size of register dependency map")
933        ;
934}
935
936const std::string&
937ElasticTrace::TraceInfo::typeToStr() const
938{
939    return Record::RecordType_Name(type);
940}
941
942const std::string
943ElasticTrace::name() const
944{
945    return ProbeListenerObject::name();
946}
947
948void
949ElasticTrace::flushTraces()
950{
951    // Write to trace all records in the depTrace.
952    writeDepTrace(depTrace.size());
953    // Delete the stream objects
954    delete dataTraceStream;
955    delete instTraceStream;
956}
957
958ElasticTrace*
959ElasticTraceParams::create()
960{
961    return new ElasticTrace(this);
962}
963