physical.cc revision 12776
12391SN/A/*
212776Snikos.nikoleris@arm.com * Copyright (c) 2012, 2014, 2018 ARM Limited
37733SAli.Saidi@ARM.com * All rights reserved
47733SAli.Saidi@ARM.com *
57733SAli.Saidi@ARM.com * The license below extends only to copyright in the software and shall
67733SAli.Saidi@ARM.com * not be construed as granting a license to any other intellectual
77733SAli.Saidi@ARM.com * property including but not limited to intellectual property relating
87733SAli.Saidi@ARM.com * to a hardware implementation of the functionality of the software
97733SAli.Saidi@ARM.com * licensed hereunder.  You may use the software subject to the license
107733SAli.Saidi@ARM.com * terms below provided that you ensure that this notice is replicated
117733SAli.Saidi@ARM.com * unmodified and in its entirety in all distributions of the software,
127733SAli.Saidi@ARM.com * modified or unmodified, in source code or in binary form.
137733SAli.Saidi@ARM.com *
142391SN/A * Redistribution and use in source and binary forms, with or without
152391SN/A * modification, are permitted provided that the following conditions are
162391SN/A * met: redistributions of source code must retain the above copyright
172391SN/A * notice, this list of conditions and the following disclaimer;
182391SN/A * redistributions in binary form must reproduce the above copyright
192391SN/A * notice, this list of conditions and the following disclaimer in the
202391SN/A * documentation and/or other materials provided with the distribution;
212391SN/A * neither the name of the copyright holders nor the names of its
222391SN/A * contributors may be used to endorse or promote products derived from
232391SN/A * this software without specific prior written permission.
242391SN/A *
252391SN/A * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
262391SN/A * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
272391SN/A * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
282391SN/A * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
292391SN/A * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
302391SN/A * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
312391SN/A * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
322391SN/A * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
332391SN/A * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
342391SN/A * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
352391SN/A * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
362665Ssaidi@eecs.umich.edu *
378931Sandreas.hansson@arm.com * Authors: Andreas Hansson
382391SN/A */
392391SN/A
4011793Sbrandon.potter@amd.com#include "mem/physical.hh"
4111793Sbrandon.potter@amd.com
4211793Sbrandon.potter@amd.com#include <fcntl.h>
439293Sandreas.hansson@arm.com#include <sys/mman.h>
449293Sandreas.hansson@arm.com#include <sys/types.h>
459293Sandreas.hansson@arm.com#include <sys/user.h>
469293Sandreas.hansson@arm.com#include <unistd.h>
479293Sandreas.hansson@arm.com#include <zlib.h>
489293Sandreas.hansson@arm.com
499293Sandreas.hansson@arm.com#include <cerrno>
509293Sandreas.hansson@arm.com#include <climits>
519293Sandreas.hansson@arm.com#include <cstdio>
529293Sandreas.hansson@arm.com#include <iostream>
539293Sandreas.hansson@arm.com#include <string>
549293Sandreas.hansson@arm.com
559356Snilay@cs.wisc.edu#include "base/trace.hh"
5610405Sandreas.hansson@arm.com#include "debug/AddrRanges.hh"
579293Sandreas.hansson@arm.com#include "debug/Checkpoint.hh"
589293Sandreas.hansson@arm.com#include "mem/abstract_mem.hh"
592394SN/A
6010700Sandreas.hansson@arm.com/**
6110700Sandreas.hansson@arm.com * On Linux, MAP_NORESERVE allow us to simulate a very large memory
6210700Sandreas.hansson@arm.com * without committing to actually providing the swap space on the
6311446Sbaz21@cam.ac.uk * host. On FreeBSD or OSX the MAP_NORESERVE flag does not exist,
6411446Sbaz21@cam.ac.uk * so simply make it 0.
6510700Sandreas.hansson@arm.com */
6611446Sbaz21@cam.ac.uk#if defined(__APPLE__) || defined(__FreeBSD__)
6710700Sandreas.hansson@arm.com#ifndef MAP_NORESERVE
6810700Sandreas.hansson@arm.com#define MAP_NORESERVE 0
6910700Sandreas.hansson@arm.com#endif
7010700Sandreas.hansson@arm.com#endif
7110700Sandreas.hansson@arm.com
722391SN/Ausing namespace std;
732391SN/A
749293Sandreas.hansson@arm.comPhysicalMemory::PhysicalMemory(const string& _name,
7510700Sandreas.hansson@arm.com                               const vector<AbstractMemory*>& _memories,
7610700Sandreas.hansson@arm.com                               bool mmap_using_noreserve) :
7710700Sandreas.hansson@arm.com    _name(_name), rangeCache(addrMap.end()), size(0),
7810700Sandreas.hansson@arm.com    mmapUsingNoReserve(mmap_using_noreserve)
792391SN/A{
8010700Sandreas.hansson@arm.com    if (mmap_using_noreserve)
8110700Sandreas.hansson@arm.com        warn("Not reserving swap space. May cause SIGSEGV on actual usage\n");
8210700Sandreas.hansson@arm.com
839293Sandreas.hansson@arm.com    // add the memories from the system to the address map as
849293Sandreas.hansson@arm.com    // appropriate
8510482Sandreas.hansson@arm.com    for (const auto& m : _memories) {
868931Sandreas.hansson@arm.com        // only add the memory if it is part of the global address map
8710482Sandreas.hansson@arm.com        if (m->isInAddrMap()) {
8810482Sandreas.hansson@arm.com            memories.push_back(m);
892391SN/A
908931Sandreas.hansson@arm.com            // calculate the total size once and for all
9110482Sandreas.hansson@arm.com            size += m->size();
928931Sandreas.hansson@arm.com
938931Sandreas.hansson@arm.com            // add the range to our interval tree and make sure it does not
948931Sandreas.hansson@arm.com            // intersect an existing range
9510482Sandreas.hansson@arm.com            fatal_if(addrMap.insert(m->getAddrRange(), m) == addrMap.end(),
9610482Sandreas.hansson@arm.com                     "Memory address range for %s is overlapping\n",
9710482Sandreas.hansson@arm.com                     m->name());
989293Sandreas.hansson@arm.com        } else {
999293Sandreas.hansson@arm.com            // this type of memory is used e.g. as reference memory by
1009293Sandreas.hansson@arm.com            // Ruby, and they also needs a backing store, but should
1019293Sandreas.hansson@arm.com            // not be part of the global address map
10210482Sandreas.hansson@arm.com            DPRINTF(AddrRanges,
10310482Sandreas.hansson@arm.com                    "Skipping memory %s that is not in global address map\n",
10410482Sandreas.hansson@arm.com                    m->name());
10510482Sandreas.hansson@arm.com
10610482Sandreas.hansson@arm.com            // sanity check
10710482Sandreas.hansson@arm.com            fatal_if(m->getAddrRange().interleaved(),
10810482Sandreas.hansson@arm.com                     "Memory %s that is not in the global address map cannot "
10910482Sandreas.hansson@arm.com                     "be interleaved\n", m->name());
1109293Sandreas.hansson@arm.com
1119293Sandreas.hansson@arm.com            // simply do it independently, also note that this kind of
1129293Sandreas.hansson@arm.com            // memories are allowed to overlap in the logic address
1139293Sandreas.hansson@arm.com            // map
11410482Sandreas.hansson@arm.com            vector<AbstractMemory*> unmapped_mems{m};
11511614Sdavid.j.hashe@gmail.com            createBackingStore(m->getAddrRange(), unmapped_mems,
11611614Sdavid.j.hashe@gmail.com                               m->isConfReported(), m->isInAddrMap(),
11711614Sdavid.j.hashe@gmail.com                               m->isKvmMap());
1188931Sandreas.hansson@arm.com        }
1199293Sandreas.hansson@arm.com    }
1209293Sandreas.hansson@arm.com
12110070Sandreas.hansson@arm.com    // iterate over the increasing addresses and chunks of contiguous
12210070Sandreas.hansson@arm.com    // space to be mapped to backing store, create it and inform the
12310070Sandreas.hansson@arm.com    // memories
1249565Sandreas.hansson@arm.com    vector<AddrRange> intlv_ranges;
1259293Sandreas.hansson@arm.com    vector<AbstractMemory*> curr_memories;
12610482Sandreas.hansson@arm.com    for (const auto& r : addrMap) {
1279293Sandreas.hansson@arm.com        // simply skip past all memories that are null and hence do
1289293Sandreas.hansson@arm.com        // not need any backing store
12910482Sandreas.hansson@arm.com        if (!r.second->isNull()) {
1309565Sandreas.hansson@arm.com            // if the range is interleaved then save it for now
13110482Sandreas.hansson@arm.com            if (r.first.interleaved()) {
1329565Sandreas.hansson@arm.com                // if we already got interleaved ranges that are not
1339565Sandreas.hansson@arm.com                // part of the same range, then first do a merge
1349565Sandreas.hansson@arm.com                // before we add the new one
1359565Sandreas.hansson@arm.com                if (!intlv_ranges.empty() &&
13610482Sandreas.hansson@arm.com                    !intlv_ranges.back().mergesWith(r.first)) {
1379565Sandreas.hansson@arm.com                    AddrRange merged_range(intlv_ranges);
13811614Sdavid.j.hashe@gmail.com
13911614Sdavid.j.hashe@gmail.com                    AbstractMemory *f = curr_memories.front();
14011614Sdavid.j.hashe@gmail.com                    for (const auto& c : curr_memories)
14111614Sdavid.j.hashe@gmail.com                        if (f->isConfReported() != c->isConfReported() ||
14211614Sdavid.j.hashe@gmail.com                            f->isInAddrMap() != c->isInAddrMap() ||
14311614Sdavid.j.hashe@gmail.com                            f->isKvmMap() != c->isKvmMap())
14411614Sdavid.j.hashe@gmail.com                            fatal("Inconsistent flags in an interleaved "
14511614Sdavid.j.hashe@gmail.com                                  "range\n");
14611614Sdavid.j.hashe@gmail.com
14711614Sdavid.j.hashe@gmail.com                    createBackingStore(merged_range, curr_memories,
14811614Sdavid.j.hashe@gmail.com                                       f->isConfReported(), f->isInAddrMap(),
14911614Sdavid.j.hashe@gmail.com                                       f->isKvmMap());
15011614Sdavid.j.hashe@gmail.com
1519565Sandreas.hansson@arm.com                    intlv_ranges.clear();
1529565Sandreas.hansson@arm.com                    curr_memories.clear();
1539565Sandreas.hansson@arm.com                }
15410482Sandreas.hansson@arm.com                intlv_ranges.push_back(r.first);
15510482Sandreas.hansson@arm.com                curr_memories.push_back(r.second);
1569565Sandreas.hansson@arm.com            } else {
15710482Sandreas.hansson@arm.com                vector<AbstractMemory*> single_memory{r.second};
15811614Sdavid.j.hashe@gmail.com                createBackingStore(r.first, single_memory,
15911614Sdavid.j.hashe@gmail.com                                   r.second->isConfReported(),
16011614Sdavid.j.hashe@gmail.com                                   r.second->isInAddrMap(),
16111614Sdavid.j.hashe@gmail.com                                   r.second->isKvmMap());
1629565Sandreas.hansson@arm.com            }
1639293Sandreas.hansson@arm.com        }
1649293Sandreas.hansson@arm.com    }
1659565Sandreas.hansson@arm.com
1669565Sandreas.hansson@arm.com    // if there is still interleaved ranges waiting to be merged, go
1679565Sandreas.hansson@arm.com    // ahead and do it
1689565Sandreas.hansson@arm.com    if (!intlv_ranges.empty()) {
1699565Sandreas.hansson@arm.com        AddrRange merged_range(intlv_ranges);
17011614Sdavid.j.hashe@gmail.com
17111614Sdavid.j.hashe@gmail.com        AbstractMemory *f = curr_memories.front();
17211614Sdavid.j.hashe@gmail.com        for (const auto& c : curr_memories)
17311614Sdavid.j.hashe@gmail.com            if (f->isConfReported() != c->isConfReported() ||
17411614Sdavid.j.hashe@gmail.com                f->isInAddrMap() != c->isInAddrMap() ||
17511614Sdavid.j.hashe@gmail.com                f->isKvmMap() != c->isKvmMap())
17611614Sdavid.j.hashe@gmail.com                fatal("Inconsistent flags in an interleaved "
17711614Sdavid.j.hashe@gmail.com                      "range\n");
17811614Sdavid.j.hashe@gmail.com
17911614Sdavid.j.hashe@gmail.com        createBackingStore(merged_range, curr_memories,
18011614Sdavid.j.hashe@gmail.com                           f->isConfReported(), f->isInAddrMap(),
18111614Sdavid.j.hashe@gmail.com                           f->isKvmMap());
1829565Sandreas.hansson@arm.com    }
1839293Sandreas.hansson@arm.com}
1849293Sandreas.hansson@arm.com
1859293Sandreas.hansson@arm.comvoid
1869293Sandreas.hansson@arm.comPhysicalMemory::createBackingStore(AddrRange range,
18711614Sdavid.j.hashe@gmail.com                                   const vector<AbstractMemory*>& _memories,
18811614Sdavid.j.hashe@gmail.com                                   bool conf_table_reported,
18911614Sdavid.j.hashe@gmail.com                                   bool in_addr_map, bool kvm_map)
1909293Sandreas.hansson@arm.com{
19110482Sandreas.hansson@arm.com    panic_if(range.interleaved(),
19210482Sandreas.hansson@arm.com             "Cannot create backing store for interleaved range %s\n",
1939411Sandreas.hansson@arm.com              range.to_string());
1949411Sandreas.hansson@arm.com
1959293Sandreas.hansson@arm.com    // perform the actual mmap
19610405Sandreas.hansson@arm.com    DPRINTF(AddrRanges, "Creating backing store for range %s with size %d\n",
1979411Sandreas.hansson@arm.com            range.to_string(), range.size());
1989293Sandreas.hansson@arm.com    int map_flags = MAP_ANON | MAP_PRIVATE;
19910700Sandreas.hansson@arm.com
20010700Sandreas.hansson@arm.com    // to be able to simulate very large memories, the user can opt to
20110700Sandreas.hansson@arm.com    // pass noreserve to mmap
20210700Sandreas.hansson@arm.com    if (mmapUsingNoReserve) {
20310700Sandreas.hansson@arm.com        map_flags |= MAP_NORESERVE;
20410700Sandreas.hansson@arm.com    }
20510700Sandreas.hansson@arm.com
2069293Sandreas.hansson@arm.com    uint8_t* pmem = (uint8_t*) mmap(NULL, range.size(),
2079293Sandreas.hansson@arm.com                                    PROT_READ | PROT_WRITE,
2089293Sandreas.hansson@arm.com                                    map_flags, -1, 0);
2099293Sandreas.hansson@arm.com
2109293Sandreas.hansson@arm.com    if (pmem == (uint8_t*) MAP_FAILED) {
2119293Sandreas.hansson@arm.com        perror("mmap");
2129405Sandreas.hansson@arm.com        fatal("Could not mmap %d bytes for range %s!\n", range.size(),
2139405Sandreas.hansson@arm.com              range.to_string());
2149293Sandreas.hansson@arm.com    }
2159293Sandreas.hansson@arm.com
2169293Sandreas.hansson@arm.com    // remember this backing store so we can checkpoint it and unmap
2179293Sandreas.hansson@arm.com    // it appropriately
21811614Sdavid.j.hashe@gmail.com    backingStore.emplace_back(range, pmem,
21911614Sdavid.j.hashe@gmail.com                              conf_table_reported, in_addr_map, kvm_map);
2209293Sandreas.hansson@arm.com
22110070Sandreas.hansson@arm.com    // point the memories to their backing store
22210482Sandreas.hansson@arm.com    for (const auto& m : _memories) {
22310405Sandreas.hansson@arm.com        DPRINTF(AddrRanges, "Mapping memory %s to backing store\n",
22410482Sandreas.hansson@arm.com                m->name());
22510482Sandreas.hansson@arm.com        m->setBackingStore(pmem);
2268931Sandreas.hansson@arm.com    }
2278931Sandreas.hansson@arm.com}
2288931Sandreas.hansson@arm.com
2299293Sandreas.hansson@arm.comPhysicalMemory::~PhysicalMemory()
2309293Sandreas.hansson@arm.com{
2319293Sandreas.hansson@arm.com    // unmap the backing store
23210482Sandreas.hansson@arm.com    for (auto& s : backingStore)
23311614Sdavid.j.hashe@gmail.com        munmap((char*)s.pmem, s.range.size());
2349293Sandreas.hansson@arm.com}
2359293Sandreas.hansson@arm.com
2368931Sandreas.hansson@arm.combool
2378931Sandreas.hansson@arm.comPhysicalMemory::isMemAddr(Addr addr) const
2388931Sandreas.hansson@arm.com{
2398931Sandreas.hansson@arm.com    // see if the address is within the last matched range
24010699Sandreas.hansson@arm.com    if (rangeCache != addrMap.end() && rangeCache->first.contains(addr)) {
24110699Sandreas.hansson@arm.com        return true;
24210699Sandreas.hansson@arm.com    } else {
2438931Sandreas.hansson@arm.com        // lookup in the interval tree
24412776Snikos.nikoleris@arm.com        const auto& r = addrMap.contains(addr);
2458931Sandreas.hansson@arm.com        if (r == addrMap.end()) {
2468931Sandreas.hansson@arm.com            // not in the cache, and not in the tree
2478931Sandreas.hansson@arm.com            return false;
2488931Sandreas.hansson@arm.com        }
2498931Sandreas.hansson@arm.com        // the range is in the tree, update the cache
25010699Sandreas.hansson@arm.com        rangeCache = r;
25110699Sandreas.hansson@arm.com        return true;
2528851Sandreas.hansson@arm.com    }
2538931Sandreas.hansson@arm.com}
2547730SAli.Saidi@ARM.com
2558931Sandreas.hansson@arm.comAddrRangeList
2568931Sandreas.hansson@arm.comPhysicalMemory::getConfAddrRanges() const
2578931Sandreas.hansson@arm.com{
2588931Sandreas.hansson@arm.com    // this could be done once in the constructor, but since it is unlikely to
2598931Sandreas.hansson@arm.com    // be called more than once the iteration should not be a problem
2608931Sandreas.hansson@arm.com    AddrRangeList ranges;
2619413Sandreas.hansson@arm.com    vector<AddrRange> intlv_ranges;
26210482Sandreas.hansson@arm.com    for (const auto& r : addrMap) {
26310482Sandreas.hansson@arm.com        if (r.second->isConfReported()) {
2649413Sandreas.hansson@arm.com            // if the range is interleaved then save it for now
26510482Sandreas.hansson@arm.com            if (r.first.interleaved()) {
2669413Sandreas.hansson@arm.com                // if we already got interleaved ranges that are not
2679413Sandreas.hansson@arm.com                // part of the same range, then first do a merge
2689413Sandreas.hansson@arm.com                // before we add the new one
2699413Sandreas.hansson@arm.com                if (!intlv_ranges.empty() &&
27010482Sandreas.hansson@arm.com                    !intlv_ranges.back().mergesWith(r.first)) {
2719413Sandreas.hansson@arm.com                    ranges.push_back(AddrRange(intlv_ranges));
2729413Sandreas.hansson@arm.com                    intlv_ranges.clear();
2739413Sandreas.hansson@arm.com                }
27410482Sandreas.hansson@arm.com                intlv_ranges.push_back(r.first);
2759413Sandreas.hansson@arm.com            } else {
2769413Sandreas.hansson@arm.com                // keep the current range
27710482Sandreas.hansson@arm.com                ranges.push_back(r.first);
2789413Sandreas.hansson@arm.com            }
2798931Sandreas.hansson@arm.com        }
2807730SAli.Saidi@ARM.com    }
2812391SN/A
2829413Sandreas.hansson@arm.com    // if there is still interleaved ranges waiting to be merged,
2839413Sandreas.hansson@arm.com    // go ahead and do it
2849413Sandreas.hansson@arm.com    if (!intlv_ranges.empty()) {
2859413Sandreas.hansson@arm.com        ranges.push_back(AddrRange(intlv_ranges));
2869413Sandreas.hansson@arm.com    }
2879413Sandreas.hansson@arm.com
2888931Sandreas.hansson@arm.com    return ranges;
2892391SN/A}
2902391SN/A
2912541SN/Avoid
2928931Sandreas.hansson@arm.comPhysicalMemory::access(PacketPtr pkt)
2932541SN/A{
2948931Sandreas.hansson@arm.com    assert(pkt->isRequest());
2958931Sandreas.hansson@arm.com    Addr addr = pkt->getAddr();
29610699Sandreas.hansson@arm.com    if (rangeCache != addrMap.end() && rangeCache->first.contains(addr)) {
29710699Sandreas.hansson@arm.com        rangeCache->second->access(pkt);
29810699Sandreas.hansson@arm.com    } else {
29910699Sandreas.hansson@arm.com        // do not update the cache here, as we typically call
30010699Sandreas.hansson@arm.com        // isMemAddr before calling access
30112776Snikos.nikoleris@arm.com        const auto& m = addrMap.contains(addr);
30210699Sandreas.hansson@arm.com        assert(m != addrMap.end());
30310699Sandreas.hansson@arm.com        m->second->access(pkt);
30410699Sandreas.hansson@arm.com    }
3052391SN/A}
3062391SN/A
3078719SAli.Saidi@ARM.comvoid
3088931Sandreas.hansson@arm.comPhysicalMemory::functionalAccess(PacketPtr pkt)
3098719SAli.Saidi@ARM.com{
3108931Sandreas.hansson@arm.com    assert(pkt->isRequest());
3118931Sandreas.hansson@arm.com    Addr addr = pkt->getAddr();
31210699Sandreas.hansson@arm.com    if (rangeCache != addrMap.end() && rangeCache->first.contains(addr)) {
31310699Sandreas.hansson@arm.com        rangeCache->second->functionalAccess(pkt);
31410699Sandreas.hansson@arm.com    } else {
31510699Sandreas.hansson@arm.com        // do not update the cache here, as we typically call
31610699Sandreas.hansson@arm.com        // isMemAddr before calling functionalAccess
31712776Snikos.nikoleris@arm.com        const auto& m = addrMap.contains(addr);
31810699Sandreas.hansson@arm.com        assert(m != addrMap.end());
31910699Sandreas.hansson@arm.com        m->second->functionalAccess(pkt);
32010699Sandreas.hansson@arm.com    }
3218719SAli.Saidi@ARM.com}
3229293Sandreas.hansson@arm.com
3239293Sandreas.hansson@arm.comvoid
32410905Sandreas.sandberg@arm.comPhysicalMemory::serialize(CheckpointOut &cp) const
3259293Sandreas.hansson@arm.com{
3269293Sandreas.hansson@arm.com    // serialize all the locked addresses and their context ids
3279293Sandreas.hansson@arm.com    vector<Addr> lal_addr;
32811005Sandreas.sandberg@arm.com    vector<ContextID> lal_cid;
3299293Sandreas.hansson@arm.com
33010482Sandreas.hansson@arm.com    for (auto& m : memories) {
33110482Sandreas.hansson@arm.com        const list<LockedAddr>& locked_addrs = m->getLockedAddrList();
33210482Sandreas.hansson@arm.com        for (const auto& l : locked_addrs) {
33310482Sandreas.hansson@arm.com            lal_addr.push_back(l.addr);
33410482Sandreas.hansson@arm.com            lal_cid.push_back(l.contextId);
3359293Sandreas.hansson@arm.com        }
3369293Sandreas.hansson@arm.com    }
3379293Sandreas.hansson@arm.com
33810905Sandreas.sandberg@arm.com    SERIALIZE_CONTAINER(lal_addr);
33910905Sandreas.sandberg@arm.com    SERIALIZE_CONTAINER(lal_cid);
3409293Sandreas.hansson@arm.com
3419293Sandreas.hansson@arm.com    // serialize the backing stores
3429293Sandreas.hansson@arm.com    unsigned int nbr_of_stores = backingStore.size();
3439293Sandreas.hansson@arm.com    SERIALIZE_SCALAR(nbr_of_stores);
3449293Sandreas.hansson@arm.com
3459293Sandreas.hansson@arm.com    unsigned int store_id = 0;
3469293Sandreas.hansson@arm.com    // store each backing store memory segment in a file
34710482Sandreas.hansson@arm.com    for (auto& s : backingStore) {
34810905Sandreas.sandberg@arm.com        ScopedCheckpointSection sec(cp, csprintf("store%d", store_id));
34911614Sdavid.j.hashe@gmail.com        serializeStore(cp, store_id++, s.range, s.pmem);
3509293Sandreas.hansson@arm.com    }
3519293Sandreas.hansson@arm.com}
3529293Sandreas.hansson@arm.com
3539293Sandreas.hansson@arm.comvoid
35410905Sandreas.sandberg@arm.comPhysicalMemory::serializeStore(CheckpointOut &cp, unsigned int store_id,
35510905Sandreas.sandberg@arm.com                               AddrRange range, uint8_t* pmem) const
3569293Sandreas.hansson@arm.com{
3579293Sandreas.hansson@arm.com    // we cannot use the address range for the name as the
3589293Sandreas.hansson@arm.com    // memories that are not part of the address map can overlap
3599386Sandreas.hansson@arm.com    string filename = name() + ".store" + to_string(store_id) + ".pmem";
3609293Sandreas.hansson@arm.com    long range_size = range.size();
3619293Sandreas.hansson@arm.com
3629293Sandreas.hansson@arm.com    DPRINTF(Checkpoint, "Serializing physical memory %s with size %d\n",
3639293Sandreas.hansson@arm.com            filename, range_size);
3649293Sandreas.hansson@arm.com
3659293Sandreas.hansson@arm.com    SERIALIZE_SCALAR(store_id);
3669293Sandreas.hansson@arm.com    SERIALIZE_SCALAR(filename);
3679293Sandreas.hansson@arm.com    SERIALIZE_SCALAR(range_size);
3689293Sandreas.hansson@arm.com
3699293Sandreas.hansson@arm.com    // write memory file
37010905Sandreas.sandberg@arm.com    string filepath = CheckpointIn::dir() + "/" + filename.c_str();
37110412Sandreas.hansson@arm.com    gzFile compressed_mem = gzopen(filepath.c_str(), "wb");
37210412Sandreas.hansson@arm.com    if (compressed_mem == NULL)
3739293Sandreas.hansson@arm.com        fatal("Can't open physical memory checkpoint file '%s'\n",
3749293Sandreas.hansson@arm.com              filename);
3759293Sandreas.hansson@arm.com
3769293Sandreas.hansson@arm.com    uint64_t pass_size = 0;
3779293Sandreas.hansson@arm.com
3789293Sandreas.hansson@arm.com    // gzwrite fails if (int)len < 0 (gzwrite returns int)
3799293Sandreas.hansson@arm.com    for (uint64_t written = 0; written < range.size();
3809293Sandreas.hansson@arm.com         written += pass_size) {
3819293Sandreas.hansson@arm.com        pass_size = (uint64_t)INT_MAX < (range.size() - written) ?
3829293Sandreas.hansson@arm.com            (uint64_t)INT_MAX : (range.size() - written);
3839293Sandreas.hansson@arm.com
3849293Sandreas.hansson@arm.com        if (gzwrite(compressed_mem, pmem + written,
3859293Sandreas.hansson@arm.com                    (unsigned int) pass_size) != (int) pass_size) {
3869293Sandreas.hansson@arm.com            fatal("Write failed on physical memory checkpoint file '%s'\n",
3879293Sandreas.hansson@arm.com                  filename);
3889293Sandreas.hansson@arm.com        }
3899293Sandreas.hansson@arm.com    }
3909293Sandreas.hansson@arm.com
3919293Sandreas.hansson@arm.com    // close the compressed stream and check that the exit status
3929293Sandreas.hansson@arm.com    // is zero
3939293Sandreas.hansson@arm.com    if (gzclose(compressed_mem))
3949293Sandreas.hansson@arm.com        fatal("Close failed on physical memory checkpoint file '%s'\n",
3959293Sandreas.hansson@arm.com              filename);
3969293Sandreas.hansson@arm.com
3979293Sandreas.hansson@arm.com}
3989293Sandreas.hansson@arm.com
3999293Sandreas.hansson@arm.comvoid
40010905Sandreas.sandberg@arm.comPhysicalMemory::unserialize(CheckpointIn &cp)
4019293Sandreas.hansson@arm.com{
4029293Sandreas.hansson@arm.com    // unserialize the locked addresses and map them to the
4039293Sandreas.hansson@arm.com    // appropriate memory controller
4049293Sandreas.hansson@arm.com    vector<Addr> lal_addr;
40511005Sandreas.sandberg@arm.com    vector<ContextID> lal_cid;
40610905Sandreas.sandberg@arm.com    UNSERIALIZE_CONTAINER(lal_addr);
40710905Sandreas.sandberg@arm.com    UNSERIALIZE_CONTAINER(lal_cid);
40811321Ssteve.reinhardt@amd.com    for (size_t i = 0; i < lal_addr.size(); ++i) {
40912776Snikos.nikoleris@arm.com        const auto& m = addrMap.contains(lal_addr[i]);
4109293Sandreas.hansson@arm.com        m->second->addLockedAddr(LockedAddr(lal_addr[i], lal_cid[i]));
4119293Sandreas.hansson@arm.com    }
4129293Sandreas.hansson@arm.com
4139293Sandreas.hansson@arm.com    // unserialize the backing stores
4149293Sandreas.hansson@arm.com    unsigned int nbr_of_stores;
4159293Sandreas.hansson@arm.com    UNSERIALIZE_SCALAR(nbr_of_stores);
4169293Sandreas.hansson@arm.com
4179293Sandreas.hansson@arm.com    for (unsigned int i = 0; i < nbr_of_stores; ++i) {
41810905Sandreas.sandberg@arm.com        ScopedCheckpointSection sec(cp, csprintf("store%d", i));
41910905Sandreas.sandberg@arm.com        unserializeStore(cp);
4209293Sandreas.hansson@arm.com    }
4219293Sandreas.hansson@arm.com
4229293Sandreas.hansson@arm.com}
4239293Sandreas.hansson@arm.com
4249293Sandreas.hansson@arm.comvoid
42510905Sandreas.sandberg@arm.comPhysicalMemory::unserializeStore(CheckpointIn &cp)
4269293Sandreas.hansson@arm.com{
4279293Sandreas.hansson@arm.com    const uint32_t chunk_size = 16384;
4289293Sandreas.hansson@arm.com
4299293Sandreas.hansson@arm.com    unsigned int store_id;
4309293Sandreas.hansson@arm.com    UNSERIALIZE_SCALAR(store_id);
4319293Sandreas.hansson@arm.com
4329293Sandreas.hansson@arm.com    string filename;
4339293Sandreas.hansson@arm.com    UNSERIALIZE_SCALAR(filename);
43410905Sandreas.sandberg@arm.com    string filepath = cp.cptDir + "/" + filename;
4359293Sandreas.hansson@arm.com
4369293Sandreas.hansson@arm.com    // mmap memoryfile
43710412Sandreas.hansson@arm.com    gzFile compressed_mem = gzopen(filepath.c_str(), "rb");
43810412Sandreas.hansson@arm.com    if (compressed_mem == NULL)
4399293Sandreas.hansson@arm.com        fatal("Can't open physical memory checkpoint file '%s'", filename);
4409293Sandreas.hansson@arm.com
44110070Sandreas.hansson@arm.com    // we've already got the actual backing store mapped
44211614Sdavid.j.hashe@gmail.com    uint8_t* pmem = backingStore[store_id].pmem;
44311614Sdavid.j.hashe@gmail.com    AddrRange range = backingStore[store_id].range;
4449293Sandreas.hansson@arm.com
4459293Sandreas.hansson@arm.com    long range_size;
4469293Sandreas.hansson@arm.com    UNSERIALIZE_SCALAR(range_size);
4479293Sandreas.hansson@arm.com
4489293Sandreas.hansson@arm.com    DPRINTF(Checkpoint, "Unserializing physical memory %s with size %d\n",
4499293Sandreas.hansson@arm.com            filename, range_size);
4509293Sandreas.hansson@arm.com
4519293Sandreas.hansson@arm.com    if (range_size != range.size())
4529293Sandreas.hansson@arm.com        fatal("Memory range size has changed! Saw %lld, expected %lld\n",
4539293Sandreas.hansson@arm.com              range_size, range.size());
4549293Sandreas.hansson@arm.com
4559293Sandreas.hansson@arm.com    uint64_t curr_size = 0;
4569293Sandreas.hansson@arm.com    long* temp_page = new long[chunk_size];
4579293Sandreas.hansson@arm.com    long* pmem_current;
4589293Sandreas.hansson@arm.com    uint32_t bytes_read;
4599293Sandreas.hansson@arm.com    while (curr_size < range.size()) {
4609293Sandreas.hansson@arm.com        bytes_read = gzread(compressed_mem, temp_page, chunk_size);
4619293Sandreas.hansson@arm.com        if (bytes_read == 0)
4629293Sandreas.hansson@arm.com            break;
4639293Sandreas.hansson@arm.com
4649293Sandreas.hansson@arm.com        assert(bytes_read % sizeof(long) == 0);
4659293Sandreas.hansson@arm.com
4669293Sandreas.hansson@arm.com        for (uint32_t x = 0; x < bytes_read / sizeof(long); x++) {
4679293Sandreas.hansson@arm.com            // Only copy bytes that are non-zero, so we don't give
4689293Sandreas.hansson@arm.com            // the VM system hell
4699293Sandreas.hansson@arm.com            if (*(temp_page + x) != 0) {
4709293Sandreas.hansson@arm.com                pmem_current = (long*)(pmem + curr_size + x * sizeof(long));
4719293Sandreas.hansson@arm.com                *pmem_current = *(temp_page + x);
4729293Sandreas.hansson@arm.com            }
4739293Sandreas.hansson@arm.com        }
4749293Sandreas.hansson@arm.com        curr_size += bytes_read;
4759293Sandreas.hansson@arm.com    }
4769293Sandreas.hansson@arm.com
4779293Sandreas.hansson@arm.com    delete[] temp_page;
4789293Sandreas.hansson@arm.com
4799293Sandreas.hansson@arm.com    if (gzclose(compressed_mem))
4809293Sandreas.hansson@arm.com        fatal("Close failed on physical memory checkpoint file '%s'\n",
4819293Sandreas.hansson@arm.com              filename);
4829293Sandreas.hansson@arm.com}
483