physical.cc revision 11005
12391SN/A/* 210482Sandreas.hansson@arm.com * Copyright (c) 2012, 2014 ARM Limited 37733SAli.Saidi@ARM.com * All rights reserved 47733SAli.Saidi@ARM.com * 57733SAli.Saidi@ARM.com * The license below extends only to copyright in the software and shall 67733SAli.Saidi@ARM.com * not be construed as granting a license to any other intellectual 77733SAli.Saidi@ARM.com * property including but not limited to intellectual property relating 87733SAli.Saidi@ARM.com * to a hardware implementation of the functionality of the software 97733SAli.Saidi@ARM.com * licensed hereunder. You may use the software subject to the license 107733SAli.Saidi@ARM.com * terms below provided that you ensure that this notice is replicated 117733SAli.Saidi@ARM.com * unmodified and in its entirety in all distributions of the software, 127733SAli.Saidi@ARM.com * modified or unmodified, in source code or in binary form. 137733SAli.Saidi@ARM.com * 142391SN/A * Redistribution and use in source and binary forms, with or without 152391SN/A * modification, are permitted provided that the following conditions are 162391SN/A * met: redistributions of source code must retain the above copyright 172391SN/A * notice, this list of conditions and the following disclaimer; 182391SN/A * redistributions in binary form must reproduce the above copyright 192391SN/A * notice, this list of conditions and the following disclaimer in the 202391SN/A * documentation and/or other materials provided with the distribution; 212391SN/A * neither the name of the copyright holders nor the names of its 222391SN/A * contributors may be used to endorse or promote products derived from 232391SN/A * this software without specific prior written permission. 242391SN/A * 252391SN/A * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 262391SN/A * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 272391SN/A * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 282391SN/A * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 292391SN/A * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 302391SN/A * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT 312391SN/A * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 322391SN/A * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 332391SN/A * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 342391SN/A * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 352391SN/A * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 362665Ssaidi@eecs.umich.edu * 378931Sandreas.hansson@arm.com * Authors: Andreas Hansson 382391SN/A */ 392391SN/A 409293Sandreas.hansson@arm.com#include <sys/mman.h> 419293Sandreas.hansson@arm.com#include <sys/types.h> 429293Sandreas.hansson@arm.com#include <sys/user.h> 439293Sandreas.hansson@arm.com#include <fcntl.h> 449293Sandreas.hansson@arm.com#include <unistd.h> 459293Sandreas.hansson@arm.com#include <zlib.h> 469293Sandreas.hansson@arm.com 479293Sandreas.hansson@arm.com#include <cerrno> 489293Sandreas.hansson@arm.com#include <climits> 499293Sandreas.hansson@arm.com#include <cstdio> 509293Sandreas.hansson@arm.com#include <iostream> 519293Sandreas.hansson@arm.com#include <string> 529293Sandreas.hansson@arm.com 539356Snilay@cs.wisc.edu#include "base/trace.hh" 5410405Sandreas.hansson@arm.com#include "debug/AddrRanges.hh" 559293Sandreas.hansson@arm.com#include "debug/Checkpoint.hh" 569293Sandreas.hansson@arm.com#include "mem/abstract_mem.hh" 572394SN/A#include "mem/physical.hh" 582394SN/A 5910700Sandreas.hansson@arm.com/** 6010700Sandreas.hansson@arm.com * On Linux, MAP_NORESERVE allow us to simulate a very large memory 6110700Sandreas.hansson@arm.com * without committing to actually providing the swap space on the 6210700Sandreas.hansson@arm.com * host. On OSX the MAP_NORESERVE flag does not exist, so simply make 6310700Sandreas.hansson@arm.com * it 0. 6410700Sandreas.hansson@arm.com */ 6510700Sandreas.hansson@arm.com#if defined(__APPLE__) 6610700Sandreas.hansson@arm.com#ifndef MAP_NORESERVE 6710700Sandreas.hansson@arm.com#define MAP_NORESERVE 0 6810700Sandreas.hansson@arm.com#endif 6910700Sandreas.hansson@arm.com#endif 7010700Sandreas.hansson@arm.com 712391SN/Ausing namespace std; 722391SN/A 739293Sandreas.hansson@arm.comPhysicalMemory::PhysicalMemory(const string& _name, 7410700Sandreas.hansson@arm.com const vector<AbstractMemory*>& _memories, 7510700Sandreas.hansson@arm.com bool mmap_using_noreserve) : 7610700Sandreas.hansson@arm.com _name(_name), rangeCache(addrMap.end()), size(0), 7710700Sandreas.hansson@arm.com mmapUsingNoReserve(mmap_using_noreserve) 782391SN/A{ 7910700Sandreas.hansson@arm.com if (mmap_using_noreserve) 8010700Sandreas.hansson@arm.com warn("Not reserving swap space. May cause SIGSEGV on actual usage\n"); 8110700Sandreas.hansson@arm.com 829293Sandreas.hansson@arm.com // add the memories from the system to the address map as 839293Sandreas.hansson@arm.com // appropriate 8410482Sandreas.hansson@arm.com for (const auto& m : _memories) { 858931Sandreas.hansson@arm.com // only add the memory if it is part of the global address map 8610482Sandreas.hansson@arm.com if (m->isInAddrMap()) { 8710482Sandreas.hansson@arm.com memories.push_back(m); 882391SN/A 898931Sandreas.hansson@arm.com // calculate the total size once and for all 9010482Sandreas.hansson@arm.com size += m->size(); 918931Sandreas.hansson@arm.com 928931Sandreas.hansson@arm.com // add the range to our interval tree and make sure it does not 938931Sandreas.hansson@arm.com // intersect an existing range 9410482Sandreas.hansson@arm.com fatal_if(addrMap.insert(m->getAddrRange(), m) == addrMap.end(), 9510482Sandreas.hansson@arm.com "Memory address range for %s is overlapping\n", 9610482Sandreas.hansson@arm.com m->name()); 979293Sandreas.hansson@arm.com } else { 989293Sandreas.hansson@arm.com // this type of memory is used e.g. as reference memory by 999293Sandreas.hansson@arm.com // Ruby, and they also needs a backing store, but should 1009293Sandreas.hansson@arm.com // not be part of the global address map 10110482Sandreas.hansson@arm.com DPRINTF(AddrRanges, 10210482Sandreas.hansson@arm.com "Skipping memory %s that is not in global address map\n", 10310482Sandreas.hansson@arm.com m->name()); 10410482Sandreas.hansson@arm.com 10510482Sandreas.hansson@arm.com // sanity check 10610482Sandreas.hansson@arm.com fatal_if(m->getAddrRange().interleaved(), 10710482Sandreas.hansson@arm.com "Memory %s that is not in the global address map cannot " 10810482Sandreas.hansson@arm.com "be interleaved\n", m->name()); 1099293Sandreas.hansson@arm.com 1109293Sandreas.hansson@arm.com // simply do it independently, also note that this kind of 1119293Sandreas.hansson@arm.com // memories are allowed to overlap in the logic address 1129293Sandreas.hansson@arm.com // map 11310482Sandreas.hansson@arm.com vector<AbstractMemory*> unmapped_mems{m}; 11410482Sandreas.hansson@arm.com createBackingStore(m->getAddrRange(), unmapped_mems); 1158931Sandreas.hansson@arm.com } 1169293Sandreas.hansson@arm.com } 1179293Sandreas.hansson@arm.com 11810070Sandreas.hansson@arm.com // iterate over the increasing addresses and chunks of contiguous 11910070Sandreas.hansson@arm.com // space to be mapped to backing store, create it and inform the 12010070Sandreas.hansson@arm.com // memories 1219565Sandreas.hansson@arm.com vector<AddrRange> intlv_ranges; 1229293Sandreas.hansson@arm.com vector<AbstractMemory*> curr_memories; 12310482Sandreas.hansson@arm.com for (const auto& r : addrMap) { 1249293Sandreas.hansson@arm.com // simply skip past all memories that are null and hence do 1259293Sandreas.hansson@arm.com // not need any backing store 12610482Sandreas.hansson@arm.com if (!r.second->isNull()) { 1279565Sandreas.hansson@arm.com // if the range is interleaved then save it for now 12810482Sandreas.hansson@arm.com if (r.first.interleaved()) { 1299565Sandreas.hansson@arm.com // if we already got interleaved ranges that are not 1309565Sandreas.hansson@arm.com // part of the same range, then first do a merge 1319565Sandreas.hansson@arm.com // before we add the new one 1329565Sandreas.hansson@arm.com if (!intlv_ranges.empty() && 13310482Sandreas.hansson@arm.com !intlv_ranges.back().mergesWith(r.first)) { 1349565Sandreas.hansson@arm.com AddrRange merged_range(intlv_ranges); 1359565Sandreas.hansson@arm.com createBackingStore(merged_range, curr_memories); 1369565Sandreas.hansson@arm.com intlv_ranges.clear(); 1379565Sandreas.hansson@arm.com curr_memories.clear(); 1389565Sandreas.hansson@arm.com } 13910482Sandreas.hansson@arm.com intlv_ranges.push_back(r.first); 14010482Sandreas.hansson@arm.com curr_memories.push_back(r.second); 1419565Sandreas.hansson@arm.com } else { 14210482Sandreas.hansson@arm.com vector<AbstractMemory*> single_memory{r.second}; 14310482Sandreas.hansson@arm.com createBackingStore(r.first, single_memory); 1449565Sandreas.hansson@arm.com } 1459293Sandreas.hansson@arm.com } 1469293Sandreas.hansson@arm.com } 1479565Sandreas.hansson@arm.com 1489565Sandreas.hansson@arm.com // if there is still interleaved ranges waiting to be merged, go 1499565Sandreas.hansson@arm.com // ahead and do it 1509565Sandreas.hansson@arm.com if (!intlv_ranges.empty()) { 1519565Sandreas.hansson@arm.com AddrRange merged_range(intlv_ranges); 1529565Sandreas.hansson@arm.com createBackingStore(merged_range, curr_memories); 1539565Sandreas.hansson@arm.com } 1549293Sandreas.hansson@arm.com} 1559293Sandreas.hansson@arm.com 1569293Sandreas.hansson@arm.comvoid 1579293Sandreas.hansson@arm.comPhysicalMemory::createBackingStore(AddrRange range, 1589293Sandreas.hansson@arm.com const vector<AbstractMemory*>& _memories) 1599293Sandreas.hansson@arm.com{ 16010482Sandreas.hansson@arm.com panic_if(range.interleaved(), 16110482Sandreas.hansson@arm.com "Cannot create backing store for interleaved range %s\n", 1629411Sandreas.hansson@arm.com range.to_string()); 1639411Sandreas.hansson@arm.com 1649293Sandreas.hansson@arm.com // perform the actual mmap 16510405Sandreas.hansson@arm.com DPRINTF(AddrRanges, "Creating backing store for range %s with size %d\n", 1669411Sandreas.hansson@arm.com range.to_string(), range.size()); 1679293Sandreas.hansson@arm.com int map_flags = MAP_ANON | MAP_PRIVATE; 16810700Sandreas.hansson@arm.com 16910700Sandreas.hansson@arm.com // to be able to simulate very large memories, the user can opt to 17010700Sandreas.hansson@arm.com // pass noreserve to mmap 17110700Sandreas.hansson@arm.com if (mmapUsingNoReserve) { 17210700Sandreas.hansson@arm.com map_flags |= MAP_NORESERVE; 17310700Sandreas.hansson@arm.com } 17410700Sandreas.hansson@arm.com 1759293Sandreas.hansson@arm.com uint8_t* pmem = (uint8_t*) mmap(NULL, range.size(), 1769293Sandreas.hansson@arm.com PROT_READ | PROT_WRITE, 1779293Sandreas.hansson@arm.com map_flags, -1, 0); 1789293Sandreas.hansson@arm.com 1799293Sandreas.hansson@arm.com if (pmem == (uint8_t*) MAP_FAILED) { 1809293Sandreas.hansson@arm.com perror("mmap"); 1819405Sandreas.hansson@arm.com fatal("Could not mmap %d bytes for range %s!\n", range.size(), 1829405Sandreas.hansson@arm.com range.to_string()); 1839293Sandreas.hansson@arm.com } 1849293Sandreas.hansson@arm.com 1859293Sandreas.hansson@arm.com // remember this backing store so we can checkpoint it and unmap 1869293Sandreas.hansson@arm.com // it appropriately 1879293Sandreas.hansson@arm.com backingStore.push_back(make_pair(range, pmem)); 1889293Sandreas.hansson@arm.com 18910070Sandreas.hansson@arm.com // point the memories to their backing store 19010482Sandreas.hansson@arm.com for (const auto& m : _memories) { 19110405Sandreas.hansson@arm.com DPRINTF(AddrRanges, "Mapping memory %s to backing store\n", 19210482Sandreas.hansson@arm.com m->name()); 19310482Sandreas.hansson@arm.com m->setBackingStore(pmem); 1948931Sandreas.hansson@arm.com } 1958931Sandreas.hansson@arm.com} 1968931Sandreas.hansson@arm.com 1979293Sandreas.hansson@arm.comPhysicalMemory::~PhysicalMemory() 1989293Sandreas.hansson@arm.com{ 1999293Sandreas.hansson@arm.com // unmap the backing store 20010482Sandreas.hansson@arm.com for (auto& s : backingStore) 20110482Sandreas.hansson@arm.com munmap((char*)s.second, s.first.size()); 2029293Sandreas.hansson@arm.com} 2039293Sandreas.hansson@arm.com 2048931Sandreas.hansson@arm.combool 2058931Sandreas.hansson@arm.comPhysicalMemory::isMemAddr(Addr addr) const 2068931Sandreas.hansson@arm.com{ 2078931Sandreas.hansson@arm.com // see if the address is within the last matched range 20810699Sandreas.hansson@arm.com if (rangeCache != addrMap.end() && rangeCache->first.contains(addr)) { 20910699Sandreas.hansson@arm.com return true; 21010699Sandreas.hansson@arm.com } else { 2118931Sandreas.hansson@arm.com // lookup in the interval tree 21210482Sandreas.hansson@arm.com const auto& r = addrMap.find(addr); 2138931Sandreas.hansson@arm.com if (r == addrMap.end()) { 2148931Sandreas.hansson@arm.com // not in the cache, and not in the tree 2158931Sandreas.hansson@arm.com return false; 2168931Sandreas.hansson@arm.com } 2178931Sandreas.hansson@arm.com // the range is in the tree, update the cache 21810699Sandreas.hansson@arm.com rangeCache = r; 21910699Sandreas.hansson@arm.com return true; 2208851Sandreas.hansson@arm.com } 2218931Sandreas.hansson@arm.com} 2227730SAli.Saidi@ARM.com 2238931Sandreas.hansson@arm.comAddrRangeList 2248931Sandreas.hansson@arm.comPhysicalMemory::getConfAddrRanges() const 2258931Sandreas.hansson@arm.com{ 2268931Sandreas.hansson@arm.com // this could be done once in the constructor, but since it is unlikely to 2278931Sandreas.hansson@arm.com // be called more than once the iteration should not be a problem 2288931Sandreas.hansson@arm.com AddrRangeList ranges; 2299413Sandreas.hansson@arm.com vector<AddrRange> intlv_ranges; 23010482Sandreas.hansson@arm.com for (const auto& r : addrMap) { 23110482Sandreas.hansson@arm.com if (r.second->isConfReported()) { 2329413Sandreas.hansson@arm.com // if the range is interleaved then save it for now 23310482Sandreas.hansson@arm.com if (r.first.interleaved()) { 2349413Sandreas.hansson@arm.com // if we already got interleaved ranges that are not 2359413Sandreas.hansson@arm.com // part of the same range, then first do a merge 2369413Sandreas.hansson@arm.com // before we add the new one 2379413Sandreas.hansson@arm.com if (!intlv_ranges.empty() && 23810482Sandreas.hansson@arm.com !intlv_ranges.back().mergesWith(r.first)) { 2399413Sandreas.hansson@arm.com ranges.push_back(AddrRange(intlv_ranges)); 2409413Sandreas.hansson@arm.com intlv_ranges.clear(); 2419413Sandreas.hansson@arm.com } 24210482Sandreas.hansson@arm.com intlv_ranges.push_back(r.first); 2439413Sandreas.hansson@arm.com } else { 2449413Sandreas.hansson@arm.com // keep the current range 24510482Sandreas.hansson@arm.com ranges.push_back(r.first); 2469413Sandreas.hansson@arm.com } 2478931Sandreas.hansson@arm.com } 2487730SAli.Saidi@ARM.com } 2492391SN/A 2509413Sandreas.hansson@arm.com // if there is still interleaved ranges waiting to be merged, 2519413Sandreas.hansson@arm.com // go ahead and do it 2529413Sandreas.hansson@arm.com if (!intlv_ranges.empty()) { 2539413Sandreas.hansson@arm.com ranges.push_back(AddrRange(intlv_ranges)); 2549413Sandreas.hansson@arm.com } 2559413Sandreas.hansson@arm.com 2568931Sandreas.hansson@arm.com return ranges; 2572391SN/A} 2582391SN/A 2592541SN/Avoid 2608931Sandreas.hansson@arm.comPhysicalMemory::access(PacketPtr pkt) 2612541SN/A{ 2628931Sandreas.hansson@arm.com assert(pkt->isRequest()); 2638931Sandreas.hansson@arm.com Addr addr = pkt->getAddr(); 26410699Sandreas.hansson@arm.com if (rangeCache != addrMap.end() && rangeCache->first.contains(addr)) { 26510699Sandreas.hansson@arm.com rangeCache->second->access(pkt); 26610699Sandreas.hansson@arm.com } else { 26710699Sandreas.hansson@arm.com // do not update the cache here, as we typically call 26810699Sandreas.hansson@arm.com // isMemAddr before calling access 26910699Sandreas.hansson@arm.com const auto& m = addrMap.find(addr); 27010699Sandreas.hansson@arm.com assert(m != addrMap.end()); 27110699Sandreas.hansson@arm.com m->second->access(pkt); 27210699Sandreas.hansson@arm.com } 2732391SN/A} 2742391SN/A 2758719SAli.Saidi@ARM.comvoid 2768931Sandreas.hansson@arm.comPhysicalMemory::functionalAccess(PacketPtr pkt) 2778719SAli.Saidi@ARM.com{ 2788931Sandreas.hansson@arm.com assert(pkt->isRequest()); 2798931Sandreas.hansson@arm.com Addr addr = pkt->getAddr(); 28010699Sandreas.hansson@arm.com if (rangeCache != addrMap.end() && rangeCache->first.contains(addr)) { 28110699Sandreas.hansson@arm.com rangeCache->second->functionalAccess(pkt); 28210699Sandreas.hansson@arm.com } else { 28310699Sandreas.hansson@arm.com // do not update the cache here, as we typically call 28410699Sandreas.hansson@arm.com // isMemAddr before calling functionalAccess 28510699Sandreas.hansson@arm.com const auto& m = addrMap.find(addr); 28610699Sandreas.hansson@arm.com assert(m != addrMap.end()); 28710699Sandreas.hansson@arm.com m->second->functionalAccess(pkt); 28810699Sandreas.hansson@arm.com } 2898719SAli.Saidi@ARM.com} 2909293Sandreas.hansson@arm.com 2919293Sandreas.hansson@arm.comvoid 29210905Sandreas.sandberg@arm.comPhysicalMemory::serialize(CheckpointOut &cp) const 2939293Sandreas.hansson@arm.com{ 2949293Sandreas.hansson@arm.com // serialize all the locked addresses and their context ids 2959293Sandreas.hansson@arm.com vector<Addr> lal_addr; 29611005Sandreas.sandberg@arm.com vector<ContextID> lal_cid; 2979293Sandreas.hansson@arm.com 29810482Sandreas.hansson@arm.com for (auto& m : memories) { 29910482Sandreas.hansson@arm.com const list<LockedAddr>& locked_addrs = m->getLockedAddrList(); 30010482Sandreas.hansson@arm.com for (const auto& l : locked_addrs) { 30110482Sandreas.hansson@arm.com lal_addr.push_back(l.addr); 30210482Sandreas.hansson@arm.com lal_cid.push_back(l.contextId); 3039293Sandreas.hansson@arm.com } 3049293Sandreas.hansson@arm.com } 3059293Sandreas.hansson@arm.com 30610905Sandreas.sandberg@arm.com SERIALIZE_CONTAINER(lal_addr); 30710905Sandreas.sandberg@arm.com SERIALIZE_CONTAINER(lal_cid); 3089293Sandreas.hansson@arm.com 3099293Sandreas.hansson@arm.com // serialize the backing stores 3109293Sandreas.hansson@arm.com unsigned int nbr_of_stores = backingStore.size(); 3119293Sandreas.hansson@arm.com SERIALIZE_SCALAR(nbr_of_stores); 3129293Sandreas.hansson@arm.com 3139293Sandreas.hansson@arm.com unsigned int store_id = 0; 3149293Sandreas.hansson@arm.com // store each backing store memory segment in a file 31510482Sandreas.hansson@arm.com for (auto& s : backingStore) { 31610905Sandreas.sandberg@arm.com ScopedCheckpointSection sec(cp, csprintf("store%d", store_id)); 31710905Sandreas.sandberg@arm.com serializeStore(cp, store_id++, s.first, s.second); 3189293Sandreas.hansson@arm.com } 3199293Sandreas.hansson@arm.com} 3209293Sandreas.hansson@arm.com 3219293Sandreas.hansson@arm.comvoid 32210905Sandreas.sandberg@arm.comPhysicalMemory::serializeStore(CheckpointOut &cp, unsigned int store_id, 32310905Sandreas.sandberg@arm.com AddrRange range, uint8_t* pmem) const 3249293Sandreas.hansson@arm.com{ 3259293Sandreas.hansson@arm.com // we cannot use the address range for the name as the 3269293Sandreas.hansson@arm.com // memories that are not part of the address map can overlap 3279386Sandreas.hansson@arm.com string filename = name() + ".store" + to_string(store_id) + ".pmem"; 3289293Sandreas.hansson@arm.com long range_size = range.size(); 3299293Sandreas.hansson@arm.com 3309293Sandreas.hansson@arm.com DPRINTF(Checkpoint, "Serializing physical memory %s with size %d\n", 3319293Sandreas.hansson@arm.com filename, range_size); 3329293Sandreas.hansson@arm.com 3339293Sandreas.hansson@arm.com SERIALIZE_SCALAR(store_id); 3349293Sandreas.hansson@arm.com SERIALIZE_SCALAR(filename); 3359293Sandreas.hansson@arm.com SERIALIZE_SCALAR(range_size); 3369293Sandreas.hansson@arm.com 3379293Sandreas.hansson@arm.com // write memory file 33810905Sandreas.sandberg@arm.com string filepath = CheckpointIn::dir() + "/" + filename.c_str(); 33910412Sandreas.hansson@arm.com gzFile compressed_mem = gzopen(filepath.c_str(), "wb"); 34010412Sandreas.hansson@arm.com if (compressed_mem == NULL) 3419293Sandreas.hansson@arm.com fatal("Can't open physical memory checkpoint file '%s'\n", 3429293Sandreas.hansson@arm.com filename); 3439293Sandreas.hansson@arm.com 3449293Sandreas.hansson@arm.com uint64_t pass_size = 0; 3459293Sandreas.hansson@arm.com 3469293Sandreas.hansson@arm.com // gzwrite fails if (int)len < 0 (gzwrite returns int) 3479293Sandreas.hansson@arm.com for (uint64_t written = 0; written < range.size(); 3489293Sandreas.hansson@arm.com written += pass_size) { 3499293Sandreas.hansson@arm.com pass_size = (uint64_t)INT_MAX < (range.size() - written) ? 3509293Sandreas.hansson@arm.com (uint64_t)INT_MAX : (range.size() - written); 3519293Sandreas.hansson@arm.com 3529293Sandreas.hansson@arm.com if (gzwrite(compressed_mem, pmem + written, 3539293Sandreas.hansson@arm.com (unsigned int) pass_size) != (int) pass_size) { 3549293Sandreas.hansson@arm.com fatal("Write failed on physical memory checkpoint file '%s'\n", 3559293Sandreas.hansson@arm.com filename); 3569293Sandreas.hansson@arm.com } 3579293Sandreas.hansson@arm.com } 3589293Sandreas.hansson@arm.com 3599293Sandreas.hansson@arm.com // close the compressed stream and check that the exit status 3609293Sandreas.hansson@arm.com // is zero 3619293Sandreas.hansson@arm.com if (gzclose(compressed_mem)) 3629293Sandreas.hansson@arm.com fatal("Close failed on physical memory checkpoint file '%s'\n", 3639293Sandreas.hansson@arm.com filename); 3649293Sandreas.hansson@arm.com 3659293Sandreas.hansson@arm.com} 3669293Sandreas.hansson@arm.com 3679293Sandreas.hansson@arm.comvoid 36810905Sandreas.sandberg@arm.comPhysicalMemory::unserialize(CheckpointIn &cp) 3699293Sandreas.hansson@arm.com{ 3709293Sandreas.hansson@arm.com // unserialize the locked addresses and map them to the 3719293Sandreas.hansson@arm.com // appropriate memory controller 3729293Sandreas.hansson@arm.com vector<Addr> lal_addr; 37311005Sandreas.sandberg@arm.com vector<ContextID> lal_cid; 37410905Sandreas.sandberg@arm.com UNSERIALIZE_CONTAINER(lal_addr); 37510905Sandreas.sandberg@arm.com UNSERIALIZE_CONTAINER(lal_cid); 3769293Sandreas.hansson@arm.com for(size_t i = 0; i < lal_addr.size(); ++i) { 37710482Sandreas.hansson@arm.com const auto& m = addrMap.find(lal_addr[i]); 3789293Sandreas.hansson@arm.com m->second->addLockedAddr(LockedAddr(lal_addr[i], lal_cid[i])); 3799293Sandreas.hansson@arm.com } 3809293Sandreas.hansson@arm.com 3819293Sandreas.hansson@arm.com // unserialize the backing stores 3829293Sandreas.hansson@arm.com unsigned int nbr_of_stores; 3839293Sandreas.hansson@arm.com UNSERIALIZE_SCALAR(nbr_of_stores); 3849293Sandreas.hansson@arm.com 3859293Sandreas.hansson@arm.com for (unsigned int i = 0; i < nbr_of_stores; ++i) { 38610905Sandreas.sandberg@arm.com ScopedCheckpointSection sec(cp, csprintf("store%d", i)); 38710905Sandreas.sandberg@arm.com unserializeStore(cp); 3889293Sandreas.hansson@arm.com } 3899293Sandreas.hansson@arm.com 3909293Sandreas.hansson@arm.com} 3919293Sandreas.hansson@arm.com 3929293Sandreas.hansson@arm.comvoid 39310905Sandreas.sandberg@arm.comPhysicalMemory::unserializeStore(CheckpointIn &cp) 3949293Sandreas.hansson@arm.com{ 3959293Sandreas.hansson@arm.com const uint32_t chunk_size = 16384; 3969293Sandreas.hansson@arm.com 3979293Sandreas.hansson@arm.com unsigned int store_id; 3989293Sandreas.hansson@arm.com UNSERIALIZE_SCALAR(store_id); 3999293Sandreas.hansson@arm.com 4009293Sandreas.hansson@arm.com string filename; 4019293Sandreas.hansson@arm.com UNSERIALIZE_SCALAR(filename); 40210905Sandreas.sandberg@arm.com string filepath = cp.cptDir + "/" + filename; 4039293Sandreas.hansson@arm.com 4049293Sandreas.hansson@arm.com // mmap memoryfile 40510412Sandreas.hansson@arm.com gzFile compressed_mem = gzopen(filepath.c_str(), "rb"); 40610412Sandreas.hansson@arm.com if (compressed_mem == NULL) 4079293Sandreas.hansson@arm.com fatal("Can't open physical memory checkpoint file '%s'", filename); 4089293Sandreas.hansson@arm.com 40910070Sandreas.hansson@arm.com // we've already got the actual backing store mapped 4109293Sandreas.hansson@arm.com uint8_t* pmem = backingStore[store_id].second; 4119293Sandreas.hansson@arm.com AddrRange range = backingStore[store_id].first; 4129293Sandreas.hansson@arm.com 4139293Sandreas.hansson@arm.com long range_size; 4149293Sandreas.hansson@arm.com UNSERIALIZE_SCALAR(range_size); 4159293Sandreas.hansson@arm.com 4169293Sandreas.hansson@arm.com DPRINTF(Checkpoint, "Unserializing physical memory %s with size %d\n", 4179293Sandreas.hansson@arm.com filename, range_size); 4189293Sandreas.hansson@arm.com 4199293Sandreas.hansson@arm.com if (range_size != range.size()) 4209293Sandreas.hansson@arm.com fatal("Memory range size has changed! Saw %lld, expected %lld\n", 4219293Sandreas.hansson@arm.com range_size, range.size()); 4229293Sandreas.hansson@arm.com 4239293Sandreas.hansson@arm.com uint64_t curr_size = 0; 4249293Sandreas.hansson@arm.com long* temp_page = new long[chunk_size]; 4259293Sandreas.hansson@arm.com long* pmem_current; 4269293Sandreas.hansson@arm.com uint32_t bytes_read; 4279293Sandreas.hansson@arm.com while (curr_size < range.size()) { 4289293Sandreas.hansson@arm.com bytes_read = gzread(compressed_mem, temp_page, chunk_size); 4299293Sandreas.hansson@arm.com if (bytes_read == 0) 4309293Sandreas.hansson@arm.com break; 4319293Sandreas.hansson@arm.com 4329293Sandreas.hansson@arm.com assert(bytes_read % sizeof(long) == 0); 4339293Sandreas.hansson@arm.com 4349293Sandreas.hansson@arm.com for (uint32_t x = 0; x < bytes_read / sizeof(long); x++) { 4359293Sandreas.hansson@arm.com // Only copy bytes that are non-zero, so we don't give 4369293Sandreas.hansson@arm.com // the VM system hell 4379293Sandreas.hansson@arm.com if (*(temp_page + x) != 0) { 4389293Sandreas.hansson@arm.com pmem_current = (long*)(pmem + curr_size + x * sizeof(long)); 4399293Sandreas.hansson@arm.com *pmem_current = *(temp_page + x); 4409293Sandreas.hansson@arm.com } 4419293Sandreas.hansson@arm.com } 4429293Sandreas.hansson@arm.com curr_size += bytes_read; 4439293Sandreas.hansson@arm.com } 4449293Sandreas.hansson@arm.com 4459293Sandreas.hansson@arm.com delete[] temp_page; 4469293Sandreas.hansson@arm.com 4479293Sandreas.hansson@arm.com if (gzclose(compressed_mem)) 4489293Sandreas.hansson@arm.com fatal("Close failed on physical memory checkpoint file '%s'\n", 4499293Sandreas.hansson@arm.com filename); 4509293Sandreas.hansson@arm.com} 451