page_table.hh revision 12432:2480d8b432f5
1/* 2 * Copyright (c) 2014 Advanced Micro Devices, Inc. 3 * Copyright (c) 2003 The Regents of The University of Michigan 4 * All rights reserved. 5 * 6 * Redistribution and use in source and binary forms, with or without 7 * modification, are permitted provided that the following conditions are 8 * met: redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer; 10 * redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution; 13 * neither the name of the copyright holders nor the names of its 14 * contributors may be used to endorse or promote products derived from 15 * this software without specific prior written permission. 16 * 17 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 18 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 19 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 20 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 21 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 22 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT 23 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 27 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 28 * 29 * Authors: Steve Reinhardt 30 */ 31 32/** 33 * @file 34 * Declarations of a non-full system Page Table. 35 */ 36 37#ifndef __MEM_PAGE_TABLE_HH__ 38#define __MEM_PAGE_TABLE_HH__ 39 40#include <string> 41#include <unordered_map> 42 43#include "arch/isa_traits.hh" 44#include "arch/tlb.hh" 45#include "base/intmath.hh" 46#include "base/types.hh" 47#include "config/the_isa.hh" 48#include "mem/request.hh" 49#include "sim/serialize.hh" 50 51class ThreadContext; 52class System; 53 54/** 55 * Declaration of base class for page table 56 */ 57class PageTableBase : public Serializable 58{ 59 protected: 60 struct cacheElement { 61 bool valid; 62 Addr vaddr; 63 TheISA::TlbEntry entry; 64 }; 65 66 struct cacheElement pTableCache[3]; 67 68 const Addr pageSize; 69 const Addr offsetMask; 70 71 const uint64_t pid; 72 const std::string _name; 73 74 public: 75 76 PageTableBase(const std::string &__name, uint64_t _pid, Addr _pageSize) 77 : pageSize(_pageSize), offsetMask(mask(floorLog2(_pageSize))), 78 pid(_pid), _name(__name) 79 { 80 assert(isPowerOf2(pageSize)); 81 pTableCache[0].valid = false; 82 pTableCache[1].valid = false; 83 pTableCache[2].valid = false; 84 } 85 86 virtual ~PageTableBase() {}; 87 88 /* generic page table mapping flags 89 * unset | set 90 * bit 0 - no-clobber | clobber 91 * bit 1 - present | not-present 92 * bit 2 - cacheable | uncacheable 93 * bit 3 - read-write | read-only 94 */ 95 enum MappingFlags : uint32_t { 96 Zero = 0, 97 Clobber = 1, 98 NotPresent = 2, 99 Uncacheable = 4, 100 ReadOnly = 8, 101 }; 102 103 virtual void initState(ThreadContext* tc) = 0; 104 105 // for DPRINTF compatibility 106 const std::string name() const { return _name; } 107 108 Addr pageAlign(Addr a) { return (a & ~offsetMask); } 109 Addr pageOffset(Addr a) { return (a & offsetMask); } 110 111 /** 112 * Maps a virtual memory region to a physical memory region. 113 * @param vaddr The starting virtual address of the region. 114 * @param paddr The starting physical address where the region is mapped. 115 * @param size The length of the region. 116 * @param flags Generic mapping flags that can be set by or-ing values 117 * from MappingFlags enum. 118 */ 119 virtual void map(Addr vaddr, Addr paddr, int64_t size, 120 uint64_t flags = 0) = 0; 121 virtual void remap(Addr vaddr, int64_t size, Addr new_vaddr) = 0; 122 virtual void unmap(Addr vaddr, int64_t size) = 0; 123 124 /** 125 * Check if any pages in a region are already allocated 126 * @param vaddr The starting virtual address of the region. 127 * @param size The length of the region. 128 * @return True if no pages in the region are mapped. 129 */ 130 virtual bool isUnmapped(Addr vaddr, int64_t size) = 0; 131 132 /** 133 * Lookup function 134 * @param vaddr The virtual address. 135 * @return entry The page table entry corresponding to vaddr. 136 */ 137 virtual bool lookup(Addr vaddr, TheISA::TlbEntry &entry) = 0; 138 139 /** 140 * Translate function 141 * @param vaddr The virtual address. 142 * @param paddr Physical address from translation. 143 * @return True if translation exists 144 */ 145 bool translate(Addr vaddr, Addr &paddr); 146 147 /** 148 * Simplified translate function (just check for translation) 149 * @param vaddr The virtual address. 150 * @return True if translation exists 151 */ 152 bool translate(Addr vaddr) { Addr dummy; return translate(vaddr, dummy); } 153 154 /** 155 * Perform a translation on the memory request, fills in paddr 156 * field of req. 157 * @param req The memory request. 158 */ 159 Fault translate(RequestPtr req); 160 161 /** 162 * Update the page table cache. 163 * @param vaddr virtual address (page aligned) to check 164 * @param pte page table entry to return 165 */ 166 inline void updateCache(Addr vaddr, TheISA::TlbEntry entry) 167 { 168 pTableCache[2].entry = pTableCache[1].entry; 169 pTableCache[2].vaddr = pTableCache[1].vaddr; 170 pTableCache[2].valid = pTableCache[1].valid; 171 172 pTableCache[1].entry = pTableCache[0].entry; 173 pTableCache[1].vaddr = pTableCache[0].vaddr; 174 pTableCache[1].valid = pTableCache[0].valid; 175 176 pTableCache[0].entry = entry; 177 pTableCache[0].vaddr = vaddr; 178 pTableCache[0].valid = true; 179 } 180 181 /** 182 * Erase an entry from the page table cache. 183 * @param vaddr virtual address (page aligned) to check 184 */ 185 inline void eraseCacheEntry(Addr vaddr) 186 { 187 // Invalidate cached entries if necessary 188 if (pTableCache[0].valid && pTableCache[0].vaddr == vaddr) { 189 pTableCache[0].valid = false; 190 } else if (pTableCache[1].valid && pTableCache[1].vaddr == vaddr) { 191 pTableCache[1].valid = false; 192 } else if (pTableCache[2].valid && pTableCache[2].vaddr == vaddr) { 193 pTableCache[2].valid = false; 194 } 195 } 196 197 virtual void getMappings(std::vector<std::pair<Addr, Addr>> 198 *addr_mappings) {}; 199}; 200 201/** 202 * Declaration of functional page table. 203 */ 204class FuncPageTable : public PageTableBase 205{ 206 private: 207 typedef std::unordered_map<Addr, TheISA::TlbEntry> PTable; 208 typedef PTable::iterator PTableItr; 209 PTable pTable; 210 211 public: 212 213 FuncPageTable(const std::string &__name, uint64_t _pid, Addr _pageSize); 214 215 ~FuncPageTable(); 216 217 void initState(ThreadContext* tc) override 218 { 219 } 220 221 void map(Addr vaddr, Addr paddr, int64_t size, 222 uint64_t flags = 0) override; 223 void remap(Addr vaddr, int64_t size, Addr new_vaddr) override; 224 void unmap(Addr vaddr, int64_t size) override; 225 226 /** 227 * Check if any pages in a region are already allocated 228 * @param vaddr The starting virtual address of the region. 229 * @param size The length of the region. 230 * @return True if no pages in the region are mapped. 231 */ 232 bool isUnmapped(Addr vaddr, int64_t size) override; 233 234 /** 235 * Lookup function 236 * @param vaddr The virtual address. 237 * @return entry The page table entry corresponding to vaddr. 238 */ 239 bool lookup(Addr vaddr, TheISA::TlbEntry &entry) override; 240 241 void serialize(CheckpointOut &cp) const override; 242 void unserialize(CheckpointIn &cp) override; 243 244 void getMappings(std::vector<std::pair<Addr, Addr>> *addr_maps) override; 245}; 246 247#endif // __MEM_PAGE_TABLE_HH__ 248