page_table.hh revision 11168:f98eb2da15a4
1/*
2 * Copyright (c) 2014 Advanced Micro Devices, Inc.
3 * Copyright (c) 2003 The Regents of The University of Michigan
4 * All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions are
8 * met: redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer;
10 * redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution;
13 * neither the name of the copyright holders nor the names of its
14 * contributors may be used to endorse or promote products derived from
15 * this software without specific prior written permission.
16 *
17 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
18 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
19 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
20 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
21 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
22 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
23 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
27 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28 *
29 * Authors: Steve Reinhardt
30 */
31
32/**
33 * @file
34 * Declarations of a non-full system Page Table.
35 */
36
37#ifndef __MEM_PAGE_TABLE_HH__
38#define __MEM_PAGE_TABLE_HH__
39
40#include <string>
41#include <unordered_map>
42
43#include "arch/isa_traits.hh"
44#include "arch/tlb.hh"
45#include "base/types.hh"
46#include "config/the_isa.hh"
47#include "mem/request.hh"
48#include "sim/serialize.hh"
49#include "sim/system.hh"
50
51class ThreadContext;
52
53/**
54 * Declaration of base class for page table
55 */
56class PageTableBase : public Serializable
57{
58  protected:
59    struct cacheElement {
60        bool valid;
61        Addr vaddr;
62        TheISA::TlbEntry entry;
63    };
64
65    struct cacheElement pTableCache[3];
66
67    const Addr pageSize;
68    const Addr offsetMask;
69
70    const uint64_t pid;
71    const std::string _name;
72
73  public:
74
75    PageTableBase(const std::string &__name, uint64_t _pid,
76              Addr _pageSize = TheISA::PageBytes)
77            : pageSize(_pageSize), offsetMask(mask(floorLog2(_pageSize))),
78              pid(_pid), _name(__name)
79    {
80        assert(isPowerOf2(pageSize));
81        pTableCache[0].valid = false;
82        pTableCache[1].valid = false;
83        pTableCache[2].valid = false;
84    }
85
86    virtual ~PageTableBase() {};
87
88    /* generic page table mapping flags
89     *              unset | set
90     * bit 0 - no-clobber | clobber
91     * bit 1 - present    | not-present
92     * bit 2 - cacheable  | uncacheable
93     * bit 3 - read-write | read-only
94     */
95    enum MappingFlags : uint32_t {
96        Clobber     = 1,
97        NotPresent  = 2,
98        Uncacheable = 4,
99        ReadOnly    = 8,
100    };
101
102    virtual void initState(ThreadContext* tc) = 0;
103
104    // for DPRINTF compatibility
105    const std::string name() const { return _name; }
106
107    Addr pageAlign(Addr a)  { return (a & ~offsetMask); }
108    Addr pageOffset(Addr a) { return (a &  offsetMask); }
109
110    /**
111     * Maps a virtual memory region to a physical memory region.
112     * @param vaddr The starting virtual address of the region.
113     * @param paddr The starting physical address where the region is mapped.
114     * @param size The length of the region.
115     * @param flags Generic mapping flags that can be set by or-ing values
116     *              from MappingFlags enum.
117     */
118    virtual void map(Addr vaddr, Addr paddr, int64_t size,
119                     uint64_t flags = 0) = 0;
120    virtual void remap(Addr vaddr, int64_t size, Addr new_vaddr) = 0;
121    virtual void unmap(Addr vaddr, int64_t size) = 0;
122
123    /**
124     * Check if any pages in a region are already allocated
125     * @param vaddr The starting virtual address of the region.
126     * @param size The length of the region.
127     * @return True if no pages in the region are mapped.
128     */
129    virtual bool isUnmapped(Addr vaddr, int64_t size) = 0;
130
131    /**
132     * Lookup function
133     * @param vaddr The virtual address.
134     * @return entry The page table entry corresponding to vaddr.
135     */
136    virtual bool lookup(Addr vaddr, TheISA::TlbEntry &entry) = 0;
137
138    /**
139     * Translate function
140     * @param vaddr The virtual address.
141     * @param paddr Physical address from translation.
142     * @return True if translation exists
143     */
144    bool translate(Addr vaddr, Addr &paddr);
145
146    /**
147     * Simplified translate function (just check for translation)
148     * @param vaddr The virtual address.
149     * @return True if translation exists
150     */
151    bool translate(Addr vaddr) { Addr dummy; return translate(vaddr, dummy); }
152
153    /**
154     * Perform a translation on the memory request, fills in paddr
155     * field of req.
156     * @param req The memory request.
157     */
158    Fault translate(RequestPtr req);
159
160    /**
161     * Update the page table cache.
162     * @param vaddr virtual address (page aligned) to check
163     * @param pte page table entry to return
164     */
165    inline void updateCache(Addr vaddr, TheISA::TlbEntry entry)
166    {
167        pTableCache[2].entry = pTableCache[1].entry;
168        pTableCache[2].vaddr = pTableCache[1].vaddr;
169        pTableCache[2].valid = pTableCache[1].valid;
170
171        pTableCache[1].entry = pTableCache[0].entry;
172        pTableCache[1].vaddr = pTableCache[0].vaddr;
173        pTableCache[1].valid = pTableCache[0].valid;
174
175        pTableCache[0].entry = entry;
176        pTableCache[0].vaddr = vaddr;
177        pTableCache[0].valid = true;
178    }
179
180    /**
181     * Erase an entry from the page table cache.
182     * @param vaddr virtual address (page aligned) to check
183     */
184    inline void eraseCacheEntry(Addr vaddr)
185    {
186        // Invalidate cached entries if necessary
187        if (pTableCache[0].valid && pTableCache[0].vaddr == vaddr) {
188            pTableCache[0].valid = false;
189        } else if (pTableCache[1].valid && pTableCache[1].vaddr == vaddr) {
190            pTableCache[1].valid = false;
191        } else if (pTableCache[2].valid && pTableCache[2].vaddr == vaddr) {
192            pTableCache[2].valid = false;
193        }
194    }
195};
196
197/**
198 * Declaration of functional page table.
199 */
200class FuncPageTable : public PageTableBase
201{
202  private:
203    typedef std::unordered_map<Addr, TheISA::TlbEntry> PTable;
204    typedef PTable::iterator PTableItr;
205    PTable pTable;
206
207  public:
208
209    FuncPageTable(const std::string &__name, uint64_t _pid,
210                  Addr _pageSize = TheISA::PageBytes);
211
212    ~FuncPageTable();
213
214    void initState(ThreadContext* tc)
215    {
216    }
217
218    void map(Addr vaddr, Addr paddr, int64_t size,
219             uint64_t flags = 0);
220    void remap(Addr vaddr, int64_t size, Addr new_vaddr);
221    void unmap(Addr vaddr, int64_t size);
222
223    /**
224     * Check if any pages in a region are already allocated
225     * @param vaddr The starting virtual address of the region.
226     * @param size The length of the region.
227     * @return True if no pages in the region are mapped.
228     */
229    bool isUnmapped(Addr vaddr, int64_t size);
230
231    /**
232     * Lookup function
233     * @param vaddr The virtual address.
234     * @return entry The page table entry corresponding to vaddr.
235     */
236    bool lookup(Addr vaddr, TheISA::TlbEntry &entry);
237
238    void serialize(CheckpointOut &cp) const override;
239    void unserialize(CheckpointIn &cp) override;
240};
241
242/**
243 * Faux page table class indended to stop the usage of
244 * an architectural page table, when there is none defined
245 * for a particular ISA.
246 */
247class NoArchPageTable : public FuncPageTable
248{
249  public:
250    NoArchPageTable(const std::string &__name, uint64_t _pid, System *_sys,
251              Addr _pageSize = TheISA::PageBytes) : FuncPageTable(__name, _pid)
252    {
253        fatal("No architectural page table defined for this ISA.\n");
254    }
255};
256
257#endif // __MEM_PAGE_TABLE_HH__
258