MOESI_hammer.py (8180:d8587c913ccf) MOESI_hammer.py (8257:7226aebb77b4)
1# Copyright (c) 2006-2007 The Regents of The University of Michigan
2# Copyright (c) 2009 Advanced Micro Devices, Inc.
3# All rights reserved.
4#
5# Redistribution and use in source and binary forms, with or without
6# modification, are permitted provided that the following conditions are
7# met: redistributions of source code must retain the above copyright
8# notice, this list of conditions and the following disclaimer;
9# redistributions in binary form must reproduce the above copyright
10# notice, this list of conditions and the following disclaimer in the
11# documentation and/or other materials provided with the distribution;
12# neither the name of the copyright holders nor the names of its
13# contributors may be used to endorse or promote products derived from
14# this software without specific prior written permission.
15#
16# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
17# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
18# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
19# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
20# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
21# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
22# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
26# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27#
28# Authors: Brad Beckmann
29
30import math
31import m5
32from m5.objects import *
33from m5.defines import buildEnv
34
35#
36# Note: the L1 Cache latency is only used by the sequencer on fast path hits
37#
38class L1Cache(RubyCache):
39 latency = 2
40
41#
42# Note: the L2 Cache latency is not currently used
43#
44class L2Cache(RubyCache):
45 latency = 10
46
47#
48# Probe filter is a cache, latency is not used
49#
50class ProbeFilter(RubyCache):
51 latency = 1
52
53def define_options(parser):
54 parser.add_option("--allow-atomic-migration", action="store_true",
55 help="allow migratory sharing for atomic only accessed blocks")
56 parser.add_option("--pf-on", action="store_true",
57 help="Hammer: enable Probe Filter")
58 parser.add_option("--dir-on", action="store_true",
59 help="Hammer: enable Full-bit Directory")
60
61def create_system(options, system, piobus, dma_devices):
62
63 if buildEnv['PROTOCOL'] != 'MOESI_hammer':
64 panic("This script requires the MOESI_hammer protocol to be built.")
65
66 cpu_sequencers = []
67
68 #
69 # The ruby network creation expects the list of nodes in the system to be
70 # consistent with the NetDest list. Therefore the l1 controller nodes must be
71 # listed before the directory nodes and directory nodes before dma nodes, etc.
72 #
73 l1_cntrl_nodes = []
74 dir_cntrl_nodes = []
75 dma_cntrl_nodes = []
76
77 #
78 # Must create the individual controllers before the network to ensure the
79 # controller constructors are called before the network constructor
80 #
81 block_size_bits = int(math.log(options.cacheline_size, 2))
1# Copyright (c) 2006-2007 The Regents of The University of Michigan
2# Copyright (c) 2009 Advanced Micro Devices, Inc.
3# All rights reserved.
4#
5# Redistribution and use in source and binary forms, with or without
6# modification, are permitted provided that the following conditions are
7# met: redistributions of source code must retain the above copyright
8# notice, this list of conditions and the following disclaimer;
9# redistributions in binary form must reproduce the above copyright
10# notice, this list of conditions and the following disclaimer in the
11# documentation and/or other materials provided with the distribution;
12# neither the name of the copyright holders nor the names of its
13# contributors may be used to endorse or promote products derived from
14# this software without specific prior written permission.
15#
16# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
17# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
18# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
19# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
20# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
21# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
22# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
26# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27#
28# Authors: Brad Beckmann
29
30import math
31import m5
32from m5.objects import *
33from m5.defines import buildEnv
34
35#
36# Note: the L1 Cache latency is only used by the sequencer on fast path hits
37#
38class L1Cache(RubyCache):
39 latency = 2
40
41#
42# Note: the L2 Cache latency is not currently used
43#
44class L2Cache(RubyCache):
45 latency = 10
46
47#
48# Probe filter is a cache, latency is not used
49#
50class ProbeFilter(RubyCache):
51 latency = 1
52
53def define_options(parser):
54 parser.add_option("--allow-atomic-migration", action="store_true",
55 help="allow migratory sharing for atomic only accessed blocks")
56 parser.add_option("--pf-on", action="store_true",
57 help="Hammer: enable Probe Filter")
58 parser.add_option("--dir-on", action="store_true",
59 help="Hammer: enable Full-bit Directory")
60
61def create_system(options, system, piobus, dma_devices):
62
63 if buildEnv['PROTOCOL'] != 'MOESI_hammer':
64 panic("This script requires the MOESI_hammer protocol to be built.")
65
66 cpu_sequencers = []
67
68 #
69 # The ruby network creation expects the list of nodes in the system to be
70 # consistent with the NetDest list. Therefore the l1 controller nodes must be
71 # listed before the directory nodes and directory nodes before dma nodes, etc.
72 #
73 l1_cntrl_nodes = []
74 dir_cntrl_nodes = []
75 dma_cntrl_nodes = []
76
77 #
78 # Must create the individual controllers before the network to ensure the
79 # controller constructors are called before the network constructor
80 #
81 block_size_bits = int(math.log(options.cacheline_size, 2))
82
83 cntrl_count = 0
82
83 for i in xrange(options.num_cpus):
84 #
85 # First create the Ruby objects associated with this cpu
86 #
87 l1i_cache = L1Cache(size = options.l1i_size,
88 assoc = options.l1i_assoc,
89 start_index_bit = block_size_bits)
90 l1d_cache = L1Cache(size = options.l1d_size,
91 assoc = options.l1d_assoc,
92 start_index_bit = block_size_bits)
93 l2_cache = L2Cache(size = options.l2_size,
94 assoc = options.l2_assoc,
95 start_index_bit = block_size_bits)
96
97 cpu_seq = RubySequencer(version = i,
98 icache = l1i_cache,
99 dcache = l1d_cache,
100 physMemPort = system.physmem.port,
101 physmem = system.physmem)
102
103 if piobus != None:
104 cpu_seq.pio_port = piobus.port
105
106 l1_cntrl = L1Cache_Controller(version = i,
84
85 for i in xrange(options.num_cpus):
86 #
87 # First create the Ruby objects associated with this cpu
88 #
89 l1i_cache = L1Cache(size = options.l1i_size,
90 assoc = options.l1i_assoc,
91 start_index_bit = block_size_bits)
92 l1d_cache = L1Cache(size = options.l1d_size,
93 assoc = options.l1d_assoc,
94 start_index_bit = block_size_bits)
95 l2_cache = L2Cache(size = options.l2_size,
96 assoc = options.l2_assoc,
97 start_index_bit = block_size_bits)
98
99 cpu_seq = RubySequencer(version = i,
100 icache = l1i_cache,
101 dcache = l1d_cache,
102 physMemPort = system.physmem.port,
103 physmem = system.physmem)
104
105 if piobus != None:
106 cpu_seq.pio_port = piobus.port
107
108 l1_cntrl = L1Cache_Controller(version = i,
109 cntrl_id = cntrl_count,
107 sequencer = cpu_seq,
108 L1IcacheMemory = l1i_cache,
109 L1DcacheMemory = l1d_cache,
110 L2cacheMemory = l2_cache,
111 no_mig_atomic = not \
112 options.allow_atomic_migration)
113
114 if options.recycle_latency:
115 l1_cntrl.recycle_latency = options.recycle_latency
116
117 exec("system.l1_cntrl%d = l1_cntrl" % i)
118 #
119 # Add controllers and sequencers to the appropriate lists
120 #
121 cpu_sequencers.append(cpu_seq)
122 l1_cntrl_nodes.append(l1_cntrl)
123
110 sequencer = cpu_seq,
111 L1IcacheMemory = l1i_cache,
112 L1DcacheMemory = l1d_cache,
113 L2cacheMemory = l2_cache,
114 no_mig_atomic = not \
115 options.allow_atomic_migration)
116
117 if options.recycle_latency:
118 l1_cntrl.recycle_latency = options.recycle_latency
119
120 exec("system.l1_cntrl%d = l1_cntrl" % i)
121 #
122 # Add controllers and sequencers to the appropriate lists
123 #
124 cpu_sequencers.append(cpu_seq)
125 l1_cntrl_nodes.append(l1_cntrl)
126
127 cntrl_count += 1
128
124 phys_mem_size = long(system.physmem.range.second) - \
125 long(system.physmem.range.first) + 1
126 mem_module_size = phys_mem_size / options.num_dirs
127
128 #
129 # determine size and index bits for probe filter
130 # By default, the probe filter size is configured to be twice the
131 # size of the L2 cache.
132 #
133 pf_size = MemorySize(options.l2_size)
134 pf_size.value = pf_size.value * 2
135 dir_bits = int(math.log(options.num_dirs, 2))
136 pf_bits = int(math.log(pf_size.value, 2))
137 if options.numa_high_bit:
138 if options.numa_high_bit > 0:
139 # if numa high bit explicitly set, make sure it does not overlap
140 # with the probe filter index
141 assert(options.numa_high_bit - dir_bits > pf_bits)
142
143 # set the probe filter start bit to just above the block offset
144 pf_start_bit = 6
145 else:
146 if dir_bits > 0:
147 pf_start_bit = dir_bits + 5
148 else:
149 pf_start_bit = 6
150
151 for i in xrange(options.num_dirs):
152 #
153 # Create the Ruby objects associated with the directory controller
154 #
155
156 mem_cntrl = RubyMemoryControl(version = i)
157
158 dir_size = MemorySize('0B')
159 dir_size.value = mem_module_size
160
161 pf = ProbeFilter(size = pf_size, assoc = 4,
162 start_index_bit = pf_start_bit)
163
164 dir_cntrl = Directory_Controller(version = i,
129 phys_mem_size = long(system.physmem.range.second) - \
130 long(system.physmem.range.first) + 1
131 mem_module_size = phys_mem_size / options.num_dirs
132
133 #
134 # determine size and index bits for probe filter
135 # By default, the probe filter size is configured to be twice the
136 # size of the L2 cache.
137 #
138 pf_size = MemorySize(options.l2_size)
139 pf_size.value = pf_size.value * 2
140 dir_bits = int(math.log(options.num_dirs, 2))
141 pf_bits = int(math.log(pf_size.value, 2))
142 if options.numa_high_bit:
143 if options.numa_high_bit > 0:
144 # if numa high bit explicitly set, make sure it does not overlap
145 # with the probe filter index
146 assert(options.numa_high_bit - dir_bits > pf_bits)
147
148 # set the probe filter start bit to just above the block offset
149 pf_start_bit = 6
150 else:
151 if dir_bits > 0:
152 pf_start_bit = dir_bits + 5
153 else:
154 pf_start_bit = 6
155
156 for i in xrange(options.num_dirs):
157 #
158 # Create the Ruby objects associated with the directory controller
159 #
160
161 mem_cntrl = RubyMemoryControl(version = i)
162
163 dir_size = MemorySize('0B')
164 dir_size.value = mem_module_size
165
166 pf = ProbeFilter(size = pf_size, assoc = 4,
167 start_index_bit = pf_start_bit)
168
169 dir_cntrl = Directory_Controller(version = i,
170 cntrl_id = cntrl_count,
165 directory = \
166 RubyDirectoryMemory( \
167 version = i,
168 size = dir_size,
169 use_map = options.use_map,
170 map_levels = \
171 options.map_levels,
172 numa_high_bit = \
173 options.numa_high_bit),
174 probeFilter = pf,
175 memBuffer = mem_cntrl,
176 probe_filter_enabled = options.pf_on,
177 full_bit_dir_enabled = options.dir_on)
178
179 if options.recycle_latency:
180 dir_cntrl.recycle_latency = options.recycle_latency
181
182 exec("system.dir_cntrl%d = dir_cntrl" % i)
183 dir_cntrl_nodes.append(dir_cntrl)
184
171 directory = \
172 RubyDirectoryMemory( \
173 version = i,
174 size = dir_size,
175 use_map = options.use_map,
176 map_levels = \
177 options.map_levels,
178 numa_high_bit = \
179 options.numa_high_bit),
180 probeFilter = pf,
181 memBuffer = mem_cntrl,
182 probe_filter_enabled = options.pf_on,
183 full_bit_dir_enabled = options.dir_on)
184
185 if options.recycle_latency:
186 dir_cntrl.recycle_latency = options.recycle_latency
187
188 exec("system.dir_cntrl%d = dir_cntrl" % i)
189 dir_cntrl_nodes.append(dir_cntrl)
190
191 cntrl_count += 1
192
185 for i, dma_device in enumerate(dma_devices):
186 #
187 # Create the Ruby objects associated with the dma controller
188 #
189 dma_seq = DMASequencer(version = i,
190 physMemPort = system.physmem.port,
191 physmem = system.physmem)
192
193 dma_cntrl = DMA_Controller(version = i,
193 for i, dma_device in enumerate(dma_devices):
194 #
195 # Create the Ruby objects associated with the dma controller
196 #
197 dma_seq = DMASequencer(version = i,
198 physMemPort = system.physmem.port,
199 physmem = system.physmem)
200
201 dma_cntrl = DMA_Controller(version = i,
202 cntrl_id = cntrl_count,
194 dma_sequencer = dma_seq)
195
196 exec("system.dma_cntrl%d = dma_cntrl" % i)
197 if dma_device.type == 'MemTest':
198 exec("system.dma_cntrl%d.dma_sequencer.port = dma_device.test" % i)
199 else:
200 exec("system.dma_cntrl%d.dma_sequencer.port = dma_device.dma" % i)
201 dma_cntrl_nodes.append(dma_cntrl)
202
203 if options.recycle_latency:
204 dma_cntrl.recycle_latency = options.recycle_latency
205
203 dma_sequencer = dma_seq)
204
205 exec("system.dma_cntrl%d = dma_cntrl" % i)
206 if dma_device.type == 'MemTest':
207 exec("system.dma_cntrl%d.dma_sequencer.port = dma_device.test" % i)
208 else:
209 exec("system.dma_cntrl%d.dma_sequencer.port = dma_device.dma" % i)
210 dma_cntrl_nodes.append(dma_cntrl)
211
212 if options.recycle_latency:
213 dma_cntrl.recycle_latency = options.recycle_latency
214
215 cntrl_count += 1
216
206 all_cntrls = l1_cntrl_nodes + dir_cntrl_nodes + dma_cntrl_nodes
207
208 return (cpu_sequencers, dir_cntrl_nodes, all_cntrls)
217 all_cntrls = l1_cntrl_nodes + dir_cntrl_nodes + dma_cntrl_nodes
218
219 return (cpu_sequencers, dir_cntrl_nodes, all_cntrls)