1# -*- mode:python -*- 2 3# Copyright (c) 2004-2006 The Regents of The University of Michigan 4# All rights reserved. 5# 6# Redistribution and use in source and binary forms, with or without 7# modification, are permitted provided that the following conditions are 8# met: redistributions of source code must retain the above copyright
--- 125 unchanged lines hidden (view full) ---
134 return 0 135 136def update_test_string(target, source, env): 137 return env.subst("Updating ${SOURCES[0].dir} from ${SOURCES[1].dir}", 138 target=target, source=source) 139 140updateAction = env.Action(update_test, update_test_string) 141
|
142def test_builder(env, category, cpu_list=[], os_list=[], refdir='ref',
143 timeout=15):
144 """Define a test.
|
142def test_builder(env, ref_dir): 143 """Define a test.""" |
144
|
146 Args:
147 category -- string describing test category (e.g., 'quick')
148 cpu_list -- list of CPUs to runs this test on (blank means all compiled CPUs)
149 os_list -- list of OSs to run this test on
150 refdir -- subdirectory containing reference output (default 'ref')
151 timeout -- test timeout in minutes (only enforced on pool)
|
145 (category, name, _ref, isa, opsys, config) = ref_dir.split('/') 146 assert(_ref == 'ref') |
147
|
153 """
|
148 # target path (where test output goes) is the same except without 149 # the 'ref' component 150 tgt_dir = os.path.join(category, name, isa, opsys, config) |
151
|
155 default_refdir = False
156 if refdir == 'ref':
157 default_refdir = True
158 valid_cpu_list = []
159 if len(cpu_list) == 0:
160 valid_cpu_list = env['CPU_MODELS']
161 else:
162 for i in cpu_list:
163 if i in env['CPU_MODELS']:
164 valid_cpu_list.append(i)
165 cpu_list = valid_cpu_list
166 if env['TEST_CPU_MODELS']:
167 valid_cpu_list = []
168 for i in env['TEST_CPU_MODELS']:
169 if i in cpu_list:
170 valid_cpu_list.append(i)
171 cpu_list = valid_cpu_list
172# Code commented out that shows the general structure if we want to test
173# different OS's as well.
174# if len(os_list) == 0:
175# for test_cpu in cpu_list:
176# build_cpu_test(env, category, '', test_cpu, refdir, timeout)
177# else:
178# for test_os in os_list:
179# for test_cpu in cpu_list:
180# build_cpu_test(env, category, test_os, test_cpu, refdir,
181# timeout)
182 # Loop through CPU models and generate proper options, ref directories
183 for cpu in cpu_list:
184 test_os = ''
185 if cpu == "AtomicSimpleCPU":
186 cpu_option = ('','atomic/')
187 elif cpu == "TimingSimpleCPU":
188 cpu_option = ('--timing','timing/')
189 elif cpu == "O3CPU":
190 cpu_option = ('--detailed','detailed/')
191 else:
192 raise TypeError, "Unknown CPU model specified"
|
152 # prepend file name with tgt_dir 153 def tgt(f): 154 return os.path.join(tgt_dir, f) |
155
|
194 if default_refdir:
195 # Reference stats located in ref/arch/os/cpu or ref/arch/cpu
196 # if no OS specified
197 test_refdir = os.path.join(refdir, env['TARGET_ISA'])
198 if test_os != '':
199 test_refdir = os.path.join(test_refdir, test_os)
200 cpu_refdir = os.path.join(test_refdir, cpu_option[1])
|
156 ref_stats = os.path.join(ref_dir, 'm5stats.txt') 157 new_stats = tgt('m5stats.txt') 158 status_file = tgt('status') |
159
|
202 ref_stats = os.path.join(cpu_refdir, 'm5stats.txt')
|
160 # Base command for running test. We mess around with indirectly 161 # referring to files via SOURCES and TARGETS so that scons can 162 # mess with paths all it wants to and we still get the right 163 # files. 164 base_cmd = '${SOURCES[0]} -d $TARGET.dir ${SOURCES[1]} %s' % tgt_dir 165 # stdout and stderr files 166 cmd_stdout = '${TARGETS[0]}' 167 cmd_stderr = '${TARGETS[1]}' |
168
|
204 # base command for running test
205 base_cmd = '${SOURCES[0]} -d $TARGET.dir ${SOURCES[1]}'
206 base_cmd = base_cmd + ' ' + cpu_option[0]
207 # stdout and stderr files
208 cmd_stdout = '${TARGETS[0]}'
209 cmd_stderr = '${TARGETS[1]}'
210
211 stdout_string = cpu_option[1] + 'stdout'
212 stderr_string = cpu_option[1] + 'stderr'
213 m5stats_string = cpu_option[1] + 'm5stats.txt'
214 outdiff_string = cpu_option[1] + 'outdiff'
215 statsdiff_string = cpu_option[1] + 'statsdiff'
216 status_string = cpu_option[1] + 'status'
217
218 # Prefix test run with batch job submission command if appropriate.
219 # Output redirection is also different for batch runs.
220 # Batch command also supports timeout arg (in seconds, not minutes).
221 if env['BATCH']:
222 cmd = [env['BATCH_CMD'], '-t', str(timeout * 60),
223 '-o', cmd_stdout, '-e', cmd_stderr, base_cmd]
224 else:
225 cmd = [base_cmd, '>', cmd_stdout, '2>', cmd_stderr]
|
169 # Prefix test run with batch job submission command if appropriate. 170 # Output redirection is also different for batch runs. 171 # Batch command also supports timeout arg (in seconds, not minutes). 172 if env['BATCH']: 173 cmd = [env['BATCH_CMD'], '-t', str(timeout * 60), 174 '-o', cmd_stdout, '-e', cmd_stderr, base_cmd] 175 else: 176 cmd = [base_cmd, '>', cmd_stdout, '2>', cmd_stderr] |
177
|
227 env.Command([stdout_string, stderr_string, m5stats_string],
228 [env.M5Binary, 'run.py'], ' '.join(cmd))
|
178 env.Command([tgt('stdout'), tgt('stderr'), new_stats], 179 [env.M5Binary, 'run.py'], ' '.join(cmd)) |
180
|
230 # order of targets is important... see check_test
231 env.Command([outdiff_string, statsdiff_string, status_string],
232 [ref_stats, m5stats_string],
233 testAction)
|
181 # order of targets is important... see check_test 182 env.Command([tgt('outdiff'), tgt('statsdiff'), status_file], 183 [ref_stats, new_stats], 184 testAction) |
185
|
235 # phony target to echo status
236 if env['update_ref']:
237 p = env.Command(cpu_option[1] + '_update',
238 [ref_stats, m5stats_string, status_string],
239 updateAction)
240 else:
241 p = env.Command(cpu_option[1] + '_print', [status_string],
242 printAction)
243 env.AlwaysBuild(p)
|
186 # phony target to echo status 187 if env['update_ref']: 188 p = env.Command(tgt('_update'), 189 [ref_stats, new_stats, status_file], 190 updateAction) 191 else: 192 p = env.Command(tgt('_print'), [status_file], printAction) |
193
|
245 env.Tests.setdefault(category, [])
246 env.Tests[category] += p
|
194 env.AlwaysBuild(p) |
195
|
248# Make test_builder a "wrapper" function. See SCons wiki page at
249# http://www.scons.org/cgi-bin/wiki/WrapperFunctions.
250SConsEnvironment.Test = test_builder
|
196 197cwd = os.getcwd() 198os.chdir(str(Dir('.').srcdir))
|
254scripts = glob.glob('*/SConscript')
|
199for config in ['simple-atomic']: 200 dirs = glob.glob('*/*/ref/%s/*/%s' % (env['TARGET_ISA'], config)) 201 for d in dirs: 202 test_builder(env, d) |
203os.chdir(cwd)
|
256
257for s in scripts:
258 SConscript(s, exports = 'env', duplicate = False)
259
260# Set up phony commands for various test categories
261allTests = []
262for (key, val) in env.Tests.iteritems():
263 env.Command(key, val, env.NoAction)
264 allTests += val
265
266# The 'all' target is redundant since just specifying the test
267# directory name (e.g., ALPHA_SE/test/opt) has the same effect.
268env.Command('all', allTests, env.NoAction)
|
|