2004-07-03 06:16:38 +02:00
|
|
|
# -*- mode:python -*-
|
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
# Copyright (c) 2009 The Hewlett-Packard Development Company
|
2005-06-05 11:16:00 +02:00
|
|
|
# Copyright (c) 2004-2005 The Regents of The University of Michigan
|
2004-07-03 06:16:38 +02:00
|
|
|
# All rights reserved.
|
|
|
|
#
|
|
|
|
# Redistribution and use in source and binary forms, with or without
|
|
|
|
# modification, are permitted provided that the following conditions are
|
|
|
|
# met: redistributions of source code must retain the above copyright
|
|
|
|
# notice, this list of conditions and the following disclaimer;
|
|
|
|
# redistributions in binary form must reproduce the above copyright
|
|
|
|
# notice, this list of conditions and the following disclaimer in the
|
|
|
|
# documentation and/or other materials provided with the distribution;
|
|
|
|
# neither the name of the copyright holders nor the names of its
|
|
|
|
# contributors may be used to endorse or promote products derived from
|
|
|
|
# this software without specific prior written permission.
|
|
|
|
#
|
|
|
|
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
|
|
# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
|
|
# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
|
|
# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
|
|
# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
|
|
# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
|
|
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
|
|
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
|
|
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
|
|
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
2006-06-01 01:26:56 +02:00
|
|
|
#
|
|
|
|
# Authors: Steve Reinhardt
|
2009-02-10 05:10:14 +01:00
|
|
|
# Nathan Binkert
|
2004-07-03 06:16:38 +02:00
|
|
|
|
|
|
|
###################################################
|
|
|
|
#
|
|
|
|
# SCons top-level build description (SConstruct) file.
|
|
|
|
#
|
2006-05-22 20:29:33 +02:00
|
|
|
# While in this directory ('m5'), just type 'scons' to build the default
|
|
|
|
# configuration (see below), or type 'scons build/<CONFIG>/<binary>'
|
|
|
|
# to build some other configuration (e.g., 'build/ALPHA_FS/m5.opt' for
|
|
|
|
# the optimized full-system version).
|
2004-07-03 06:16:38 +02:00
|
|
|
#
|
2006-05-22 20:29:33 +02:00
|
|
|
# You can build M5 in a different directory as long as there is a
|
|
|
|
# 'build/<CONFIG>' somewhere along the target path. The build system
|
2006-06-17 15:26:08 +02:00
|
|
|
# expects that all configs under the same build directory are being
|
2006-05-22 20:29:33 +02:00
|
|
|
# built for the same host system.
|
|
|
|
#
|
|
|
|
# Examples:
|
2006-06-17 15:26:08 +02:00
|
|
|
#
|
|
|
|
# The following two commands are equivalent. The '-u' option tells
|
|
|
|
# scons to search up the directory tree for this SConstruct file.
|
2006-05-22 20:29:33 +02:00
|
|
|
# % cd <path-to-src>/m5 ; scons build/ALPHA_FS/m5.debug
|
|
|
|
# % cd <path-to-src>/m5/build/ALPHA_FS; scons -u m5.debug
|
2006-06-17 15:26:08 +02:00
|
|
|
#
|
|
|
|
# The following two commands are equivalent and demonstrate building
|
|
|
|
# in a directory outside of the source tree. The '-C' option tells
|
|
|
|
# scons to chdir to the specified directory to find this SConstruct
|
|
|
|
# file.
|
2006-05-22 20:29:33 +02:00
|
|
|
# % cd <path-to-src>/m5 ; scons /local/foo/build/ALPHA_FS/m5.debug
|
|
|
|
# % cd /local/foo/build/ALPHA_FS; scons -C <path-to-src>/m5 m5.debug
|
|
|
|
#
|
|
|
|
# You can use 'scons -H' to print scons options. If you're in this
|
|
|
|
# 'm5' directory (or use -u or -C to tell scons where to find this
|
|
|
|
# file), you can use 'scons -h' to print all the M5-specific build
|
|
|
|
# options as well.
|
2004-07-03 06:16:38 +02:00
|
|
|
#
|
|
|
|
###################################################
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Check for recent-enough Python and SCons versions.
|
|
|
|
try:
|
|
|
|
# Really old versions of scons only take two options for the
|
|
|
|
# function, so check once without the revision and once with the
|
|
|
|
# revision, the first instance will fail for stuff other than
|
|
|
|
# 0.98, and the second will fail for 0.98.0
|
|
|
|
EnsureSConsVersion(0, 98)
|
|
|
|
EnsureSConsVersion(0, 98, 1)
|
|
|
|
except SystemExit, e:
|
|
|
|
print """
|
|
|
|
For more details, see:
|
|
|
|
http://m5sim.org/wiki/index.php/Compiling_M5
|
|
|
|
"""
|
|
|
|
raise
|
|
|
|
|
|
|
|
# We ensure the python version early because we have stuff that
|
|
|
|
# requires python 2.4
|
|
|
|
try:
|
|
|
|
EnsurePythonVersion(2, 4)
|
|
|
|
except SystemExit, e:
|
|
|
|
print """
|
|
|
|
You can use a non-default installation of the Python interpreter by
|
|
|
|
either (1) rearranging your PATH so that scons finds the non-default
|
|
|
|
'python' first or (2) explicitly invoking an alternative interpreter
|
|
|
|
on the scons script.
|
|
|
|
|
|
|
|
For more details, see:
|
|
|
|
http://m5sim.org/wiki/index.php/Using_a_non-default_Python_installation
|
|
|
|
"""
|
|
|
|
raise
|
|
|
|
|
2004-07-03 06:16:38 +02:00
|
|
|
import os
|
2008-04-08 05:40:23 +02:00
|
|
|
import re
|
2009-02-10 05:10:14 +01:00
|
|
|
import subprocess
|
|
|
|
import sys
|
2007-03-11 08:00:54 +01:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
from os import mkdir, environ
|
|
|
|
from os.path import abspath, basename, dirname, expanduser, normpath
|
|
|
|
from os.path import exists, isdir, isfile
|
|
|
|
from os.path import join as joinpath, split as splitpath
|
2004-07-03 06:16:38 +02:00
|
|
|
|
2007-11-17 05:10:33 +01:00
|
|
|
import SCons
|
2009-02-12 01:58:51 +01:00
|
|
|
import SCons.Node
|
2007-11-17 05:10:33 +01:00
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
def read_command(cmd, **kwargs):
|
2009-02-10 05:10:14 +01:00
|
|
|
"""run the command cmd, read the results and return them
|
|
|
|
this is sorta like `cmd` in shell"""
|
|
|
|
from subprocess import Popen, PIPE, STDOUT
|
2009-02-12 01:58:51 +01:00
|
|
|
|
2009-02-16 05:38:49 +01:00
|
|
|
if isinstance(cmd, str):
|
|
|
|
cmd = cmd.split()
|
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
no_exception = 'exception' in kwargs
|
|
|
|
exception = kwargs.pop('exception', None)
|
|
|
|
|
|
|
|
kwargs.setdefault('shell', False)
|
|
|
|
kwargs.setdefault('stdout', PIPE)
|
|
|
|
kwargs.setdefault('stderr', STDOUT)
|
|
|
|
kwargs.setdefault('close_fds', True)
|
|
|
|
try:
|
|
|
|
subp = Popen(cmd, **kwargs)
|
|
|
|
except Exception, e:
|
|
|
|
if no_exception:
|
|
|
|
return exception
|
|
|
|
raise
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
return subp.communicate()[0]
|
2007-11-08 23:45:58 +01:00
|
|
|
|
2008-04-08 05:40:23 +02:00
|
|
|
# helper function: compare arrays or strings of version numbers.
|
|
|
|
# E.g., compare_version((1,3,25), (1,4,1)')
|
|
|
|
# returns -1, 0, 1 if v1 is <, ==, > v2
|
|
|
|
def compare_versions(v1, v2):
|
|
|
|
def make_version_list(v):
|
|
|
|
if isinstance(v, (list,tuple)):
|
|
|
|
return v
|
|
|
|
elif isinstance(v, str):
|
2008-10-09 00:34:19 +02:00
|
|
|
return map(lambda x: int(re.match('\d+', x).group()), v.split('.'))
|
2008-04-08 05:40:23 +02:00
|
|
|
else:
|
|
|
|
raise TypeError
|
|
|
|
|
|
|
|
v1 = make_version_list(v1)
|
|
|
|
v2 = make_version_list(v2)
|
|
|
|
# Compare corresponding elements of lists
|
|
|
|
for n1,n2 in zip(v1, v2):
|
|
|
|
if n1 < n2: return -1
|
|
|
|
if n1 > n2: return 1
|
|
|
|
# all corresponding values are equal... see if one has extra values
|
|
|
|
if len(v1) < len(v2): return -1
|
|
|
|
if len(v1) > len(v2): return 1
|
|
|
|
return 0
|
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
########################################################################
|
|
|
|
#
|
|
|
|
# Set up the base build environment.
|
|
|
|
#
|
|
|
|
########################################################################
|
|
|
|
use_vars = set([ 'AS', 'AR', 'CC', 'CXX', 'HOME', 'PATH', 'RANLIB' ])
|
2004-07-03 06:16:38 +02:00
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
use_env = {}
|
|
|
|
for key,val in os.environ.iteritems():
|
|
|
|
if key in use_vars or key.startswith("M5"):
|
|
|
|
use_env[key] = val
|
2004-07-03 06:16:38 +02:00
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
env = Environment(ENV=use_env)
|
|
|
|
env.root = Dir(".") # The current directory (where this file lives).
|
|
|
|
env.srcdir = Dir("src") # The source directory
|
2005-03-14 20:43:10 +01:00
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
########################################################################
|
|
|
|
#
|
2009-02-10 05:10:14 +01:00
|
|
|
# Mercurial Stuff.
|
2009-02-12 01:58:51 +01:00
|
|
|
#
|
|
|
|
# If the M5 directory is a mercurial repository, we should do some
|
|
|
|
# extra things.
|
|
|
|
#
|
|
|
|
########################################################################
|
2009-02-10 05:10:14 +01:00
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
hgdir = env.root.Dir(".hg")
|
|
|
|
|
|
|
|
mercurial_style_message = """
|
2007-07-22 07:33:08 +02:00
|
|
|
You're missing the M5 style hook.
|
|
|
|
Please install the hook so we can ensure that all code fits a common style.
|
|
|
|
|
|
|
|
All you'd need to do is add the following lines to your repository .hg/hgrc
|
|
|
|
or your personal .hgrc
|
|
|
|
----------------
|
|
|
|
|
|
|
|
[extensions]
|
|
|
|
style = %s/util/style.py
|
|
|
|
|
|
|
|
[hooks]
|
|
|
|
pretxncommit.style = python:style.check_whitespace
|
2009-02-12 01:58:51 +01:00
|
|
|
""" % (env.root)
|
|
|
|
|
|
|
|
mercurial_bin_not_found = """
|
|
|
|
Mercurial binary cannot be found, unfortunately this means that we
|
|
|
|
cannot easily determine the version of M5 that you are running and
|
|
|
|
this makes error messages more difficult to collect. Please consider
|
|
|
|
installing mercurial if you choose to post an error message
|
|
|
|
"""
|
|
|
|
|
|
|
|
mercurial_lib_not_found = """
|
|
|
|
Mercurial libraries cannot be found, ignoring style hook
|
|
|
|
If you are actually a M5 developer, please fix this and
|
|
|
|
run the style hook. It is important.
|
|
|
|
"""
|
|
|
|
|
2009-03-05 23:05:06 +01:00
|
|
|
hg_info = "Unknown"
|
2009-02-12 01:58:51 +01:00
|
|
|
if hgdir.exists():
|
|
|
|
# 1) Grab repository revision if we know it.
|
|
|
|
cmd = "hg id -n -i -t -b"
|
|
|
|
try:
|
|
|
|
hg_info = read_command(cmd, cwd=env.root.abspath).strip()
|
|
|
|
except OSError:
|
|
|
|
print mercurial_bin_not_found
|
|
|
|
|
|
|
|
# 2) Ensure that the style hook is in place.
|
2007-07-22 07:33:08 +02:00
|
|
|
try:
|
2009-02-12 01:58:51 +01:00
|
|
|
ui = None
|
|
|
|
if ARGUMENTS.get('IGNORE_STYLE') != 'True':
|
|
|
|
from mercurial import ui
|
|
|
|
ui = ui.ui()
|
2007-07-22 07:33:08 +02:00
|
|
|
except ImportError:
|
2009-02-12 01:58:51 +01:00
|
|
|
print mercurial_lib_not_found
|
2007-07-22 07:33:08 +02:00
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
if ui is not None:
|
|
|
|
ui.readconfig(hgdir.File('hgrc').abspath)
|
|
|
|
style_hook = ui.config('hooks', 'pretxncommit.style', None)
|
|
|
|
|
|
|
|
if not style_hook:
|
|
|
|
print mercurial_style_message
|
|
|
|
sys.exit(1)
|
|
|
|
else:
|
|
|
|
print ".hg directory not found"
|
2009-03-05 23:05:06 +01:00
|
|
|
env['HG_INFO'] = hg_info
|
2009-02-10 05:10:14 +01:00
|
|
|
|
2004-07-03 06:16:38 +02:00
|
|
|
###################################################
|
|
|
|
#
|
2006-05-22 20:29:33 +02:00
|
|
|
# Figure out which configurations to set up based on the path(s) of
|
|
|
|
# the target(s).
|
2004-07-03 06:16:38 +02:00
|
|
|
#
|
|
|
|
###################################################
|
|
|
|
|
|
|
|
# Find default configuration & binary.
|
2009-02-10 05:10:14 +01:00
|
|
|
Default(environ.get('M5_DEFAULT_BINARY', 'build/ALPHA_SE/m5.debug'))
|
2004-07-03 06:16:38 +02:00
|
|
|
|
2006-05-22 20:29:33 +02:00
|
|
|
# helper function: find last occurrence of element in list
|
|
|
|
def rfind(l, elt, offs = -1):
|
|
|
|
for i in range(len(l)+offs, 0, -1):
|
|
|
|
if l[i] == elt:
|
|
|
|
return i
|
|
|
|
raise ValueError, "element not found"
|
|
|
|
|
|
|
|
# Each target must have 'build' in the interior of the path; the
|
|
|
|
# directory below this will determine the build parameters. For
|
|
|
|
# example, for target 'foo/bar/build/ALPHA_SE/arch/alpha/blah.do' we
|
|
|
|
# recognize that ALPHA_SE specifies the configuration because it
|
|
|
|
# follow 'build' in the bulid path.
|
|
|
|
|
2006-12-04 18:09:36 +01:00
|
|
|
# Generate absolute paths to targets so we can see where the build dir is
|
|
|
|
if COMMAND_LINE_TARGETS:
|
|
|
|
# Ask SCons which directory it was invoked from
|
|
|
|
launch_dir = GetLaunchDir()
|
|
|
|
# Make targets relative to invocation directory
|
2009-02-10 05:10:14 +01:00
|
|
|
abs_targets = [ normpath(joinpath(launch_dir, str(x))) for x in \
|
|
|
|
COMMAND_LINE_TARGETS]
|
2006-12-04 18:09:36 +01:00
|
|
|
else:
|
|
|
|
# Default targets are relative to root of tree
|
2009-02-10 05:10:14 +01:00
|
|
|
abs_targets = [ normpath(joinpath(ROOT, str(x))) for x in \
|
|
|
|
DEFAULT_TARGETS]
|
2006-12-04 18:09:36 +01:00
|
|
|
|
|
|
|
|
2006-05-23 03:51:59 +02:00
|
|
|
# Generate a list of the unique build roots and configs that the
|
|
|
|
# collected targets reference.
|
2009-02-10 05:10:14 +01:00
|
|
|
variant_paths = []
|
2006-05-23 04:37:56 +02:00
|
|
|
build_root = None
|
2006-05-22 20:29:33 +02:00
|
|
|
for t in abs_targets:
|
|
|
|
path_dirs = t.split('/')
|
|
|
|
try:
|
|
|
|
build_top = rfind(path_dirs, 'build', -2)
|
|
|
|
except:
|
|
|
|
print "Error: no non-leaf 'build' dir found on target path", t
|
2005-08-30 19:18:54 +02:00
|
|
|
Exit(1)
|
2006-12-04 17:55:06 +01:00
|
|
|
this_build_root = joinpath('/',*path_dirs[:build_top+1])
|
2006-05-23 04:37:56 +02:00
|
|
|
if not build_root:
|
|
|
|
build_root = this_build_root
|
|
|
|
else:
|
|
|
|
if this_build_root != build_root:
|
|
|
|
print "Error: build targets not under same build root\n"\
|
|
|
|
" %s\n %s" % (build_root, this_build_root)
|
|
|
|
Exit(1)
|
2009-02-10 05:10:14 +01:00
|
|
|
variant_path = joinpath('/',*path_dirs[:build_top+2])
|
|
|
|
if variant_path not in variant_paths:
|
|
|
|
variant_paths.append(variant_path)
|
2004-07-03 06:16:38 +02:00
|
|
|
|
2008-02-06 02:40:08 +01:00
|
|
|
# Make sure build_root exists (might not if this is the first build there)
|
|
|
|
if not isdir(build_root):
|
2009-02-10 05:10:14 +01:00
|
|
|
mkdir(build_root)
|
2008-02-06 02:40:08 +01:00
|
|
|
|
2007-05-30 23:08:12 +02:00
|
|
|
Export('env')
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
env.SConsignFile(joinpath(build_root, "sconsign"))
|
2004-10-22 08:11:51 +02:00
|
|
|
|
2006-06-10 05:01:31 +02:00
|
|
|
# Default duplicate option is to use hard links, but this messes up
|
|
|
|
# when you use emacs to edit a file in the target dir, as emacs moves
|
|
|
|
# file to file~ then copies to file, breaking the link. Symbolic
|
|
|
|
# (soft) links work better.
|
|
|
|
env.SetOption('duplicate', 'soft-copy')
|
|
|
|
|
2008-02-06 02:40:08 +01:00
|
|
|
#
|
2009-02-10 05:10:14 +01:00
|
|
|
# Set up global sticky variables... these are common to an entire build
|
2008-02-06 02:40:08 +01:00
|
|
|
# tree (not specific to a particular build like ALPHA_SE)
|
|
|
|
#
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Variable validators & converters for global sticky variables
|
2008-02-06 02:40:08 +01:00
|
|
|
def PathListMakeAbsolute(val):
|
|
|
|
if not val:
|
|
|
|
return val
|
2009-02-10 05:10:14 +01:00
|
|
|
f = lambda p: abspath(expanduser(p))
|
2008-02-06 02:40:08 +01:00
|
|
|
return ':'.join(map(f, val.split(':')))
|
|
|
|
|
|
|
|
def PathListAllExist(key, val, env):
|
|
|
|
if not val:
|
|
|
|
return
|
|
|
|
paths = val.split(':')
|
|
|
|
for path in paths:
|
|
|
|
if not isdir(path):
|
|
|
|
raise SCons.Errors.UserError("Path does not exist: '%s'" % path)
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
global_sticky_vars_file = joinpath(build_root, 'variables.global')
|
2008-02-06 02:40:08 +01:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
global_sticky_vars = Variables(global_sticky_vars_file, args=ARGUMENTS)
|
2008-02-06 02:40:08 +01:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
global_sticky_vars.AddVariables(
|
|
|
|
('CC', 'C compiler', environ.get('CC', env['CC'])),
|
|
|
|
('CXX', 'C++ compiler', environ.get('CXX', env['CXX'])),
|
2008-04-08 05:40:24 +02:00
|
|
|
('BATCH', 'Use batch pool for build and tests', False),
|
|
|
|
('BATCH_CMD', 'Batch pool submission command name', 'qdo'),
|
2008-02-06 02:40:08 +01:00
|
|
|
('EXTRAS', 'Add Extra directories to the compilation', '',
|
|
|
|
PathListAllExist, PathListMakeAbsolute)
|
|
|
|
)
|
|
|
|
|
|
|
|
# base help text
|
|
|
|
help_text = '''
|
|
|
|
Usage: scons [scons options] [build options] [target(s)]
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
Global sticky options:
|
2008-02-06 02:40:08 +01:00
|
|
|
'''
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
help_text += global_sticky_vars.GenerateHelpText(env)
|
2008-02-06 02:40:08 +01:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Update env with values from ARGUMENTS & file global_sticky_vars_file
|
|
|
|
global_sticky_vars.Update(env)
|
2008-02-06 02:40:08 +01:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Save sticky variable settings back to current variables file
|
|
|
|
global_sticky_vars.Save(global_sticky_vars_file, env)
|
2008-02-06 02:40:08 +01:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Parse EXTRAS variable to build list of all directories where we're
|
2008-02-06 02:40:08 +01:00
|
|
|
# look for sources etc. This list is exported as base_dir_list.
|
2009-02-12 01:58:51 +01:00
|
|
|
base_dir = env.srcdir.abspath
|
2008-02-06 02:40:08 +01:00
|
|
|
if env['EXTRAS']:
|
2008-11-10 20:51:18 +01:00
|
|
|
extras_dir_list = env['EXTRAS'].split(':')
|
|
|
|
else:
|
|
|
|
extras_dir_list = []
|
2008-02-06 02:40:08 +01:00
|
|
|
|
2008-11-10 20:51:18 +01:00
|
|
|
Export('base_dir')
|
|
|
|
Export('extras_dir_list')
|
2008-02-06 02:40:08 +01:00
|
|
|
|
2006-05-22 20:29:33 +02:00
|
|
|
# M5_PLY is used by isa_parser.py to find the PLY package.
|
2007-10-31 08:44:51 +01:00
|
|
|
env.Append(ENV = { 'M5_PLY' : str(Dir('ext/ply')) })
|
2009-02-12 01:58:51 +01:00
|
|
|
|
|
|
|
CXX_version = read_command([env['CXX'],'--version'], exception=False)
|
|
|
|
CXX_V = read_command([env['CXX'],'-V'], exception=False)
|
|
|
|
|
|
|
|
env['GCC'] = CXX_version and CXX_version.find('g++') >= 0
|
|
|
|
env['SUNCC'] = CXX_V and CXX_V.find('Sun C++') >= 0
|
|
|
|
env['ICC'] = CXX_V and CXX_V.find('Intel') >= 0
|
2007-01-27 21:47:18 +01:00
|
|
|
if env['GCC'] + env['SUNCC'] + env['ICC'] > 1:
|
2007-01-27 21:38:04 +01:00
|
|
|
print 'Error: How can we have two at the same time?'
|
2007-01-27 00:48:51 +01:00
|
|
|
Exit(1)
|
|
|
|
|
2005-08-30 19:18:54 +02:00
|
|
|
# Set up default C++ compiler flags
|
2007-01-27 00:48:51 +01:00
|
|
|
if env['GCC']:
|
|
|
|
env.Append(CCFLAGS='-pipe')
|
|
|
|
env.Append(CCFLAGS='-fno-strict-aliasing')
|
|
|
|
env.Append(CCFLAGS=Split('-Wall -Wno-sign-compare -Werror -Wundef'))
|
2008-09-28 06:03:50 +02:00
|
|
|
env.Append(CXXFLAGS='-Wno-deprecated')
|
2007-01-27 21:38:04 +01:00
|
|
|
elif env['ICC']:
|
|
|
|
pass #Fix me... add warning flags once we clean up icc warnings
|
2007-01-27 00:48:51 +01:00
|
|
|
elif env['SUNCC']:
|
|
|
|
env.Append(CCFLAGS='-Qoption ccfe')
|
|
|
|
env.Append(CCFLAGS='-features=gcc')
|
|
|
|
env.Append(CCFLAGS='-features=extensions')
|
|
|
|
env.Append(CCFLAGS='-library=stlport4')
|
|
|
|
env.Append(CCFLAGS='-xar')
|
2009-02-12 01:58:51 +01:00
|
|
|
#env.Append(CCFLAGS='-instances=semiexplicit')
|
2007-01-27 00:48:51 +01:00
|
|
|
else:
|
|
|
|
print 'Error: Don\'t know what compiler options to use for your compiler.'
|
2007-01-27 21:38:04 +01:00
|
|
|
print ' Please fix SConstruct and src/SConscript and try again.'
|
2007-01-27 00:48:51 +01:00
|
|
|
Exit(1)
|
|
|
|
|
2008-04-08 05:40:24 +02:00
|
|
|
# Do this after we save setting back, or else we'll tack on an
|
|
|
|
# extra 'qdo' every time we run scons.
|
|
|
|
if env['BATCH']:
|
2008-10-26 19:45:47 +01:00
|
|
|
env['CC'] = env['BATCH_CMD'] + ' ' + env['CC']
|
|
|
|
env['CXX'] = env['BATCH_CMD'] + ' ' + env['CXX']
|
|
|
|
env['AS'] = env['BATCH_CMD'] + ' ' + env['AS']
|
|
|
|
env['AR'] = env['BATCH_CMD'] + ' ' + env['AR']
|
|
|
|
env['RANLIB'] = env['BATCH_CMD'] + ' ' + env['RANLIB']
|
2008-04-08 05:40:24 +02:00
|
|
|
|
2005-08-19 23:10:17 +02:00
|
|
|
if sys.platform == 'cygwin':
|
|
|
|
# cygwin has some header file issues...
|
2005-08-30 19:18:54 +02:00
|
|
|
env.Append(CCFLAGS=Split("-Wno-uninitialized"))
|
2007-10-31 08:55:32 +01:00
|
|
|
env.Append(CPPPATH=[Dir('ext/dnet')])
|
2004-07-03 06:16:38 +02:00
|
|
|
|
2006-08-22 00:25:33 +02:00
|
|
|
# Check for SWIG
|
|
|
|
if not env.has_key('SWIG'):
|
|
|
|
print 'Error: SWIG utility not found.'
|
|
|
|
print ' Please install (see http://www.swig.org) and retry.'
|
|
|
|
Exit(1)
|
|
|
|
|
|
|
|
# Check for appropriate SWIG version
|
2009-02-12 01:58:51 +01:00
|
|
|
swig_version = read_command(('swig', '-version'), exception='').split()
|
2006-08-22 00:25:33 +02:00
|
|
|
# First 3 words should be "SWIG Version x.y.z"
|
2007-07-15 03:35:26 +02:00
|
|
|
if len(swig_version) < 3 or \
|
|
|
|
swig_version[0] != 'SWIG' or swig_version[1] != 'Version':
|
2006-08-22 00:25:33 +02:00
|
|
|
print 'Error determining SWIG version.'
|
|
|
|
Exit(1)
|
|
|
|
|
|
|
|
min_swig_version = '1.3.28'
|
|
|
|
if compare_versions(swig_version[2], min_swig_version) < 0:
|
|
|
|
print 'Error: SWIG version', min_swig_version, 'or newer required.'
|
|
|
|
print ' Installed version:', swig_version[2]
|
|
|
|
Exit(1)
|
2006-06-10 05:01:31 +02:00
|
|
|
|
2006-08-22 00:25:33 +02:00
|
|
|
# Set up SWIG flags & scanner
|
2007-06-10 22:54:59 +02:00
|
|
|
swig_flags=Split('-c++ -python -modern -templatereduce $_CPPINCFLAGS')
|
|
|
|
env.Append(SWIGFLAGS=swig_flags)
|
|
|
|
|
|
|
|
# filter out all existing swig scanners, they mess up the dependency
|
|
|
|
# stuff for some reason
|
|
|
|
scanners = []
|
|
|
|
for scanner in env['SCANNERS']:
|
|
|
|
skeys = scanner.skeys
|
|
|
|
if skeys == '.i':
|
|
|
|
continue
|
2007-07-29 01:49:20 +02:00
|
|
|
|
2007-06-10 22:54:59 +02:00
|
|
|
if isinstance(skeys, (list, tuple)) and '.i' in skeys:
|
|
|
|
continue
|
2006-06-10 05:01:31 +02:00
|
|
|
|
2007-06-10 22:54:59 +02:00
|
|
|
scanners.append(scanner)
|
2006-06-10 05:01:31 +02:00
|
|
|
|
2007-06-10 22:54:59 +02:00
|
|
|
# add the new swig scanner that we like better
|
|
|
|
from SCons.Scanner import ClassicCPP as CPPScanner
|
2006-06-10 05:01:31 +02:00
|
|
|
swig_inc_re = '^[ \t]*[%,#][ \t]*(?:include|import)[ \t]*(<|")([^>"]+)(>|")'
|
2007-06-10 22:54:59 +02:00
|
|
|
scanners.append(CPPScanner("SwigScan", [ ".i" ], "CPPPATH", swig_inc_re))
|
2006-06-10 05:01:31 +02:00
|
|
|
|
2007-06-10 22:54:59 +02:00
|
|
|
# replace the scanners list that has what we want
|
|
|
|
env['SCANNERS'] = scanners
|
2006-06-10 05:01:31 +02:00
|
|
|
|
libm5: Create a libm5 static library for embedding m5.
This should allow m5 to be more easily embedded into other simulators.
The m5 binary adds a simple main function which then calls into the m5
libarary to start the simulation. In order to make this work
correctly, it was necessary embed python code directly into the
library instead of the zipfile hack. This is because you can't just
append the zipfile to the end of a library the way you can a binary.
As a result, Python files that are part of the m5 simulator are now
compile, marshalled, compressed, and then inserted into the library's
data section with a certain symbol name. Additionally, a new Importer
was needed to allow python to get at the embedded python code.
Small additional changes include:
- Get rid of the PYTHONHOME stuff since I don't think anyone ever used
it, and it just confuses things. Easy enough to add back if I'm wrong.
- Create a few new functions that are key to initializing and running
the simulator: initSignals, initM5Python, m5Main.
The original code for creating libm5 was inspired by a patch Michael
Adler, though the code here was done by me.
2008-08-04 03:19:54 +02:00
|
|
|
# Add a custom Check function to the Configure context so that we can
|
|
|
|
# figure out if the compiler adds leading underscores to global
|
|
|
|
# variables. This is needed for the autogenerated asm files that we
|
|
|
|
# use for embedding the python code.
|
|
|
|
def CheckLeading(context):
|
|
|
|
context.Message("Checking for leading underscore in global variables...")
|
|
|
|
# 1) Define a global variable called x from asm so the C compiler
|
|
|
|
# won't change the symbol at all.
|
|
|
|
# 2) Declare that variable.
|
|
|
|
# 3) Use the variable
|
|
|
|
#
|
|
|
|
# If the compiler prepends an underscore, this will successfully
|
|
|
|
# link because the external symbol 'x' will be called '_x' which
|
|
|
|
# was defined by the asm statement. If the compiler does not
|
|
|
|
# prepend an underscore, this will not successfully link because
|
|
|
|
# '_x' will have been defined by assembly, while the C portion of
|
|
|
|
# the code will be trying to use 'x'
|
|
|
|
ret = context.TryLink('''
|
|
|
|
asm(".globl _x; _x: .byte 0");
|
|
|
|
extern int x;
|
|
|
|
int main() { return x; }
|
|
|
|
''', extension=".c")
|
|
|
|
context.env.Append(LEADING_UNDERSCORE=ret)
|
|
|
|
context.Result(ret)
|
|
|
|
return ret
|
|
|
|
|
2006-05-23 04:37:56 +02:00
|
|
|
# Platform-specific configuration. Note again that we assume that all
|
|
|
|
# builds under a given build root run on the same host platform.
|
|
|
|
conf = Configure(env,
|
2006-12-04 17:55:06 +01:00
|
|
|
conf_dir = joinpath(build_root, '.scons_config'),
|
libm5: Create a libm5 static library for embedding m5.
This should allow m5 to be more easily embedded into other simulators.
The m5 binary adds a simple main function which then calls into the m5
libarary to start the simulation. In order to make this work
correctly, it was necessary embed python code directly into the
library instead of the zipfile hack. This is because you can't just
append the zipfile to the end of a library the way you can a binary.
As a result, Python files that are part of the m5 simulator are now
compile, marshalled, compressed, and then inserted into the library's
data section with a certain symbol name. Additionally, a new Importer
was needed to allow python to get at the embedded python code.
Small additional changes include:
- Get rid of the PYTHONHOME stuff since I don't think anyone ever used
it, and it just confuses things. Easy enough to add back if I'm wrong.
- Create a few new functions that are key to initializing and running
the simulator: initSignals, initM5Python, m5Main.
The original code for creating libm5 was inspired by a patch Michael
Adler, though the code here was done by me.
2008-08-04 03:19:54 +02:00
|
|
|
log_file = joinpath(build_root, 'scons_config.log'),
|
|
|
|
custom_tests = { 'CheckLeading' : CheckLeading })
|
|
|
|
|
|
|
|
# Check for leading underscores. Don't really need to worry either
|
|
|
|
# way so don't need to check the return code.
|
|
|
|
conf.CheckLeading()
|
2005-08-30 19:18:54 +02:00
|
|
|
|
2007-11-08 23:45:58 +01:00
|
|
|
# Check if we should compile a 64 bit binary on Mac OS X/Darwin
|
|
|
|
try:
|
|
|
|
import platform
|
|
|
|
uname = platform.uname()
|
|
|
|
if uname[0] == 'Darwin' and compare_versions(uname[2], '9.0.0') >= 0:
|
2009-02-10 05:10:14 +01:00
|
|
|
if int(read_command('sysctl -n hw.cpu64bit_capable')[0]):
|
2007-11-08 23:45:58 +01:00
|
|
|
env.Append(CCFLAGS='-arch x86_64')
|
|
|
|
env.Append(CFLAGS='-arch x86_64')
|
|
|
|
env.Append(LINKFLAGS='-arch x86_64')
|
|
|
|
env.Append(ASFLAGS='-arch x86_64')
|
|
|
|
except:
|
|
|
|
pass
|
|
|
|
|
2007-11-01 22:28:59 +01:00
|
|
|
# Recent versions of scons substitute a "Null" object for Configure()
|
|
|
|
# when configuration isn't necessary, e.g., if the "--help" option is
|
|
|
|
# present. Unfortuantely this Null object always returns false,
|
|
|
|
# breaking all our configuration checks. We replace it with our own
|
|
|
|
# more optimistic null object that returns True instead.
|
|
|
|
if not conf:
|
|
|
|
def NullCheck(*args, **kwargs):
|
|
|
|
return True
|
|
|
|
|
|
|
|
class NullConf:
|
|
|
|
def __init__(self, env):
|
|
|
|
self.env = env
|
|
|
|
def Finish(self):
|
|
|
|
return self.env
|
|
|
|
def __getattr__(self, mname):
|
|
|
|
return NullCheck
|
|
|
|
|
|
|
|
conf = NullConf(env)
|
|
|
|
|
2006-10-01 07:42:18 +02:00
|
|
|
# Find Python include and library directories for embedding the
|
|
|
|
# interpreter. For consistency, we will use the same Python
|
|
|
|
# installation used to run scons (and thus this script). If you want
|
|
|
|
# to link in an alternate version, see above for instructions on how
|
|
|
|
# to invoke scons with a different copy of the Python interpreter.
|
2009-02-10 05:10:14 +01:00
|
|
|
from distutils import sysconfig
|
|
|
|
|
|
|
|
py_getvar = sysconfig.get_config_var
|
2006-10-01 07:42:18 +02:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
py_version = 'python' + py_getvar('VERSION')
|
2006-10-01 07:42:18 +02:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
py_general_include = sysconfig.get_python_inc()
|
|
|
|
py_platform_include = sysconfig.get_python_inc(plat_specific=True)
|
|
|
|
py_includes = [ py_general_include ]
|
|
|
|
if py_platform_include != py_general_include:
|
|
|
|
py_includes.append(py_platform_include)
|
|
|
|
|
|
|
|
py_lib_path = []
|
|
|
|
# add the prefix/lib/pythonX.Y/config dir, but only if there is no
|
|
|
|
# shared library in prefix/lib/.
|
|
|
|
if not py_getvar('Py_ENABLE_SHARED'):
|
|
|
|
py_lib_path.append('-L' + py_getvar('LIBPL'))
|
|
|
|
|
|
|
|
py_libs = []
|
|
|
|
for lib in py_getvar('LIBS').split() + py_getvar('SYSLIBS').split():
|
|
|
|
if lib not in py_libs:
|
|
|
|
py_libs.append(lib)
|
|
|
|
py_libs.append('-l' + py_version)
|
|
|
|
|
|
|
|
env.Append(CPPPATH=py_includes)
|
|
|
|
env.Append(LIBPATH=py_lib_path)
|
|
|
|
#env.Append(LIBS=py_libs)
|
|
|
|
|
|
|
|
# verify that this stuff works
|
2006-10-01 07:42:18 +02:00
|
|
|
if not conf.CheckHeader('Python.h', '<>'):
|
2009-02-10 05:10:14 +01:00
|
|
|
print "Error: can't find Python.h header in", py_includes
|
2006-10-01 07:42:18 +02:00
|
|
|
Exit(1)
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
for lib in py_libs:
|
|
|
|
assert lib.startswith('-l')
|
|
|
|
lib = lib[2:]
|
|
|
|
if not conf.CheckLib(lib):
|
|
|
|
print "Error: can't find library %s required by python" % lib
|
|
|
|
Exit(1)
|
2006-10-01 07:42:18 +02:00
|
|
|
|
2006-11-05 03:41:01 +01:00
|
|
|
# On Solaris you need to use libsocket for socket ops
|
2006-11-06 16:15:27 +01:00
|
|
|
if not conf.CheckLibWithHeader(None, 'sys/socket.h', 'C++', 'accept(0,0,0);'):
|
|
|
|
if not conf.CheckLibWithHeader('socket', 'sys/socket.h', 'C++', 'accept(0,0,0);'):
|
2006-11-05 03:41:01 +01:00
|
|
|
print "Can't find library with socket calls (e.g. accept())"
|
|
|
|
Exit(1)
|
|
|
|
|
2006-08-22 00:25:33 +02:00
|
|
|
# Check for zlib. If the check passes, libz will be automatically
|
|
|
|
# added to the LIBS environment variable.
|
2007-01-27 00:48:51 +01:00
|
|
|
if not conf.CheckLibWithHeader('z', 'zlib.h', 'C++','zlibVersion();'):
|
2006-08-22 00:25:33 +02:00
|
|
|
print 'Error: did not find needed zlib compression library '\
|
|
|
|
'and/or zlib.h header file.'
|
|
|
|
print ' Please install zlib and try again.'
|
|
|
|
Exit(1)
|
|
|
|
|
2005-08-30 19:18:54 +02:00
|
|
|
# Check for <fenv.h> (C99 FP environment control)
|
|
|
|
have_fenv = conf.CheckHeader('fenv.h', '<>')
|
|
|
|
if not have_fenv:
|
|
|
|
print "Warning: Header file <fenv.h> not found."
|
|
|
|
print " This host has no IEEE FP rounding mode control."
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
######################################################################
|
|
|
|
#
|
2005-08-31 05:34:36 +02:00
|
|
|
# Check for mysql.
|
2009-02-10 05:10:14 +01:00
|
|
|
#
|
2005-08-30 19:18:54 +02:00
|
|
|
mysql_config = WhereIs('mysql_config')
|
2009-02-10 05:10:14 +01:00
|
|
|
have_mysql = bool(mysql_config)
|
2005-08-30 19:18:54 +02:00
|
|
|
|
2005-08-31 05:34:36 +02:00
|
|
|
# Check MySQL version.
|
|
|
|
if have_mysql:
|
2009-02-10 05:10:14 +01:00
|
|
|
mysql_version = read_command(mysql_config + ' --version')
|
2006-08-22 00:25:33 +02:00
|
|
|
min_mysql_version = '4.1'
|
|
|
|
if compare_versions(mysql_version, min_mysql_version) < 0:
|
|
|
|
print 'Warning: MySQL', min_mysql_version, 'or newer required.'
|
|
|
|
print ' Version', mysql_version, 'detected.'
|
2005-08-31 05:34:36 +02:00
|
|
|
have_mysql = False
|
|
|
|
|
|
|
|
# Set up mysql_config commands.
|
|
|
|
if have_mysql:
|
|
|
|
mysql_config_include = mysql_config + ' --include'
|
|
|
|
if os.system(mysql_config_include + ' > /dev/null') != 0:
|
|
|
|
# older mysql_config versions don't support --include, use
|
|
|
|
# --cflags instead
|
2005-08-31 16:00:42 +02:00
|
|
|
mysql_config_include = mysql_config + ' --cflags | sed s/\\\'//g'
|
2005-08-31 05:34:36 +02:00
|
|
|
# This seems to work in all versions
|
|
|
|
mysql_config_libs = mysql_config + ' --libs'
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
######################################################################
|
|
|
|
#
|
|
|
|
# Finish the configuration
|
|
|
|
#
|
2005-08-30 19:18:54 +02:00
|
|
|
env = conf.Finish()
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
######################################################################
|
|
|
|
#
|
|
|
|
# Collect all non-global variables
|
|
|
|
#
|
|
|
|
|
2009-02-12 01:58:51 +01:00
|
|
|
Export('env')
|
|
|
|
|
2006-02-23 04:22:06 +01:00
|
|
|
# Define the universe of supported ISAs
|
2007-03-11 08:00:54 +01:00
|
|
|
all_isa_list = [ ]
|
|
|
|
Export('all_isa_list')
|
2006-02-23 04:22:06 +01:00
|
|
|
|
2006-02-23 23:00:29 +01:00
|
|
|
# Define the universe of supported CPU models
|
2007-03-11 08:00:54 +01:00
|
|
|
all_cpu_list = [ ]
|
|
|
|
default_cpus = [ ]
|
|
|
|
Export('all_cpu_list', 'default_cpus')
|
2006-02-23 23:00:29 +01:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Sticky variables get saved in the variables file so they persist from
|
2005-09-05 22:31:27 +02:00
|
|
|
# one invocation to the next (unless overridden, in which case the new
|
|
|
|
# value becomes sticky).
|
2009-02-10 05:10:14 +01:00
|
|
|
sticky_vars = Variables(args=ARGUMENTS)
|
|
|
|
Export('sticky_vars')
|
2007-03-11 08:00:54 +01:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Non-sticky variables only apply to the current build.
|
|
|
|
nonsticky_vars = Variables(args=ARGUMENTS)
|
|
|
|
Export('nonsticky_vars')
|
2007-03-11 08:00:54 +01:00
|
|
|
|
|
|
|
# Walk the tree and execute all SConsopts scripts that wil add to the
|
2009-02-10 05:10:14 +01:00
|
|
|
# above variables
|
2008-11-10 20:51:18 +01:00
|
|
|
for bdir in [ base_dir ] + extras_dir_list:
|
|
|
|
for root, dirs, files in os.walk(bdir):
|
2008-02-06 02:40:08 +01:00
|
|
|
if 'SConsopts' in files:
|
2008-02-06 02:43:45 +01:00
|
|
|
print "Reading", joinpath(root, 'SConsopts')
|
|
|
|
SConscript(joinpath(root, 'SConsopts'))
|
2007-03-11 08:00:54 +01:00
|
|
|
|
|
|
|
all_isa_list.sort()
|
|
|
|
all_cpu_list.sort()
|
|
|
|
default_cpus.sort()
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
sticky_vars.AddVariables(
|
|
|
|
EnumVariable('TARGET_ISA', 'Target ISA', 'alpha', all_isa_list),
|
|
|
|
BoolVariable('FULL_SYSTEM', 'Full-system support', False),
|
|
|
|
ListVariable('CPU_MODELS', 'CPU models', default_cpus, all_cpu_list),
|
|
|
|
BoolVariable('NO_FAST_ALLOC', 'Disable fast object allocator', False),
|
|
|
|
BoolVariable('FAST_ALLOC_DEBUG', 'Enable fast object allocator debugging',
|
|
|
|
False),
|
|
|
|
BoolVariable('FAST_ALLOC_STATS', 'Enable fast object allocator statistics',
|
|
|
|
False),
|
|
|
|
BoolVariable('EFENCE', 'Link with Electric Fence malloc debugger',
|
|
|
|
False),
|
|
|
|
BoolVariable('SS_COMPATIBLE_FP',
|
|
|
|
'Make floating-point results compatible with SimpleScalar',
|
|
|
|
False),
|
|
|
|
BoolVariable('USE_SSE2',
|
|
|
|
'Compile for SSE2 (-msse2) to get IEEE FP on x86 hosts',
|
|
|
|
False),
|
|
|
|
BoolVariable('USE_MYSQL', 'Use MySQL for stats output', have_mysql),
|
|
|
|
BoolVariable('USE_FENV', 'Use <fenv.h> IEEE mode control', have_fenv),
|
|
|
|
BoolVariable('USE_CHECKER', 'Use checker for detailed CPU models', False),
|
2009-02-27 01:29:17 +01:00
|
|
|
BoolVariable('CP_ANNOTATE', 'Enable critical path annotation capability', False),
|
2005-09-05 22:31:27 +02:00
|
|
|
)
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
nonsticky_vars.AddVariables(
|
|
|
|
BoolVariable('update_ref', 'Update test reference outputs', False)
|
2005-09-05 22:31:27 +02:00
|
|
|
)
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# These variables get exported to #defines in config/*.hh (see src/SConscript).
|
|
|
|
env.ExportVariables = ['FULL_SYSTEM', 'ALPHA_TLASER', 'USE_FENV', \
|
|
|
|
'USE_MYSQL', 'NO_FAST_ALLOC', 'FAST_ALLOC_DEBUG', \
|
|
|
|
'FAST_ALLOC_STATS', 'SS_COMPATIBLE_FP', \
|
2009-02-27 01:29:17 +01:00
|
|
|
'USE_CHECKER', 'TARGET_ISA', 'CP_ANNOTATE']
|
2005-09-05 22:31:27 +02:00
|
|
|
|
|
|
|
###################################################
|
|
|
|
#
|
|
|
|
# Define a SCons builder for configuration flag headers.
|
|
|
|
#
|
|
|
|
###################################################
|
|
|
|
|
|
|
|
# This function generates a config header file that #defines the
|
2009-02-10 05:10:14 +01:00
|
|
|
# variable symbol to the current variable setting (0 or 1). The source
|
|
|
|
# operands are the name of the variable and a Value node containing the
|
|
|
|
# value of the variable.
|
2005-09-05 22:31:27 +02:00
|
|
|
def build_config_file(target, source, env):
|
2009-02-10 05:10:14 +01:00
|
|
|
(variable, value) = [s.get_contents() for s in source]
|
2005-08-30 19:18:54 +02:00
|
|
|
f = file(str(target[0]), 'w')
|
2009-02-10 05:10:14 +01:00
|
|
|
print >> f, '#define', variable, value
|
2005-08-30 19:18:54 +02:00
|
|
|
f.close()
|
|
|
|
return None
|
|
|
|
|
2005-09-05 22:31:27 +02:00
|
|
|
# Generate the message to be printed when building the config file.
|
|
|
|
def build_config_file_string(target, source, env):
|
2009-02-10 05:10:14 +01:00
|
|
|
(variable, value) = [s.get_contents() for s in source]
|
|
|
|
return "Defining %s as %s in %s." % (variable, value, target[0])
|
2005-09-05 22:31:27 +02:00
|
|
|
|
|
|
|
# Combine the two functions into a scons Action object.
|
|
|
|
config_action = Action(build_config_file, build_config_file_string)
|
|
|
|
|
|
|
|
# The emitter munges the source & target node lists to reflect what
|
|
|
|
# we're really doing.
|
|
|
|
def config_emitter(target, source, env):
|
2009-02-10 05:10:14 +01:00
|
|
|
# extract variable name from Builder arg
|
|
|
|
variable = str(target[0])
|
2005-09-05 22:31:27 +02:00
|
|
|
# True target is config header file
|
2009-02-10 05:10:14 +01:00
|
|
|
target = joinpath('config', variable.lower() + '.hh')
|
|
|
|
val = env[variable]
|
2006-10-20 20:37:59 +02:00
|
|
|
if isinstance(val, bool):
|
|
|
|
# Force value to 0/1
|
|
|
|
val = int(val)
|
|
|
|
elif isinstance(val, str):
|
|
|
|
val = '"' + val + '"'
|
2007-07-29 01:49:20 +02:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Sources are variable name & value (packaged in SCons Value nodes)
|
|
|
|
return ([target], [Value(variable), Value(val)])
|
2005-08-30 19:18:54 +02:00
|
|
|
|
2005-09-05 22:31:27 +02:00
|
|
|
config_builder = Builder(emitter = config_emitter, action = config_action)
|
2005-08-30 19:18:54 +02:00
|
|
|
|
2005-09-05 22:31:27 +02:00
|
|
|
env.Append(BUILDERS = { 'ConfigFile' : config_builder })
|
2004-08-07 23:23:01 +02:00
|
|
|
|
2006-05-23 04:37:56 +02:00
|
|
|
# libelf build is shared across all configs in the build root.
|
|
|
|
env.SConscript('ext/libelf/SConscript',
|
2009-02-12 01:58:51 +01:00
|
|
|
variant_dir = joinpath(build_root, 'libelf'))
|
2006-05-23 04:37:56 +02:00
|
|
|
|
2008-11-16 05:42:11 +01:00
|
|
|
# gzstream build is shared across all configs in the build root.
|
|
|
|
env.SConscript('ext/gzstream/SConscript',
|
2009-02-12 01:58:51 +01:00
|
|
|
variant_dir = joinpath(build_root, 'gzstream'))
|
2008-11-16 05:42:11 +01:00
|
|
|
|
2006-11-07 11:33:21 +01:00
|
|
|
###################################################
|
|
|
|
#
|
|
|
|
# This function is used to set up a directory with switching headers
|
|
|
|
#
|
|
|
|
###################################################
|
|
|
|
|
2007-03-11 08:00:54 +01:00
|
|
|
env['ALL_ISA_LIST'] = all_isa_list
|
2009-02-10 05:10:14 +01:00
|
|
|
def make_switching_dir(dname, switch_headers, env):
|
2006-11-07 11:33:21 +01:00
|
|
|
# Generate the header. target[0] is the full path of the output
|
|
|
|
# header to generate. 'source' is a dummy variable, since we get the
|
|
|
|
# list of ISAs from env['ALL_ISA_LIST'].
|
|
|
|
def gen_switch_hdr(target, source, env):
|
2007-07-29 01:49:20 +02:00
|
|
|
fname = str(target[0])
|
2009-02-10 05:10:14 +01:00
|
|
|
bname = basename(fname)
|
2007-07-29 01:49:20 +02:00
|
|
|
f = open(fname, 'w')
|
|
|
|
f.write('#include "arch/isa_specific.hh"\n')
|
|
|
|
cond = '#if'
|
|
|
|
for isa in all_isa_list:
|
|
|
|
f.write('%s THE_ISA == %s_ISA\n#include "%s/%s/%s"\n'
|
2009-02-10 05:10:14 +01:00
|
|
|
% (cond, isa.upper(), dname, isa, bname))
|
2007-07-29 01:49:20 +02:00
|
|
|
cond = '#elif'
|
|
|
|
f.write('#else\n#error "THE_ISA not set"\n#endif\n')
|
|
|
|
f.close()
|
|
|
|
return 0
|
2006-11-07 11:33:21 +01:00
|
|
|
|
|
|
|
# String to print when generating header
|
|
|
|
def gen_switch_hdr_string(target, source, env):
|
2007-07-29 01:49:20 +02:00
|
|
|
return "Generating switch header " + str(target[0])
|
2006-11-07 11:33:21 +01:00
|
|
|
|
|
|
|
# Build SCons Action object. 'varlist' specifies env vars that this
|
|
|
|
# action depends on; when env['ALL_ISA_LIST'] changes these actions
|
|
|
|
# should get re-executed.
|
|
|
|
switch_hdr_action = Action(gen_switch_hdr, gen_switch_hdr_string,
|
|
|
|
varlist=['ALL_ISA_LIST'])
|
|
|
|
|
|
|
|
# Instantiate actions for each header
|
|
|
|
for hdr in switch_headers:
|
|
|
|
env.Command(hdr, [], switch_hdr_action)
|
2007-03-11 08:00:54 +01:00
|
|
|
Export('make_switching_dir')
|
2006-11-07 11:33:21 +01:00
|
|
|
|
2004-07-03 06:16:38 +02:00
|
|
|
###################################################
|
|
|
|
#
|
|
|
|
# Define build environments for selected configurations.
|
|
|
|
#
|
|
|
|
###################################################
|
|
|
|
|
2005-08-30 19:18:54 +02:00
|
|
|
# rename base env
|
|
|
|
base_env = env
|
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
for variant_path in variant_paths:
|
|
|
|
print "Building in", variant_path
|
2008-02-11 16:47:44 +01:00
|
|
|
|
|
|
|
# Make a copy of the build-root environment to use for this config.
|
2009-02-10 05:10:14 +01:00
|
|
|
env = base_env.Clone()
|
|
|
|
env['BUILDDIR'] = variant_path
|
2007-07-26 03:21:11 +02:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# variant_dir is the tail component of build path, and is used to
|
2006-05-22 20:29:33 +02:00
|
|
|
# determine the build parameters (e.g., 'ALPHA_SE')
|
2009-02-10 05:10:14 +01:00
|
|
|
(build_root, variant_dir) = splitpath(variant_path)
|
|
|
|
|
|
|
|
# Set env variables according to the build directory config.
|
|
|
|
sticky_vars.files = []
|
|
|
|
# Variables for $BUILD_ROOT/$VARIANT_DIR are stored in
|
|
|
|
# $BUILD_ROOT/variables/$VARIANT_DIR so you can nuke
|
|
|
|
# $BUILD_ROOT/$VARIANT_DIR without losing your variables settings.
|
|
|
|
current_vars_file = joinpath(build_root, 'variables', variant_dir)
|
|
|
|
if isfile(current_vars_file):
|
|
|
|
sticky_vars.files.append(current_vars_file)
|
|
|
|
print "Using saved variables file %s" % current_vars_file
|
2005-09-25 03:24:44 +02:00
|
|
|
else:
|
2009-02-10 05:10:14 +01:00
|
|
|
# Build dir-specific variables file doesn't exist.
|
2006-05-22 20:29:33 +02:00
|
|
|
|
|
|
|
# Make sure the directory is there so we can create it later
|
2009-02-10 05:10:14 +01:00
|
|
|
opt_dir = dirname(current_vars_file)
|
2008-02-06 02:43:45 +01:00
|
|
|
if not isdir(opt_dir):
|
2009-02-10 05:10:14 +01:00
|
|
|
mkdir(opt_dir)
|
2006-05-22 20:29:33 +02:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Get default build variables from source tree. Variables are
|
|
|
|
# normally determined by name of $VARIANT_DIR, but can be
|
2006-05-22 20:29:33 +02:00
|
|
|
# overriden by 'default=' arg on command line.
|
2009-02-10 05:10:14 +01:00
|
|
|
default_vars_file = joinpath('build_opts',
|
|
|
|
ARGUMENTS.get('default', variant_dir))
|
|
|
|
if isfile(default_vars_file):
|
|
|
|
sticky_vars.files.append(default_vars_file)
|
|
|
|
print "Variables file %s not found,\n using defaults in %s" \
|
|
|
|
% (current_vars_file, default_vars_file)
|
2006-05-22 20:29:33 +02:00
|
|
|
else:
|
2009-02-10 05:10:14 +01:00
|
|
|
print "Error: cannot find variables file %s or %s" \
|
|
|
|
% (current_vars_file, default_vars_file)
|
2006-05-22 20:29:33 +02:00
|
|
|
Exit(1)
|
2005-09-25 03:16:02 +02:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Apply current variable settings to env
|
|
|
|
sticky_vars.Update(env)
|
|
|
|
nonsticky_vars.Update(env)
|
2005-08-30 19:18:54 +02:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
help_text += "\nSticky variables for %s:\n" % variant_dir \
|
|
|
|
+ sticky_vars.GenerateHelpText(env) \
|
|
|
|
+ "\nNon-sticky variables for %s:\n" % variant_dir \
|
|
|
|
+ nonsticky_vars.GenerateHelpText(env)
|
2006-05-09 18:43:01 +02:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Process variable settings.
|
2005-08-30 19:18:54 +02:00
|
|
|
|
|
|
|
if not have_fenv and env['USE_FENV']:
|
|
|
|
print "Warning: <fenv.h> not available; " \
|
2009-02-10 05:10:14 +01:00
|
|
|
"forcing USE_FENV to False in", variant_dir + "."
|
2005-08-30 19:18:54 +02:00
|
|
|
env['USE_FENV'] = False
|
|
|
|
|
|
|
|
if not env['USE_FENV']:
|
2009-02-10 05:10:14 +01:00
|
|
|
print "Warning: No IEEE FP rounding mode control in", variant_dir + "."
|
2005-09-12 14:59:14 +02:00
|
|
|
print " FP results may deviate slightly from other platforms."
|
2005-08-30 19:18:54 +02:00
|
|
|
|
|
|
|
if env['EFENCE']:
|
|
|
|
env.Append(LIBS=['efence'])
|
|
|
|
|
|
|
|
if env['USE_MYSQL']:
|
|
|
|
if not have_mysql:
|
|
|
|
print "Warning: MySQL not available; " \
|
2009-02-10 05:10:14 +01:00
|
|
|
"forcing USE_MYSQL to False in", variant_dir + "."
|
2005-08-30 19:18:54 +02:00
|
|
|
env['USE_MYSQL'] = False
|
|
|
|
else:
|
2009-02-10 05:10:14 +01:00
|
|
|
print "Compiling in", variant_dir, "with MySQL support."
|
2005-08-31 05:34:36 +02:00
|
|
|
env.ParseConfig(mysql_config_libs)
|
|
|
|
env.ParseConfig(mysql_config_include)
|
2005-09-05 22:31:27 +02:00
|
|
|
|
2009-02-10 05:10:14 +01:00
|
|
|
# Save sticky variable settings back to current variables file
|
|
|
|
sticky_vars.Save(current_vars_file, env)
|
2005-09-05 22:31:27 +02:00
|
|
|
|
2005-11-09 15:52:07 +01:00
|
|
|
if env['USE_SSE2']:
|
|
|
|
env.Append(CCFLAGS='-msse2')
|
|
|
|
|
2006-06-17 15:26:08 +02:00
|
|
|
# The src/SConscript file sets up the build rules in 'env' according
|
2009-02-10 05:10:14 +01:00
|
|
|
# to the configured variables. It returns a list of environments,
|
2005-09-05 22:31:27 +02:00
|
|
|
# one for each variant build (debug, opt, etc.)
|
2009-02-10 05:10:14 +01:00
|
|
|
envList = SConscript('src/SConscript', variant_dir = variant_path,
|
2006-06-10 05:01:31 +02:00
|
|
|
exports = 'env')
|
2005-09-05 22:31:27 +02:00
|
|
|
|
|
|
|
# Set up the regression tests for each build.
|
2006-07-19 22:07:25 +02:00
|
|
|
for e in envList:
|
|
|
|
SConscript('tests/SConscript',
|
2009-02-10 05:10:14 +01:00
|
|
|
variant_dir = joinpath(variant_path, 'tests', e.Label),
|
2006-07-19 22:07:25 +02:00
|
|
|
exports = { 'env' : e }, duplicate = False)
|
2004-07-03 06:16:38 +02:00
|
|
|
|
2006-05-09 18:43:01 +02:00
|
|
|
Help(help_text)
|