#!/usr/bin/env python """CI build script for Linux/MacOS/Windows on Travis/AppVeyor """ from __future__ import print_function import sys, os, stat, shutil import fileinput import logging import re import subprocess as sp import distutils.util logger = logging.getLogger(__name__) # Detect the service and set up context hash accordingly def detect_context(): buildconfig = 'default' if 'TRAVIS' in os.environ: ci['service'] = 'travis' ci['os'] = os.environ['TRAVIS_OS_NAME'] ci['platform'] = 'x64' ci['compiler'] = os.environ['TRAVIS_COMPILER'] if ci['os'] == 'windows': ci['choco'] += ['strawberryperl'] if re.match(r'^vs', ci['compiler']): # Only Visual Studio 2017 available ci['compiler'] = 'vs2017' if 'BCFG' in os.environ: buildconfig = os.environ['BCFG'].lower() if 'APPVEYOR' in os.environ: ci['service'] = 'appveyor' if re.match(r'^Visual', os.environ['APPVEYOR_BUILD_WORKER_IMAGE']): ci['os'] = 'windows' elif re.match(r'^Ubuntu', os.environ['APPVEYOR_BUILD_WORKER_IMAGE']): ci['os'] = 'linux' elif re.match(r'^macOS', os.environ['APPVEYOR_BUILD_WORKER_IMAGE']): ci['os'] = 'osx' ci['platform'] = os.environ['PLATFORM'].lower() if 'CMP' in os.environ: ci['compiler'] = os.environ['CMP'] buildconfig = os.environ['CONFIGURATION'].lower() if 'STATIC' in os.environ: print("{0}WARNING: Variable 'STATIC' not supported anymore; use 'BCFG' instead{1}" .format(ANSI_RED, ANSI_RESET)) sys.stdout.flush() if not re.match(r'^((default|static|shared|dynamic|optimized|debug)-?)+$', buildconfig): print("{0}WARNING: Unrecognized build configuration setting '{1}'{2}" .format(ANSI_RED, buildconfig, ANSI_RESET)) sys.stdout.flush() if re.search('static', buildconfig): ci['static'] = True if re.search('debug', buildconfig): ci['debug'] = True if ci['static']: ci['configuration'] = 'static' else: ci['configuration'] = 'shared' if ci['debug']: ci['configuration'] += '-debug' else: ci['configuration'] += '-optimized' ci['scriptsdir'] = os.path.abspath(os.path.dirname(sys.argv[0])) if 'CHOCO' in os.environ: ci['choco'].extend(os.environ['CHOCO'].split()) if 'APT' in os.environ: ci['apt'].extend(os.environ['APT'].split()) ci['test'] = True if 'TEST' in os.environ and os.environ['TEST'].lower() == 'no': ci['test'] = False ci['parallel_make'] = 2 if 'PARALLEL_MAKE' in os.environ: ci['parallel_make'] = os.environ['PARALLEL_MAKE'] ci['clean_deps'] = True if 'CLEAN_DEPS' in os.environ and os.environ['CLEAN_DEPS'].lower() == 'no': ci['clean_deps'] = False logger.debug('Detected a build hosted on %s, using %s on %s (%s) configured as %s ' + '(test: %s, clean_deps: %s)', ci['service'], ci['compiler'], ci['os'], ci['platform'], ci['configuration'], ci['test'], ci['clean_deps']) curdir = os.getcwd() ci = {} seen_setups = [] modules_to_compile = [] setup = {} places = {} extra_makeargs = [] is_base314 = False is_make3 = False has_test_results = False silent_dep_builds = True do_recompile = False def clear_lists(): global is_base314, has_test_results, silent_dep_builds, is_make3 del seen_setups[:] del modules_to_compile[:] del extra_makeargs[:] setup.clear() places.clear() is_base314 = False is_make3 = False has_test_results = False silent_dep_builds = True do_recompile = False ci['service'] = '' ci['os'] = '' ci['platform'] = '' ci['compiler'] = '' ci['static'] = False ci['debug'] = False ci['configuration'] = '' ci['scriptsdir'] = '' ci['choco'] = ['make'] ci['apt'] = [] clear_lists() if 'BASE' in os.environ and os.environ['BASE'] == 'SELF': building_base = True places['EPICS_BASE'] = curdir else: building_base = False # Setup ANSI Colors ANSI_RED = "\033[31;1m" ANSI_GREEN = "\033[32;1m" ANSI_YELLOW = "\033[33;1m" ANSI_BLUE = "\033[34;1m" ANSI_MAGENTA = "\033[35;1m" ANSI_CYAN = "\033[36;1m" ANSI_RESET = "\033[0m" ANSI_CLEAR = "\033[0K" # Travis log fold control # from https://github.com/travis-ci/travis-rubies/blob/build/build.sh def fold_start(tag, title): if ci['service'] == 'travis': print('travis_fold:start:{0}{1}{2}{3}' .format(tag, ANSI_YELLOW, title, ANSI_RESET)) elif ci['service'] == 'appveyor': print('{0}===== \\/ \\/ \\/ ===== START: {1} ====={2}' .format(ANSI_YELLOW, title, ANSI_RESET)) sys.stdout.flush() def fold_end(tag, title): if ci['service'] == 'travis': print('\ntravis_fold:end:{0}\r' .format(tag), end='') elif ci['service'] == 'appveyor': print('{0}----- /\\ /\\ /\\ ----- END: {1} -----{2}' .format(ANSI_YELLOW, title, ANSI_RESET)) sys.stdout.flush() homedir = curdir if 'HomeDrive' in os.environ: homedir = os.path.join(os.getenv('HomeDrive'), os.getenv('HomePath')) elif 'HOME' in os.environ: homedir = os.getenv('HOME') cachedir = os.path.join(homedir, '.cache') toolsdir = os.path.join(homedir, '.tools') rtemsdir = os.path.join(homedir, '.rtems') if 'CACHEDIR' in os.environ: cachedir = os.environ['CACHEDIR'] vcvars_table = { # https://en.wikipedia.org/wiki/Microsoft_Visual_Studio#History 'vs2019': [r'C:\Program Files (x86)\Microsoft Visual Studio\2019\Community\VC\Auxiliary\Build\vcvarsall.bat'], 'vs2017': [r'C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\VC\Auxiliary\Build\vcvarsall.bat', r'C:\Program Files (x86)\Microsoft Visual Studio\2017\BuildTools\VC\Auxiliary\Build\vcvarsall.bat'], 'vs2015': [r'C:\Program Files (x86)\Microsoft Visual Studio 14.0\VC\vcvarsall.bat'], 'vs2013': [r'C:\Program Files (x86)\Microsoft Visual Studio 12.0\VC\vcvarsall.bat'], 'vs2012': [r'C:\Program Files (x86)\Microsoft Visual Studio 11.0\VC\vcvarsall.bat'], 'vs2010': [r'C:\Program Files (x86)\Microsoft Visual Studio 10.0\VC\vcvarsall.bat'], 'vs2008': [r'C:\Program Files (x86)\Microsoft Visual Studio 9.0\VC\vcvarsall.bat'], } vcvars_found = {} for key in vcvars_table: for dir in vcvars_table[key]: if os.path.exists(dir): vcvars_found[key] = dir def modlist(): if building_base: ret = [] else: for var in ['ADD_MODULES', 'MODULES']: setup.setdefault(var, '') if var in os.environ: setup[var] = os.environ[var] logger.debug('ENV assignment: %s = %s', var, setup[var]) ret = ['BASE'] + setup['ADD_MODULES'].upper().split() + setup['MODULES'].upper().split() return ret def host_info(): print('{0}Build using {1} compiler on {2} ({3}) hosted by {4}{5}' .format(ANSI_CYAN, ci['compiler'], ci['os'], ci['platform'], ci['service'], ANSI_RESET)) print('{0}Python setup{1}'.format(ANSI_CYAN, ANSI_RESET)) print(sys.version) print('PYTHONPATH') for dname in sys.path: print(' ', dname) print('platform =', distutils.util.get_platform()) if ci['os'] == 'windows': print('{0}Available Visual Studio versions{1}'.format(ANSI_CYAN, ANSI_RESET)) for comp in vcvars_found: print(comp, 'in', vcvars_found[comp]) sys.stdout.flush() # Error-handler to make shutil.rmtree delete read-only files on Windows def remove_readonly(func, path, excinfo): os.chmod(path, stat.S_IWRITE) func(path) # source_set(setup) # # Source a settings file (extension .set) found in the setup_dirs path # May be called recursively (from within a setup file) def source_set(name): # allowed separators: colon or whitespace setup_dirs = os.getenv('SETUP_PATH', "").replace(':', ' ').split() if len(setup_dirs) == 0: raise NameError("{0}Search path for setup files (SETUP_PATH) is empty{1}".format(ANSI_RED, ANSI_RESET)) for set_dir in setup_dirs: set_file = os.path.join(set_dir, name) + ".set" if set_file in seen_setups: print("Ignoring already included setup file {0}".format(set_file)) return if os.path.isfile(set_file): seen_setups.append(set_file) print("Opening setup file {0}".format(set_file)) sys.stdout.flush() with open(set_file) as fp: for line in fp: if not line.strip() or line.strip()[0] == '#': continue if line.startswith("include"): logger.debug('%s: Found include directive, reading %s next', set_file, line.split()[1]) source_set(line.split()[1]) continue assign = line.replace('"', '').strip().split("=", 1) setup.setdefault(assign[0], os.getenv(assign[0], "")) if not setup[assign[0]].strip(): logger.debug('%s: setup[%s] = %s', set_file, assign[0], assign[1]) setup[assign[0]] = assign[1] logger.debug('Done with setup file %s', set_file) break else: raise NameError("{0}Setup file {1}.set does not exist in SETUP_PATH search path ({2}){3}" .format(ANSI_RED, name, setup_dirs, ANSI_RESET)) # update_release_local(var, location) # var name of the variable to set in RELEASE.local # location location (absolute path) of where variable should point to # # Manipulate RELEASE.local in the cache location: # - replace "$var=$location" line if it exists and has changed # - otherwise add "$var=$location" line and possibly move EPICS_BASE=... line to the end # Set places[var] = location def update_release_local(var, location): release_local = os.path.join(cachedir, 'RELEASE.local') updated_line = '{0}={1}'.format(var, location.replace('\\', '/')) places[var] = location if not os.path.exists(release_local): logger.debug('RELEASE.local does not exist, creating it') try: os.makedirs(cachedir) except: pass touch = open(release_local, 'w') touch.close() base_line = '' found = False logger.debug("Opening RELEASE.local for adding '%s'", updated_line) for line in fileinput.input(release_local, inplace=1): output_line = line.strip() if 'EPICS_BASE=' in line: base_line = line.strip() logger.debug("Found EPICS_BASE line '%s', not writing it", base_line) continue elif '{0}='.format(var) in line: logger.debug("Found '%s=' line, replacing", var) found = True output_line = updated_line logger.debug("Writing line to RELEASE.local: '%s'", output_line) print(output_line) fileinput.close() release_local = open(release_local, "a") if not found: logger.debug("Adding new definition: '%s'", updated_line) print(updated_line, file=release_local) if base_line: logger.debug("Writing EPICS_BASE line: '%s'", base_line) print(base_line, file=release_local) release_local.close() def set_setup_from_env(dep): for postf in ['', '_DIRNAME', '_REPONAME', '_REPOOWNER', '_REPOURL', '_VARNAME', '_RECURSIVE', '_DEPTH', '_HOOK']: if dep + postf in os.environ: setup[dep + postf] = os.environ[dep + postf] logger.debug('ENV assignment: %s = %s', dep + postf, setup[dep + postf]) def call_git(args, **kws): if 'cwd' in kws: place = kws['cwd'] else: place = os.getcwd() logger.debug("EXEC '%s' in %s", ' '.join(['git'] + args), place) sys.stdout.flush() exitcode = sp.call(['git'] + args, **kws) logger.debug('EXEC DONE') return exitcode def call_make(args=[], **kws): place = kws.get('cwd', os.getcwd()) parallel = kws.pop('parallel', ci['parallel_make']) silent = kws.pop('silent', False) use_extra = kws.pop('use_extra', False) # no parallel make for Base 3.14 if parallel <= 0 or is_base314: makeargs = [] else: makeargs = ['-j{0}'.format(parallel)] if not is_make3: makeargs += ['-Otarget'] if silent: makeargs += ['-s'] if use_extra: makeargs += extra_makeargs logger.debug("EXEC '%s' in %s", ' '.join(['make'] + makeargs + args), place) sys.stdout.flush() exitcode = sp.call(['make'] + makeargs + args, **kws) logger.debug('EXEC DONE') if exitcode != 0: sys.exit(exitcode) def get_git_hash(place): logger.debug("EXEC 'git log -n1 --pretty=format:%%H' in %s", place) sys.stdout.flush() head = sp.check_output(['git', 'log', '-n1', '--pretty=format:%H'], cwd=place).decode() logger.debug('EXEC DONE') return head def complete_setup(dep): set_setup_from_env(dep) setup.setdefault(dep, 'master') setup.setdefault(dep + "_DIRNAME", dep.lower()) setup.setdefault(dep + "_REPONAME", dep.lower()) setup.setdefault('REPOOWNER', 'epics-modules') setup.setdefault(dep + "_REPOOWNER", setup['REPOOWNER']) setup.setdefault(dep + "_REPOURL", 'https://github.com/{0}/{1}.git' .format(setup[dep + '_REPOOWNER'], setup[dep + '_REPONAME'])) setup.setdefault(dep + "_VARNAME", dep) setup.setdefault(dep + "_RECURSIVE", 'YES') setup.setdefault(dep + "_DEPTH", -1) # add_dependency(dep, tag) # # Add a dependency to the cache area: # - check out (recursive if configured) in the CACHE area unless it already exists and the # required commit has been built # - Defaults: # $dep_DIRNAME = lower case ($dep) # $dep_REPONAME = lower case ($dep) # $dep_REPOURL = GitHub / $dep_REPOOWNER (or $REPOOWNER or epics-modules) / $dep_REPONAME .git # $dep_VARNAME = $dep # $dep_DEPTH = 5 # $dep_RECURSIVE = 1/YES (0/NO to for a flat clone) # - Add $dep_VARNAME line to the RELEASE.local file in the cache area (unless already there) # - Add full path to $modules_to_compile def add_dependency(dep): global do_recompile recurse = setup[dep + '_RECURSIVE'].lower() if recurse not in ['0', 'no']: recursearg = ["--recursive"] elif recurse not in ['1', 'yes']: recursearg = [] else: raise RuntimeError("Invalid value for {}_RECURSIVE='{}' not 0/NO/1/YES".format(dep, recurse)) deptharg = { '-1': ['--depth', '5'], '0': [], }.get(str(setup[dep + '_DEPTH']), ['--depth', str(setup[dep + '_DEPTH'])]) tag = setup[dep] logger.debug('Adding dependency %s with tag %s', dep, setup[dep]) # determine if dep points to a valid release or branch if call_git(['ls-remote', '--quiet', '--exit-code', '--refs', setup[dep + '_REPOURL'], tag]): raise RuntimeError("{0}{1} is neither a tag nor a branch name for {2} ({3}){4}" .format(ANSI_RED, tag, dep, setup[dep + '_REPOURL'], ANSI_RESET)) dirname = setup[dep + '_DIRNAME'] + '-{0}'.format(tag) place = os.path.join(cachedir, dirname) checked_file = os.path.join(place, "checked_out") if os.path.isdir(place): logger.debug('Dependency %s: directory %s exists, comparing checked-out commit', dep, place) # check HEAD commit against the hash in marker file if os.path.exists(checked_file): with open(checked_file, 'r') as bfile: checked_out = bfile.read().strip() bfile.close() else: checked_out = 'never' head = get_git_hash(place) logger.debug('Found checked_out commit %s, git head is %s', checked_out, head) if head != checked_out: logger.debug('Dependency %s out of date - removing', dep) shutil.rmtree(place, onerror=remove_readonly) else: print('Found {0} of dependency {1} up-to-date in {2}'.format(tag, dep, place)) sys.stdout.flush() if not os.path.isdir(place): if not os.path.isdir(cachedir): os.makedirs(cachedir) # clone dependency print('Cloning {0} of dependency {1} into {2}' .format(tag, dep, place)) sys.stdout.flush() call_git(['clone', '--quiet'] + deptharg + recursearg + ['--branch', tag, setup[dep + '_REPOURL'], dirname], cwd=cachedir) sp.check_call(['git', 'log', '-n1'], cwd=place) logger.debug('Setting do_recompile = True (all following modules will be recompiled') do_recompile = True if dep == 'BASE': # add MSI 1.7 to Base 3.14 versionfile = os.path.join(place, 'configure', 'CONFIG_BASE_VERSION') if os.path.exists(versionfile): with open(versionfile) as f: if 'BASE_3_14=YES' in f.read(): print('Adding MSI 1.7 to {0}'.format(place)) sys.stdout.flush() sp.check_call(['patch', '-p1', '-i', os.path.join(ci['scriptsdir'], 'add-msi-to-314.patch')], cwd=place) else: # force including RELEASE.local for non-base modules by overwriting their configure/RELEASE release = os.path.join(place, "configure", "RELEASE") if os.path.exists(release): with open(release, 'w') as fout: print('-include $(TOP)/../RELEASE.local', file=fout) # run hook if defined if dep + '_HOOK' in setup: hook = os.path.join(place, setup[dep + '_HOOK']) if os.path.exists(hook): print('Running hook {0} in {1}'.format(setup[dep + '_HOOK'], place)) sys.stdout.flush() sp.check_call(hook, shell=True, cwd=place) # write checked out commit hash to marker file head = get_git_hash(place) logger.debug('Writing hash of checked-out dependency (%s) to marker file', head) with open(checked_file, "w") as fout: print(head, file=fout) fout.close() if do_recompile: modules_to_compile.append(dep) update_release_local(setup[dep + "_VARNAME"], place) def detect_epics_host_arch(): if ci['os'] == 'windows': if re.match(r'^vs', ci['compiler']): # there is no combined static and debug EPICS_HOST_ARCH target, # so a combined debug and static target will appear to be just static # but debug will have been specified in CONFIG_SITE by prepare() hostarchsuffix = '' if ci['debug']: hostarchsuffix = '-debug' if ci['static']: hostarchsuffix = '-static' if ci['platform'] == 'x86': os.environ['EPICS_HOST_ARCH'] = 'win32-x86' + hostarchsuffix elif ci['platform'] == 'x64': os.environ['EPICS_HOST_ARCH'] = 'windows-x64' + hostarchsuffix elif ci['compiler'] == 'gcc': if ci['platform'] == 'x86': os.environ['EPICS_HOST_ARCH'] = 'win32-x86-mingw' elif ci['platform'] == 'x64': os.environ['EPICS_HOST_ARCH'] = 'windows-x64-mingw' if 'EPICS_HOST_ARCH' not in os.environ: logger.debug('Running script to detect EPICS host architecture in %s', places['EPICS_BASE']) os.environ['EPICS_HOST_ARCH'] = 'unknown' eha_scripts = [ os.path.join(places['EPICS_BASE'], 'src', 'tools', 'EpicsHostArch.pl'), os.path.join(places['EPICS_BASE'], 'startup', 'EpicsHostArch.pl'), ] for eha in eha_scripts: if os.path.exists(eha): os.environ['EPICS_HOST_ARCH'] = sp.check_output(['perl', eha]).decode('ascii').strip() logger.debug('%s returned: %s', eha, os.environ['EPICS_HOST_ARCH']) break def setup_for_build(args): global is_base314, has_test_results, is_make3 dllpaths = [] if ci['os'] == 'windows': if ci['service'] == 'appveyor': if ci['compiler'] == 'vs2019': # put strawberry perl in the PATH os.environ['PATH'] = os.pathsep.join([os.path.join(r'C:\Strawberry\perl\site\bin'), os.path.join(r'C:\Strawberry\perl\bin'), os.environ['PATH']]) if ci['compiler'] == 'gcc': if 'INCLUDE' not in os.environ: os.environ['INCLUDE'] = '' if ci['platform'] == 'x86': os.environ['INCLUDE'] = os.pathsep.join( [r'C:\mingw-w64\i686-6.3.0-posix-dwarf-rt_v5-rev1\mingw32\include', os.environ['INCLUDE']]) os.environ['PATH'] = os.pathsep.join([r'C:\mingw-w64\i686-6.3.0-posix-dwarf-rt_v5-rev1\mingw32\bin', os.environ['PATH']]) elif ci['platform'] == 'x64': os.environ['INCLUDE'] = os.pathsep.join( [r'C:\mingw-w64\x86_64-8.1.0-posix-seh-rt_v6-rev0\mingw64\include', os.environ['INCLUDE']]) os.environ['PATH'] = os.pathsep.join([r'C:\mingw-w64\x86_64-8.1.0-posix-seh-rt_v6-rev0\mingw64\bin', os.environ['PATH']]) if ci['service'] == 'travis': os.environ['PATH'] = os.pathsep.join([r'C:\Strawberry\perl\site\bin', r'C:\Strawberry\perl\bin', os.environ['PATH']]) # Find BASE location if not building_base: with open(os.path.join(cachedir, 'RELEASE.local'), 'r') as f: lines = f.readlines() for line in lines: (mod, place) = line.strip().split('=') if mod == 'EPICS_BASE': places['EPICS_BASE'] = place else: places['EPICS_BASE'] = '.' detect_epics_host_arch() if ci['os'] == 'windows': if not building_base: with open(os.path.join(cachedir, 'RELEASE.local'), 'r') as f: lines = f.readlines() for line in lines: (mod, place) = line.strip().split('=') bin_dir = os.path.join(place, 'bin', os.environ['EPICS_HOST_ARCH']) if os.path.isdir(bin_dir): dllpaths.append(bin_dir) # Add DLL location to PATH bin_dir = os.path.join(os.getcwd(), 'bin', os.environ['EPICS_HOST_ARCH']) if os.path.isdir(bin_dir): dllpaths.append(bin_dir) os.environ['PATH'] = os.pathsep.join(dllpaths + [os.environ['PATH']]) logger.debug('DLL paths added to PATH: %s', os.pathsep.join(dllpaths)) cfg_base_version = os.path.join(places['EPICS_BASE'], 'configure', 'CONFIG_BASE_VERSION') if os.path.exists(cfg_base_version): with open(cfg_base_version) as myfile: if 'BASE_3_14=YES' in myfile.read(): is_base314 = True if not is_base314: rules_build = os.path.join(places['EPICS_BASE'], 'configure', 'RULES_BUILD') if os.path.exists(rules_build): with open(rules_build) as myfile: for line in myfile: if re.match('^test-results:', line): has_test_results = True # Check make version if re.match(r'^GNU Make 3', sp.check_output(['make', '-v']).decode('ascii')): is_make3 = True # apparently %CD% is handled automagically os.environ['TOP'] = os.getcwd() addpaths = [] for path in args.paths: try: addpaths.append(path.format(**os.environ)) except KeyError: print('Environment') [print(' ', K, '=', repr(V)) for K, V in os.environ.items()] raise os.environ['PATH'] = os.pathsep.join([os.environ['PATH']] + addpaths) # Add EXTRA make arguments for tag in ['EXTRA', 'EXTRA1', 'EXTRA2', 'EXTRA3', 'EXTRA4', 'EXTRA5']: if tag in os.environ: extra_makeargs.append(os.environ[tag]) def fix_etc_hosts(): # Several travis-ci images throw us a curveball in /etc/hosts # by including two entries for localhost. The first for 127.0.1.1 # causes epicsSockResolveTest to fail. # cat /etc/hosts # ... # 127.0.1.1 localhost localhost ip4-loopback # 127.0.0.1 localhost nettuno travis vagrant travis-job-.... logger.debug("EXEC sudo sed -ie '/^127\.0\.1\.1/ s|localhost\s*||g' /etc/hosts") sys.stdout.flush() exitcode = sp.call(['sudo', 'sed', '-ie', '/^127\.0\.1\.1/ s|localhost\s*||g', '/etc/hosts']) logger.debug('EXEC DONE') def prepare(args): host_info() fold_start('load.setup', 'Loading setup files') if 'SET' in os.environ: source_set(os.environ['SET']) source_set('defaults') [complete_setup(mod) for mod in modlist()] fold_end('load.setup', 'Loading setup files') logger.debug('Loaded setup') kvs = list(setup.items()) kvs.sort() [logger.debug(' %s = "%s"', *kv) for kv in kvs] logger.debug('Effective module list: %s', modlist()) if ci['service'] == 'travis' and ci['os'] == 'linux': fix_etc_hosts() # we're working with tags (detached heads) a lot: suppress advice call_git(['config', '--global', 'advice.detachedHead', 'false']) fold_start('check.out.dependencies', 'Checking/cloning dependencies') [add_dependency(mod) for mod in modlist()] if not building_base: if os.path.isdir('configure'): targetdir = 'configure' else: targetdir = '.' shutil.copy(os.path.join(cachedir, 'RELEASE.local'), targetdir) fold_end('check.out.dependencies', 'Checking/cloning dependencies') if 'BASE' in modules_to_compile or building_base: fold_start('set.up.epics_build', 'Configuring EPICS build system') detect_epics_host_arch() # Set static/debug in CONFIG_SITE with open(os.path.join(places['EPICS_BASE'], 'configure', 'CONFIG_SITE'), 'a') as f: if ci['static']: f.write('SHARED_LIBRARIES=NO\n') f.write('STATIC_BUILD=YES\n') linktype = 'static' else: linktype = 'shared (DLL)' if ci['debug']: f.write('HOST_OPT=NO\n') optitype = 'debug' else: optitype = 'optimized' print('EPICS Base build system set up for {0} build with {1} linking' .format(optitype, linktype)) # Enable/fix parallel build for VisualStudio compiler on older Base versions if ci['os'] == 'windows' and re.match(r'^vs', ci['compiler']): add_vs_fix = True config_win = os.path.join(places['EPICS_BASE'], 'configure', 'os', 'CONFIG.win32-x86.win32-x86') with open(config_win) as f: for line in f: if re.match(r'^ifneq \(\$\(VisualStudioVersion\),11\.0\)', line): add_vs_fix = False if add_vs_fix: logger.debug('Adding parallel build fix for VisualStudio to %s', config_win) with open(config_win, 'a') as f: f.write(''' # Fix parallel build for some VisualStudio versions ifneq ($(VisualStudioVersion),) ifneq ($(VisualStudioVersion),11.0) ifeq ($(findstring -FS,$(OPT_CXXFLAGS_NO)),) OPT_CXXFLAGS_NO += -FS OPT_CFLAGS_NO += -FS endif else OPT_CXXFLAGS_NO := $(filter-out -FS,$(OPT_CXXFLAGS_NO)) OPT_CFLAGS_NO := $(filter-out -FS,$(OPT_CFLAGS_NO)) endif endif''') # Cross-compilations from Linux platform if ci['os'] == 'linux': # Cross compilation to Windows/Wine (set WINE to architecture "32", "64") # requires wine and g++-mingw-w64-i686 / g++-mingw-w64-x86-64 if 'WINE' in os.environ: if os.environ['WINE'] == '32': print('Cross compiler mingw32 / Wine') with open(os.path.join(places['EPICS_BASE'], 'configure', 'os', 'CONFIG.linux-x86.win32-x86-mingw'), 'a') as f: f.write(''' CMPLR_PREFIX=i686-w64-mingw32-''') with open(os.path.join(places['EPICS_BASE'], 'configure', 'CONFIG_SITE'), 'a') as f: f.write(''' CROSS_COMPILER_TARGET_ARCHS+=win32-x86-mingw''') if os.environ['WINE'] == '64': print('Cross compiler mingw64 / Wine') with open(os.path.join(places['EPICS_BASE'], 'configure', 'os', 'CONFIG.linux-x86.windows-x64-mingw'), 'a') as f: f.write(''' CMPLR_PREFIX=x86_64-w64-mingw32-''') with open(os.path.join(places['EPICS_BASE'], 'configure', 'CONFIG_SITE'), 'a') as f: f.write(''' CROSS_COMPILER_TARGET_ARCHS += windows-x64-mingw''') # Cross compilation on Linux to RTEMS (set RTEMS to version "4.9", "4.10") # requires qemu, bison, flex, texinfo, install-info if 'RTEMS' in os.environ: print('Cross compiler RTEMS{0} @ pc386',format(os.environ['RTEMS'])) with open(os.path.join(places['EPICS_BASE'], 'configure', 'os', 'CONFIG_SITE.Common.RTEMS'), 'a') as f: f.write(''' RTEMS_VERSION={0} RTEMS_BASE={1}'''.format(os.environ['RTEMS'], rtemsdir)) # Base 3.15 doesn't have -qemu target architecture qemu_suffix = '' if os.path.exists(os.path.join(places['EPICS_BASE'], 'configure', 'os', 'CONFIG.Common.RTEMS-pc386-qemu')): qemu_suffix = '-qemu' with open(os.path.join(places['EPICS_BASE'], 'configure', 'CONFIG_SITE'), 'a') as f: f.write(''' CROSS_COMPILER_TARGET_ARCHS += RTEMS-pc386{0}'''.format(qemu_suffix)) host_ccmplr_name = re.sub(r'^([a-zA-Z][^-]*(-[a-zA-Z][^-]*)*)+(-[0-9.]|)$', r'\1', ci['compiler']) host_cmplr_ver_suffix = re.sub(r'^([a-zA-Z][^-]*(-[a-zA-Z][^-]*)*)+(-[0-9.]|)$', r'\3', ci['compiler']) host_cmpl_ver = host_cmplr_ver_suffix[1:] if host_ccmplr_name == 'clang': print('Host compiler clang') host_cppcmplr_name = re.sub(r'clang', r'clang++', host_ccmplr_name) with open(os.path.join(places['EPICS_BASE'], 'configure', 'os', 'CONFIG_SITE.Common.'+os.environ['EPICS_HOST_ARCH']), 'a') as f: f.write(''' GNU = NO CMPLR_CLASS = clang CC = {0}{2} CCC = {1}{2}'''.format(host_ccmplr_name, host_cppcmplr_name, host_cmplr_ver_suffix)) # hack with open(os.path.join(places['EPICS_BASE'], 'configure', 'CONFIG.gnuCommon'), 'a') as f: f.write(''' CMPLR_CLASS = clang''') if host_ccmplr_name == 'gcc': print('Host compiler gcc') host_cppcmplr_name = re.sub(r'gcc', r'g++', host_ccmplr_name) with open(os.path.join(places['EPICS_BASE'], 'configure', 'os', 'CONFIG_SITE.Common.' + os.environ['EPICS_HOST_ARCH']), 'a') as f: f.write(''' CC = {0}{2} CCC = {1}{2}'''.format(host_ccmplr_name, host_cppcmplr_name, host_cmplr_ver_suffix)) # Add additional flags to CONFIG_SITE flags_text = '' if 'USR_CPPFLAGS' in os.environ: flags_text += ''' USR_CPPFLAGS += {0}'''.format(os.environ['USR_CPPFLAGS']) if 'USR_CFLAGS' in os.environ: flags_text += ''' USR_CFLAGS += {0}'''.format(os.environ['USR_CFLAGS']) if 'USR_CXXFLAGS' in os.environ: flags_text += ''' USR_CXXFLAGS += {0}'''.format(os.environ['USR_CXXFLAGS']) if flags_text: with open(os.path.join(places['EPICS_BASE'], 'configure', 'CONFIG_SITE'), 'a') as f: f.write(flags_text) fold_end('set.up.epics_build', 'Configuring EPICS build system') if not os.path.isdir(toolsdir): os.makedirs(toolsdir) if ci['os'] == 'windows' and ci['choco']: fold_start('install.choco', 'Installing CHOCO packages') sp.check_call(['choco', 'install'] + ci['choco']) fold_end('install.choco', 'Installing CHOCO packages') if ci['os'] == 'linux' and ci['apt']: fold_start('install.apt', 'Installing APT packages') sp.check_call(['sudo', 'apt-get', '-y', 'install'] + ci['apt']) fold_end('install.apt', 'Installing APT packages') if ci['os'] == 'linux' and 'RTEMS' in os.environ: tar_name = 'i386-rtems{0}-trusty-20171203-{0}.tar.bz2'.format(os.environ['RTEMS']) print('Downloading RTEMS {0} cross compiler: {1}' .format(os.environ['RTEMS'], tar_name)) sys.stdout.flush() sp.check_call(['curl', '-fsSL', '--retry', '3', '-o', tar_name, 'https://github.com/mdavidsaver/rsb/releases/download/20171203-{0}/{1}' .format(os.environ['RTEMS'], tar_name)], cwd=toolsdir) sp.check_call(['tar', '-C', '/', '-xmj', '-f', os.path.join(toolsdir, tar_name)]) os.remove(os.path.join(toolsdir, tar_name)) setup_for_build(args) print('{0}EPICS_HOST_ARCH = {1}{2}'.format(ANSI_CYAN, os.environ['EPICS_HOST_ARCH'], ANSI_RESET)) print('{0}$ make --version{1}'.format(ANSI_CYAN, ANSI_RESET)) sys.stdout.flush() call_make(['--version'], parallel=0) print('{0}$ perl --version{1}'.format(ANSI_CYAN, ANSI_RESET)) sys.stdout.flush() sp.check_call(['perl', '--version']) if re.match(r'^vs', ci['compiler']): print('{0}$ cl{1}'.format(ANSI_CYAN, ANSI_RESET)) sys.stdout.flush() sp.check_call(['cl']) else: cc = ci['compiler'] print('{0}$ {1} --version{2}'.format(ANSI_CYAN, cc, ANSI_RESET)) sys.stdout.flush() sp.check_call([cc, '--version']) if not building_base: fold_start('build.dependencies', 'Build missing/outdated dependencies') for mod in modules_to_compile: place = places[setup[mod + "_VARNAME"]] print('{0}Building dependency {1} in {2}{3}'.format(ANSI_YELLOW, mod, place, ANSI_RESET)) call_make(cwd=place, silent=silent_dep_builds) if ci['clean_deps']: call_make(args=['clean'], cwd=place, silent=silent_dep_builds) fold_end('build.dependencies', 'Build missing/outdated dependencies') print('{0}Dependency module information{1}'.format(ANSI_CYAN, ANSI_RESET)) print('Module Tag Binaries Commit') print(100 * '-') for mod in modlist(): if mod in modules_to_compile: stat = 'rebuilt' else: stat = 'from cache' commit = sp.check_output(['git', 'log', '-n1', '--oneline'], cwd=places[setup[mod + "_VARNAME"]])\ .decode('ascii').strip() print("%-10s %-12s %-11s %s" % (mod, setup[mod], stat, commit)) print('{0}Contents of RELEASE.local{1}'.format(ANSI_CYAN, ANSI_RESET)) with open(os.path.join(cachedir, 'RELEASE.local'), 'r') as f: print(f.read().strip()) def build(args): setup_for_build(args) fold_start('build.module', 'Build the main module') call_make(args.makeargs, use_extra=True) fold_end('build.module', 'Build the main module') def test(args): if ci['test']: setup_for_build(args) fold_start('test.module', 'Run the main module tests') if has_test_results: call_make(['tapfiles']) else: call_make(['runtests']) fold_end('test.module', 'Run the main module tests') else: print("{0}Action 'test' skipped as per configuration{1}" .format(ANSI_YELLOW, ANSI_RESET)) def test_results(args): if ci['test']: setup_for_build(args) fold_start('test.results', 'Sum up main module test results') if has_test_results: call_make(['test-results'], parallel=0, silent=True) else: print("{0}Base in {1} does not implement 'test-results' target{2}" .format(ANSI_YELLOW, places['EPICS_BASE'], ANSI_RESET)) fold_end('test.results', 'Sum up main module test results') else: print("{0}Action 'test-results' skipped as per configuration{1}" .format(ANSI_YELLOW, ANSI_RESET)) def doExec(args): 'exec user command with vcvars' setup_for_build(args) os.environ['MAKE'] = 'make' fold_start('exec.command', 'Execute command {}'.format(args.cmd)) sp.check_call(' '.join(args.cmd), shell=True) fold_end('exec.command', 'Execute command {}'.format(args.cmd)) def with_vcvars(cmd): '''re-exec main script with a (hopefully different) command ''' CC = ci['compiler'] # cf. https://docs.microsoft.com/en-us/cpp/build/building-on-the-command-line info = { 'python': sys.executable, 'self': sys.argv[0], 'cmd': cmd, } info['arch'] = { 'x86': 'x86', # 'amd64_x86' ?? 'x64': 'amd64', }[ci['platform']] # 'x86' or 'x64' info['vcvars'] = vcvars_found[CC] script = ''' call "{vcvars}" {arch} "{python}" "{self}" {cmd} '''.format(**info) print('{0}Calling vcvars-trampoline.bat to set environment for {1} on {2}{3}' .format(ANSI_YELLOW, CC, ci['platform'], ANSI_RESET)) sys.stdout.flush() logger.debug('----- Creating vcvars-trampoline.bat -----') for line in script.split('\n'): logger.debug(line) logger.debug('----- snip -----') with open('vcvars-trampoline.bat', 'w') as F: F.write(script) returncode = sp.call('vcvars-trampoline.bat', shell=True) if returncode != 0: sys.exit(returncode) def getargs(): from argparse import ArgumentParser, REMAINDER p = ArgumentParser() p.add_argument('--no-vcvars', dest='vcvars', default=True, action='store_false', help='Assume vcvarsall.bat has already been run') p.add_argument('--add-path', dest='paths', default=[], action='append', help='Append directory to %PATH%. Expands {ENVVAR}') subp = p.add_subparsers() cmd = subp.add_parser('prepare') cmd.set_defaults(func=prepare) cmd = subp.add_parser('build') cmd.add_argument('makeargs', nargs=REMAINDER) cmd.set_defaults(func=build) cmd = subp.add_parser('test') cmd.set_defaults(func=test) cmd = subp.add_parser('test-results') cmd.set_defaults(func=test_results) cmd = subp.add_parser('exec') cmd.add_argument('cmd', nargs=REMAINDER) cmd.set_defaults(func=doExec) return p def main(raw): global silent_dep_builds args = getargs().parse_args(raw) if 'VV' in os.environ and os.environ['VV'] == '1': logging.basicConfig(level=logging.DEBUG) silent_dep_builds = False detect_context() if args.vcvars and ci['compiler'].startswith('vs'): # re-exec with MSVC in PATH with_vcvars(' '.join(['--no-vcvars'] + raw)) else: args.func(args) if __name__ == '__main__': main(sys.argv[1:])