#!/usr/bin/python #---------------------------------------------------------------------- # Name: build.py # Purpose: Master build controller script. # This script is used to run through the commands used for the # various stages of building Phoenix, and can also be a front-end # for building wxWidgets and the wxPython distribution files. # # Author: Robin Dunn # # Created: 3-Dec-2010 # Copyright: (c) 2010-2020 by Total Control Software # License: wxWindows License #---------------------------------------------------------------------- import sys import glob import hashlib import optparse import os import re import shutil import subprocess import traceback from io import BytesIO import bz2 import tarfile import tempfile import datetime import shlex import textwrap import warnings from pathlib import Path from shutil import which try: from setuptools.modified import newer, newer_group except ImportError: from distutils.dep_util import newer, newer_group from buildtools.config import Config, msg, opj, posixjoin, loadETG, etg2sip, findCmd, \ phoenixDir, wxDir, copyIfNewer, copyFile, \ macSetLoaderNames, \ getVcsRev, runcmd, getSipFiles, \ getVisCVersion, getToolsPlatformName, updateLicenseFiles, \ TemporaryDirectory, getMSVCInfo, generateVersionFiles import buildtools.version as version # defaults PYVER = '3.9' PYSHORTVER = '39' PYTHON = None # it will be set later PYTHON_ARCH = 'UNKNOWN' PYTHON_MACHINE = None MSVC_ARCH = None # convenience access to the wxPython version digits version2 = "%d.%d" % (version.VER_MAJOR, version.VER_MINOR) version3 = "%d.%d.%d" % (version.VER_MAJOR, version.VER_MINOR, version.VER_RELEASE) version2_nodot = version2.replace(".", "") version3_nodot = version3.replace(".", "") # same for the wxWidgets version wxversion2 = "%d.%d" % (version.wxVER_MAJOR, version.wxVER_MINOR) wxversion3 = "%d.%d.%d" % (version.wxVER_MAJOR, version.wxVER_MINOR, version.wxVER_RELEASE) wxversion2_nodot = wxversion2.replace(".", "") wxversion3_nodot = wxversion3.replace(".", "") unstable_series = (version.wxVER_MINOR % 2) == 1 # is the minor version odd or even? isWindows = sys.platform.startswith('win') isDarwin = sys.platform == "darwin" devMode = False baseName = version.PROJECT_NAME eggInfoName = baseName + '.egg-info' defaultMask='%s-%s*' % (baseName, version.VER_MAJOR) pyICON = 'packaging/docset/Vippi-blocks-icon-32.png' wxICON = 'packaging/docset/mondrian.png' # Some tools will be downloaded for the builds. These are the versions and # MD5s of the tool binaries currently in use. wafCurrentVersion = '2.1.5' wafMD5 = '2e7b2166c030dbac3e21891048df10aa' doxygenCurrentVersion = '1.9.1' doxygenMD5 = { 'darwin' : '6912d41cef5971fb07573190849a8a84', 'win32' : '90439896025dc8ddcd5d767ab2c2c489', 'linux' : 'ed2c35099165fce0d07d9a1809935928', } # And the location where they can be downloaded from toolsURL = 'https://wxwidgets.github.io/wxPython-tools' # MS Edge code and DLLs needed for the wxWEBVIEW_BACKEND_EDGE backend MS_edge_version = '1.0.1185.39' MS_edge_url = 'https://www.nuget.org/api/v2/package/Microsoft.Web.WebView2/{}'.format(MS_edge_version) #--------------------------------------------------------------------------- def usage(): print ("""\ Usage: ./build.py [command(s)] [options] Commands: N.N NN Major.Minor version number of the Python to use to run the other commands. Default is 3.9. Or you can use --python to specify the actual Python executable to use. dox Run Doxygen to produce the XML file used by ETG scripts doxhtml Run Doxygen to create the HTML documentation for wx touch 'touch' the etg files so they will all get run the next time the etg command is run. etg Run the ETG scripts that are out of date to update their SIP files and their Sphinx input files sip Run sip to generate the C++ wrapper source wxlib Build the Sphinx input files for wx.lib sphinx Run the documentation building process using Sphinx docset Build Dash or Zeal compatible docsets build Build both wxWidgets and wxPython build_wx Do only the wxWidgets part of the build build_py Build wxPython only install Install both wxWidgets and wxPython install_wx Install wxWidgets (but only if this tool was used to build it) install_py Install wxPython only sdist Build a tarball containing all source files sdist_demo Build a tarball containing just the demo and samples folders bdist Create a binary tarball release of wxPython Phoenix bdist_docs Build a tarball containing the documentation bdist_egg Build a Python egg. Requires magic. bdist_wheel Build a Python wheel. Requires magic. test Run the unit test suite test_* Run just the one named test module clean_wx Clean the wx parts of the build clean_py Clean the wxPython parts of the build clean_sphinx Clean the sphinx files clean Clean both wx and wxPython cleanall Clean all and do a little extra scrubbing too """) parser = makeOptionParser() parser.print_help() def main(args): setDevModeOptions(args) options, commands = parseArgs(args) setPythonVersion(args) os.environ['PYTHONPATH'] = os.environ.get('PYTHONPATH', '') + os.pathsep + phoenixDir() os.environ['PYTHONUNBUFFERED'] = 'yes' os.environ['WXWIN'] = wxDir() # Ensure that PYTHON's containing folder is on the PATH. Normally it will # already be there, and probably at the beginning. But if this script was # invoked with a venv Python without activating the venv (like we do on the # BuildBot) then it may not be in the PATH at all. os.environ['PATH'] += os.pathsep + os.path.dirname(PYTHON) wxpydir = os.path.join(phoenixDir(), "wx") if not os.path.exists(wxpydir): os.makedirs(wxpydir) if not args or 'help' in args or '--help' in args or '-h' in args: usage() sys.exit(1) cfg = Config(noWxConfig=True) msg('cfg.VERSION: %s' % cfg.VERSION) msg('') # For collecting test names or files, to be run after all have been pulled # off the command line test_names = [] while commands: # ensure that each command starts with the CWD being the phoenix dir. os.chdir(phoenixDir()) cmd = commands.pop(0) if not cmd: continue # ignore empty command-line args (possible with the buildbot) elif cmd.startswith('test_'): test_names.append('unittests/%s.py' % cmd) elif cmd.startswith('unittests/test_'): test_names.append(cmd) elif 'cmd_'+cmd in globals(): function = globals()['cmd_'+cmd] function(options, args) else: print('*** Unknown command: ' + cmd) usage() sys.exit(1) # Now run the collected tests names, if any if test_names: cmd_test(options, args, test_names) msg("Done!") #--------------------------------------------------------------------------- # Helper functions (see also buildtools.config for more) #--------------------------------------------------------------------------- def setPythonVersion(args): global PYVER global PYSHORTVER global PYTHON global PYTHON_ARCH global PYTHON_MACHINE global MSVC_ARCH havePyVer = False havePyPath = False for idx, arg in enumerate(args): if re.match(r'^[0-9]\.[0-9][0-9]?$', arg): havePyVer = True PYVER = arg PYSHORTVER = arg[0] + arg[2:] del args[idx] break if re.match(r'^[0-9][0-9][0-9]?$', arg): havePyVer = True PYVER = '%s.%s' % (arg[0], arg[1:]) PYSHORTVER = arg del args[idx] break if arg.startswith('--python'): havePyPath = True if '=' in arg: PYTHON = arg.split('=')[1] del args[idx] else: PYTHON = args[idx+1] del args[idx:idx+2] PYVER = runcmd([PYTHON, '-c', 'import sys; print(sys.version[:3])'], getOutput=True, echoCmd=False) PYSHORTVER = PYVER[0] + PYVER[2:] break if havePyVer: if isWindows and os.environ.get('TOOLS'): # Use $TOOLS to find the correct Python. If set then it should be # the install root of all Python's on the system, with the 64-bit # ones in an amd64 subfolder, like this: # # $TOOLS\Python27\python.exe # $TOOLS\Python33\python.exe # $TOOLS\amd64\Python27\python.exe # $TOOLS\amd64\Python33\python.exe # TOOLS = os.environ.get('TOOLS') if 'cygdrive' in TOOLS: TOOLS = bash2dosPath(TOOLS) use64flag = '--x64' in args if use64flag: args.remove('--x64') CPU = os.environ.get('CPU') if use64flag or CPU in ['AMD64', 'X64', 'amd64', 'x64']: TOOLS = posixjoin(TOOLS, 'amd64') PYTHON = posixjoin(TOOLS, 'python%s' % PYSHORTVER, 'python.exe') elif isWindows: # Otherwise check if the invoking Python is the right version if sys.version[:3] != PYVER: msg('ERROR: The invoking Python is not the requested version. Perhaps you should use --python') sys.exit(1) PYTHON = sys.executable PYVER = sys.version[:3] PYSHORTVER = PYVER[0] + PYVER[2] elif not isWindows: # find a pythonX.Y on the PATH PYTHON = runcmd("which python%s" % PYVER, True, False) if not PYTHON: # If no version or path were specified then default to the python # that invoked this script PYTHON = sys.executable PYVER = '{}.{}'.format(sys.version_info.major, sys.version_info.minor) PYSHORTVER = '{}{}'.format(sys.version_info.major, sys.version_info.minor) PYTHON = os.path.abspath(PYTHON) msg('Will build using: "%s"' % PYTHON) msg(runcmd([PYTHON, '-c', 'import sys; print(sys.version)'], True, False)) PYTHON_ARCH = runcmd( [PYTHON, '-c', 'import platform; print(platform.architecture()[0])'], True, False) if isWindows: PYTHON_MACHINE = runcmd( [PYTHON, '-c', 'import platform; print(platform.machine())'], True, False) if PYTHON_MACHINE == 'ARM64': MSVC_ARCH = 'arm64' elif PYTHON_ARCH == '64bit': MSVC_ARCH = 'x64' else: MSVC_ARCH = 'x86' msg('Python\'s architecture is %s' % PYTHON_ARCH) os.environ['PYTHON'] = PYTHON if PYTHON_ARCH == '64bit': # Make sure this is set in case it wasn't above. if isWindows and PYTHON_MACHINE == 'ARM64': os.environ['CPU'] = 'ARM64' else: os.environ['CPU'] = 'X64' def setDevModeOptions(args): # Using --dev is a shortcut for setting several build options that I use # while working on the code in my local workspaces. Most people will # probably not use this so it is not part for the documented options and # is explicitly handled here before the options parser is created. If # anybody besides Robin is using this option do not depend on the options # it inserts into the args list being consistent. They could change at any # update from the repository. myDevModeOptions = [ #'--build_dir=../bld', #'--prefix=/opt/wx/2.9', '--jobs={}'.format(max(2, int(numCPUs()/2))), # These will be ignored on the other platforms so it is okay to # include them unconditionally '--osx_cocoa', '--mac_arch=arm64,x86_64', '--no_allmo', ] if not isWindows: myDevModeOptions.append('--debug') if isWindows: myDevModeOptions.append('--cairo') if '--dev' in args: global devMode devMode = True idx = args.index('--dev') # replace the --dev item with the items from the list args[idx:idx+1] = myDevModeOptions def numCPUs(): """ Detects the number of CPUs on a system. This approach is from detectCPUs here: http://www.artima.com/weblogs/viewpost.jsp?thread=230001 """ # Linux, Unix and MacOS: if hasattr(os, "sysconf"): if "SC_NPROCESSORS_ONLN" in os.sysconf_names: # Linux & Unix: ncpus = os.sysconf("SC_NPROCESSORS_ONLN") if isinstance(ncpus, int) and ncpus > 0: return ncpus else: # OSX: p = subprocess.Popen("sysctl -n hw.ncpu", shell=True, stdout=subprocess.PIPE) return int(p.stdout.read()) # Windows: if "NUMBER_OF_PROCESSORS" in os.environ: ncpus = int(os.environ["NUMBER_OF_PROCESSORS"]) if ncpus > 0: return ncpus return 1 # Default def getMSWSettings(options): checkCompiler(quiet=True) class MSWsettings(object): pass msw = MSWsettings() msw.CPU = os.environ.get('CPU') if msw.CPU in ['AMD64', 'X64', 'ARM64'] or PYTHON_ARCH == '64bit': msw.dllArch = 'arm64' if msw.CPU == 'ARM64' else 'x64' msw.dllDir = posixjoin(wxDir(), "lib", "vc%s_%s_dll" % (getVisCVersion(), msw.dllArch)) else: msw.dllDir = posixjoin(wxDir(), "lib", "vc%s_dll" % getVisCVersion()) msw.buildDir = posixjoin(wxDir(), "build", "msw") msw.dll_type = "u" if options.debug: msw.dll_type = "ud" return msw def makeOptionParser(): OPTS = [ ("python", ("", "The python executable to build for.")), ("debug", (False, "Build wxPython with debug symbols")), ("relwithdebug", (False, "Turn on the generation of debug info for release builds on MSW.")), ("release", (False, "Turn off some development options for a release build.")), ("keep_hash_lines",(False, "Don't remove the '#line N' lines from the SIP generated code")), ("gtk2", (False, "On Linux build for gtk2 (default gtk3)")), ("gtk3", (True, "On Linux build for gtk3")), ("osx_cocoa", (True, "Build the OSX Cocoa port on Mac (default)")), ("osx_carbon", (False, "Build the OSX Carbon port on Mac (unsupported)")), ("mac_framework", (False, "Build wxWidgets as a Mac framework.")), ("mac_arch", ("", "Comma separated list of architectures to build on Mac")), ("use_syswx", (False, "Try to use an installed wx rather than building the " "one in this source tree. The wx-config in {prefix}/bin " "or the first found on the PATH determines which wx is " "used. Implies --no_magic.")), ("force_config", (False, "Run configure when building even if the script " "determines it's not necessary.")), ("no_config", (False, "Turn off configure step on autoconf builds")), ("no_magic", (False, "Do NOT use the magic that will enable the wxWidgets " "libraries to be bundled with wxPython. (Such as when " "using an uninstalled wx/wxPython from the build dir, " "or when distributing wxPython as an egg.) When using " "this flag you should either build with an already " "installed wxWidgets, or allow this script to build and " "install wxWidgets independently of wxPython.")), ("build_dir", ("", "Directory to store wx build files. (Not used on Windows)")), ("prefix", ("", "Prefix value to pass to the wx build.")), ("destdir", ("", "Installation root for wxWidgets, files will go to {destdir}/{prefix}")), ("extra_setup", ("", "Extra args to pass on setup.py's command line.")), ("extra_make", ("", "Extra args to pass on [n]make's command line.")), ("extra_waf", ("", "Extra args to pass on waf's command line.")), ("extra_pytest", ("", "Extra args to pass on py.test's command line.")), (("j","jobs"), ("", "Number of parallel compile jobs to do, if supported.")), ("both", (False, "Build both a debug and release version. (Only used on Windows)")), ("unicode", (True, "Build wxPython with unicode support (always on for wx2.9+)")), (("v", "verbose"), (False, "Print out more information during the build.")), ("nodoc", (False, "Do not run the default docs generator")), ("upload", (False, "Upload bdist and/or sdist packages to snapshot server.")), ("cairo", (False, "Allow Cairo use with wxGraphicsContext (Windows only)")), ("x64", (False, "Use and build for the 64bit version of Python on Windows")), ("jom", (False, "Use jom instead of nmake for the wxMSW build")), ("pytest_timeout", ("0", "Timeout, in seconds, for stopping stuck test cases. (Currently not working as expected, so disabled by default.)")), ("pytest_jobs", ("", "Number of parallel processes py.test should run")), ("docker_img", ("all", "Comma separated list of image tags to use for the build_docker command. Defaults to \"all\"")), ("dump_waf_log", (False, "If the waf build tool fails then using this option will cause waf's configure log to be printed")), ("regenerate_sysconfig", (False, "Waf uses Python's sysconfig and related tools to configure the build. In some cases that info can be incorrect, so this option regenerates it. Must have write access to Python's lib folder.")), ("no_allmo", (False, "Skip regenerating the wxWidgets message catalogs")), ("no_msedge", (False, "Do not include the MS Edge backend for wx.html2.WebView. (Windows only)")), ("quiet", (False, "Silence some of the messages from build.py")) ] parser = optparse.OptionParser("build options:") for opt, info in OPTS: default, txt = info action = 'store' if type(default) == bool: action = 'store_true' if isinstance(opt, str): opts = ('--'+opt, ) dest = opt else: opts = ('-'+opt[0], '--'+opt[1]) dest = opt[1] parser.add_option(*opts, default=default, action=action, dest=dest, help=txt) return parser def parseArgs(args): # If WXPYTHON_BUILD_ARGS is set in the environment, split it and add to args if os.environ.get('WXPYTHON_BUILD_ARGS', None): args += shlex.split(os.environ.get('WXPYTHON_BUILD_ARGS')) # Parse the args into options parser = makeOptionParser() options, args = parser.parse_args(args) if isWindows: # We always use magic on Windows options.no_magic = False options.use_syswx = False elif options.use_syswx: # Turn off magic if using the system wx options.no_magic = True # Some options don't make sense for release builds if options.release: options.debug = False options.both = False if os.path.exists('REV.txt'): os.unlink('REV.txt') if options.gtk2: options.gtk3 = False if options.quiet: import buildtools.config buildtools.config.runSilently = True return options, args class pushDir(object): def __init__(self, newDir): self.cwd = os.getcwd() os.chdir(newDir) def __del__(self): # pop back to the original dir os.chdir(self.cwd) def getBuildDir(options): if not isDarwin and not isWindows: BUILD_DIR = opj(phoenixDir(), 'build', 'wxbld', 'gtk3' if options.gtk3 else 'gtk2') else: BUILD_DIR = opj(phoenixDir(), 'build', 'wxbld') if options.build_dir: BUILD_DIR = os.path.abspath(options.build_dir) return BUILD_DIR def deleteIfExists(deldir, verbose=True): if os.path.exists(deldir) and os.path.isdir(deldir): try: if verbose: msg("Removing folder: %s" % deldir) shutil.rmtree(deldir) except Exception: if verbose: msg("Error: %s" % traceback.format_exc(1)) else: if verbose: msg("Unable to delete: '%s' (it doesn't exist or is not a folder)" % deldir) def delFiles(fileList, verbose=True): for afile in fileList: if verbose: print("Removing file: %s" % afile) os.remove(afile) def errorMsg(txt, cmdname, envvar): msg('ERROR: ' + txt) msg(' Set %s in the environment to use a local build of %s instead' % (envvar, cmdname)) def downloadTool(cmd, cmdname, envvar): msg('Not found. Attempting to download...') url = f"{toolsURL}/{os.path.basename(cmd)}.bz2" try: import requests resp = requests.get(url) resp.raise_for_status() except requests.exceptions.HTTPError: # The files could be packed as a tarball url = url.replace(".bz2", ".tar.bz2") try: resp = requests.get(url) resp.raise_for_status() except requests.exceptions.HTTPError: errorMsg('Unable to download ' + url, cmdname, envvar) traceback.print_exc() sys.exit(1) msg('Connection successful...') data = resp.content msg('Data downloaded...') if ".tar." in url: with tarfile.open('r:bz2', fileobj=BytesIO(data)) as tar: # Extraction filters are only available since Python 3.12 tar.extraction_filter = getattr(tarfile, 'data_filter', (lambda member, path: member)) filelist = [] for file in tar.getmembers(): if file.name.startswith("doxygen") or \ file.name.startswith("libclang"): filelist.append(file) tar.extractall(members=filelist, path=os.path.dirname(cmd)) else: data = bz2.decompress(data) Path(cmd).write_bytes(data) os.chmod(cmd, 0o755) def getTool(cmdName, version, MD5, envVar, platformBinary, linuxBits=False): # Check in the bin dir for the specified version of the tool command. If # it's not there then attempt to download it. Validity of the binary is # checked with an MD5 hash. if os.environ.get(envVar): # Setting a value in the environment overrides other options return os.environ.get(envVar) # setup if platformBinary: platform = getToolsPlatformName(linuxBits) ext = '' if platform == 'win32': ext = '.exe' cmd = opj(phoenixDir(), 'bin', '%s-%s-%s%s' % (cmdName, version, platform, ext)) md5 = MD5[platform] else: cmd = opj(phoenixDir(), 'bin', '%s-%s' % (cmdName, version)) md5 = MD5 msg('Checking for %s...' % cmd) if os.path.exists(cmd): # if the file exists run some verification checks on it # first make sure it is a normal file if not os.path.isfile(cmd): errorMsg('%s exists but is not a regular file.' % cmd, cmdName, envVar) sys.exit(1) # now check the MD5 if not in dev mode and it's set to None if not (devMode and md5 is None): m = hashlib.md5() m.update(Path(cmd).read_bytes()) if m.hexdigest() != md5: errorMsg('MD5 mismatch, got "%s"\n ' 'expected "%s"' % (m.hexdigest(), md5), cmdName, envVar) sys.exit(1) # If the cmd is a script run by some interpreter, or similar, # then we don't need to check anything else if not platformBinary: return cmd # Ensure that commands that are platform binaries are executable if not os.access(cmd, os.R_OK | os.X_OK): errorMsg('Cannot execute %s due to permissions error' % cmd, cmdName, envVar) sys.exit(1) try: p = subprocess.Popen([cmd, '--help'], stdout=subprocess.PIPE, stderr=subprocess.PIPE, env=os.environ) p.communicate() except OSError as e: errorMsg('Could not execute %s, got "%s"' % (cmd, e), cmdName, envVar) sys.exit(1) # if we get this far then all is well, the cmd is good to go return cmd downloadTool(cmd, cmdName, envVar) # Recursive call so the MD5 value will be double-checked on what was # just downloaded return getTool(cmdName, version, MD5, envVar, platformBinary, linuxBits) # The download and MD5 check only needs to happen once per run, cache the sip # cmd value here the first time through. _wafCmd = None def getWafCmd(): global _wafCmd if _wafCmd is None: _wafCmd = getTool('waf', wafCurrentVersion, wafMD5, 'WAF', False) return _wafCmd # and Doxygen _doxCmd = None def getDoxCmd(): global _doxCmd if _doxCmd is None: _doxCmd = getTool('doxygen', doxygenCurrentVersion, doxygenMD5, 'DOXYGEN', True) return _doxCmd def getMSWebView2(): fname = '{}.zip'.format(MS_edge_version) dest = opj(wxDir(), '3rdparty', 'webview2') if not os.path.exists(dest) or not os.path.exists(opj(dest, fname)): if os.path.exists(dest): shutil.rmtree(dest) os.makedirs(dest) msg('Downloading microsoft.web.webview2 {}...'.format(MS_edge_version)) try: import requests resp = requests.get(MS_edge_url) resp.raise_for_status() msg('Connection successful...') data = resp.content msg('Data downloaded...') except Exception: traceback.print_exc() sys.exit(1) # Write the downloaded data to a local file Path(dest, fname).write_bytes(data) # Unzip it from zipfile import ZipFile with ZipFile(opj(dest, fname)) as zip: zip.extractall(dest) class CommandTimer(object): def __init__(self, name): self.name = name self.startTime = datetime.datetime.now() msg('Running command: %s' % self.name) def __del__(self): delta = datetime.datetime.now() - self.startTime time = "" if delta.seconds / 60 > 0: time = "%dm" % (delta.seconds / 60) time += "%d.%ds" % (delta.seconds % 60, delta.microseconds / 1000) msg('Finished command: %s (%s)' % (self.name, time)) def uploadPackage(fileName, options, mask=defaultMask, keep=75): """ Upload the given filename to the configured package server location. Only the `keep` most recent files matching `mask` will be kept so the server space is not overly consumed. It is assumed that if the files are in sorted order then the end of the list will be the newest files. """ fileName = os.path.relpath(fileName) fileName = fileName.replace('\\', '/') msg("Uploading %s..." % fileName) # NOTE: It is expected that there will be a host entry defined in # ~/.ssh/config named wxpython-rbot, with the proper host, user, identity # file, etc. needed for making an SSH connection to the snapshots server. # Release builds work similarly with their own host configuration defined # in the ~/.ssh/config file. if options.release: host = 'wxpython-release' uploadDir = 'release-builds' else: host = 'wxpython-rbot' uploadDir = 'snapshot-builds' # copy the new file to the server cmd = 'scp {} {}:{}'.format(fileName, host, uploadDir) runcmd(cmd) # Make sure it is readable by all, and writable by rbot cmd = 'ssh {} "cd {}; chmod 644 {}"'.format(host, uploadDir, os.path.basename(fileName)) runcmd(cmd) if not options.release: # get the list of all snapshot files on the server cmd = 'ssh {} "cd {}; ls {}"'.format(host, uploadDir, mask) allFiles = runcmd(cmd, getOutput=True) allFiles = allFiles.strip().split('\n') allFiles.sort() # Leave the last keep number of builds, including this new one, on the server. # Delete the rest. rmFiles = allFiles[:-keep] if rmFiles: msg("Deleting %s" % ", ".join(rmFiles)) cmd = 'ssh {} "cd {}; rm {}"'.format(host, uploadDir, " ".join(rmFiles)) runcmd(cmd) msg("Upload complete!") def uploadTree(srcPath, destPath, options, days=30): """ Similar to the above but uploads a tree of files. """ msg("Uploading tree at {}...".format(srcPath)) if options.release: host = 'wxpython-release' uploadDir = opj('release-builds', destPath) else: host = 'wxpython-rbot' uploadDir = opj('snapshot-builds', destPath) # Ensure the destination exists cmd = 'ssh {0} "if [ ! -d {1} ]; then mkdir -p {1}; fi"'.format(host, uploadDir) runcmd(cmd) # Upload the tree cmd = 'scp -r {} {}:{}'.format(srcPath, host, uploadDir) runcmd(cmd) # Make sure it is readable by all cmd = 'ssh {} "chmod -R a+r {}"'.format(host, uploadDir) runcmd(cmd) if not options.release: # Remove files that were last modified more than `days` days ago msg("Cleaning up old builds.") cmd = 'ssh {} "find {} -type f -mtime +{} -delete"'.format(host, uploadDir, days) runcmd(cmd) msg("Tree upload and cleanup complete!") def checkCompiler(quiet=False): if isWindows: # Set up the PATH and other environment variables so the proper version # of MSVC will be used. The setuptools package is used to find the # needed info, so the target python shoudl have a recent version of # setuptools installed. info = getMSVCInfo(PYTHON, MSVC_ARCH, set_env=True) # # Just needed for debugging # if not quiet: # msg('MSVCinfo:') # msg(f' vc_ver: {info.vc_ver}') # msg(f' vs_ver: {info.vs_ver}') # msg(f' arch: {info.arch}') # msg(f' include: {info.include}') # msg(f' lib: {info.lib}') # msg(f' libpath: {info.libpath}') # Make sure there is now a cl.exe on the PATH CL = 'NOT FOUND' for d in os.environ['PATH'].split(os.pathsep): p = Path(d, 'cl.exe') if p.exists(): CL = p break if not quiet: msg(f' CL.exe: {CL}') # # Just needed for debugging # for d in info.include.split(os.pathsep): # p = Path(d, 'tchar.h') # if p.exists(): # msg(f' tchar.h: {p}') # break # else: # msg('**** tchar.h NOT FOUND!') # NOTE: SIP is now generating code with scoped-enums. Older linux # platforms like what we're using for builds, and also TravisCI for # example, are using GCC versions that are still defaulting to C++98, # so this flag is needed to turn on the C++11 mode. If this flag # causes problems with other non-Windows, non-Darwin compilers then # we'll need to make this a little smarter about what flag (if any) # needs to be used. # # NOTE 2: SIP changed its output such that this doesn't appear to be # needed anymore, but we'll leave the code in place to make it easy to # turn it back on again if/when needed. if False and not isWindows and not isDarwin: stdflag = '-std=c++11' curflags = os.environ.get('CXXFLAGS', '') if stdflag not in curflags: os.environ['CXXFLAGS'] = '{} {}'.format(stdflag, curflags) #print('**** Using CXXFLAGS:', os.environ.get('CXXFLAGS', '')) def getWafBuildBase(): base = posixjoin('build', 'waf', PYVER) if isWindows: if PYTHON_ARCH == '64bit': base = posixjoin(base, 'x64') else: base = posixjoin(base, 'x86') return base def getBashPath(): """Check if there is a bash.exe on the PATH""" bash = which('bash.exe') return bash def dos2bashPath(path): """ Convert an absolute dos-style path to one bash.exe can understand. """ path = path.replace('\\', '/') cygpath = which('cygpath') wsl = which('wsl') # If we have cygwin then we can use cygpath to convert the path. # Note that MSYS2 (and Git Bash) now also have cygpath so this should # work there too. if cygpath: path = runcmd('"{}" -u "{}"'.format(cygpath, path), getOutput=True, echoCmd=False) return path elif wsl: # Are we using Windows System for Linux? (untested) path = runcmd('"{}" wslpath -a -u "{}"'.format(wsl, path), getOutput=True, echoCmd=False) return path else: # Otherwise, do a simple translate and hope for the best? # c:/foo --> /c/foo # TODO: Check this!! drive, rest = os.path.splitdrive(path) path = '/{}/{}'.format(drive[0], rest) return path def bash2dosPath(path): """ Convert an absolute unix-style path to one Windows can understand. """ cygpath = which('cygpath') wsl = which('wsl') # If we have cygwin then we can use cygpath to convert the path. # Note that MSYS2 (and Git Bash) now also have cygpath so this should # work there too. if cygpath: path = runcmd('"{}" -w "{}"'.format(cygpath, path), getOutput=True, echoCmd=False) return path elif wsl: # Are we using Windows System for Linux? (untested) path = runcmd('"{}" wslpath -a -w "{}"'.format(wsl, path), getOutput=True, echoCmd=False) return path else: # Otherwise, do a simple translate and hope for the best? # /c/foo --> c:/foo # There's also paths like /cygdrive/c/foo or /mnt/c/foo, but in those # cases cygpath or wsl should be available. components = path.split('/') assert components[0] == '' and len(components[1]) == 1, "Expecting a path like /c/foo" path = components[1] + ':/' + '/'.join(components[2:]) return path def do_regenerate_sysconfig(): """ If a Python environment has been relocated to a new folder then it's possible that the sysconfig can still be using paths for the original location. Since wxPython's build uses WAF, which uses the sysconfig (via python-config, distutils.sysconfig, etc.) then we need to ensure that these paths match the current environment. TODO: Can this be done in a way that doesn't require overwriting a file in the environment? """ with TemporaryDirectory() as td: pwd = pushDir(td) # generate a new sysconfig data file cmd = [PYTHON, '-m', 'sysconfig', '--generate-posix-vars'] runcmd(cmd) # On success the new data module will have been written to a subfolder # of the current folder, which is recorded in ./pybuilddir.txt pybd = Path('pybuilddir.txt').read_text() # grab the file in that folder and copy it into the Python lib p = opj(td, pybd, '*') datafile = min(glob.glob(opj(td, pybd, '*'))) cmd = [PYTHON, '-c', 'import sysconfig; print(sysconfig.get_path("stdlib"))'] stdlib = runcmd(cmd, getOutput=True) shutil.copy(datafile, stdlib) del pwd def _setTarItemPerms(tarinfo): """ Used to set permissions of the files and fodlers in the source tarball """ if tarinfo.isdir(): tarinfo.mode = 0o755 else: tarinfo.mode |= 0o644 return tarinfo #--------------------------------------------------------------------------- # Command functions and helpers #--------------------------------------------------------------------------- def _doDox(arg): doxCmd = getDoxCmd() doxCmd = os.path.abspath(doxCmd) if isWindows: bash = getBashPath() if not bash: raise RuntimeError("ERROR: Unable to find bash.exe, needed for running regen.sh") doxCmd = dos2bashPath(doxCmd) print(doxCmd) os.environ['DOXYGEN'] = doxCmd os.environ['WX_SKIP_DOXYGEN_VERSION_CHECK'] = '1' d = posixjoin(wxDir(), 'docs/doxygen') d = d.replace('\\', '/') cmd = '"{}" -l -c "cd {} && ./regen.sh {}"'.format(bash, d, arg) else: os.environ['DOXYGEN'] = doxCmd os.environ['WX_SKIP_DOXYGEN_VERSION_CHECK'] = '1' pwd = pushDir(posixjoin(wxDir(), 'docs/doxygen')) cmd = 'bash ./regen.sh %s' % arg runcmd(cmd) def cmd_dox(options, args): cmdTimer = CommandTimer('dox') _doDox('xml') def cmd_doxhtml(options, args): cmdTimer = CommandTimer('doxhtml') _doDox('html') _doDox('chm') def cmd_docset_wx(options, args): cmdTimer = CommandTimer('docset_wx') cfg = Config() # Use Doxygen to generate the docset _doDox('docset') # Remove any existing docset in the dist dir and move the new docset in srcname = posixjoin(wxDir(), 'docs/doxygen/out/docset', 'wxWidgets-%s.docset' % wxversion2) destname = 'dist/wxWidgets-%s.docset' % wxversion3 if not os.path.isdir(srcname): msg('ERROR: %s not found' % srcname) sys.exit(1) if os.path.isdir(destname): shutil.rmtree(destname) shutil.move(srcname, destname) shutil.copyfile(wxICON, posixjoin(destname, 'icon.png')) def cmd_docset_py(options, args): cmdTimer = CommandTimer('docset_py') cfg = Config(noWxConfig=True) if not os.path.isdir('docs/html'): msg('ERROR: No docs/html dir found, has the sphinx build command been run?') sys.exit(1) # clear out any old docset build name = 'wxPython-{}.docset'.format(cfg.VERSION) docset = posixjoin('dist', name) if os.path.isdir(docset): shutil.rmtree(docset) # run the docset generator VERBOSE = '--verbose' if options.verbose else '' URL = 'https://docs.wxpython.org/' if options.release else 'https://wxpython.org/Phoenix/docs/html/' cmd = [PYTHON, '-m doc2dash', VERBOSE, '--name', name, '--icon', pyICON, '--index-page index.html', '--enable-js', '--online-redirect-url', URL, '--destination dist docs/html'] runcmd(' '.join(cmd)) # Remove the sidebar from the pages in the docset msg('Removing sidebar from docset pages...') _removeSidebar(opj('dist', name, 'Contents', 'Resources', 'Documents')) # build the tarball msg('Archiving Phoenix docset...') rootname = "wxPython-docset-{}".format(cfg.VERSION) tarfilename = posixjoin(phoenixDir(), 'dist', '{}.tar.gz'.format(rootname)) if os.path.exists(tarfilename): os.remove(tarfilename) with tarfile.open(name=tarfilename, mode="w:gz") as tarball: tarball.add(opj('dist', name), name, filter=_setTarItemPerms) if options.upload: uploadPackage(tarfilename, options, keep=5, mask='%s-docset-%s*' % (baseName, cfg.VER_MAJOR)) msg("Docset file built at %s" % tarfilename) def _removeSidebar(path): """ Remove the sidebar
from the pages going into the docset """ from bs4 import BeautifulSoup for filename in sorted(glob.glob(opj(path, '*.html'))): text = Path(filename).read_text(encoding="utf-8") text = text.replace('', '') soup = BeautifulSoup(text, 'html.parser') tag = soup.find('div', 'sphinxsidebar') if tag: tag.extract() tag = soup.find('div', 'document') if tag: tag.attrs['class'] = ['document-no-sidebar'] text = str(soup) Path(filename).write_text(text, encoding="utf-8") def cmd_docset(options, args): cmd_docset_wx(options, args) cmd_docset_py(options, args) def cmd_version(options, args): cfg = Config() print(cfg.VERSION) def cmd_etg(options, args): cmdTimer = CommandTimer('etg') cfg = Config() assert os.path.exists(cfg.DOXY_XML_DIR), "Doxygen XML folder not found: " + cfg.DOXY_XML_DIR pwd = pushDir(cfg.ROOT_DIR) # TODO: Better support for selecting etg cmd-line flags... flags = '--sip' if options.nodoc: flags += ' --nodoc' # get the files to run, moving _core the to the front of the list etgfiles = sorted(glob.glob(opj('etg', '_*.py'))) core_file = opj('etg', '_core.py') if core_file in etgfiles: etgfiles.remove(core_file) etgfiles.insert(0, core_file) for script in etgfiles: sipfile = etg2sip(script) deps = [script] ns = loadETG(script) if hasattr(ns, 'ETGFILES'): etgfiles += ns.ETGFILES[1:] # all but itself if hasattr(ns, 'DEPENDS'): deps += ns.DEPENDS if hasattr(ns, 'OTHERDEPS'): deps += ns.OTHERDEPS # run the script only if any dependencies are newer if newer_group(deps, sipfile): runcmd('"%s" %s %s' % (PYTHON, script, flags)) def cmd_sphinx(options, args): from sphinxtools.postprocess import genIndexes, makeHeadings, postProcess, genGallery from sphinxtools.stc_doc_postprocess import stc_categorise_methods cmdTimer = CommandTimer('sphinx') pwd = pushDir(phoenixDir()) sphinxDir = os.path.join(phoenixDir(), 'docs', 'sphinx') if not os.path.isdir(sphinxDir): raise Exception('Missing sphinx folder in the distribution') textFiles = sorted(glob.glob(sphinxDir + '/*.txt')) if not textFiles: raise Exception('No documentation files found. Please run "build.py touch etg" first') # Sort all wx.stc.StyledTextCtrl methods into categories stc_categorise_methods() # Copy the rst files into txt files restDir = os.path.join(sphinxDir, 'rest_substitutions', 'overviews') rstFiles = sorted(glob.glob(restDir + '/*.rst')) for rst in rstFiles: rstName = os.path.split(rst)[1] txt = os.path.join(sphinxDir, os.path.splitext(rstName)[0] + '.txt') copyIfNewer(rst, txt) genIndexes(sphinxDir) genGallery() # Copy the hand-edited top level doc files too rstFiles = sorted(glob.glob(os.path.join(phoenixDir(), 'docs', '*.rst'))) for rst in rstFiles: txt = os.path.join(sphinxDir, os.path.splitext(os.path.basename(rst))[0] + '.txt') copyIfNewer(rst, txt) makeHeadings() pwd2 = pushDir(sphinxDir) buildDir = os.path.join(sphinxDir, 'build') htmlDir = os.path.join(phoenixDir(), 'docs', 'html') sphinx_log = os.path.join(htmlDir, 'warnings', 'sphinx_warnings.log') runcmd('"{}" -m sphinx --builder html --color --warning-file {} \ --doctree-dir {}/doctrees . {}'.format(PYTHON, sphinx_log, buildDir, htmlDir)) del pwd2 msg('Postprocessing sphinx output...') postProcess(htmlDir, options) def cmd_wxlib(options, args): from sphinxtools.modulehunter import ModuleHunter cmdTimer = CommandTimer('wxlib') pwd = pushDir(phoenixDir()) for wx_pkg in ['lib', 'py', 'svg', 'tools']: libDir = os.path.join(phoenixDir(), 'wx', wx_pkg) if not os.path.isdir(libDir): raise Exception('Missing wx.{} folder in the distribution'.format(wx_pkg)) init_name = os.path.join(libDir, '__init__.py') import_name = 'wx.{}'.format(wx_pkg) ModuleHunter(init_name, import_name, version3) def cmd_wxpy(options, args): msg('Command wxpy has been folded into command wxlib.') def cmd_wxtools(options, args): msg('Command wxtools has been folded into command wxlib.') def cmd_bdist_docs(options, args): cmdTimer = CommandTimer('bdist_docs') pwd = pushDir(phoenixDir()) cfg = Config() msg("Archiving wxPython Phoenix documentation...") rootname = "%s-docs-%s" % (baseName, cfg.VERSION) tarfilename = posixjoin(phoenixDir(), 'dist', '%s.tar.gz' % rootname) if not os.path.exists('dist'): os.makedirs('dist') if os.path.exists(tarfilename): os.remove(tarfilename) with tarfile.open(name=tarfilename, mode="w:gz") as tarball: tarball.add('docs/html', os.path.join(rootname, 'docs/html'), filter=_setTarItemPerms) if options.upload: uploadPackage(tarfilename, options, keep=5, mask='%s-docs-%s*' % (baseName, cfg.VER_MAJOR)) msg('Documentation tarball built at %s' % tarfilename) # # pythonhosted.org can host the wxPython documentation for us, so let's # # use it for the docs associated with the latest release of wxPython. It # # requires that the docs be in a .zip file with an index.html file at the # # top level. To build this we'll just need to do like the above tarball # # code, except add the files from within the docs/html folder so they will # # all be at the top level of the archive. shutil.make_archive can be used # # in this case because we don't need to rewrite the pathnames in the # # archive. # if options.release: # msg("Archiving wxPython Phoenix documentation for pythonhosted.org...") # rootname = "%s-docs-pythonhosted-%s" % (baseName, cfg.VERSION) # zipfilename = "dist/%s.zip" % rootname # # if os.path.exists(zipfilename): # os.remove(zipfilename) # # # with zipfile.ZipFile(zipfilename, 'w', zipfile.ZIP_DEFLATED) as zip: # # pwd2 = pushDir('docs/html') # # zipfilename = shutil.make_archive(base_name=os.path.splitext(zipfilename)[0], # format="zip", # root_dir="docs/html") # zipfilename = os.path.relpath(zipfilename) # zipfilename = zipfilename.replace('\\', '/') # # if options.upload: # uploadPackage(zipfilename, options, keep=5, # mask='%s-docs-pythonhosted-%s*' % (baseName, cfg.VER_MAJOR)) # # msg('Pythonhosted zip file built at %s' % zipfilename) def cmd_sip(options, args): cmdTimer = CommandTimer('sip') cfg = Config() pwd = pushDir(cfg.ROOT_DIR) modules = sorted(glob.glob(opj(cfg.SIPGEN, '_*.sip'))) # move _core the to the front of the list core_file = opj(cfg.SIPGEN, '_core.sip') if core_file in modules: modules.remove(core_file) modules.insert(0, core_file) for src_name in modules: tmpdir = tempfile.mkdtemp() tmpdir = tmpdir.replace('\\', '/') src_name = src_name.replace('\\', '/') base = os.path.basename(os.path.splitext(src_name)[0]) sbf = posixjoin(cfg.ROOT_DIR, cfg.SIPOUT, base) + '.sbf' pycode = base[1:] # remove the leading _ pycode = opj(cfg.ROOT_DIR, cfg.PKGDIR, pycode) + '.py' # Check if any of the included files are newer than the .sbf file # produced by the previous run of sip. If not then we don't need to # run sip again. etg = loadETG(posixjoin('etg', base + '.py')) sipFiles = getSipFiles(etg.INCLUDES) + [opj(cfg.SIPGEN, base+'.sip')] if not newer_group(sipFiles, sbf) and os.path.exists(pycode): continue # Leave it turned off for now. TODO: Experiment with this... # pyi_extract = posixjoin(cfg.PKGDIR, base[1:]) + '.pyi' pyi_extract = None # SIP extracts are used to pull python snippets and put them into the # module's .py file pycode = 'pycode'+base+':'+pycode # Write out a pyproject.toml to configure sip pyproject_toml = textwrap.dedent("""\ [build-system] requires = ["sip >=6.6.2, <7"] build-backend = "sipbuild.api" [project] name = "{base}" [tool.sip.bindings.{base}] docstrings = true release-gil = true exceptions = false tracing = {tracing} protected-is-public = false generate-extracts = [\'{extracts}\'] pep484-pyi = false [tool.sip.project] abi-version = "{abi_version}" sip-files-dir = '{sip_gen_dir}' sip-include-dirs = ['{src_dir}'] sip-module = "wx.siplib" """.format( base=base, abi_version=cfg.SIP_ABI, tracing=str(cfg.SIP_TRACE).lower(), extracts=pycode, src_dir=opj(phoenixDir(), 'src'), sip_gen_dir=opj(phoenixDir(), 'sip', 'gen'), ) ) Path(tmpdir, 'pyproject.toml').write_text(pyproject_toml) sip_pwd = pushDir(tmpdir) cmd = 'sip-build --no-compile' runcmd(cmd) del sip_pwd # Write out a sip build file (no longer done by sip itself) sip_tmp_out_dir = opj(tmpdir, 'build', base) sip_pwd = pushDir(sip_tmp_out_dir) header = min(glob.glob('*.h')) sources = sorted(glob.glob('*.cpp')) del sip_pwd Path(sbf).write_text(f"sources = {' '.join(sources)}\nheaders = {header}\n") classesNeedingClassInfo = { 'sip_corewxTreeCtrl.cpp' : 'wxTreeCtrl', } def processSrc(src, keepHashLines=False): with open(src, 'rt', encoding="utf-8") as f: srcTxt = f.read() if keepHashLines: # Either just fix the pathnames in the #line lines... srcTxt = srcTxt.replace(sip_tmp_out_dir, cfg.SIPOUT) else: # ...or totally remove them by replacing those lines with '' import re srcTxt = re.sub(r'^\s*#line.*\n', '', srcTxt, flags=re.MULTILINE) className = classesNeedingClassInfo.get(os.path.basename(src)) if className: srcTxt = injectClassInfo(className, srcTxt) return srcTxt def injectClassInfo(className, srcTxt): # inject wxClassInfo macros into the sip generated wrapper class lines = srcTxt.splitlines() # find the beginning of the class declaration for idx, line in enumerate(lines): if line.startswith('class sip{}'.format(className)): # next line is '{', insert after that lines[idx+2:idx+2] = ['wxDECLARE_ABSTRACT_CLASS(sip{});'.format(className)] break # find the '};' that terminates the class for idx, line in enumerate(lines[idx+2:], idx+2): if line.startswith('};'): lines[idx+1:idx+1] = ['\nwxIMPLEMENT_ABSTRACT_CLASS(sip{0}, {0});'.format(className)] break # join it back up and return return '\n'.join(lines) # Check each file in tmpdir to see if it is different than the same file # in cfg.SIPOUT. If so then copy the new one to cfg.SIPOUT, otherwise # ignore it. for src in sorted(glob.glob(sip_tmp_out_dir + '/*')): dest = Path(opj(cfg.SIPOUT, os.path.basename(src))) if not dest.exists(): msg('%s is a new file, copying...' % os.path.basename(src)) srcTxt = processSrc(src, options.keep_hash_lines) dest.write_text(srcTxt, encoding="utf-8") continue srcTxt = processSrc(src, options.keep_hash_lines) destTxt = dest.read_text(encoding="utf-8") if srcTxt == destTxt: pass else: msg('%s is changed, copying...' % os.path.basename(src)) dest.write_text(srcTxt, encoding="utf-8") # Remove tmpdir and its contents shutil.rmtree(tmpdir) # Generate sip module code deleteIfExists(cfg.SIPINC) with tempfile.TemporaryDirectory() as tmpdir: cmd = 'sip-module --sdist --abi-version {} --target-dir {} wx.siplib'.format(cfg.SIP_ABI, tmpdir) runcmd(cmd) tf_name = min(glob.glob(tmpdir + '/*.tar*')) tf_dir = os.path.splitext(os.path.splitext(tf_name)[0])[0] with tarfile.open(tf_name) as tf: try: tf.extractall(tmpdir, filter='data') except TypeError: warnings.warn('Falling back to less safe tarfile.extractall') tf.extractall(tmpdir) shutil.move(tf_dir, cfg.SIPINC) # Copy sip's sip.h for distribution with wxPython's header copyFile('sip/siplib/sip.h', 'wx/include/wxPython', verbose=True) def cmd_touch(options, args): cmdTimer = CommandTimer('touch') pwd = pushDir(phoenixDir()) etg = Path('etg') for item in sorted(etg.glob('*.py')): item.touch() cmd_touch_others(options, args) def cmd_test(options, args, tests=None): # Run all tests cmdTimer = CommandTimer('test') pwd = pushDir(phoenixDir()) # --boxed runs each test in a new process (only for posix *&##$#@$^!!) # -n is the number of processes to run in parallel # --timeout will kill the test process if it gets stuck jobs = '-n{}'.format(options.pytest_jobs) if options.pytest_jobs else '' boxed = '--forked' if not isWindows else '' sec = options.pytest_timeout timeout = '--timeout={}'.format(sec) if sec and sec != "0" else '' cmd = '"{}" -m pytest {} {} {} {} {} '.format( PYTHON, '-v' if options.verbose else '', boxed, jobs, timeout, options.extra_pytest) if not tests: # let pytest find all tests in the unittest folder cmd += 'unittests' else: # otherwise, run only the test modules given cmd += ' '.join(tests) runcmd(cmd, fatal=False) def cmd_build(options, args): cmdTimer = CommandTimer('build') cmd_build_wx(options, args) cmd_build_py(options, args) def cmd_build_wx(options, args): cmdTimer = CommandTimer('build_wx') if not isWindows and options.use_syswx: msg("use_syswx option specified, skipping wxWidgets build") return checkCompiler() build_options = ['--wxpython', '--unicode'] if options.jobs: build_options.append('--jobs=%s' % options.jobs) if isWindows: # Windows-specific pre build stuff if not options.no_msedge: getMSWebView2() else: build_options.append('--no_msedge') if options.cairo: build_options.append('--cairo') cairo_root = os.path.join(phoenixDir(), 'packaging', 'msw-cairo') os.environ['CAIRO_ROOT'] = cairo_root if options.jom: build_options.append('--jom') else: # Platform is something other than MSW if options.osx_carbon: options.osx_cocoa = False BUILD_DIR = getBuildDir(options) DESTDIR = options.destdir PREFIX = options.prefix if options.mac_framework and isDarwin: # TODO: Don't hard-code this path PREFIX = "/Library/Frameworks/wx.framework/Versions/%s" % wxversion2 if PREFIX: build_options.append('--prefix=%s' % PREFIX) if not os.path.exists(BUILD_DIR): os.makedirs(BUILD_DIR) if isDarwin: if options.osx_cocoa: build_options.append("--osx_cocoa") if options.mac_arch: build_options.append("--mac_universal_binary=%s" % options.mac_arch) if options.no_config: build_options.append('--no_config') elif not options.force_config: dependencies = [ os.path.join(wxDir(), 'Makefile.in'), os.path.join(wxDir(), 'configure'), os.path.join(wxDir(), 'setup.h.in'), os.path.join(wxDir(), 'version-script.in'), os.path.join(wxDir(), 'wx-config.in'), ] for dep in dependencies: if newer(dep, os.path.join(BUILD_DIR, "Makefile")): break else: build_options.append("--no_config") #if options.install: # build_options.append('--installdir=%s' % DESTDIR) # build_options.append("--install") if options.mac_framework and isDarwin: build_options.append("--mac_framework") if not isDarwin: if options.gtk2: build_options.append('--gtk2') if options.gtk3: build_options.append('--gtk3') # Change to what will be the wxWidgets build folder # (Note, this needs to be after any testing for file/path existence, etc. # because they may be specified as relative paths.) pwd = pushDir(BUILD_DIR) if options.debug or (isWindows and options.both): build_options.append('--debug') if options.extra_make: build_options.append('--extra_make="%s"' % options.extra_make) if not isWindows and not isDarwin and not options.no_magic and not options.use_syswx: # Using $ORIGIN in the rpath will cause the dynamic linker to look # for shared libraries in a folder relative to the loading binary's # location. Here we'll use just $ORIGIN so it should look in the same # folder as the wxPython extension modules. os.environ['LD_RUN_PATH'] = '$ORIGIN' try: # Import and run the wxWidgets build script from buildtools import build_wxwidgets as wxbuild print('wxWidgets build options: ' + str(build_options)) wxbuild.main(wxDir(), build_options) # build again without the --debug flag? if isWindows and options.both: build_options.remove('--debug') print('wxWidgets build options: ' + str(build_options)) wxbuild.main(wxDir(), build_options) except Exception: print("ERROR: failed building wxWidgets") traceback.print_exc() sys.exit(1) if not options.no_allmo: make_allmo() def make_allmo(): # Build the wx message catalogs, but first check that there is a msgfmt # command available if findCmd('msgfmt') and findCmd('make'): locale_pwd = pushDir(posixjoin(wxDir(), 'locale')) print('Building message catalogs in ' + os.getcwd()) runcmd('make allmo') del locale_pwd else: print("WARNING: msgfmt and/or make commands not found, message catalogs not \n" " rebuilt. Please install gettext and associated tools.") def copyWxDlls(options): if options.no_magic or options.use_syswx: return if isWindows: # Copy the wxWidgets DLLs to the wxPython package folder msw = getMSWSettings(options) cfg = Config() ver = wxversion3_nodot if unstable_series else wxversion2_nodot arch = msw.dllArch if PYTHON_ARCH == '64bit' else 'x86' dlls = list() if not options.debug or options.both: dlls += sorted(glob.glob(os.path.join(msw.dllDir, "wx*%su_*.dll" % ver))) if options.relwithdebug: dlls += sorted(glob.glob(os.path.join(msw.dllDir, "wx*%su_*.pdb" % ver))) if options.debug or options.both: dlls += sorted(glob.glob(os.path.join(msw.dllDir, "wx*%sud_*.dll" % ver))) dlls += sorted(glob.glob(os.path.join(msw.dllDir, "wx*%sud_*.pdb" % ver))) # Also copy the cairo DLLs if needed if options.cairo: cairo_root = os.path.join(phoenixDir(), 'packaging', 'msw-cairo') dlls += sorted(glob.glob(os.path.join(cairo_root, arch, 'bin', '*.dll'))) # And the webview2 (MS EDGE) DLL wv2_root = os.path.join(phoenixDir(), 'packaging', 'msw-webview2') dlls += sorted(glob.glob(os.path.join(wv2_root, arch, '*.dll'))) # For Python 3.5 and 3.6 builds we also need to copy some VC14 redist DLLs. # NOTE: Do it for 3.7+ too for now. But when we fully switch over to VS 2017 # this may need to change. See notes in wscript about it. if PYVER in ['3.5', '3.6', '3.7', '3.8', '3.9', '3.10']: redist_dir = os.path.join( phoenixDir(), 'packaging', 'msw-vcredist', arch, 'Microsoft.VC140.CRT', '*.dll') dlls += sorted(glob.glob(redist_dir)) for dll in dlls: copyIfNewer(dll, posixjoin(phoenixDir(), cfg.PKGDIR, os.path.basename(dll)), verbose=True) elif isDarwin: # Copy the wxWidgets dylibs cfg = Config() wxlibdir = os.path.join(getBuildDir(options), "lib") dlls = sorted(glob.glob(wxlibdir + '/*.dylib')) for dll in dlls: copyIfNewer(dll, posixjoin(phoenixDir(), cfg.PKGDIR, os.path.basename(dll)), verbose=True) # Now use install_name_tool to change the extension modules to look # in the same folder for the wx libs, instead of the build dir. Also # change the wx libs the same way. macSetLoaderNames(sorted(glob.glob(opj(phoenixDir(), cfg.PKGDIR, '*.so'))) + sorted(glob.glob(opj(phoenixDir(), cfg.PKGDIR, '*.dylib')))) else: # Not Windows and not OSX. For now that means that we'll assume it's wxGTK. cfg = Config() wxlibdir = os.path.join(getBuildDir(options), "lib") dlls = sorted(glob.glob(wxlibdir + '/libwx_*.so')) dlls += sorted(glob.glob(wxlibdir + '/libwx_*.so.[0-9]*')) for dll in dlls: copyIfNewer(dll, posixjoin(phoenixDir(), cfg.PKGDIR, os.path.basename(dll)), verbose=True) # If all went well the wxPython extensions should already be looking # for the wxlib's in $ORIGIN, so there is nothing else to do here def cmd_build_py(options, args): cmdTimer = CommandTimer('build_py') waf = getWafCmd() checkCompiler() BUILD_DIR = getBuildDir(options) if options.release: os.environ['WXPYTHON_RELEASE'] = 'yes' if not isWindows: WX_CONFIG = os.environ.get('WX_CONFIG', None) if WX_CONFIG is None: WX_CONFIG = posixjoin(BUILD_DIR, 'wx-config') if options.use_syswx: wxcfg = posixjoin(options.prefix, 'bin', 'wx-config') if options.prefix and os.path.exists(wxcfg): WX_CONFIG = wxcfg else: WX_CONFIG = 'wx-config' # hope it is on the PATH if isDarwin: # WAF does a test build as part of the configuration phase, but the # default compiler and flags it uses are not even close to how we'll # configure it later in the configuration process. At a minimum we need # to add the -isysroot for builds on Darwin, and wxWidgets configure is # adding this to the compiler command so we can fetch it there. def _getWxCompiler(flag, compName, flagName): cmd = "%s %s" % (WX_CONFIG, flag) value = os.popen(cmd, 'r').read()[:-1] cmd = shlex.split(value) compiler = cmd[0] flags = cmd[1:] #os.environ[compName] = compiler # don't reset the compiler here, it will be done later os.environ[flagName] = ' '.join(flags) _getWxCompiler('--cc', 'CC', 'CFLAGS') _getWxCompiler('--cxx', 'CXX', 'CXXFLAGS') wafBuildBase = wafBuildDir = getWafBuildBase() if isWindows: wafBuildDir = posixjoin(wafBuildBase, 'release') build_options = list() if options.verbose: build_options.append('--verbose') if options.debug or (isWindows and options.both): build_options.append("--debug") if isWindows: wafBuildDir = posixjoin(wafBuildBase, 'debug') if isDarwin and options.mac_arch: build_options.append("--mac_arch=%s" % options.mac_arch) if isWindows: build_options.append('--msvc_arch=%s' % MSVC_ARCH) if not isWindows: build_options.append('--wx_config=%s' % WX_CONFIG) if options.jobs: build_options.append('--jobs=%s' % options.jobs) if options.relwithdebug: build_options.append('--msvc_relwithdebug') if not isDarwin and not isWindows: if options.gtk2: build_options.append('--gtk2') wafBuildDir = posixjoin(wafBuildBase, 'gtk2') if options.gtk3: build_options.append('--gtk3') wafBuildDir = posixjoin(wafBuildBase, 'gtk3') build_options.append('--python="%s"' % PYTHON) build_options.append('--out=%s' % wafBuildDir) # this needs to be the last option if not isWindows and not isDarwin and not options.no_magic and not options.use_syswx: # Using $ORIGIN in the rpath will cause the dynamic linker to look # for shared libraries in a folder relative to the loading binary's # location. Here we'll use just $ORIGIN so it should look in the same # folder as the wxPython extension modules. os.environ['LD_RUN_PATH'] = '$ORIGIN' # Regenerate the _sysconfigdata module? if options.regenerate_sysconfig: do_regenerate_sysconfig() # Run waf to perform the builds pwd = pushDir(phoenixDir()) cmd = '"%s" %s %s configure build %s' % (PYTHON, waf, ' '.join(build_options), options.extra_waf) def _onWafError(): if options.dump_waf_log: logfilename = opj(wafBuildDir, 'config.log') if not os.path.exists(logfilename): msg('WARNING: waf log "{}" not found!'.format(logfilename)) return msg('*-'*40) msg('WAF config log "{}":'.format(logfilename)) msg(Path(logfilename).read_text()) msg('*-'*40) runcmd(cmd, onError=_onWafError) if isWindows and options.both: build_options.remove('--debug') del build_options[-1] wafBuildDir = posixjoin(wafBuildBase, 'release') build_options.append('--out=%s' % wafBuildDir) cmd = '"%s" %s %s configure build %s' % (PYTHON, waf, ' '.join(build_options), options.extra_waf) runcmd(cmd, onError=_onWafError) copyWxDlls(options) cmd_build_others(options, args) cfg = Config() cfg.build_locale_dir(opj(cfg.PKGDIR, 'locale')) print("\n------------ BUILD FINISHED ------------") print("To use wxPython from the build folder (without installing):") print(" - Set your PYTHONPATH variable to %s." % phoenixDir()) if not isWindows: print(" - You may also need to set your (DY)LD_LIBRARY_PATH to %s/lib,\n or wherever the wxWidgets libs have been installed." % BUILD_DIR) print("") def cmd_build_docker(options, args): # Uses docker images defined by the Dockerfiles under ./docker/build for # building wxPython wheels for a few versions of Linux. # # Requirements: Docker # Internet connection for downloading docker images, if needed # One wxPython source archive in ./dist (created with cmd_sdist) cmdTimer = CommandTimer('build_docker') cfg = Config(noWxConfig=True) cmd = ['inv', 'build-wxpython'] if options.docker_img != 'all': for img in options.docker_img.split(','): cmd.extend(('-i', img)) # Do just the gtk2 builds? if options.gtk2: cmd.extend(['--port', 'gtk2']) # TODO: Instead of the simple options.gtk2 test above, do something like the # following to select both. But currently if gtk2 is selected then # options.gtk3 is explicitly set to False... That needs to be made a little # smarter. # if options.gtk2 and options.gtk3: # cmd.extend(['--port', 'all']) # elif options.gtk2: # cmd.extend(['--port', 'gtk2']) # else: # cmd.extend(['--port', 'gtk3']) # 'none' can be used to skip building and go straight to uploading if options.docker_img != 'none': pwd = pushDir('docker') runcmd(cmd, echoCmd=True) if options.upload: for tag in ['gtk2', 'gtk3']: src = opj(phoenixDir(), 'dist', 'linux', tag) if os.path.isdir(src): uploadTree(src, 'linux', options) def cmd_build_others(options, args): # Build other stuff that may have their own separate build commands instead # of the (ab)normal etg/tweak/generate/sip/compile sequence that the rest of # wxPython uses. So far, it's just the wx.svg package cmdTimer = CommandTimer('build_others') cmd = [PYTHON, 'setup-wxsvg.py', 'build_ext', '--inplace'] if options.verbose: cmd.append('--verbose') runcmd(cmd) def cmd_touch_others(options, args): cmdTimer = CommandTimer('touch_others') pwd = pushDir(phoenixDir()) cfg = Config(noWxConfig=True) pth = Path(opj(cfg.PKGDIR, 'svg')) for item in sorted(pth.glob('*.pyx')): item.touch() def cmd_clean_others(options, args): cmdTimer = CommandTimer('clean_others') pwd = pushDir(phoenixDir()) cfg = Config(noWxConfig=True) files = [] for wc in ['*.pyd', '*.so']: files += sorted(glob.glob(opj(cfg.PKGDIR, 'svg', wc))) delFiles(files) def cmd_install(options, args): cmdTimer = CommandTimer('install') cmd_install_wx(options, args) cmd_install_py(options, args) def cmd_install_wx(options, args): cmdTimer = CommandTimer('install_wx') if isWindows: msg('No wxWidgets installation required on Windows') return if options.use_syswx: msg("use_syswx option specified, skipping wxWidgets install") return if not options.no_magic: msg('Magic is in use, wxWidgets will be installed with wxPython') return # Otherwise, run 'make install' in the wx build folder BUILD_DIR = getBuildDir(options) DESTDIR = '' if not options.destdir else 'DESTDIR=' + options.destdir pwd = pushDir(BUILD_DIR) runcmd("make install %s %s" % (DESTDIR, options.extra_make)) def cmd_install_py(options, args): cmdTimer = CommandTimer('install_py') DESTDIR = '' if not options.destdir else '--root=' + options.destdir VERBOSE = '--verbose' if options.verbose else '' cmd = '"%s" setup.py install --skip-build %s %s %s' % ( PYTHON, DESTDIR, VERBOSE, options.extra_setup) runcmd(cmd) def cmd_build_pdbzip(options, args): if isWindows and options.relwithdebug: cmdTimer = CommandTimer('build_pdbzip') if not os.path.exists('dist'): os.mkdir('dist') cfg = Config() filenames = sorted(glob.glob('./wx/*.pdb')) if not filenames: msg('No PDB files found in ./wx!') return arch = 'win_amd64' if PYTHON_ARCH == '64bit' else 'win32' pyver = 'py{}'.format(PYSHORTVER) zipname = 'dist/{}-pdb-{}-{}-{}.zip'.format(baseName, cfg.VERSION, pyver, arch) from zipfile import ZipFile, ZIP_DEFLATED with ZipFile(zipname, 'w', ZIP_DEFLATED) as zip: for name in filenames: zip.write(name) msg('PDB zip file created at: {}'.format(zipname)) for name in filenames: os.unlink(name) msg('PDB files removed from ./wx') return zipname def _doSimpleSetupCmd(options, args, setupCmd): cmdTimer = CommandTimer(setupCmd) VERBOSE = '--verbose' if options.verbose else '' cmd = '"%s" setup.py %s --skip-build %s %s' % (PYTHON, setupCmd, VERBOSE, options.extra_setup) runcmd(cmd) def cmd_bdist_egg(options, args): pdbzip = cmd_build_pdbzip(options, args) _doSimpleSetupCmd(options, args, 'bdist_egg') cfg = Config() if options.upload: filemask = "dist/%s-%s-*.egg" % (baseName, cfg.VERSION) filenames = sorted(glob.glob(filemask)) assert len(filenames) == 1, "Unknown files found:"+repr(filenames) uploadPackage(filenames[0], options) if pdbzip: uploadPackage(pdbzip, options, keep=24, mask='%s-pdb-%s*' % (baseName, cfg.VER_MAJOR)) def cmd_bdist_wheel(options, args): pdbzip = cmd_build_pdbzip(options, args) _doSimpleSetupCmd(options, args, 'bdist_wheel') cfg = Config() if options.upload: filemask = "dist/%s-%s-*.whl" % (baseName, cfg.VERSION) filenames = sorted(glob.glob(filemask)) print(f'**** filemask: {filemask}') print(f'**** matched: {filenames}') print(f'**** all dist: {sorted(glob.glob("dist/*"))}') assert len(filenames) == 1, "Unknown files found:"+repr(filenames) uploadPackage(filenames[0], options) if pdbzip: uploadPackage(pdbzip, options, keep=24, mask='%s-pdb-%s*' % (baseName, cfg.VER_MAJOR)) def cmd_bdist_wininst(options, args): pdbzip = cmd_build_pdbzip(options, args) _doSimpleSetupCmd(options, args, 'bdist_wininst') cfg = Config() if options.upload: filemask = "dist/%s-%s-*.exe" % (baseName, cfg.VERSION) filenames = sorted(glob.glob(filemask)) assert len(filenames) == 1, "Unknown files found:"+repr(filenames) uploadPackage(filenames[0], options) if pdbzip: uploadPackage(pdbzip, options, keep=24, mask='%s-pdb-%s*' % (baseName, cfg.VER_MAJOR)) def cmd_bdist_msi(options, args): cmd_build_pdbzip(options, args) _doSimpleSetupCmd(options, args, 'bdist_msi') def cmd_egg_info(options, args, egg_base=None): cmdTimer = CommandTimer('egg_info') VERBOSE = '--verbose' if options.verbose else '' BASE = '--egg-base '+egg_base if egg_base is not None else '' cmd = '"%s" setup.py egg_info %s %s' % (PYTHON, VERBOSE, BASE) runcmd(cmd) def cmd_clean_wx(options, args): cmdTimer = CommandTimer('clean_wx') if isWindows: if options.both: options.debug = True msw = getMSWSettings(options) deleteIfExists(opj(msw.dllDir, 'msw'+msw.dll_type)) delFiles(sorted(glob.glob(opj(msw.dllDir, 'wx*%s%s*' % (wxversion2_nodot, msw.dll_type))))) delFiles(sorted(glob.glob(opj(msw.dllDir, 'wx*%s%s*' % (wxversion3_nodot, msw.dll_type))))) if PYTHON_ARCH == '64bit': deleteIfExists(opj(msw.buildDir, 'vc%s_%s_msw%sdll' % (getVisCVersion(), msw.dllArch, msw.dll_type))) else: deleteIfExists(opj(msw.buildDir, 'vc%s_msw%sdll' % (getVisCVersion(), msw.dll_type))) if options.both: options.debug = False options.both = False cmd_clean_wx(options, args) options.both = True else: BUILD_DIR = getBuildDir(options) deleteIfExists(BUILD_DIR) def cmd_clean_py(options, args): cmdTimer = CommandTimer('clean_py') assert os.getcwd() == phoenixDir() if isWindows and options.both: options.debug = True cfg = Config() deleteIfExists(getWafBuildBase()) files = list() for wc in ['*.py', '*.pyc', '*.so', '*.dylib', '*.pyd', '*.pdb', '*.pi', '*.pyi']: files += sorted(glob.glob(opj(cfg.PKGDIR, wc))) if isWindows: msw = getMSWSettings(options) for wc in [ 'wx*' + wxversion2_nodot + msw.dll_type + '*.dll', 'wx*' + wxversion3_nodot + msw.dll_type + '*.dll']: files += sorted(glob.glob(opj(cfg.PKGDIR, wc))) delFiles(files) # Also remove any remaining DLLs just to make sure. This includes the C++ # runtime DLLs, Cairo, etc. # TODO: Check for specific files, not just *.dll if isWindows: files = sorted(glob.glob(opj(cfg.PKGDIR, '*.dll'))) delFiles(files) if options.both: options.debug = False options.both = False cmd_clean_py(options, args) options.both = True cmd_clean_others(options, args) def cmd_clean_sphinx(options, args): cmdTimer = CommandTimer('clean_sphinx') assert os.getcwd() == phoenixDir() sphinxDir = opj(phoenixDir(), 'docs', 'sphinx') globs = [ opj(sphinxDir, '*.txt'), opj(sphinxDir, '*.rst'), opj(sphinxDir, '*.inc'), opj(sphinxDir, '*.pkl'), opj(sphinxDir, '*.lst'), opj(sphinxDir, '_templates/gallery.html'), opj(sphinxDir, 'rest_substitutions/snippets/python/*.py'), opj(sphinxDir, 'rest_substitutions/snippets/cpp/*.cpp'), opj(sphinxDir, '_static/images/inheritance/*.*'), ] for wc in globs: for f in sorted(glob.glob(wc)): os.remove(f) dirs = [opj(sphinxDir, 'build'), opj(phoenixDir(), 'docs/html'), ] for d in dirs: if os.path.exists(d): shutil.rmtree(d) def cmd_clean(options, args): cmd_clean_wx(options, args) cmd_clean_py(options, args) def cmd_clean_docker(options, args): cmdTimer = CommandTimer('clean_docker') assert os.getcwd() == phoenixDir() d = opj(phoenixDir(), 'dist', 'linux') if os.path.exists(d): shutil.rmtree(d) def cmd_clean_all(options, args): cmd_cleanall(options, args) def cmd_cleanall(options, args): # These take care of all the object, lib, shared lib files created by the # compilation part of build cmd_clean_wx(options, args) cmd_clean_py(options, args) # Clean all the intermediate and generated files from the sphinx command cmd_clean_sphinx(options, args) # Now also scrub out all of the SIP and C++ source files that are # generated by the Phoenix ETG system. cmdTimer = CommandTimer('cleanall') assert os.getcwd() == phoenixDir() files = list() for wc in ['sip/cpp/*.h', 'sip/cpp/*.cpp', 'sip/cpp/*.sbf', 'sip/gen/*.sip']: files += sorted(glob.glob(wc)) delFiles(files) cmd_clean_docker(options, args) def cmd_buildall(options, args): # (re)build everything cmd_build_wx(options, args) cmd_dox(options, args) cmd_touch(options, args) cmd_etg(options, args) cmd_sip(options, args) cmd_build_py(options, args) cmd_test(options, args) def cmd_sdist(options, args): # Build a source tarball that includes the generated SIP and CPP files. cmdTimer = CommandTimer('sdist') assert os.getcwd() == phoenixDir() cfg = Config() isGit = os.path.exists('.git') if not isGit: msg("Sorry, I don't know what to do in this source tree, no git workspace found.") return # Make a place to export everything to PDEST = 'build/sdist' if os.path.exists(PDEST): shutil.rmtree(PDEST) os.makedirs(PDEST) TMP = os.path.abspath('build') # and a place to put the final tarball if not os.path.exists('dist'): os.mkdir('dist') # recursively export a git archive of this repo and submodules def _archive_submodules(root, dest): msg('Exporting {}...'.format(root)) if not os.path.exists(dest): os.path.makedirs(dest) pwd = pushDir(root) #runcmd('git archive HEAD | tar -x -C %s' % dest, echoCmd=False) archive = opj(TMP, 'export.tar') runcmd('git archive --format=tar -o %s HEAD' % archive) runcmd('tar -C %s -xf %s' %(dest, archive)) os.unlink(archive) if os.path.exists('.gitmodules'): with open('.gitmodules', 'rt') as fid: for line in fid: line = line.strip() if line.startswith('path = '): sub = line[7:] _archive_submodules(sub, opj(dest, sub)) _archive_submodules('.', os.path.abspath(PDEST)) generateVersionFiles(cfg) # copy .py files that need to go into the root wx package dir for name in ['src/__init__.py', 'src/gizmos.py',]: copyFile(name, cfg.PKGDIR, verbose=True) # copy Phoenix's generated code into the archive tree msg('Copying generated files...') os.makedirs(posixjoin(PDEST, 'sip', 'siplib'), exist_ok=True) for srcdir in ['cpp', 'gen', 'siplib']: destdir = posixjoin(PDEST, 'sip', srcdir) for name in sorted(glob.glob(posixjoin('sip', srcdir, '*'))): if not os.path.isdir(name): copyFile(name, destdir) sip_h_dir = posixjoin(cfg.PKGDIR, 'include', 'wxPython') copyFile(posixjoin(sip_h_dir, 'sip.h'), posixjoin(PDEST, sip_h_dir)) for wc in ['*.py', '*.pi', '*.pyi']: destdir = posixjoin(PDEST, cfg.PKGDIR) for name in sorted(glob.glob(posixjoin(cfg.PKGDIR, wc))): copyFile(name, destdir) copyFile('demo/version.py', posixjoin(PDEST, 'demo')) # Copy the license files from wxWidgets msg('Copying license files...') updateLicenseFiles(cfg) shutil.copytree('license', opj(PDEST, 'license')) copyFile('LICENSE.txt', PDEST) # Copy the locale message catalogs msg('Copying message catalog files...') if not glob.glob(opj(cfg.WXDIR, 'locale', '*.mo')): make_allmo() cfg.build_locale_dir(opj(cfg.PKGDIR, 'locale')) shutil.copytree(opj(cfg.PKGDIR, 'locale'), opj(PDEST, cfg.PKGDIR, 'locale')) # Also add the waf executable, fetching it first if we don't already have it getWafCmd() copyFile('bin/waf-%s' % wafCurrentVersion, os.path.join(PDEST, 'bin')) # And the REV.txt if there is one if os.path.exists('REV.txt'): copyFile('REV.txt', PDEST) # Copy the Sphinx source files in the docs tree, excluding the html and # sphinx/build folders, if present. shutil.rmtree(opj(PDEST, 'docs'), ignore_errors=True) if options.nodoc: os.makedirs(opj(PDEST, 'docs')) Path(PDEST, 'docs', 'README.txt').write_text("The sphinx files and generated docs are not included in this archive.\n") else: shutil.copytree('docs', opj(PDEST, 'docs'), ignore=shutil.ignore_patterns('html', 'build', '__pycache__', 'cpp')) # Add some extra stuff to the root folder cmd_egg_info(options, args, egg_base=PDEST) copyFile(opj(PDEST, '{}.egg-info/PKG-INFO'.format(baseName)), opj(PDEST, 'PKG-INFO')) # (TEMP) apply patch to wxWidgets for building newer libraries msg('Applying patches...') pwd = pushDir(PDEST) runcmd('dos2unix ext/wxWidgets/build/msw/makefile.vc') runcmd('patch -p1 -d ext/wxWidgets -i ../../buildtools/newer_libtiff_pcre.patch') runcmd('unix2dos ext/wxWidgets/build/msw/makefile.vc') del pwd # build the tarball msg('Archiving Phoenix source...') rootname = "%s-%s" % (baseName, cfg.VERSION) tarfilename = posixjoin(phoenixDir(), 'dist', '%s.tar.gz' % rootname) if os.path.exists(tarfilename): os.remove(tarfilename) pwd = pushDir(PDEST) with tarfile.open(name=tarfilename, mode="w:gz") as tarball: for name in sorted(glob.glob('*')): tarball.add(name, os.path.join(rootname, name), filter=_setTarItemPerms) msg('Cleaning up...') del pwd os.chdir(phoenixDir()) shutil.rmtree(PDEST) if options.upload: uploadPackage(tarfilename, options) msg("Source release built at %s" % tarfilename) def cmd_sdist_demo(options, args): # Build a tarball containing the demo and samples cmdTimer = CommandTimer('sdist_demo') assert os.getcwd() == phoenixDir() cfg = Config() isGit = os.path.exists('.git') if not isGit: msg("Sorry, I don't know what to do in this source tree, no git workspace found.") return # make a tree for building up the archive files PDEST = 'build/sdist_demo' if not os.path.exists(PDEST): os.makedirs(PDEST) # and a place to put the final tarball if not os.path.exists('dist'): os.mkdir('dist') # pull out an archive copy of the repo files msg('Exporting Phoenix/demo...') runcmd('git archive HEAD demo | tar -x -C %s' % PDEST, echoCmd=False) msg('Exporting Phoenix/demo/samples...') runcmd('git archive HEAD samples | tar -x -C %s' % PDEST, echoCmd=False) # Add in the README file copyFile('packaging/README-sdist_demo.txt', posixjoin(PDEST, 'README.txt')) copyFile('demo/version.py', posixjoin(PDEST, 'demo')) # build the tarball msg('Archiving Phoenix demo and samples...') rootname = "%s-demo-%s" % (baseName, cfg.VERSION) tarfilename = posixjoin(phoenixDir(), 'dist', '%s.tar.gz' % rootname) if os.path.exists(tarfilename): os.remove(tarfilename) pwd = pushDir(PDEST) with tarfile.open(name=tarfilename, mode="w:gz") as tarball: for name in sorted(glob.glob('*')): tarball.add(name, os.path.join(rootname, name), filter=_setTarItemPerms) msg('Cleaning up...') del pwd shutil.rmtree(PDEST) if options.upload: uploadPackage(tarfilename, options, keep=5, mask='%s-demo-%s*' % (baseName, cfg.VER_MAJOR)) msg("demo and samples tarball built at %s" % tarfilename) def cmd_bdist(options, args): # Build a tarball and/or installer that includes all the files needed at # runtime for the current platform and the current version of Python. cmdTimer = CommandTimer('bdist') assert os.getcwd() == phoenixDir() cmd_egg_info(options, args) cfg = Config() dllext = ".so" wxlibdir = os.path.join(getBuildDir(options), "lib") if sys.platform.startswith('darwin'): dllext = ".dylib" platform = sys.platform if isWindows and PYTHON_ARCH == '64bit': platform = 'win64' rootname = "%s-%s-%s-py%s" % (baseName, cfg.VERSION, platform, PYVER) tarfilename = posixjoin(phoenixDir(), 'dist', '%s.tar.gz' % rootname) if not os.path.exists('dist'): os.makedirs('dist') if os.path.exists(tarfilename): os.remove(tarfilename) msg("Archiving Phoenix binaries to %s..." % tarfilename) with tarfile.open(name=tarfilename, mode="w:gz") as tarball: tarball.add('wx', opj(rootname, 'wx'), filter=lambda info: None if '.svn' in info.name \ or info.name.endswith('.pyc') \ or '__pycache__' in info.name else info) tarball.add(eggInfoName, opj(rootname, eggInfoName)) if not isDarwin and not isWindows and not options.no_magic and not options.use_syswx: # If the DLLs are not already in the wx package folder then go fetch # them now. msg("Archiving wxWidgets shared libraries...") dlls = sorted(glob.glob(os.path.join(wxlibdir, "*%s" % dllext))) for dll in dlls: tarball.add(dll, os.path.join(rootname, 'wx', os.path.basename(dll))) tarball.add('packaging/README-bdist.txt', os.path.join(rootname, 'README.txt')) if options.upload: uploadPackage(tarfilename, options) msg("Binary release built at %s" % tarfilename) def cmd_setrev(options, args): # Grab the current revision number from the version control system # (if possible) and write it to a file we'll use later for # building the package version number cmdTimer = CommandTimer('setrev') assert os.getcwd() == phoenixDir() if options.release: # Ignore this command for release builds, they are not supposed to # include current VCS revision info in the version number. msg('This is a release build, setting REV.txt skipped') else: svnrev = getVcsRev() Path('REV.txt').write_text(svnrev) msg('REV.txt set to "%s"' % svnrev) cfg = Config() cfg.resetVersion() msg('cfg.VERSION: %s' % cfg.VERSION) def cmd_setpythonpath(options, args): cmdTimer = CommandTimer('setpythonpath') assert os.getcwd() == phoenixDir() sys.path.insert(0, phoenixDir()) os.environ['PYTHONPATH'] = os.environ.get('PYTHONPATH', '') + os.pathsep + phoenixDir() #--------------------------------------------------------------------------- if __name__ == '__main__': main(sys.argv[1:])