| Server IP : 123.56.80.60 / Your IP : 216.73.216.78 Web Server : Apache/2.4.54 (Win32) OpenSSL/1.1.1s PHP/7.4.33 mod_fcgid/2.3.10-dev System : Windows NT iZhx3sob14hnz7Z 10.0 build 14393 (Windows Server 2016) i586 User : SYSTEM ( 0) PHP Version : 7.4.33 Disable Function : NONE MySQL : OFF | cURL : ON | WGET : OFF | Perl : OFF | Python : OFF | Sudo : OFF | Pkexec : OFF Directory : C:/Program Files/python/Lib/site-packages/PyInstaller/building/ |
Upload File : |
#-----------------------------------------------------------------------------
# Copyright (c) 2005-2020, PyInstaller Development Team.
#
# Distributed under the terms of the GNU General Public License (version 2
# or later) with exception for distributing the bootloader.
#
# The full license is in the file COPYING.txt, distributed with this software.
#
# SPDX-License-Identifier: (GPL-2.0-or-later WITH Bootloader-exception)
#-----------------------------------------------------------------------------
from __future__ import print_function
"""
Build packages using spec files.
NOTE: All global variables, classes and imported modules create API
for .spec files.
"""
import glob
import os
import pprint
import shutil
import sys
# Relative imports to PyInstaller modules.
from .. import HOMEPATH, DEFAULT_DISTPATH, DEFAULT_WORKPATH
from .. import compat
from .. import log as logging
from ..utils.misc import absnormpath, compile_py_files
from ..compat import is_py2, is_win, PYDYLIB_NAMES, \
open_file, text_type, unicode_writer
from ..depend import bindepend
from ..depend.analysis import initialize_modgraph
from .api import PYZ, EXE, COLLECT, MERGE
from .datastruct import TOC, Target, Tree, _check_guts_eq
from .osx import BUNDLE
from .toc_conversion import DependencyProcessor
from .utils import _check_guts_toc_mtime, format_binaries_and_datas
from ..depend.utils import create_py3_base_library, scan_code_for_ctypes
from ..archive import pyz_crypto
from ..utils.misc import get_path_to_toplevel_modules, get_unicode_modules, mtime
from ..configure import get_importhooks_dir
if is_win:
from ..utils.win32 import winmanifest
logger = logging.getLogger(__name__)
STRINGTYPE = type('')
TUPLETYPE = type((None,))
rthooks = {}
# place where the loader modules and initialization scripts live
_init_code_path = os.path.join(HOMEPATH, 'PyInstaller', 'loader')
IMPORT_TYPES = ['top-level', 'conditional', 'delayed', 'delayed, conditional',
'optional', 'conditional, optional', 'delayed, optional',
'delayed, conditional, optional']
WARNFILE_HEADER = """\
This file lists modules PyInstaller was not able to find. This does not
necessarily mean this module is required for running you program. Python and
Python 3rd-party packages include a lot of conditional or optional modules. For
example the module 'ntpath' only exists on Windows, whereas the module
'posixpath' only exists on Posix systems.
Types if import:
* top-level: imported at the top-level - look at these first
* conditional: imported within an if-statement
* delayed: imported from within a function
* optional: imported within a try-except-statement
IMPORTANT: Do NOT post this list to the issue-tracker. Use it as a basis for
yourself tracking down the missing module. Thanks!
"""
def _old_api_error(obj_name):
"""
Cause PyInstall to exit when .spec file uses old api.
:param obj_name: Name of the old api that is no longer suppored.
"""
raise SystemExit('%s has been removed in PyInstaller 2.0. '
'Please update your spec-file. See '
'http://www.pyinstaller.org/wiki/MigrateTo2.0 '
'for details' % obj_name)
# TODO find better place for function.
def setupUPXFlags():
f = compat.getenv("UPX", "")
if is_win:
# Binaries built with Visual Studio 7.1 require --strip-loadconf
# or they won't compress. Configure.py makes sure that UPX is new
# enough to support --strip-loadconf.
f = "--strip-loadconf " + f
# Do not compress any icon, so that additional icons in the executable
# can still be externally bound
f = "--compress-icons=0 " + f
f = "--best " + f
compat.setenv("UPX", f)
class Analysis(Target):
"""
Class does analysis of the user's main Python scripts.
An Analysis has five outputs, all TOCs (Table of Contents) accessed as
attributes of the analysis.
scripts
The scripts you gave Analysis as input, with any runtime hook scripts
prepended.
pure
The pure Python modules.
binaries
The extensionmodules and their dependencies. The secondary dependecies
are filtered. On Windows files from C:\\Windows are excluded by default.
On Linux/Unix only system libraries from /lib or /usr/lib are excluded.
datas
Data-file dependencies. These are data-file that are found to be needed
by modules. They can be anything: plugins, font files, images, translations,
etc.
zipfiles
The zipfiles dependencies (usually .egg files).
"""
_old_scripts = set((
absnormpath(os.path.join(HOMEPATH, "support", "_mountzlib.py")),
absnormpath(os.path.join(HOMEPATH, "support", "useUnicode.py")),
absnormpath(os.path.join(HOMEPATH, "support", "useTK.py")),
absnormpath(os.path.join(HOMEPATH, "support", "unpackTK.py")),
absnormpath(os.path.join(HOMEPATH, "support", "removeTK.py")),
))
def __init__(self, scripts, pathex=None, binaries=None, datas=None,
hiddenimports=None, hookspath=None, excludes=None, runtime_hooks=None,
cipher=None, win_no_prefer_redirects=False, win_private_assemblies=False,
noarchive=False):
"""
scripts
A list of scripts specified as file names.
pathex
An optional list of paths to be searched before sys.path.
binaries
An optional list of additional binaries (dlls, etc.) to include.
datas
An optional list of additional data files to include.
hiddenimport
An optional list of additional (hidden) modules to include.
hookspath
An optional list of additional paths to search for hooks.
(hook-modules).
excludes
An optional list of module or package names (their Python names,
not path names) that will be ignored (as though they were not found).
runtime_hooks
An optional list of scripts to use as users' runtime hooks. Specified
as file names.
win_no_prefer_redirects
If True, prefers not to follow version redirects when searching for
Windows SxS Assemblies.
win_private_assemblies
If True, changes all bundled Windows SxS Assemblies into Private
Assemblies to enforce assembly versions.
noarchive
If True, don't place source files in a archive, but keep them as
individual files.
"""
super(Analysis, self).__init__()
from ..config import CONF
self.inputs = []
spec_dir = os.path.dirname(CONF['spec'])
for script in scripts:
# If path is relative, it is relative to the location of .spec file.
if not os.path.isabs(script):
script = os.path.join(spec_dir, script)
if absnormpath(script) in self._old_scripts:
logger.warning('Ignoring obsolete auto-added script %s', script)
continue
# Normalize script path.
script = os.path.normpath(script)
if not os.path.exists(script):
raise ValueError("script '%s' not found" % script)
self.inputs.append(script)
# Django hook requires this variable to find the script manage.py.
CONF['main_script'] = self.inputs[0]
self.pathex = self._extend_pathex(pathex, self.inputs)
# Set global config variable 'pathex' to make it available for
# PyInstaller.utils.hooks and import hooks. Path extensions for module
# search.
CONF['pathex'] = self.pathex
# Extend sys.path so PyInstaller could find all necessary modules.
logger.info('Extending PYTHONPATH with paths\n' + pprint.pformat(self.pathex))
sys.path.extend(self.pathex)
# Set global variable to hold assembly binding redirects
CONF['binding_redirects'] = []
self.hiddenimports = hiddenimports or []
# Include modules detected when parsing options, like 'codecs' and encodings.
self.hiddenimports.extend(CONF['hiddenimports'])
self.hookspath = hookspath
# Custom runtime hook files that should be included and started before
# any existing PyInstaller runtime hooks.
self.custom_runtime_hooks = runtime_hooks or []
if cipher:
logger.info('Will encrypt Python bytecode with key: %s', cipher.key)
# Create a Python module which contains the decryption key which will
# be used at runtime by pyi_crypto.PyiBlockCipher.
pyi_crypto_key_path = os.path.join(CONF['workpath'], 'pyimod00_crypto_key.py')
with open_file(pyi_crypto_key_path, 'w', encoding='utf-8') as f:
f.write(text_type('# -*- coding: utf-8 -*-\n'
'key = %r\n' % cipher.key))
logger.info('Adding dependencies on pyi_crypto.py module')
self.hiddenimports.append(pyz_crypto.get_crypto_hiddenimports())
self.excludes = excludes or []
self.scripts = TOC()
self.pure = TOC()
self.binaries = TOC()
self.zipfiles = TOC()
self.zipped_data = TOC()
self.datas = TOC()
self.dependencies = TOC()
self.binding_redirects = CONF['binding_redirects'] = []
self.win_no_prefer_redirects = win_no_prefer_redirects
self.win_private_assemblies = win_private_assemblies
self._python_version = sys.version
self.noarchive = noarchive
self.__postinit__()
# TODO create function to convert datas/binaries from 'hook format' to TOC.
# Initialise 'binaries' and 'datas' with lists specified in .spec file.
if binaries:
logger.info("Appending 'binaries' from .spec")
for name, pth in format_binaries_and_datas(binaries, workingdir=spec_dir):
self.binaries.append((name, pth, 'BINARY'))
if datas:
logger.info("Appending 'datas' from .spec")
for name, pth in format_binaries_and_datas(datas, workingdir=spec_dir):
self.datas.append((name, pth, 'DATA'))
_GUTS = (# input parameters
('inputs', _check_guts_eq), # parameter `scripts`
('pathex', _check_guts_eq),
('hiddenimports', _check_guts_eq),
('hookspath', _check_guts_eq),
('excludes', _check_guts_eq),
('custom_runtime_hooks', _check_guts_eq),
('win_no_prefer_redirects', _check_guts_eq),
('win_private_assemblies', _check_guts_eq),
#'cipher': no need to check as it is implied by an
# additional hidden import
#calculated/analysed values
('_python_version', _check_guts_eq),
('scripts', _check_guts_toc_mtime),
('pure', lambda *args: _check_guts_toc_mtime(*args, **{'pyc': 1})),
('binaries', _check_guts_toc_mtime),
('zipfiles', _check_guts_toc_mtime),
('zipped_data', None), # TODO check this, too
('datas', _check_guts_toc_mtime),
# TODO: Need to add "dependencies"?
# cached binding redirects - loaded into CONF for PYZ/COLLECT to find.
('binding_redirects', None),
)
def _extend_pathex(self, spec_pathex, scripts):
"""
Normalize additional paths where PyInstaller will look for modules and
add paths with scripts to the list of paths.
:param spec_pathex: Additional paths defined defined in .spec file.
:param scripts: Scripts to create executable from.
:return: list of updated paths
"""
# Based on main supplied script - add top-level modules directory to PYTHONPATH.
# Sometimes the main app script is not top-level module but submodule like 'mymodule.mainscript.py'.
# In that case PyInstaller will not be able find modules in the directory containing 'mymodule'.
# Add this directory to PYTHONPATH so PyInstaller could find it.
pathex = []
# Add scripts paths first.
for script in scripts:
logger.debug('script: %s' % script)
script_toplevel_dir = get_path_to_toplevel_modules(script)
if script_toplevel_dir:
pathex.append(script_toplevel_dir)
# Append paths from .spec.
if spec_pathex is not None:
pathex.extend(spec_pathex)
# Normalize paths in pathex and make them absolute.
return [absnormpath(p) for p in pathex]
def _check_guts(self, data, last_build):
if Target._check_guts(self, data, last_build):
return True
for fnm in self.inputs:
if mtime(fnm) > last_build:
logger.info("Building because %s changed", fnm)
return True
# Now we know that none of the input parameters and none of
# the input files has changed. So take the values calculated
# resp. analysed in the last run and store them in `self`.
self.scripts = TOC(data['scripts'])
self.pure = TOC(data['pure'])
self.binaries = TOC(data['binaries'])
self.zipfiles = TOC(data['zipfiles'])
self.zipped_data = TOC(data['zipped_data'])
self.datas = TOC(data['datas'])
# Store previously found binding redirects in CONF for later use by PKG/COLLECT
from ..config import CONF
self.binding_redirects = CONF['binding_redirects'] = data['binding_redirects']
return False
def assemble(self):
"""
This method is the MAIN method for finding all necessary files to be bundled.
"""
from ..config import CONF
for m in self.excludes:
logger.debug("Excluding module '%s'" % m)
self.graph = initialize_modgraph(
excludes=self.excludes, user_hook_dirs=self.hookspath)
# TODO Find a better place where to put 'base_library.zip' and when to created it.
# For Python 3 it is necessary to create file 'base_library.zip'
# containing core Python modules. In Python 3 some built-in modules
# are written in pure Python. base_library.zip is a way how to have
# those modules as "built-in".
if not is_py2:
libzip_filename = os.path.join(CONF['workpath'], 'base_library.zip')
create_py3_base_library(libzip_filename, graph=self.graph)
# Bundle base_library.zip as data file.
# Data format of TOC item: ('relative_path_in_dist_dir', 'absolute_path_on_disk', 'DATA')
self.datas.append((os.path.basename(libzip_filename), libzip_filename, 'DATA'))
# Expand sys.path of module graph.
# The attribute is the set of paths to use for imports: sys.path,
# plus our loader, plus other paths from e.g. --path option).
self.graph.path = self.pathex + self.graph.path
self.graph.set_setuptools_nspackages()
logger.info("running Analysis %s", self.tocbasename)
# Get paths to Python and, in Windows, the manifest.
python = sys.executable
if not is_win:
# Linux/MacOS: get a real, non-link path to the running Python executable.
while os.path.islink(python):
python = os.path.join(os.path.dirname(python), os.readlink(python))
depmanifest = None
else:
# Windows: Create a manifest to embed into built .exe, containing the same
# dependencies as python.exe.
depmanifest = winmanifest.Manifest(type_="win32", name=CONF['specnm'],
processorArchitecture=winmanifest.processor_architecture(),
version=(1, 0, 0, 0))
depmanifest.filename = os.path.join(CONF['workpath'],
CONF['specnm'] + ".exe.manifest")
# We record "binaries" separately from the modulegraph, as there
# is no way to record those dependencies in the graph. These include
# the python executable and any binaries added by hooks later.
# "binaries" are not the same as "extensions" which are .so or .dylib
# that are found and recorded as extension nodes in the graph.
# Reset seen variable before running bindepend. We use bindepend only for
# the python executable.
bindepend.seen.clear()
# Add binary and assembly dependencies of Python.exe.
# This also ensures that its assembly depencies under Windows get added to the
# built .exe's manifest. Python 2.7 extension modules have no assembly
# dependencies, and rely on the app-global dependencies set by the .exe.
self.binaries.extend(bindepend.Dependencies([('', python, '')],
manifest=depmanifest,
redirects=self.binding_redirects)[1:])
if is_win:
depmanifest.writeprettyxml()
### Module graph.
#
# Construct the module graph of import relationships between modules
# required by this user's application. For each entry point (top-level
# user-defined Python script), all imports originating from this entry
# point are recursively parsed into a subgraph of the module graph. This
# subgraph is then connected to this graph's root node, ensuring
# imported module nodes will be reachable from the root node -- which is
# is (arbitrarily) chosen to be the first entry point's node.
# List to hold graph nodes of scripts and runtime hooks in use order.
priority_scripts = []
# Assume that if the script does not exist, Modulegraph will raise error.
# Save the graph nodes of each in sequence.
for script in self.inputs:
logger.info("Analyzing %s", script)
priority_scripts.append(self.graph.run_script(script))
# Analyze the script's hidden imports (named on the command line)
self.graph.add_hiddenimports(self.hiddenimports)
### Post-graph hooks.
self.graph.process_post_graph_hooks()
# Update 'binaries' TOC and 'datas' TOC.
deps_proc = DependencyProcessor(self.graph,
self.graph._additional_files_cache)
self.binaries.extend(deps_proc.make_binaries_toc())
self.datas.extend(deps_proc.make_datas_toc())
self.zipped_data.extend(deps_proc.make_zipped_data_toc())
# Note: zipped eggs are collected below
### Look for dlls that are imported by Python 'ctypes' module.
# First get code objects of all modules that import 'ctypes'.
logger.info('Looking for ctypes DLLs')
ctypes_code_objs = self.graph.get_co_using_ctypes() # dict like: {'module1': code_obj, 'module2': code_obj}
for name, co in ctypes_code_objs.items():
# Get dlls that might be needed by ctypes.
logger.debug('Scanning %s for shared libraries or dlls', name)
ctypes_binaries = scan_code_for_ctypes(co)
self.binaries.extend(set(ctypes_binaries))
# Analyze run-time hooks.
# Run-time hooks has to be executed before user scripts. Add them
# to the beginning of 'priority_scripts'.
priority_scripts = self.graph.analyze_runtime_hooks(self.custom_runtime_hooks) + priority_scripts
# 'priority_scripts' is now a list of the graph nodes of custom runtime
# hooks, then regular runtime hooks, then the PyI loader scripts.
# Further on, we will make sure they end up at the front of self.scripts
### Extract the nodes of the graph as TOCs for further processing.
# Initialize the scripts list with priority scripts in the proper order.
self.scripts = self.graph.nodes_to_toc(priority_scripts)
# Extend the binaries list with all the Extensions modulegraph has found.
self.binaries = self.graph.make_binaries_toc(self.binaries)
# Fill the "pure" list with pure Python modules.
assert len(self.pure) == 0
self.pure = self.graph.make_pure_toc()
# And get references to module code objects constructed by ModuleGraph
# to avoid writing .pyc/pyo files to hdd.
self.pure._code_cache = self.graph.get_code_objects()
# Add remaining binary dependencies - analyze Python C-extensions and what
# DLLs they depend on.
logger.info('Looking for dynamic libraries')
self.binaries.extend(bindepend.Dependencies(self.binaries,
redirects=self.binding_redirects))
### Include zipped Python eggs.
logger.info('Looking for eggs')
self.zipfiles.extend(deps_proc.make_zipfiles_toc())
# Verify that Python dynamic library can be found.
# Without dynamic Python library PyInstaller cannot continue.
self._check_python_library(self.binaries)
if is_win:
# Remove duplicate redirects
self.binding_redirects[:] = list(set(self.binding_redirects))
logger.info("Found binding redirects: \n%s", self.binding_redirects)
# Place Python source in data files for the noarchive case.
if self.noarchive:
# Create a new TOC of ``(dest path for .pyc, source for .py, type)``.
new_toc = TOC()
for name, path, typecode in self.pure:
assert typecode == 'PYMODULE'
# Transform a python module name into a file name.
name = name.replace('.', os.sep)
# Special case: modules have an implied filename to add.
if os.path.splitext(os.path.basename(path))[0] == '__init__':
name += os.sep + '__init__'
# Append the extension for the compiled result.
name += '.py' + ('o' if sys.flags.optimize else 'c')
new_toc.append((name, path, typecode))
# Put the result of byte-compiling this TOC in datas. Mark all entries as data.
for name, path, typecode in compile_py_files(new_toc, CONF['workpath']):
self.datas.append((name, path, 'DATA'))
# Store no source in the archive.
self.pure = TOC()
# Write warnings about missing modules.
self._write_warnings()
# Write debug information about hte graph
self._write_graph_debug()
def _write_warnings(self):
"""
Write warnings about missing modules. Get them from the graph
and use the graph to figure out who tried to import them.
"""
def dependency_description(name, depInfo):
if not depInfo or depInfo == 'direct':
imptype = 0
else:
imptype = (depInfo.conditional
+ 2 * depInfo.function
+ 4 * depInfo.tryexcept)
return '%s (%s)' % (name, IMPORT_TYPES[imptype])
from ..config import CONF
miss_toc = self.graph.make_missing_toc()
with open_file(CONF['warnfile'], 'w', encoding='utf-8') as wf:
wf_unicode = unicode_writer(wf)
wf_unicode.write(WARNFILE_HEADER)
for (n, p, status) in miss_toc:
importers = self.graph.get_importers(n)
print(status, 'module named', n, '- imported by',
', '.join(dependency_description(name, data)
for name, data in importers),
file=wf_unicode)
logger.info("Warnings written to %s", CONF['warnfile'])
def _write_graph_debug(self):
"""Write a xref (in html) and with `--log-level DEBUG` a dot-drawing
of the graph.
"""
from ..config import CONF
with open_file(CONF['xref-file'], 'w', encoding='utf-8') as fh:
self.graph.create_xref(unicode_writer(fh))
logger.info("Graph cross-reference written to %s", CONF['xref-file'])
if logger.getEffectiveLevel() > logging.DEBUG:
return
# The `DOT language's <https://www.graphviz.org/doc/info/lang.html>`_
# default character encoding (see the end of the linked page) is UTF-8.
with open_file(CONF['dot-file'], 'w', encoding='utf-8') as fh:
self.graph.graphreport(unicode_writer(fh))
logger.info("Graph drawing written to %s", CONF['dot-file'])
def _check_python_library(self, binaries):
"""
Verify presence of the Python dynamic library in the binary dependencies.
Python library is an essential piece that has to be always included.
"""
# First check that libpython is in resolved binary dependencies.
for (nm, filename, typ) in binaries:
if typ == 'BINARY' and nm in PYDYLIB_NAMES:
# Just print its filename and return.
logger.info('Using Python library %s', filename)
# Checking was successful - end of function.
return
# Python lib not in dependencies - try to find it.
logger.info('Python library not in binary dependencies. Doing additional searching...')
python_lib = bindepend.get_python_library_path()
logger.debug('Adding Python library to binary dependencies')
binaries.append((os.path.basename(python_lib), python_lib, 'BINARY'))
logger.info('Using Python library %s', python_lib)
class ExecutableBuilder(object):
"""
Class that constructs the executable.
"""
# TODO wrap the 'main' and 'build' function into this class.
def build(spec, distpath, workpath, clean_build):
"""
Build the executable according to the created SPEC file.
"""
from ..config import CONF
# For combatibility with Python < 2.7.9 we can not use `lambda`,
# but need to declare _old_api_error as beeing global, see issue #1408
def TkPKG(*args, **kwargs):
global _old_api_error
_old_api_error('TkPKG')
def TkTree(*args, **kwargs):
global _old_api_error
_old_api_error('TkTree')
# Ensure starting tilde and environment variables get expanded in distpath / workpath.
# '~/path/abc', '${env_var_name}/path/abc/def'
distpath = compat.expand_path(distpath)
workpath = compat.expand_path(workpath)
CONF['spec'] = compat.expand_path(spec)
CONF['specpath'], CONF['specnm'] = os.path.split(spec)
CONF['specnm'] = os.path.splitext(CONF['specnm'])[0]
# Add 'specname' to workpath and distpath if they point to PyInstaller homepath.
if os.path.dirname(distpath) == HOMEPATH:
distpath = os.path.join(HOMEPATH, CONF['specnm'], os.path.basename(distpath))
CONF['distpath'] = distpath
if os.path.dirname(workpath) == HOMEPATH:
workpath = os.path.join(HOMEPATH, CONF['specnm'], os.path.basename(workpath), CONF['specnm'])
else:
workpath = os.path.join(workpath, CONF['specnm'])
CONF['warnfile'] = os.path.join(workpath, 'warn-%s.txt' % CONF['specnm'])
CONF['dot-file'] = os.path.join(workpath, 'graph-%s.dot' % CONF['specnm'])
CONF['xref-file'] = os.path.join(workpath, 'xref-%s.html' % CONF['specnm'])
# Clean PyInstaller cache (CONF['cachedir']) and temporary files (workpath)
# to be able start a clean build.
if clean_build:
logger.info('Removing temporary files and cleaning cache in %s', CONF['cachedir'])
for pth in (CONF['cachedir'], workpath):
if os.path.exists(pth):
# Remove all files in 'pth'.
for f in glob.glob(pth + '/*'):
# Remove dirs recursively.
if os.path.isdir(f):
shutil.rmtree(f)
else:
os.remove(f)
# Create DISTPATH and workpath if they does not exist.
for pth in (CONF['distpath'], workpath):
if not os.path.exists(pth):
os.makedirs(pth)
# Construct NAMESPACE for running the Python code from .SPEC file.
# NOTE: Passing NAMESPACE allows to avoid having global variables in this
# module and makes isolated environment for running tests.
# NOTE: Defining NAMESPACE allows to map any class to a apecific name for .SPEC.
# FIXME: Some symbols might be missing. Add them if there are some failures.
# TODO: What from this .spec API is deprecated and could be removed?
spec_namespace = {
# Set of global variables that can be used while processing .spec file.
# Some of them act as configuration options.
'DISTPATH': CONF['distpath'],
'HOMEPATH': HOMEPATH,
'SPEC': CONF['spec'],
'specnm': CONF['specnm'],
'SPECPATH': CONF['specpath'],
'WARNFILE': CONF['warnfile'],
'workpath': workpath,
# PyInstaller classes for .spec.
'TOC': TOC,
'Analysis': Analysis,
'BUNDLE': BUNDLE,
'COLLECT': COLLECT,
'EXE': EXE,
'MERGE': MERGE,
'PYZ': PYZ,
'Tree': Tree,
# Old classes for .spec - raise Exception for user.
'TkPKG': TkPKG,
'TkTree': TkTree,
# Python modules available for .spec.
'os': os,
'pyi_crypto': pyz_crypto,
}
# Set up module PyInstaller.config for passing some arguments to 'exec'
# function.
from ..config import CONF
CONF['workpath'] = workpath
# Execute the specfile. Read it as a binary file...
with open(spec, 'rU' if is_py2 else 'rb') as f:
# ... then let Python determine the encoding, since ``compile`` accepts
# byte strings.
code = compile(f.read(), spec, 'exec')
exec(code, spec_namespace)
def __add_options(parser):
parser.add_argument("--distpath", metavar="DIR",
default=DEFAULT_DISTPATH,
help=('Where to put the bundled app (default: %s)' %
os.path.join(os.curdir, 'dist')))
parser.add_argument('--workpath', default=DEFAULT_WORKPATH,
help=('Where to put all the temporary work files, '
'.log, .pyz and etc. (default: %s)' %
os.path.join(os.curdir, 'build')))
parser.add_argument('-y', '--noconfirm',
action="store_true", default=False,
help='Replace output directory (default: %s) without '
'asking for confirmation' % os.path.join('SPECPATH', 'dist', 'SPECNAME'))
parser.add_argument('--upx-dir', default=None,
help='Path to UPX utility (default: search the execution path)')
parser.add_argument("-a", "--ascii", action="store_true",
help="Do not include unicode encoding support "
"(default: included if available)")
parser.add_argument('--clean', dest='clean_build', action='store_true',
default=False,
help='Clean PyInstaller cache and remove temporary '
'files before building.')
def main(pyi_config, specfile, noconfirm, ascii=False, **kw):
from ..config import CONF
CONF['noconfirm'] = noconfirm
# Some modules are included if they are detected at build-time or
# if a command-line argument is specified. (e.g. --ascii)
if CONF.get('hiddenimports') is None:
CONF['hiddenimports'] = []
# Test unicode support.
if not ascii:
CONF['hiddenimports'].extend(get_unicode_modules())
# FIXME: this should be a global import, but can't due to recursive imports
# If configuration dict is supplied - skip configuration step.
if pyi_config is None:
import PyInstaller.configure as configure
CONF.update(configure.get_config(kw.get('upx_dir')))
else:
CONF.update(pyi_config)
if CONF['hasUPX']:
setupUPXFlags()
CONF['ui_admin'] = kw.get('ui_admin', False)
CONF['ui_access'] = kw.get('ui_uiaccess', False)
build(specfile, kw.get('distpath'), kw.get('workpath'), kw.get('clean_build'))