mirror of
https://github.com/mozilla/gecko-dev.git
synced 2024-11-25 05:41:12 +00:00
725792bd05
The existing terminology had two issues: * `VirtualenvManager` wasn't always associated with an on-disk `virtualenv`: for example, when running in automation, Mach "activates" a `VirtualenvManager`, updating its import scope, but without ever creating an on-disk `virtualenv`. * An upcoming patch splits the `VirtualenvManager` class, pulling "on-disk virtualenv-handling functions" from the project-wide interface for managing Python's import scope. After some good discussion with Ahal, I think we've struck the terminology that handles this distinction well: we'll call the "import scope"-handling part the "site", and we'll continue to call on-disk virtualenvs (and their representative classes) as, well, virtualenvs. Differential Revision: https://phabricator.services.mozilla.com/D130391
1153 lines
34 KiB
Python
Executable File
1153 lines
34 KiB
Python
Executable File
# -*- Mode: python; indent-tabs-mode: nil; tab-width: 40 -*-
|
|
# vim: set filetype=python:
|
|
# This Source Code Form is subject to the terms of the Mozilla Public
|
|
# License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
|
|
|
include("build/moz.configure/init.configure")
|
|
|
|
# Note:
|
|
# - Gecko-specific options and rules should go in toolkit/moz.configure.
|
|
# - Firefox-specific options and rules should go in browser/moz.configure.
|
|
# - Fennec-specific options and rules should go in
|
|
# mobile/android/moz.configure.
|
|
# - Spidermonkey-specific options and rules should go in js/moz.configure.
|
|
# - etc.
|
|
|
|
option(
|
|
"--enable-artifact-builds",
|
|
env="MOZ_ARTIFACT_BUILDS",
|
|
help="Download and use prebuilt binary artifacts.",
|
|
)
|
|
|
|
|
|
@depends("--enable-artifact-builds")
|
|
def artifact_builds(value):
|
|
if value:
|
|
return True
|
|
|
|
|
|
set_config("MOZ_ARTIFACT_BUILDS", artifact_builds)
|
|
|
|
imply_option(
|
|
"--enable-artifact-build-symbols",
|
|
depends(artifact_builds)(lambda v: False if v is None else None),
|
|
reason="--disable-artifact-builds",
|
|
)
|
|
|
|
option(
|
|
"--enable-artifact-build-symbols",
|
|
nargs="?",
|
|
choices=("full",),
|
|
help="Download symbols when artifact builds are enabled.",
|
|
)
|
|
|
|
|
|
@depends("--enable-artifact-build-symbols", "MOZ_AUTOMATION", target)
|
|
def enable_artifact_build_symbols(value, automation, target):
|
|
if len(value):
|
|
return value[0]
|
|
if bool(value):
|
|
if target.os == "Android" and not automation:
|
|
return "full"
|
|
return True
|
|
return None
|
|
|
|
|
|
set_config("MOZ_ARTIFACT_BUILD_SYMBOLS", enable_artifact_build_symbols)
|
|
|
|
|
|
@depends("--enable-artifact-builds")
|
|
def imply_disable_compile_environment(value):
|
|
if value:
|
|
return False
|
|
|
|
|
|
option(
|
|
env="MOZ_COPY_PDBS",
|
|
help="For builds that do not support symbols in the normal fashion,"
|
|
" generate and copy them into the resulting build archive.",
|
|
)
|
|
|
|
set_config("MOZ_COPY_PDBS", depends_if("MOZ_COPY_PDBS")(lambda _: True))
|
|
|
|
imply_option("--enable-compile-environment", imply_disable_compile_environment)
|
|
|
|
option("--disable-compile-environment", help="Disable compiler/library checks")
|
|
|
|
|
|
@depends("--disable-compile-environment")
|
|
def compile_environment(compile_env):
|
|
if compile_env:
|
|
return True
|
|
|
|
|
|
set_config("COMPILE_ENVIRONMENT", compile_environment)
|
|
add_old_configure_assignment("COMPILE_ENVIRONMENT", compile_environment)
|
|
|
|
option("--disable-tests", help="Do not build test libraries & programs")
|
|
|
|
|
|
@depends("--disable-tests")
|
|
def enable_tests(value):
|
|
if value:
|
|
return True
|
|
|
|
|
|
set_config("ENABLE_TESTS", enable_tests)
|
|
set_define("ENABLE_TESTS", enable_tests)
|
|
|
|
|
|
@depends(enable_tests)
|
|
def gtest_has_rtti(value):
|
|
if value:
|
|
return "0"
|
|
|
|
|
|
set_define("GTEST_HAS_RTTI", gtest_has_rtti)
|
|
|
|
|
|
@depends(target, enable_tests)
|
|
def linux_gtest_defines(target, enable_tests):
|
|
if enable_tests and target.os == "Android":
|
|
return namespace(os_linux_android=True, use_own_tr1_tuple=True, has_clone="0")
|
|
|
|
|
|
set_define("GTEST_OS_LINUX_ANDROID", linux_gtest_defines.os_linux_android)
|
|
set_define("GTEST_USE_OWN_TR1_TUPLE", linux_gtest_defines.use_own_tr1_tuple)
|
|
set_define("GTEST_HAS_CLONE", linux_gtest_defines.has_clone)
|
|
|
|
option(
|
|
"--enable-debug",
|
|
nargs="?",
|
|
help="Enable building with developer debug info "
|
|
"(using the given compiler flags).",
|
|
)
|
|
|
|
|
|
@depends("--enable-debug")
|
|
def moz_debug(debug):
|
|
if debug:
|
|
return bool(debug)
|
|
|
|
|
|
set_config("MOZ_DEBUG", moz_debug)
|
|
set_define("MOZ_DEBUG", moz_debug)
|
|
# Override any value MOZ_DEBUG may have from the environment when passing it
|
|
# down to old-configure.
|
|
add_old_configure_assignment("MOZ_DEBUG", depends("--enable-debug")(lambda x: bool(x)))
|
|
|
|
option(
|
|
"--with-debug-label",
|
|
nargs="+",
|
|
help="Debug DEBUG_<value> for each comma-separated value given",
|
|
)
|
|
|
|
|
|
@depends(moz_debug, "--with-debug-label")
|
|
def debug_defines(debug, labels):
|
|
if debug:
|
|
return ["DEBUG"] + ["DEBUG_%s" % label for label in labels]
|
|
return ["NDEBUG", "TRIMMED"]
|
|
|
|
|
|
set_config("MOZ_DEBUG_DEFINES", debug_defines)
|
|
|
|
option(env="MOZ_PGO", help="Build with profile guided optimizations")
|
|
|
|
set_config("MOZ_PGO", depends("MOZ_PGO")(lambda x: bool(x)))
|
|
|
|
|
|
imply_option("--enable-release", mozilla_official)
|
|
imply_option("--enable-release", depends_if("MOZ_AUTOMATION")(lambda x: True))
|
|
|
|
option(
|
|
"--enable-release",
|
|
default=milestone.is_release_or_beta,
|
|
help="{Build|Do not build} with more conservative, release "
|
|
"engineering-oriented options.{ This may slow down builds.|}",
|
|
)
|
|
|
|
|
|
@depends("--enable-release")
|
|
def developer_options(value):
|
|
if not value:
|
|
return True
|
|
|
|
|
|
add_old_configure_assignment("DEVELOPER_OPTIONS", developer_options)
|
|
set_config("DEVELOPER_OPTIONS", developer_options)
|
|
|
|
|
|
# hybrid build handling
|
|
# ==============================================================
|
|
|
|
option(
|
|
"--disable-unified-build",
|
|
help="Enable building modules that are not marked with `REQUIRES_UNIFIED_BUILD` in non unified context",
|
|
)
|
|
|
|
set_config("ENABLE_UNIFIED_BUILD", True, when="--disable-unified-build")
|
|
|
|
|
|
option(
|
|
env="MOZ_FETCHES_DIR",
|
|
nargs=1,
|
|
when="MOZ_AUTOMATION",
|
|
help="Directory containing fetched artifacts",
|
|
)
|
|
|
|
|
|
@depends("MOZ_FETCHES_DIR", when="MOZ_AUTOMATION")
|
|
def moz_fetches_dir(value):
|
|
if value:
|
|
return value[0]
|
|
|
|
|
|
@depends(vcs_checkout_type, milestone.is_nightly, "MOZ_AUTOMATION")
|
|
def bootstrap_default(vcs_checkout_type, is_nightly, automation):
|
|
if automation:
|
|
return False
|
|
# We only enable if building off a VCS checkout of central.
|
|
if is_nightly and vcs_checkout_type:
|
|
return True
|
|
|
|
|
|
option(
|
|
"--enable-bootstrap",
|
|
default=bootstrap_default,
|
|
help="{Automatically bootstrap or update some toolchains|Disable bootstrap or update of toolchains}",
|
|
)
|
|
|
|
|
|
@depends(developer_options, "--enable-bootstrap", moz_fetches_dir)
|
|
def bootstrap_search_path_order(developer_options, bootstrap, moz_fetches_dir):
|
|
if moz_fetches_dir:
|
|
log.debug("Prioritizing MOZ_FETCHES_DIR in toolchain path.")
|
|
return "prepend"
|
|
|
|
if bootstrap:
|
|
log.debug(
|
|
"Prioritizing mozbuild state dir in toolchain paths because "
|
|
"bootstrap mode is enabled."
|
|
)
|
|
return "prepend"
|
|
|
|
if developer_options:
|
|
log.debug(
|
|
"Prioritizing mozbuild state dir in toolchain paths because "
|
|
"you are not building in release mode."
|
|
)
|
|
return "prepend"
|
|
|
|
log.debug(
|
|
"Prioritizing system over mozbuild state dir in "
|
|
"toolchain paths because you are building in "
|
|
"release mode."
|
|
)
|
|
return "append"
|
|
|
|
|
|
toolchains_base_dir = moz_fetches_dir | mozbuild_state_path
|
|
|
|
|
|
@dependable
|
|
@imports("os")
|
|
@imports(_from="os", _import="environ")
|
|
def original_path():
|
|
return environ["PATH"].split(os.pathsep)
|
|
|
|
|
|
@depends(host, when="--enable-bootstrap")
|
|
@imports("os")
|
|
@imports("traceback")
|
|
@imports(_from="mozbuild.toolchains", _import="toolchain_task_definitions")
|
|
@imports(_from="__builtin__", _import="Exception")
|
|
def bootstrap_toolchain_tasks(host):
|
|
prefix = {
|
|
("x86_64", "GNU", "Linux"): "linux64",
|
|
("x86_64", "OSX", "Darwin"): "macosx64",
|
|
("aarch64", "OSX", "Darwin"): "macosx64-aarch64",
|
|
("x86_64", "WINNT", "WINNT"): "win64",
|
|
}.get((host.cpu, host.os, host.kernel))
|
|
try:
|
|
return namespace(prefix=prefix, tasks=toolchain_task_definitions())
|
|
except Exception as e:
|
|
message = traceback.format_exc()
|
|
log.warning(str(e))
|
|
log.debug(message)
|
|
return None
|
|
|
|
|
|
@template
|
|
def bootstrap_path(path, **kwargs):
|
|
when = kwargs.pop("when", None)
|
|
if kwargs:
|
|
configure_error("bootstrap_path only takes `when` as a keyword argument")
|
|
|
|
@depends(
|
|
"--enable-bootstrap",
|
|
toolchains_base_dir,
|
|
bootstrap_toolchain_tasks,
|
|
shell,
|
|
check_build_environment,
|
|
dependable(path),
|
|
when=when,
|
|
)
|
|
@imports("os")
|
|
@imports("subprocess")
|
|
@imports(_from="mozbuild.util", _import="ensureParentDir")
|
|
@imports(_from="__builtin__", _import="open")
|
|
@imports(_from="__builtin__", _import="Exception")
|
|
def bootstrap_path(bootstrap, toolchains_base_dir, tasks, shell, build_env, path):
|
|
path_parts = path.split("/")
|
|
|
|
def try_bootstrap(exists):
|
|
if not tasks:
|
|
return False
|
|
prefixes = [""]
|
|
if tasks.prefix:
|
|
prefixes.insert(0, "{}-".format(tasks.prefix))
|
|
for prefix in prefixes:
|
|
label = "toolchain-{}{}".format(prefix, path_parts[0])
|
|
task = tasks.tasks.get(label)
|
|
if task:
|
|
break
|
|
log.debug("Trying to bootstrap %s", label)
|
|
if not task:
|
|
return False
|
|
task_index = task.optimization.get("index-search")
|
|
if not task_index:
|
|
return False
|
|
log.debug("Resolved %s to %s", label, task_index[0])
|
|
task_index = task_index[0].split(".")[-1]
|
|
artifact = task.attributes["toolchain-artifact"]
|
|
# `mach artifact toolchain` doesn't support authentication for
|
|
# private artifacts.
|
|
if not artifact.startswith("public/"):
|
|
log.debug("Cannot bootstrap %s: not a public artifact", label)
|
|
return False
|
|
index_file = os.path.join(toolchains_base_dir, "indices", path_parts[0])
|
|
try:
|
|
with open(index_file) as fh:
|
|
index = fh.read().strip()
|
|
except Exception:
|
|
index = None
|
|
if index == task_index and exists:
|
|
log.debug("%s is up-to-date", label)
|
|
return True
|
|
log.info(
|
|
"%s bootstrapped toolchain in %s",
|
|
"Updating" if exists else "Installing",
|
|
os.path.join(toolchains_base_dir, path_parts[0]),
|
|
)
|
|
subprocess.run(
|
|
[
|
|
shell,
|
|
os.path.join(build_env.topsrcdir, "mach"),
|
|
"--log-no-times",
|
|
"artifact",
|
|
"toolchain",
|
|
"--from-build",
|
|
label,
|
|
],
|
|
cwd=toolchains_base_dir,
|
|
check=True,
|
|
)
|
|
ensureParentDir(index_file)
|
|
with open(index_file, "w") as fh:
|
|
fh.write(task_index)
|
|
return True
|
|
|
|
path = os.path.join(toolchains_base_dir, *path_parts)
|
|
if bootstrap:
|
|
try:
|
|
if not try_bootstrap(os.path.exists(path)):
|
|
# If there aren't toolchain artifacts to use for this build,
|
|
# don't return a path.
|
|
return None
|
|
except Exception as e:
|
|
log.error("%s", e)
|
|
die("If you can't fix the above, retry with --disable-bootstrap.")
|
|
# We re-test whether the path exists because it may have been created by
|
|
# try_bootstrap. Automation will not have gone through the bootstrap
|
|
# process, but we want to return the path if it exists.
|
|
if os.path.exists(path):
|
|
return path
|
|
|
|
return bootstrap_path
|
|
|
|
|
|
@template
|
|
def bootstrap_search_path(path, paths=original_path, **kwargs):
|
|
@depends(
|
|
bootstrap_path(path, **kwargs),
|
|
bootstrap_search_path_order,
|
|
paths,
|
|
original_path,
|
|
)
|
|
def bootstrap_search_path(path, order, paths, original_path):
|
|
if paths is None:
|
|
paths = original_path
|
|
if not path:
|
|
return paths
|
|
if order == "prepend":
|
|
return [path] + paths
|
|
return paths + [path]
|
|
|
|
return bootstrap_search_path
|
|
|
|
|
|
# The execution model of the configure sandbox doesn't allow for
|
|
# check_prog to use bootstrap_search_path directly because check_prog
|
|
# comes first, so we use a trick to allow it. Uses of check_prog
|
|
# happening before here won't allow bootstrap.
|
|
@template
|
|
def check_prog(*args, **kwargs):
|
|
kwargs["bootstrap_search_path"] = bootstrap_search_path
|
|
return check_prog(*args, **kwargs)
|
|
|
|
|
|
check_prog("WGET", ("wget",), allow_missing=True)
|
|
|
|
|
|
include("build/moz.configure/toolchain.configure", when="--enable-compile-environment")
|
|
|
|
include("build/moz.configure/pkg.configure")
|
|
# Make this assignment here rather than in pkg.configure to avoid
|
|
# requiring this file in unit tests.
|
|
add_old_configure_assignment("PKG_CONFIG", pkg_config)
|
|
|
|
include("build/moz.configure/memory.configure", when="--enable-compile-environment")
|
|
include("build/moz.configure/headers.configure", when="--enable-compile-environment")
|
|
include("build/moz.configure/warnings.configure", when="--enable-compile-environment")
|
|
include("build/moz.configure/flags.configure", when="--enable-compile-environment")
|
|
include("build/moz.configure/lto-pgo.configure", when="--enable-compile-environment")
|
|
# rust.configure is included by js/moz.configure.
|
|
|
|
option("--enable-valgrind", help="Enable Valgrind integration hooks")
|
|
|
|
valgrind_h = check_header("valgrind/valgrind.h", when="--enable-valgrind")
|
|
|
|
|
|
@depends("--enable-valgrind", valgrind_h)
|
|
def check_valgrind(valgrind, valgrind_h):
|
|
if valgrind:
|
|
if not valgrind_h:
|
|
die("--enable-valgrind specified but Valgrind is not installed")
|
|
return True
|
|
|
|
|
|
set_define("MOZ_VALGRIND", check_valgrind)
|
|
set_config("MOZ_VALGRIND", check_valgrind)
|
|
|
|
|
|
@depends(target, host)
|
|
def is_openbsd(target, host):
|
|
return target.kernel == "OpenBSD" or host.kernel == "OpenBSD"
|
|
|
|
|
|
option(
|
|
env="SO_VERSION",
|
|
nargs=1,
|
|
default="1.0",
|
|
when=is_openbsd,
|
|
help="Shared library version for OpenBSD systems",
|
|
)
|
|
|
|
|
|
@depends("SO_VERSION", when=is_openbsd)
|
|
def so_version(value):
|
|
return value
|
|
|
|
|
|
@template
|
|
def library_name_info_template(host_or_target):
|
|
assert host_or_target in {host, target}
|
|
compiler = {
|
|
host: host_c_compiler,
|
|
target: c_compiler,
|
|
}[host_or_target]
|
|
|
|
@depends(host_or_target, compiler, so_version)
|
|
def library_name_info_impl(host_or_target, compiler, so_version):
|
|
if host_or_target.kernel == "WINNT":
|
|
# There aren't artifacts for mingw builds, so it's OK that the
|
|
# results are inaccurate in that case.
|
|
if compiler and compiler.type != "clang-cl":
|
|
return namespace(
|
|
dll=namespace(prefix="", suffix=".dll"),
|
|
lib=namespace(prefix="lib", suffix="a"),
|
|
import_lib=namespace(prefix="lib", suffix="a"),
|
|
obj=namespace(prefix="", suffix="o"),
|
|
)
|
|
|
|
return namespace(
|
|
dll=namespace(prefix="", suffix=".dll"),
|
|
lib=namespace(prefix="", suffix="lib"),
|
|
import_lib=namespace(prefix="", suffix="lib"),
|
|
obj=namespace(prefix="", suffix="obj"),
|
|
)
|
|
|
|
elif host_or_target.kernel == "Darwin":
|
|
return namespace(
|
|
dll=namespace(prefix="lib", suffix=".dylib"),
|
|
lib=namespace(prefix="lib", suffix="a"),
|
|
import_lib=namespace(prefix=None, suffix=""),
|
|
obj=namespace(prefix="", suffix="o"),
|
|
)
|
|
elif so_version:
|
|
so = ".so.%s" % so_version
|
|
else:
|
|
so = ".so"
|
|
|
|
return namespace(
|
|
dll=namespace(prefix="lib", suffix=so),
|
|
lib=namespace(prefix="lib", suffix="a"),
|
|
import_lib=namespace(prefix=None, suffix=""),
|
|
obj=namespace(prefix="", suffix="o"),
|
|
)
|
|
|
|
return library_name_info_impl
|
|
|
|
|
|
host_library_name_info = library_name_info_template(host)
|
|
library_name_info = library_name_info_template(target)
|
|
|
|
set_config("DLL_PREFIX", library_name_info.dll.prefix)
|
|
set_config("DLL_SUFFIX", library_name_info.dll.suffix)
|
|
set_config("HOST_DLL_PREFIX", host_library_name_info.dll.prefix)
|
|
set_config("HOST_DLL_SUFFIX", host_library_name_info.dll.suffix)
|
|
set_config("LIB_PREFIX", library_name_info.lib.prefix)
|
|
set_config("LIB_SUFFIX", library_name_info.lib.suffix)
|
|
set_config("OBJ_SUFFIX", library_name_info.obj.suffix)
|
|
# Lots of compilation tests depend on this variable being present.
|
|
add_old_configure_assignment("OBJ_SUFFIX", library_name_info.obj.suffix)
|
|
set_config("IMPORT_LIB_SUFFIX", library_name_info.import_lib.suffix)
|
|
set_define(
|
|
"MOZ_DLL_PREFIX", depends(library_name_info.dll.prefix)(lambda s: '"%s"' % s)
|
|
)
|
|
set_define(
|
|
"MOZ_DLL_SUFFIX", depends(library_name_info.dll.suffix)(lambda s: '"%s"' % s)
|
|
)
|
|
set_config("WASM_OBJ_SUFFIX", "wasm")
|
|
|
|
# Make `profiling` available to this file even when js/moz.configure
|
|
# doesn't end up included.
|
|
profiling = dependable(False)
|
|
# Same for js_standalone
|
|
js_standalone = dependable(False)
|
|
# Same for fold_libs
|
|
fold_libs = dependable(False)
|
|
|
|
include(include_project_configure)
|
|
|
|
|
|
@depends("--help")
|
|
@imports(_from="mozbuild.backend", _import="backends")
|
|
def build_backends_choices(_):
|
|
return tuple(backends)
|
|
|
|
|
|
@deprecated_option("--enable-build-backend", nargs="+", choices=build_backends_choices)
|
|
def build_backend(backends):
|
|
if backends:
|
|
return tuple("+%s" % b for b in backends)
|
|
|
|
|
|
imply_option("--build-backends", build_backend)
|
|
|
|
|
|
@depends(
|
|
"--enable-artifact-builds",
|
|
"--disable-compile-environment",
|
|
"--enable-build-backend",
|
|
"--enable-project",
|
|
"--enable-application",
|
|
"--help",
|
|
)
|
|
@imports("sys")
|
|
def build_backend_defaults(
|
|
artifact_builds, compile_environment, requested_backends, project, application, _
|
|
):
|
|
if application:
|
|
project = application[0]
|
|
elif project:
|
|
project = project[0]
|
|
|
|
if "Tup" in requested_backends:
|
|
# As a special case, if Tup was requested, do not combine it with any
|
|
# Make based backend by default.
|
|
all_backends = []
|
|
elif artifact_builds:
|
|
all_backends = ["FasterMake+RecursiveMake"]
|
|
else:
|
|
all_backends = ["RecursiveMake", "FasterMake"]
|
|
# Normally, we'd use target.os == 'WINNT', but a dependency on target
|
|
# would require target to depend on --help, as well as host and shell,
|
|
# and this is not a can of worms we can open at the moment.
|
|
if (
|
|
sys.platform == "win32"
|
|
and compile_environment
|
|
and project not in ("mobile/android", "memory", "tools/update-programs")
|
|
):
|
|
all_backends.append("VisualStudio")
|
|
return tuple(all_backends) or None
|
|
|
|
|
|
option(
|
|
"--build-backends",
|
|
nargs="+",
|
|
default=build_backend_defaults,
|
|
choices=build_backends_choices,
|
|
help="Build backends to generate",
|
|
)
|
|
|
|
|
|
@depends("--build-backends")
|
|
def build_backends(backends):
|
|
return backends
|
|
|
|
|
|
set_config("BUILD_BACKENDS", build_backends)
|
|
|
|
|
|
@depends(check_build_environment, build_backends)
|
|
@imports("glob")
|
|
def check_objdir_backend_reuse(build_env, backends):
|
|
# "Make based" might be RecursiveMake or a hybrid backend, so "Make" is
|
|
# intentionally vague for use with the substring match below.
|
|
incompatible_backends = (("Tup", "Make"), ("Make", "Tup"))
|
|
for backend_file in glob.iglob(
|
|
os.path.join(build_env.topobjdir, "backend.*Backend")
|
|
):
|
|
for prev, curr in incompatible_backends:
|
|
if prev in backend_file and any(curr in b for b in backends):
|
|
die(
|
|
"The active objdir, %s, was previously "
|
|
"used to build with a %s based backend. "
|
|
"Change objdirs (by setting MOZ_OBJDIR in "
|
|
"your mozconfig) or clobber to continue.\n",
|
|
build_env.topobjdir,
|
|
prev,
|
|
)
|
|
|
|
|
|
# Determine whether to build the gtest xul. This happens in automation
|
|
# on Android and Desktop platforms with the exception of:
|
|
# - Windows PGO, where linking xul-gtest.dll takes too long;
|
|
# - Android other than x86_64, where gtest is not required.
|
|
@depends(
|
|
"MOZ_PGO",
|
|
build_project,
|
|
target,
|
|
"MOZ_AUTOMATION",
|
|
enable_tests,
|
|
when="--enable-compile-environment",
|
|
)
|
|
def build_gtest(pgo, build_project, target, automation, enable_tests):
|
|
return bool(
|
|
enable_tests
|
|
and automation
|
|
and build_project in ("browser", "comm/mail", "mobile/android")
|
|
and not (
|
|
(pgo and target.os == "WINNT")
|
|
or (target.os == "Android" and target.cpu != "x86_64")
|
|
)
|
|
)
|
|
|
|
|
|
option(
|
|
"--enable-gtest-in-build",
|
|
default=build_gtest,
|
|
help="{Enable|Force disable} building the gtest libxul during the build.",
|
|
when="--enable-compile-environment",
|
|
)
|
|
|
|
set_config("LINK_GTEST_DURING_COMPILE", True, when="--enable-gtest-in-build")
|
|
|
|
# Localization
|
|
# ==============================================================
|
|
option(
|
|
"--enable-ui-locale",
|
|
default="en-US",
|
|
help="Select the user interface locale (default: en-US)",
|
|
)
|
|
|
|
set_config("MOZ_UI_LOCALE", depends("--enable-ui-locale")(lambda x: x))
|
|
|
|
option(
|
|
"--enable-icu4x",
|
|
help="An experiment to use ICU4X instead of ICU4C. See intl/ICU4X.md",
|
|
)
|
|
|
|
set_config("MOZ_ICU4X", True, when="--enable-icu4x")
|
|
|
|
# clang-plugin location
|
|
# ==============================================================
|
|
|
|
|
|
@depends(host_library_name_info, check_build_environment, when="--enable-clang-plugin")
|
|
def clang_plugin_path(library_name_info, build_env):
|
|
topobjdir = build_env.topobjdir
|
|
if topobjdir.endswith("/js/src"):
|
|
topobjdir = topobjdir[:-7]
|
|
return os.path.abspath(
|
|
os.path.join(
|
|
topobjdir,
|
|
"build",
|
|
"clang-plugin",
|
|
"%sclang-plugin%s"
|
|
% (library_name_info.dll.prefix, library_name_info.dll.suffix),
|
|
)
|
|
)
|
|
|
|
|
|
set_config("CLANG_PLUGIN", clang_plugin_path)
|
|
add_old_configure_assignment("CLANG_PLUGIN", clang_plugin_path)
|
|
|
|
|
|
# Awk detection
|
|
# ==============================================================
|
|
awk = check_prog("AWK", ("gawk", "mawk", "nawk", "awk"))
|
|
|
|
# Until the AWK variable is not necessary in old-configure
|
|
|
|
|
|
@depends(awk)
|
|
def awk_for_old_configure(value):
|
|
return value
|
|
|
|
|
|
add_old_configure_assignment("AWK", awk_for_old_configure)
|
|
|
|
|
|
# Perl detection
|
|
# ==============================================================
|
|
perl = check_prog("PERL", ("perl5", "perl"))
|
|
|
|
# Until the PERL variable is not necessary in old-configure
|
|
|
|
|
|
@depends(perl)
|
|
def perl_for_old_configure(value):
|
|
return value
|
|
|
|
|
|
add_old_configure_assignment("PERL", perl_for_old_configure)
|
|
|
|
|
|
@template
|
|
def perl_version_check(min_version):
|
|
@depends(perl)
|
|
@checking("for minimum required perl version >= %s" % min_version)
|
|
def get_perl_version(perl):
|
|
return Version(
|
|
check_cmd_output(
|
|
perl,
|
|
"-e",
|
|
"print $]",
|
|
onerror=lambda: die("Failed to get perl version."),
|
|
)
|
|
)
|
|
|
|
@depends(get_perl_version)
|
|
def check_perl_version(version):
|
|
if version < min_version:
|
|
die("Perl %s or higher is required.", min_version)
|
|
|
|
@depends(perl)
|
|
@checking("for full perl installation")
|
|
@imports("subprocess")
|
|
def has_full_perl_installation(perl):
|
|
ret = subprocess.call([perl, "-e", "use Config; exit(!-d $Config{archlib})"])
|
|
return ret == 0
|
|
|
|
@depends(has_full_perl_installation)
|
|
def require_full_perl_installation(has_full_perl_installation):
|
|
if not has_full_perl_installation:
|
|
die(
|
|
"Cannot find Config.pm or $Config{archlib}. "
|
|
"A full perl installation is required."
|
|
)
|
|
|
|
|
|
perl_version_check("5.006")
|
|
|
|
|
|
# GNU make detection
|
|
# ==============================================================
|
|
option(env="MAKE", nargs=1, help="Path to GNU make")
|
|
|
|
|
|
@depends("MAKE", host)
|
|
def possible_makes(make, host):
|
|
candidates = []
|
|
if host.kernel == "WINNT":
|
|
candidates.append("mingw32-make")
|
|
if make:
|
|
candidates.append(make[0])
|
|
if host.kernel == "WINNT":
|
|
candidates.extend(("mozmake", "make", "gmake"))
|
|
else:
|
|
candidates.extend(("gmake", "make"))
|
|
return candidates
|
|
|
|
|
|
check_prog("GMAKE", possible_makes, bootstrap="mozmake")
|
|
|
|
# watchman detection
|
|
# ==============================================================
|
|
|
|
option(env="WATCHMAN", nargs=1, help="Path to the watchman program")
|
|
|
|
|
|
@depends(host, "WATCHMAN")
|
|
@checking("for watchman", callback=lambda w: w.path if w else "not found")
|
|
def watchman(host, prog):
|
|
# On Windows, `watchman` is only supported on 64-bit hosts.
|
|
if host.os == "WINNT" and host.cpu != "x86_64":
|
|
return
|
|
|
|
if not prog:
|
|
prog = find_program("watchman")
|
|
|
|
if not prog:
|
|
return
|
|
|
|
# `watchman version` will talk to the Watchman daemon service.
|
|
# This can hang due to permissions problems. e.g.
|
|
# https://github.com/facebook/watchman/issues/376. So use
|
|
# `watchman --version` to prevent a class of failures.
|
|
out = check_cmd_output(prog, "--version", onerror=lambda: None)
|
|
if out is None:
|
|
return
|
|
|
|
return namespace(path=prog, version=Version(out.strip()))
|
|
|
|
|
|
@depends_if(watchman)
|
|
@checking("for watchman version")
|
|
def watchman_version(w):
|
|
return w.version
|
|
|
|
|
|
set_config("WATCHMAN", watchman.path)
|
|
|
|
|
|
@depends_all(hg_version, hg_config, watchman)
|
|
@checking("for watchman Mercurial integration")
|
|
@imports("os")
|
|
def watchman_hg(hg_version, hg_config, watchman):
|
|
if hg_version < Version("3.8"):
|
|
return "no (Mercurial 3.8+ required)"
|
|
|
|
ext_enabled = False
|
|
mode_disabled = False
|
|
|
|
for k in ("extensions.fsmonitor", "extensions.hgext.fsmonitor"):
|
|
if k in hg_config and hg_config[k] != "!":
|
|
ext_enabled = True
|
|
|
|
mode_disabled = hg_config.get("fsmonitor.mode") == "off"
|
|
|
|
if not ext_enabled:
|
|
return "no (fsmonitor extension not enabled)"
|
|
if mode_disabled:
|
|
return "no (fsmonitor.mode=off disables fsmonitor)"
|
|
|
|
return True
|
|
|
|
|
|
# Miscellaneous programs
|
|
# ==============================================================
|
|
check_prog("XARGS", ("xargs",))
|
|
|
|
|
|
@depends(target)
|
|
def extra_programs(target):
|
|
if target.kernel == "Darwin":
|
|
return namespace(
|
|
DSYMUTIL=("dsymutil", "llvm-dsymutil"),
|
|
MKFSHFS=("newfs_hfs", "mkfs.hfsplus"),
|
|
HFS_TOOL=("hfsplus",),
|
|
)
|
|
if target.os == "GNU" and target.kernel == "Linux":
|
|
return namespace(RPMBUILD=("rpmbuild",))
|
|
|
|
|
|
check_prog("DSYMUTIL", extra_programs.DSYMUTIL, allow_missing=True)
|
|
check_prog("MKFSHFS", extra_programs.MKFSHFS, allow_missing=True)
|
|
check_prog("HFS_TOOL", extra_programs.HFS_TOOL, allow_missing=True)
|
|
check_prog("RPMBUILD", extra_programs.RPMBUILD, allow_missing=True)
|
|
|
|
|
|
nsis = check_prog(
|
|
"MAKENSISU",
|
|
("makensis",),
|
|
bootstrap="nsis/bin",
|
|
allow_missing=True,
|
|
when=target_is_windows,
|
|
)
|
|
|
|
# Make sure the version of makensis is up to date.
|
|
|
|
|
|
@depends_if(nsis)
|
|
@checking("for NSIS version")
|
|
@imports("re")
|
|
def nsis_version(nsis):
|
|
nsis_min_version = "3.0b1"
|
|
|
|
def onerror():
|
|
return die("Failed to get nsis version.")
|
|
|
|
out = check_cmd_output(nsis, "-version", onerror=onerror)
|
|
|
|
m = re.search(r"(?<=v)[0-9]+\.[0-9]+((a|b|rc)[0-9]+)?", out)
|
|
|
|
if not m:
|
|
raise FatalCheckError("Unknown version of makensis")
|
|
ver = Version(m.group(0))
|
|
|
|
# Versions comparisons don't quite work well with beta versions, so ensure
|
|
# it works for the non-beta version.
|
|
if ver < nsis_min_version and (ver >= "3.0a" or ver < "3"):
|
|
raise FatalCheckError(
|
|
"To build the installer you must have NSIS"
|
|
" version %s or greater in your path" % nsis_min_version
|
|
)
|
|
|
|
return ver
|
|
|
|
|
|
# And that makensis is 32-bit (but only on Windows).
|
|
@depends_if(nsis, when=depends(host)(lambda h: h.kernel == "WINNT"))
|
|
@checking("for 32-bit NSIS")
|
|
def nsis_binary_type(nsis):
|
|
bin_type = windows_binary_type(nsis)
|
|
if bin_type != "win32":
|
|
raise FatalCheckError("%s is not a 32-bit Windows application" % nsis)
|
|
|
|
return "yes"
|
|
|
|
|
|
# And any flags we have to give to makensis
|
|
@depends(host)
|
|
def nsis_flags(host):
|
|
if host.kernel != "WINNT":
|
|
return "-nocd"
|
|
return ""
|
|
|
|
|
|
set_config("MAKENSISU_FLAGS", nsis_flags)
|
|
|
|
check_prog("7Z", ("7z", "7za"), allow_missing=True, when=target_is_windows)
|
|
check_prog("UPX", ("upx",), allow_missing=True, when=target_is_windows)
|
|
|
|
|
|
@depends(host_c_compiler, c_compiler, bindgen_config_paths)
|
|
def llvm_objdump(host_c_compiler, c_compiler, bindgen_config_paths):
|
|
clang = None
|
|
for compiler in (host_c_compiler, c_compiler):
|
|
if compiler and compiler.type == "clang":
|
|
clang = compiler.compiler
|
|
break
|
|
elif compiler and compiler.type == "clang-cl":
|
|
clang = os.path.join(os.path.dirname(compiler.compiler), "clang")
|
|
break
|
|
|
|
if not clang and bindgen_config_paths:
|
|
clang = bindgen_config_paths.clang_path
|
|
llvm_objdump = "llvm-objdump"
|
|
if clang:
|
|
out = check_cmd_output(
|
|
clang, "--print-prog-name=llvm-objdump", onerror=lambda: None
|
|
)
|
|
if out:
|
|
llvm_objdump = out.rstrip()
|
|
return (llvm_objdump,)
|
|
|
|
|
|
llvm_objdump = check_prog(
|
|
"LLVM_OBJDUMP",
|
|
llvm_objdump,
|
|
what="llvm-objdump",
|
|
when="--enable-compile-environment",
|
|
paths=clang_search_path,
|
|
)
|
|
|
|
add_old_configure_assignment("LLVM_OBJDUMP", llvm_objdump)
|
|
|
|
|
|
option("--enable-dtrace", help="Build with dtrace support")
|
|
|
|
dtrace = check_header(
|
|
"sys/sdt.h",
|
|
when="--enable-dtrace",
|
|
onerror=lambda: die("dtrace enabled but sys/sdt.h not found"),
|
|
)
|
|
|
|
set_config("HAVE_DTRACE", True, when=dtrace)
|
|
set_define("INCLUDE_MOZILLA_DTRACE", True, when=dtrace)
|
|
add_old_configure_assignment("enable_dtrace", "yes", when=dtrace)
|
|
|
|
|
|
option("--disable-icf", help="Disable Identical Code Folding")
|
|
|
|
add_old_configure_assignment(
|
|
"MOZ_DISABLE_ICF", "1", when=depends("--enable-icf")(lambda x: not x)
|
|
)
|
|
|
|
|
|
option(
|
|
"--enable-strip",
|
|
when=compile_environment,
|
|
help="Enable stripping of libs & executables",
|
|
)
|
|
|
|
# This should be handled as a `when` once bug 1617793 is fixed.
|
|
|
|
|
|
@depends("--enable-strip", c_compiler, when=compile_environment)
|
|
def enable_strip(strip, c_compiler):
|
|
if strip and c_compiler.type != "clang-cl":
|
|
return True
|
|
|
|
|
|
set_config("ENABLE_STRIP", enable_strip)
|
|
|
|
option(
|
|
"--disable-install-strip",
|
|
when=compile_environment,
|
|
help="Enable stripping of libs & executables when packaging",
|
|
)
|
|
|
|
# This should be handled as a `when` once bug 1617793 is fixed.
|
|
|
|
|
|
@depends("--enable-install-strip", c_compiler, when=compile_environment)
|
|
def enable_install_strip(strip, c_compiler):
|
|
if strip and c_compiler.type != "clang-cl":
|
|
return True
|
|
|
|
|
|
set_config("PKG_STRIP", enable_install_strip)
|
|
|
|
|
|
@depends("--enable-strip", "--enable-install-strip", when=compile_environment)
|
|
def strip(strip, install_strip):
|
|
return strip or install_strip
|
|
|
|
|
|
option(env="STRIP_FLAGS", nargs=1, when=strip, help="Flags for the strip command")
|
|
|
|
|
|
@depends("STRIP_FLAGS", profiling, target, when=strip)
|
|
def strip_flags(flags, profiling, target):
|
|
if flags:
|
|
return flags[0].split()
|
|
if profiling:
|
|
# Only strip debug info and symbols when profiling is enabled, keeping
|
|
# local symbols.
|
|
if target.kernel == "Darwin":
|
|
return ["-S"]
|
|
elif target.os == "Android":
|
|
# The tooling we use with Android supports detached symbols, and the
|
|
# size increase caused by local symbols are too much for mobile. So,
|
|
# don't restrict the amount of stripping with a flag.
|
|
return
|
|
else:
|
|
return ["--strip-debug"]
|
|
# Otherwise strip everything we can, which happens without flags on non-Darwin.
|
|
# On Darwin, it tries to strip things it can't, so we need to limit its scope.
|
|
elif target.kernel == "Darwin":
|
|
return ["-x", "-S"]
|
|
|
|
|
|
set_config("STRIP_FLAGS", strip_flags)
|
|
|
|
|
|
@depends(js_standalone, target)
|
|
def system_zlib_default(js_standalone, target):
|
|
return js_standalone and target.kernel not in ("WINNT", "Darwin")
|
|
|
|
|
|
option(
|
|
"--with-system-zlib",
|
|
nargs="?",
|
|
default=system_zlib_default,
|
|
help="{Use|Do not use} system libz",
|
|
)
|
|
|
|
|
|
@depends("--with-system-zlib")
|
|
def deprecated_system_zlib_path(value):
|
|
if len(value) == 1:
|
|
die(
|
|
"--with-system-zlib=PATH is not supported anymore. Please use "
|
|
"--with-system-zlib and set any necessary pkg-config environment variable."
|
|
)
|
|
|
|
|
|
pkg_check_modules("MOZ_ZLIB", "zlib >= 1.2.3", when="--with-system-zlib")
|
|
|
|
set_config("MOZ_SYSTEM_ZLIB", True, when="--with-system-zlib")
|
|
add_old_configure_assignment("MOZ_SYSTEM_ZLIB", True, when="--with-system-zlib")
|
|
|
|
|
|
# Please do not add configure checks from here on.
|
|
|
|
# Fallthrough to autoconf-based configure
|
|
include("build/moz.configure/old.configure")
|
|
|
|
# JS Subconfigure.
|
|
include("js/sub.configure", when=compile_environment & toolkit)
|
|
|
|
|
|
@depends(check_build_environment, build_project)
|
|
@imports("__sandbox__")
|
|
@imports("glob")
|
|
@imports(_from="os.path", _import="exists")
|
|
def config_status_deps(build_env, build_project):
|
|
|
|
topsrcdir = build_env.topsrcdir
|
|
topobjdir = build_env.topobjdir
|
|
|
|
if not topobjdir.endswith("js/src"):
|
|
extra_deps = [os.path.join(topobjdir, ".mozconfig.json")]
|
|
else:
|
|
# mozconfig changes may impact js configure.
|
|
extra_deps = [os.path.join(topobjdir[:-7], ".mozconfig.json")]
|
|
|
|
confvars = os.path.join(topsrcdir, build_project, "confvars.sh")
|
|
if exists(confvars):
|
|
extra_deps.append(confvars)
|
|
|
|
return (
|
|
list(__sandbox__._all_paths)
|
|
+ extra_deps
|
|
+ [
|
|
os.path.join(topsrcdir, "CLOBBER"),
|
|
os.path.join(topsrcdir, "configure.in"),
|
|
os.path.join(topsrcdir, "js", "src", "configure.in"),
|
|
os.path.join(topsrcdir, "nsprpub", "configure"),
|
|
os.path.join(topsrcdir, "config", "milestone.txt"),
|
|
os.path.join(topsrcdir, "browser", "config", "version.txt"),
|
|
os.path.join(topsrcdir, "browser", "config", "version_display.txt"),
|
|
os.path.join(topsrcdir, "build", "build_virtualenv_packages.txt"),
|
|
os.path.join(topsrcdir, "build", "common_virtualenv_packages.txt"),
|
|
os.path.join(topsrcdir, "build", "mach_virtualenv_packages.txt"),
|
|
os.path.join(topsrcdir, "python", "mach", "mach", "site.py"),
|
|
os.path.join(topsrcdir, "aclocal.m4"),
|
|
os.path.join(topsrcdir, "old-configure.in"),
|
|
os.path.join(topsrcdir, "js", "src", "aclocal.m4"),
|
|
os.path.join(topsrcdir, "js", "src", "old-configure.in"),
|
|
]
|
|
+ glob.glob(os.path.join(topsrcdir, "build", "autoconf", "*.m4"))
|
|
)
|
|
|
|
|
|
set_config("CONFIG_STATUS_DEPS", config_status_deps)
|
|
# Please do not add anything after setting config_dep_paths.
|