2024-09-24 01:17:42 +00:00
|
|
|
###
|
|
|
|
# decomp-toolkit project generator
|
|
|
|
# Generates build.ninja and objdiff.json.
|
|
|
|
#
|
|
|
|
# This generator is intentionally project-agnostic
|
|
|
|
# and shared between multiple projects. Any configuration
|
|
|
|
# specific to a project should be added to `configure.py`.
|
|
|
|
#
|
|
|
|
# If changes are made, please submit a PR to
|
|
|
|
# https://github.com/encounter/dtk-template
|
|
|
|
###
|
|
|
|
|
|
|
|
import io
|
|
|
|
import json
|
|
|
|
import math
|
|
|
|
import os
|
|
|
|
import platform
|
|
|
|
import sys
|
|
|
|
from pathlib import Path
|
2024-10-14 16:58:15 +00:00
|
|
|
from typing import IO, Any, Dict, Iterable, List, Optional, Set, Tuple, Union, cast
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
from . import ninja_syntax
|
|
|
|
from .ninja_syntax import serialize_path
|
|
|
|
|
|
|
|
if sys.platform == "cygwin":
|
|
|
|
sys.exit(
|
|
|
|
f"Cygwin/MSYS2 is not supported."
|
|
|
|
f"\nPlease use native Windows Python instead."
|
|
|
|
f"\n(Current path: {sys.executable})"
|
|
|
|
)
|
|
|
|
|
|
|
|
Library = Dict[str, Any]
|
|
|
|
|
|
|
|
|
|
|
|
class Object:
|
|
|
|
def __init__(self, completed: bool, name: str, **options: Any) -> None:
|
|
|
|
self.name = name
|
|
|
|
self.completed = completed
|
|
|
|
self.options: Dict[str, Any] = {
|
|
|
|
"add_to_all": None,
|
|
|
|
"asflags": None,
|
|
|
|
"asm_dir": None,
|
|
|
|
"cflags": None,
|
|
|
|
"extra_asflags": None,
|
|
|
|
"extra_cflags": None,
|
|
|
|
"host": None,
|
|
|
|
"lib": None,
|
|
|
|
"mw_version": None,
|
|
|
|
"progress_category": None,
|
2024-11-21 01:58:20 +00:00
|
|
|
"scratch_preset_id": None,
|
2024-09-24 01:17:42 +00:00
|
|
|
"shift_jis": None,
|
|
|
|
"source": name,
|
|
|
|
"src_dir": None,
|
|
|
|
}
|
|
|
|
self.options.update(options)
|
|
|
|
|
|
|
|
# Internal
|
|
|
|
self.src_path: Optional[Path] = None
|
|
|
|
self.asm_path: Optional[Path] = None
|
|
|
|
self.src_obj_path: Optional[Path] = None
|
|
|
|
self.asm_obj_path: Optional[Path] = None
|
|
|
|
self.host_obj_path: Optional[Path] = None
|
|
|
|
self.ctx_path: Optional[Path] = None
|
|
|
|
|
|
|
|
def resolve(self, config: "ProjectConfig", lib: Library) -> "Object":
|
|
|
|
# Use object options, then library options
|
|
|
|
obj = Object(self.completed, self.name, **lib)
|
|
|
|
for key, value in self.options.items():
|
|
|
|
if value is not None or key not in obj.options:
|
|
|
|
obj.options[key] = value
|
|
|
|
|
|
|
|
# Use default options from config
|
|
|
|
def set_default(key: str, value: Any) -> None:
|
|
|
|
if obj.options[key] is None:
|
|
|
|
obj.options[key] = value
|
|
|
|
|
|
|
|
set_default("add_to_all", True)
|
|
|
|
set_default("asflags", config.asflags)
|
|
|
|
set_default("asm_dir", config.asm_dir)
|
|
|
|
set_default("host", False)
|
|
|
|
set_default("mw_version", config.linker_version)
|
2024-11-21 01:58:20 +00:00
|
|
|
set_default("scratch_preset_id", config.scratch_preset_id)
|
2024-09-24 01:17:42 +00:00
|
|
|
set_default("shift_jis", config.shift_jis)
|
|
|
|
set_default("src_dir", config.src_dir)
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Validate progress categories
|
|
|
|
def check_category(category: str):
|
|
|
|
if not any(category == c.id for c in config.progress_categories):
|
|
|
|
sys.exit(
|
|
|
|
f"Progress category '{category}' missing from config.progress_categories"
|
|
|
|
)
|
|
|
|
|
|
|
|
progress_category = obj.options["progress_category"]
|
|
|
|
if isinstance(progress_category, list):
|
|
|
|
for category in progress_category:
|
|
|
|
check_category(category)
|
|
|
|
elif progress_category is not None:
|
|
|
|
check_category(progress_category)
|
|
|
|
|
2024-09-24 01:17:42 +00:00
|
|
|
# Resolve paths
|
|
|
|
build_dir = config.out_path()
|
|
|
|
obj.src_path = Path(obj.options["src_dir"]) / obj.options["source"]
|
|
|
|
if obj.options["asm_dir"] is not None:
|
|
|
|
obj.asm_path = (
|
|
|
|
Path(obj.options["asm_dir"]) / obj.options["source"]
|
|
|
|
).with_suffix(".s")
|
|
|
|
base_name = Path(self.name).with_suffix("")
|
|
|
|
obj.src_obj_path = build_dir / "src" / f"{base_name}.o"
|
|
|
|
obj.asm_obj_path = build_dir / "mod" / f"{base_name}.o"
|
|
|
|
obj.host_obj_path = build_dir / "host" / f"{base_name}.o"
|
|
|
|
obj.ctx_path = build_dir / "src" / f"{base_name}.ctx"
|
|
|
|
return obj
|
|
|
|
|
|
|
|
|
|
|
|
class ProgressCategory:
|
|
|
|
def __init__(self, id: str, name: str) -> None:
|
|
|
|
self.id = id
|
|
|
|
self.name = name
|
|
|
|
|
|
|
|
|
|
|
|
class ProjectConfig:
|
|
|
|
def __init__(self) -> None:
|
|
|
|
# Paths
|
|
|
|
self.build_dir: Path = Path("build") # Output build files
|
|
|
|
self.src_dir: Path = Path("src") # C/C++/asm source files
|
|
|
|
self.tools_dir: Path = Path("tools") # Python scripts
|
|
|
|
self.asm_dir: Optional[Path] = Path(
|
|
|
|
"asm"
|
|
|
|
) # Override incomplete objects (for modding)
|
|
|
|
|
|
|
|
# Tooling
|
|
|
|
self.binutils_tag: Optional[str] = None # Git tag
|
|
|
|
self.binutils_path: Optional[Path] = None # If None, download
|
|
|
|
self.dtk_tag: Optional[str] = None # Git tag
|
|
|
|
self.dtk_path: Optional[Path] = None # If None, download
|
|
|
|
self.compilers_tag: Optional[str] = None # 1
|
|
|
|
self.compilers_path: Optional[Path] = None # If None, download
|
|
|
|
self.wibo_tag: Optional[str] = None # Git tag
|
|
|
|
self.wrapper: Optional[Path] = None # If None, download wibo on Linux
|
|
|
|
self.sjiswrap_tag: Optional[str] = None # Git tag
|
|
|
|
self.sjiswrap_path: Optional[Path] = None # If None, download
|
|
|
|
self.objdiff_tag: Optional[str] = None # Git tag
|
|
|
|
self.objdiff_path: Optional[Path] = None # If None, download
|
|
|
|
|
|
|
|
# Project config
|
|
|
|
self.non_matching: bool = False
|
|
|
|
self.build_rels: bool = True # Build REL files
|
|
|
|
self.check_sha_path: Optional[Path] = None # Path to version.sha1
|
|
|
|
self.config_path: Optional[Path] = None # Path to config.yml
|
|
|
|
self.generate_map: bool = False # Generate map file(s)
|
|
|
|
self.asflags: Optional[List[str]] = None # Assembler flags
|
|
|
|
self.ldflags: Optional[List[str]] = None # Linker flags
|
|
|
|
self.libs: Optional[List[Library]] = None # List of libraries
|
|
|
|
self.linker_version: Optional[str] = None # mwld version
|
|
|
|
self.version: Optional[str] = None # Version name
|
|
|
|
self.warn_missing_config: bool = False # Warn on missing unit configuration
|
|
|
|
self.warn_missing_source: bool = False # Warn on missing source file
|
|
|
|
self.rel_strip_partial: bool = True # Generate PLFs with -strip_partial
|
|
|
|
self.rel_empty_file: Optional[str] = (
|
|
|
|
None # Object name for generating empty RELs
|
|
|
|
)
|
|
|
|
self.shift_jis = (
|
|
|
|
True # Convert source files from UTF-8 to Shift JIS automatically
|
|
|
|
)
|
|
|
|
self.reconfig_deps: Optional[List[Path]] = (
|
|
|
|
None # Additional re-configuration dependency files
|
|
|
|
)
|
|
|
|
self.custom_build_rules: Optional[List[Dict[str, Any]]] = (
|
|
|
|
None # Custom ninja build rules
|
|
|
|
)
|
|
|
|
self.custom_build_steps: Optional[Dict[str, List[Dict[str, Any]]]] = (
|
|
|
|
None # Custom build steps, types are ["pre-compile", "post-compile", "post-link", "post-build"]
|
|
|
|
)
|
2024-10-14 16:58:15 +00:00
|
|
|
self.generate_compile_commands: bool = (
|
|
|
|
True # Generate compile_commands.json for clangd
|
|
|
|
)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-11-21 01:58:20 +00:00
|
|
|
self.scratch_preset_id: Optional[int] = None # Default decomp.me preset ID for scratches
|
|
|
|
|
2024-09-24 01:17:42 +00:00
|
|
|
# Progress output, progress.json and report.json config
|
2024-10-14 16:58:15 +00:00
|
|
|
self.progress = True # Enable progress output
|
2024-09-24 01:17:42 +00:00
|
|
|
self.progress_all: bool = True # Include combined "all" category
|
|
|
|
self.progress_modules: bool = True # Include combined "modules" category
|
|
|
|
self.progress_each_module: bool = (
|
|
|
|
False # Include individual modules, disable for large numbers of modules
|
|
|
|
)
|
|
|
|
self.progress_categories: List[ProgressCategory] = [] # Additional categories
|
2024-10-14 16:58:15 +00:00
|
|
|
self.print_progress_categories: Union[bool, List[str]] = (
|
|
|
|
True # Print additional progress categories in the CLI progress output
|
|
|
|
)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
# Progress fancy printing
|
|
|
|
self.progress_use_fancy: bool = False
|
|
|
|
self.progress_code_fancy_frac: int = 0
|
|
|
|
self.progress_code_fancy_item: str = ""
|
|
|
|
self.progress_data_fancy_frac: int = 0
|
|
|
|
self.progress_data_fancy_item: str = ""
|
|
|
|
|
|
|
|
def validate(self) -> None:
|
|
|
|
required_attrs = [
|
|
|
|
"build_dir",
|
|
|
|
"src_dir",
|
|
|
|
"tools_dir",
|
|
|
|
"check_sha_path",
|
|
|
|
"config_path",
|
|
|
|
"ldflags",
|
|
|
|
"linker_version",
|
|
|
|
"libs",
|
|
|
|
"version",
|
|
|
|
]
|
|
|
|
for attr in required_attrs:
|
|
|
|
if getattr(self, attr) is None:
|
|
|
|
sys.exit(f"ProjectConfig.{attr} missing")
|
|
|
|
|
|
|
|
# Creates a map of object names to Object instances
|
|
|
|
# Options are fully resolved from the library and object
|
|
|
|
def objects(self) -> Dict[str, Object]:
|
|
|
|
out = {}
|
|
|
|
for lib in self.libs or {}:
|
|
|
|
objects: List[Object] = lib["objects"]
|
|
|
|
for obj in objects:
|
|
|
|
if obj.name in out:
|
|
|
|
sys.exit(f"Duplicate object name {obj.name}")
|
|
|
|
out[obj.name] = obj.resolve(self, lib)
|
|
|
|
return out
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Gets the output path for build-related files.
|
2024-09-24 01:17:42 +00:00
|
|
|
def out_path(self) -> Path:
|
|
|
|
return self.build_dir / str(self.version)
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Gets the path to the compilers directory.
|
|
|
|
# Exits the program if neither `compilers_path` nor `compilers_tag` is provided.
|
|
|
|
def compilers(self) -> Path:
|
|
|
|
if self.compilers_path:
|
|
|
|
return self.compilers_path
|
|
|
|
elif self.compilers_tag:
|
|
|
|
return self.build_dir / "compilers"
|
|
|
|
else:
|
|
|
|
sys.exit("ProjectConfig.compilers_tag missing")
|
|
|
|
|
|
|
|
# Gets the wrapper to use for compiler commands, if set.
|
|
|
|
def compiler_wrapper(self) -> Optional[Path]:
|
|
|
|
wrapper = self.wrapper
|
|
|
|
|
|
|
|
if self.use_wibo():
|
|
|
|
wrapper = self.build_dir / "tools" / "wibo"
|
|
|
|
if not is_windows() and wrapper is None:
|
|
|
|
wrapper = Path("wine")
|
|
|
|
|
|
|
|
return wrapper
|
|
|
|
|
|
|
|
# Determines whether or not to use wibo as the compiler wrapper.
|
|
|
|
def use_wibo(self) -> bool:
|
|
|
|
return (
|
|
|
|
self.wibo_tag is not None
|
|
|
|
and sys.platform == "linux"
|
|
|
|
and platform.machine() in ("i386", "x86_64")
|
|
|
|
and self.wrapper is None
|
|
|
|
)
|
|
|
|
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
def is_windows() -> bool:
|
|
|
|
return os.name == "nt"
|
|
|
|
|
|
|
|
|
|
|
|
# On Windows, we need this to use && in commands
|
|
|
|
CHAIN = "cmd /c " if is_windows() else ""
|
|
|
|
# Native executable extension
|
|
|
|
EXE = ".exe" if is_windows() else ""
|
|
|
|
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
def file_is_asm(path: Path) -> bool:
|
|
|
|
return path.suffix.lower() == ".s"
|
|
|
|
|
|
|
|
|
|
|
|
def file_is_c(path: Path) -> bool:
|
|
|
|
return path.suffix.lower() == ".c"
|
|
|
|
|
|
|
|
|
|
|
|
def file_is_cpp(path: Path) -> bool:
|
|
|
|
return path.suffix.lower() in (".cc", ".cp", ".cpp", ".cxx")
|
|
|
|
|
|
|
|
|
|
|
|
def file_is_c_cpp(path: Path) -> bool:
|
|
|
|
return file_is_c(path) or file_is_cpp(path)
|
|
|
|
|
|
|
|
|
|
|
|
def make_flags_str(flags: Optional[List[str]]) -> str:
|
2024-09-24 01:17:42 +00:00
|
|
|
if flags is None:
|
|
|
|
return ""
|
2024-10-14 16:58:15 +00:00
|
|
|
return " ".join(flags)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
|
|
|
|
# Load decomp-toolkit generated config.json
|
|
|
|
def load_build_config(
|
|
|
|
config: ProjectConfig, build_config_path: Path
|
|
|
|
) -> Optional[Dict[str, Any]]:
|
|
|
|
if not build_config_path.is_file():
|
|
|
|
return None
|
|
|
|
|
|
|
|
def versiontuple(v: str) -> Tuple[int, ...]:
|
|
|
|
return tuple(map(int, (v.split("."))))
|
|
|
|
|
|
|
|
f = open(build_config_path, "r", encoding="utf-8")
|
|
|
|
build_config: Dict[str, Any] = json.load(f)
|
|
|
|
config_version = build_config.get("version")
|
|
|
|
if config_version is None:
|
|
|
|
print("Invalid config.json, regenerating...")
|
|
|
|
f.close()
|
|
|
|
os.remove(build_config_path)
|
|
|
|
return None
|
|
|
|
|
|
|
|
dtk_version = str(config.dtk_tag)[1:] # Strip v
|
|
|
|
if versiontuple(config_version) < versiontuple(dtk_version):
|
|
|
|
print("Outdated config.json, regenerating...")
|
|
|
|
f.close()
|
|
|
|
os.remove(build_config_path)
|
|
|
|
return None
|
|
|
|
|
|
|
|
f.close()
|
|
|
|
return build_config
|
|
|
|
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Generate build.ninja, objdiff.json and compile_commands.json
|
2024-09-24 01:17:42 +00:00
|
|
|
def generate_build(config: ProjectConfig) -> None:
|
|
|
|
config.validate()
|
|
|
|
objects = config.objects()
|
|
|
|
build_config = load_build_config(config, config.out_path() / "config.json")
|
|
|
|
generate_build_ninja(config, objects, build_config)
|
|
|
|
generate_objdiff_config(config, objects, build_config)
|
2024-10-14 16:58:15 +00:00
|
|
|
generate_compile_commands(config, objects, build_config)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
|
|
|
|
# Generate build.ninja
|
|
|
|
def generate_build_ninja(
|
|
|
|
config: ProjectConfig,
|
|
|
|
objects: Dict[str, Object],
|
|
|
|
build_config: Optional[Dict[str, Any]],
|
|
|
|
) -> None:
|
|
|
|
out = io.StringIO()
|
|
|
|
n = ninja_syntax.Writer(out)
|
|
|
|
n.variable("ninja_required_version", "1.3")
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
configure_script = Path(os.path.relpath(os.path.abspath(sys.argv[0])))
|
|
|
|
python_lib = Path(os.path.relpath(__file__))
|
|
|
|
python_lib_dir = python_lib.parent
|
|
|
|
n.comment("The arguments passed to configure.py, for rerunning it.")
|
|
|
|
n.variable("configure_args", sys.argv[1:])
|
|
|
|
n.variable("python", f'"{sys.executable}"')
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Variables
|
|
|
|
###
|
|
|
|
n.comment("Variables")
|
|
|
|
n.variable("ldflags", make_flags_str(config.ldflags))
|
|
|
|
if config.linker_version is None:
|
|
|
|
sys.exit("ProjectConfig.linker_version missing")
|
|
|
|
n.variable("mw_version", Path(config.linker_version))
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Tooling
|
|
|
|
###
|
|
|
|
n.comment("Tooling")
|
|
|
|
|
|
|
|
build_path = config.out_path()
|
|
|
|
progress_path = build_path / "progress.json"
|
|
|
|
report_path = build_path / "report.json"
|
|
|
|
build_tools_path = config.build_dir / "tools"
|
|
|
|
download_tool = config.tools_dir / "download_tool.py"
|
|
|
|
n.rule(
|
|
|
|
name="download_tool",
|
|
|
|
command=f"$python {download_tool} $tool $out --tag $tag",
|
|
|
|
description="TOOL $out",
|
|
|
|
)
|
|
|
|
|
|
|
|
decompctx = config.tools_dir / "decompctx.py"
|
|
|
|
n.rule(
|
|
|
|
name="decompctx",
|
|
|
|
command=f"$python {decompctx} $in -o $out -d $out.d",
|
|
|
|
description="CTX $in",
|
|
|
|
depfile="$out.d",
|
|
|
|
deps="gcc",
|
|
|
|
)
|
|
|
|
|
|
|
|
cargo_rule_written = False
|
|
|
|
|
|
|
|
def write_cargo_rule():
|
|
|
|
nonlocal cargo_rule_written
|
|
|
|
if not cargo_rule_written:
|
|
|
|
n.pool("cargo", 1)
|
|
|
|
n.rule(
|
|
|
|
name="cargo",
|
|
|
|
command="cargo build --release --manifest-path $in --bin $bin --target-dir $target",
|
|
|
|
description="CARGO $bin",
|
|
|
|
pool="cargo",
|
|
|
|
depfile=Path("$target") / "release" / "$bin.d",
|
|
|
|
deps="gcc",
|
|
|
|
)
|
|
|
|
cargo_rule_written = True
|
|
|
|
|
|
|
|
if config.dtk_path is not None and config.dtk_path.is_file():
|
|
|
|
dtk = config.dtk_path
|
|
|
|
elif config.dtk_path is not None:
|
|
|
|
dtk = build_tools_path / "release" / f"dtk{EXE}"
|
|
|
|
write_cargo_rule()
|
|
|
|
n.build(
|
|
|
|
outputs=dtk,
|
|
|
|
rule="cargo",
|
|
|
|
inputs=config.dtk_path / "Cargo.toml",
|
|
|
|
implicit=config.dtk_path / "Cargo.lock",
|
|
|
|
variables={
|
|
|
|
"bin": "dtk",
|
|
|
|
"target": build_tools_path,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
elif config.dtk_tag:
|
|
|
|
dtk = build_tools_path / f"dtk{EXE}"
|
|
|
|
n.build(
|
|
|
|
outputs=dtk,
|
|
|
|
rule="download_tool",
|
|
|
|
implicit=download_tool,
|
|
|
|
variables={
|
|
|
|
"tool": "dtk",
|
|
|
|
"tag": config.dtk_tag,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
sys.exit("ProjectConfig.dtk_tag missing")
|
|
|
|
|
|
|
|
if config.objdiff_path is not None and config.objdiff_path.is_file():
|
|
|
|
objdiff = config.objdiff_path
|
|
|
|
elif config.objdiff_path is not None:
|
|
|
|
objdiff = build_tools_path / "release" / f"objdiff-cli{EXE}"
|
|
|
|
write_cargo_rule()
|
|
|
|
n.build(
|
|
|
|
outputs=objdiff,
|
|
|
|
rule="cargo",
|
|
|
|
inputs=config.objdiff_path / "Cargo.toml",
|
|
|
|
implicit=config.objdiff_path / "Cargo.lock",
|
|
|
|
variables={
|
|
|
|
"bin": "objdiff-cli",
|
|
|
|
"target": build_tools_path,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
elif config.objdiff_tag:
|
|
|
|
objdiff = build_tools_path / f"objdiff-cli{EXE}"
|
|
|
|
n.build(
|
|
|
|
outputs=objdiff,
|
|
|
|
rule="download_tool",
|
|
|
|
implicit=download_tool,
|
|
|
|
variables={
|
|
|
|
"tool": "objdiff-cli",
|
|
|
|
"tag": config.objdiff_tag,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
sys.exit("ProjectConfig.objdiff_tag missing")
|
|
|
|
|
|
|
|
if config.sjiswrap_path:
|
|
|
|
sjiswrap = config.sjiswrap_path
|
|
|
|
elif config.sjiswrap_tag:
|
|
|
|
sjiswrap = build_tools_path / "sjiswrap.exe"
|
|
|
|
n.build(
|
|
|
|
outputs=sjiswrap,
|
|
|
|
rule="download_tool",
|
|
|
|
implicit=download_tool,
|
|
|
|
variables={
|
|
|
|
"tool": "sjiswrap",
|
|
|
|
"tag": config.sjiswrap_tag,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
sys.exit("ProjectConfig.sjiswrap_tag missing")
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
wrapper = config.compiler_wrapper()
|
2024-09-24 01:17:42 +00:00
|
|
|
# Only add an implicit dependency on wibo if we download it
|
|
|
|
wrapper_implicit: Optional[Path] = None
|
2024-10-14 16:58:15 +00:00
|
|
|
if wrapper is not None and config.use_wibo():
|
2024-09-24 01:17:42 +00:00
|
|
|
wrapper_implicit = wrapper
|
|
|
|
n.build(
|
|
|
|
outputs=wrapper,
|
|
|
|
rule="download_tool",
|
|
|
|
implicit=download_tool,
|
|
|
|
variables={
|
|
|
|
"tool": "wibo",
|
|
|
|
"tag": config.wibo_tag,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
wrapper_cmd = f"{wrapper} " if wrapper else ""
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
compilers = config.compilers()
|
2024-09-24 01:17:42 +00:00
|
|
|
compilers_implicit: Optional[Path] = None
|
2024-10-14 16:58:15 +00:00
|
|
|
if config.compilers_path is None and config.compilers_tag is not None:
|
2024-09-24 01:17:42 +00:00
|
|
|
compilers_implicit = compilers
|
|
|
|
n.build(
|
|
|
|
outputs=compilers,
|
|
|
|
rule="download_tool",
|
|
|
|
implicit=download_tool,
|
|
|
|
variables={
|
|
|
|
"tool": "compilers",
|
|
|
|
"tag": config.compilers_tag,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
binutils_implicit = None
|
|
|
|
if config.binutils_path:
|
|
|
|
binutils = config.binutils_path
|
|
|
|
elif config.binutils_tag:
|
|
|
|
binutils = config.build_dir / "binutils"
|
|
|
|
binutils_implicit = binutils
|
|
|
|
n.build(
|
|
|
|
outputs=binutils,
|
|
|
|
rule="download_tool",
|
|
|
|
implicit=download_tool,
|
|
|
|
variables={
|
|
|
|
"tool": "binutils",
|
|
|
|
"tag": config.binutils_tag,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
sys.exit("ProjectConfig.binutils_tag missing")
|
|
|
|
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Helper rule for downloading all tools
|
|
|
|
###
|
|
|
|
n.comment("Download all tools")
|
|
|
|
n.build(
|
|
|
|
outputs="tools",
|
|
|
|
rule="phony",
|
|
|
|
inputs=[dtk, sjiswrap, wrapper, compilers, binutils, objdiff],
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Build rules
|
|
|
|
###
|
|
|
|
compiler_path = compilers / "$mw_version"
|
|
|
|
|
|
|
|
# MWCC
|
|
|
|
mwcc = compiler_path / "mwcceppc.exe"
|
|
|
|
mwcc_cmd = f"{wrapper_cmd}{mwcc} $cflags -MMD -c $in -o $basedir"
|
|
|
|
mwcc_implicit: List[Optional[Path]] = [compilers_implicit or mwcc, wrapper_implicit]
|
|
|
|
|
|
|
|
# MWCC with UTF-8 to Shift JIS wrapper
|
|
|
|
mwcc_sjis_cmd = f"{wrapper_cmd}{sjiswrap} {mwcc} $cflags -MMD -c $in -o $basedir"
|
|
|
|
mwcc_sjis_implicit: List[Optional[Path]] = [*mwcc_implicit, sjiswrap]
|
|
|
|
|
|
|
|
# MWLD
|
|
|
|
mwld = compiler_path / "mwldeppc.exe"
|
|
|
|
mwld_cmd = f"{wrapper_cmd}{mwld} $ldflags -o $out @$out.rsp"
|
|
|
|
mwld_implicit: List[Optional[Path]] = [compilers_implicit or mwld, wrapper_implicit]
|
|
|
|
|
|
|
|
# GNU as
|
|
|
|
gnu_as = binutils / f"powerpc-eabi-as{EXE}"
|
|
|
|
gnu_as_cmd = (
|
|
|
|
f"{CHAIN}{gnu_as} $asflags -o $out $in -MD $out.d"
|
|
|
|
+ f" && {dtk} elf fixup $out $out"
|
|
|
|
)
|
|
|
|
gnu_as_implicit = [binutils_implicit or gnu_as, dtk]
|
|
|
|
|
|
|
|
if os.name != "nt":
|
|
|
|
transform_dep = config.tools_dir / "transform_dep.py"
|
|
|
|
mwcc_cmd += f" && $python {transform_dep} $basefile.d $basefile.d"
|
|
|
|
mwcc_sjis_cmd += f" && $python {transform_dep} $basefile.d $basefile.d"
|
|
|
|
mwcc_implicit.append(transform_dep)
|
|
|
|
mwcc_sjis_implicit.append(transform_dep)
|
|
|
|
|
|
|
|
n.comment("Link ELF file")
|
|
|
|
n.rule(
|
|
|
|
name="link",
|
|
|
|
command=mwld_cmd,
|
|
|
|
description="LINK $out",
|
|
|
|
rspfile="$out.rsp",
|
|
|
|
rspfile_content="$in_newline",
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
n.comment("Generate DOL")
|
|
|
|
n.rule(
|
|
|
|
name="elf2dol",
|
|
|
|
command=f"{dtk} elf2dol $in $out",
|
|
|
|
description="DOL $out",
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
n.comment("MWCC build")
|
|
|
|
n.rule(
|
|
|
|
name="mwcc",
|
|
|
|
command=mwcc_cmd,
|
|
|
|
description="MWCC $out",
|
|
|
|
depfile="$basefile.d",
|
|
|
|
deps="gcc",
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
n.comment("MWCC build (with UTF-8 to Shift JIS wrapper)")
|
|
|
|
n.rule(
|
|
|
|
name="mwcc_sjis",
|
|
|
|
command=mwcc_sjis_cmd,
|
|
|
|
description="MWCC $out",
|
|
|
|
depfile="$basefile.d",
|
|
|
|
deps="gcc",
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
n.comment("Assemble asm")
|
|
|
|
n.rule(
|
|
|
|
name="as",
|
|
|
|
command=gnu_as_cmd,
|
|
|
|
description="AS $out",
|
|
|
|
depfile="$out.d",
|
|
|
|
deps="gcc",
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
if len(config.custom_build_rules or {}) > 0:
|
|
|
|
n.comment("Custom project build rules (pre/post-processing)")
|
|
|
|
for rule in config.custom_build_rules or {}:
|
|
|
|
n.rule(
|
|
|
|
name=cast(str, rule.get("name")),
|
|
|
|
command=cast(str, rule.get("command")),
|
|
|
|
description=rule.get("description", None),
|
|
|
|
depfile=rule.get("depfile", None),
|
|
|
|
generator=rule.get("generator", False),
|
|
|
|
pool=rule.get("pool", None),
|
|
|
|
restat=rule.get("restat", False),
|
|
|
|
rspfile=rule.get("rspfile", None),
|
|
|
|
rspfile_content=rule.get("rspfile_content", None),
|
|
|
|
deps=rule.get("deps", None),
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
def write_custom_step(step: str) -> List[str | Path]:
|
|
|
|
implicit: List[str | Path] = []
|
|
|
|
if config.custom_build_steps and step in config.custom_build_steps:
|
|
|
|
n.comment(f"Custom build steps ({step})")
|
|
|
|
for custom_step in config.custom_build_steps[step]:
|
|
|
|
outputs = cast(List[str | Path], custom_step.get("outputs"))
|
|
|
|
|
|
|
|
if isinstance(outputs, list):
|
|
|
|
implicit.extend(outputs)
|
|
|
|
else:
|
|
|
|
implicit.append(outputs)
|
|
|
|
|
|
|
|
n.build(
|
|
|
|
outputs=outputs,
|
|
|
|
rule=cast(str, custom_step.get("rule")),
|
|
|
|
inputs=custom_step.get("inputs", None),
|
|
|
|
implicit=custom_step.get("implicit", None),
|
|
|
|
order_only=custom_step.get("order_only", None),
|
|
|
|
variables=custom_step.get("variables", None),
|
|
|
|
implicit_outputs=custom_step.get("implicit_outputs", None),
|
|
|
|
pool=custom_step.get("pool", None),
|
|
|
|
dyndep=custom_step.get("dyndep", None),
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
return implicit
|
|
|
|
|
|
|
|
n.comment("Host build")
|
|
|
|
n.variable("host_cflags", "-I include -Wno-trigraphs")
|
|
|
|
n.variable(
|
|
|
|
"host_cppflags",
|
|
|
|
"-std=c++98 -I include -fno-exceptions -fno-rtti -D_CRT_SECURE_NO_WARNINGS -Wno-trigraphs -Wno-c++11-extensions",
|
|
|
|
)
|
|
|
|
n.rule(
|
|
|
|
name="host_cc",
|
|
|
|
command="clang $host_cflags -c -o $out $in",
|
|
|
|
description="CC $out",
|
|
|
|
)
|
|
|
|
n.rule(
|
|
|
|
name="host_cpp",
|
|
|
|
command="clang++ $host_cppflags -c -o $out $in",
|
|
|
|
description="CXX $out",
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
# Add all build steps needed before we compile (e.g. processing assets)
|
|
|
|
precompile_implicit = write_custom_step("pre-compile")
|
|
|
|
|
|
|
|
###
|
|
|
|
# Source files
|
|
|
|
###
|
|
|
|
n.comment("Source files")
|
|
|
|
|
|
|
|
def map_path(path: Path) -> Path:
|
|
|
|
return path.parent / (path.name + ".MAP")
|
|
|
|
|
|
|
|
class LinkStep:
|
|
|
|
def __init__(self, config: Dict[str, Any]) -> None:
|
|
|
|
self.name: str = config["name"]
|
|
|
|
self.module_id: int = config["module_id"]
|
|
|
|
self.ldscript: Optional[Path] = Path(config["ldscript"])
|
|
|
|
self.entry = config["entry"]
|
|
|
|
self.inputs: List[str] = []
|
|
|
|
|
|
|
|
def add(self, obj: Path) -> None:
|
|
|
|
self.inputs.append(serialize_path(obj))
|
|
|
|
|
|
|
|
def output(self) -> Path:
|
|
|
|
if self.module_id == 0:
|
|
|
|
return build_path / f"{self.name}.dol"
|
|
|
|
else:
|
|
|
|
return build_path / self.name / f"{self.name}.rel"
|
|
|
|
|
|
|
|
def partial_output(self) -> Path:
|
|
|
|
if self.module_id == 0:
|
|
|
|
return build_path / f"{self.name}.elf"
|
|
|
|
else:
|
|
|
|
return build_path / self.name / f"{self.name}.plf"
|
|
|
|
|
|
|
|
def write(self, n: ninja_syntax.Writer) -> None:
|
|
|
|
n.comment(f"Link {self.name}")
|
|
|
|
if self.module_id == 0:
|
|
|
|
elf_path = build_path / f"{self.name}.elf"
|
|
|
|
elf_ldflags = f"$ldflags -lcf {serialize_path(self.ldscript)}"
|
|
|
|
if config.generate_map:
|
|
|
|
elf_map = map_path(elf_path)
|
|
|
|
elf_ldflags += f" -map {serialize_path(elf_map)}"
|
|
|
|
else:
|
|
|
|
elf_map = None
|
|
|
|
n.build(
|
|
|
|
outputs=elf_path,
|
|
|
|
rule="link",
|
|
|
|
inputs=self.inputs,
|
|
|
|
implicit=[
|
|
|
|
*precompile_implicit,
|
|
|
|
self.ldscript,
|
|
|
|
*mwld_implicit,
|
|
|
|
*postcompile_implicit,
|
|
|
|
],
|
|
|
|
implicit_outputs=elf_map,
|
|
|
|
variables={"ldflags": elf_ldflags},
|
|
|
|
)
|
|
|
|
else:
|
|
|
|
preplf_path = build_path / self.name / f"{self.name}.preplf"
|
|
|
|
plf_path = build_path / self.name / f"{self.name}.plf"
|
|
|
|
preplf_ldflags = "$ldflags -sdata 0 -sdata2 0 -r"
|
|
|
|
plf_ldflags = f"$ldflags -sdata 0 -sdata2 0 -r1 -lcf {serialize_path(self.ldscript)}"
|
|
|
|
if self.entry:
|
|
|
|
plf_ldflags += f" -m {self.entry}"
|
|
|
|
# -strip_partial is only valid with -m
|
|
|
|
if config.rel_strip_partial:
|
|
|
|
plf_ldflags += " -strip_partial"
|
|
|
|
if config.generate_map:
|
|
|
|
preplf_map = map_path(preplf_path)
|
|
|
|
preplf_ldflags += f" -map {serialize_path(preplf_map)}"
|
|
|
|
plf_map = map_path(plf_path)
|
|
|
|
plf_ldflags += f" -map {serialize_path(plf_map)}"
|
|
|
|
else:
|
|
|
|
preplf_map = None
|
|
|
|
plf_map = None
|
|
|
|
n.build(
|
|
|
|
outputs=preplf_path,
|
|
|
|
rule="link",
|
|
|
|
inputs=self.inputs,
|
|
|
|
implicit=mwld_implicit,
|
|
|
|
implicit_outputs=preplf_map,
|
|
|
|
variables={"ldflags": preplf_ldflags},
|
|
|
|
)
|
|
|
|
n.build(
|
|
|
|
outputs=plf_path,
|
|
|
|
rule="link",
|
|
|
|
inputs=self.inputs,
|
|
|
|
implicit=[self.ldscript, preplf_path, *mwld_implicit],
|
|
|
|
implicit_outputs=plf_map,
|
|
|
|
variables={"ldflags": plf_ldflags},
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
link_outputs: List[Path] = []
|
|
|
|
if build_config:
|
|
|
|
link_steps: List[LinkStep] = []
|
|
|
|
used_compiler_versions: Set[str] = set()
|
|
|
|
source_inputs: List[Path] = []
|
|
|
|
host_source_inputs: List[Path] = []
|
|
|
|
source_added: Set[Path] = set()
|
|
|
|
|
|
|
|
def c_build(obj: Object, src_path: Path) -> Optional[Path]:
|
|
|
|
# Avoid creating duplicate build rules
|
|
|
|
if obj.src_obj_path is None or obj.src_obj_path in source_added:
|
|
|
|
return obj.src_obj_path
|
|
|
|
source_added.add(obj.src_obj_path)
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
cflags = obj.options["cflags"]
|
|
|
|
extra_cflags = obj.options["extra_cflags"]
|
|
|
|
|
|
|
|
# Add appropriate language flag if it doesn't exist already
|
|
|
|
# Added directly to the source so it flows to other generation tasks
|
|
|
|
if not any(flag.startswith("-lang") for flag in cflags) and (
|
|
|
|
extra_cflags is None
|
|
|
|
or not any(flag.startswith("-lang") for flag in extra_cflags)
|
|
|
|
):
|
|
|
|
# Ensure extra_cflags is a unique instance,
|
|
|
|
# and insert into there to avoid modifying shared sets of flags
|
|
|
|
if extra_cflags is None:
|
|
|
|
extra_cflags = []
|
|
|
|
extra_cflags = obj.options["extra_cflags"] = list(extra_cflags)
|
|
|
|
if file_is_cpp(src_path):
|
|
|
|
extra_cflags.insert(0, "-lang=c++")
|
|
|
|
else:
|
|
|
|
extra_cflags.insert(0, "-lang=c")
|
|
|
|
|
|
|
|
cflags_str = make_flags_str(cflags)
|
|
|
|
if extra_cflags is not None:
|
|
|
|
extra_cflags_str = make_flags_str(extra_cflags)
|
|
|
|
cflags_str += " " + extra_cflags_str
|
|
|
|
used_compiler_versions.add(obj.options["mw_version"])
|
|
|
|
|
2024-09-24 01:17:42 +00:00
|
|
|
# Add MWCC build rule
|
|
|
|
lib_name = obj.options["lib"]
|
|
|
|
n.comment(f"{obj.name}: {lib_name} (linked {obj.completed})")
|
|
|
|
n.build(
|
|
|
|
outputs=obj.src_obj_path,
|
|
|
|
rule="mwcc_sjis" if obj.options["shift_jis"] else "mwcc",
|
|
|
|
inputs=src_path,
|
|
|
|
variables={
|
|
|
|
"mw_version": Path(obj.options["mw_version"]),
|
|
|
|
"cflags": cflags_str,
|
|
|
|
"basedir": os.path.dirname(obj.src_obj_path),
|
|
|
|
"basefile": obj.src_obj_path.with_suffix(""),
|
|
|
|
},
|
|
|
|
implicit=(
|
|
|
|
mwcc_sjis_implicit if obj.options["shift_jis"] else mwcc_implicit
|
|
|
|
),
|
|
|
|
)
|
|
|
|
|
|
|
|
# Add ctx build rule
|
|
|
|
if obj.ctx_path is not None:
|
|
|
|
n.build(
|
|
|
|
outputs=obj.ctx_path,
|
|
|
|
rule="decompctx",
|
|
|
|
inputs=src_path,
|
|
|
|
implicit=decompctx,
|
|
|
|
)
|
|
|
|
|
|
|
|
# Add host build rule
|
|
|
|
if obj.options["host"] and obj.host_obj_path is not None:
|
|
|
|
n.build(
|
|
|
|
outputs=obj.host_obj_path,
|
2024-10-14 16:58:15 +00:00
|
|
|
rule="host_cc" if file_is_c(src_path) else "host_cpp",
|
2024-09-24 01:17:42 +00:00
|
|
|
inputs=src_path,
|
|
|
|
variables={
|
|
|
|
"basedir": os.path.dirname(obj.host_obj_path),
|
|
|
|
"basefile": obj.host_obj_path.with_suffix(""),
|
|
|
|
},
|
|
|
|
)
|
|
|
|
if obj.options["add_to_all"]:
|
|
|
|
host_source_inputs.append(obj.host_obj_path)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
if obj.options["add_to_all"]:
|
|
|
|
source_inputs.append(obj.src_obj_path)
|
|
|
|
|
|
|
|
return obj.src_obj_path
|
|
|
|
|
|
|
|
def asm_build(
|
|
|
|
obj: Object, src_path: Path, obj_path: Optional[Path]
|
|
|
|
) -> Optional[Path]:
|
|
|
|
if obj.options["asflags"] is None:
|
|
|
|
sys.exit("ProjectConfig.asflags missing")
|
|
|
|
asflags_str = make_flags_str(obj.options["asflags"])
|
|
|
|
if obj.options["extra_asflags"] is not None:
|
|
|
|
extra_asflags_str = make_flags_str(obj.options["extra_asflags"])
|
|
|
|
asflags_str += " " + extra_asflags_str
|
|
|
|
|
|
|
|
# Avoid creating duplicate build rules
|
|
|
|
if obj_path is None or obj_path in source_added:
|
|
|
|
return obj_path
|
|
|
|
source_added.add(obj_path)
|
|
|
|
|
|
|
|
# Add assembler build rule
|
|
|
|
lib_name = obj.options["lib"]
|
|
|
|
n.comment(f"{obj.name}: {lib_name} (linked {obj.completed})")
|
|
|
|
n.build(
|
|
|
|
outputs=obj_path,
|
|
|
|
rule="as",
|
|
|
|
inputs=src_path,
|
|
|
|
variables={"asflags": asflags_str},
|
|
|
|
implicit=gnu_as_implicit,
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
if obj.options["add_to_all"]:
|
|
|
|
source_inputs.append(obj_path)
|
|
|
|
|
|
|
|
return obj_path
|
|
|
|
|
|
|
|
def add_unit(build_obj, link_step: LinkStep):
|
|
|
|
obj_path, obj_name = build_obj["object"], build_obj["name"]
|
|
|
|
obj = objects.get(obj_name)
|
|
|
|
if obj is None:
|
|
|
|
if config.warn_missing_config and not build_obj["autogenerated"]:
|
|
|
|
print(f"Missing configuration for {obj_name}")
|
|
|
|
link_step.add(obj_path)
|
|
|
|
return
|
|
|
|
|
|
|
|
link_built_obj = obj.completed
|
|
|
|
built_obj_path: Optional[Path] = None
|
|
|
|
if obj.src_path is not None and obj.src_path.exists():
|
2024-10-14 16:58:15 +00:00
|
|
|
if file_is_c_cpp(obj.src_path):
|
2024-09-24 01:17:42 +00:00
|
|
|
# Add MWCC & host build rules
|
|
|
|
built_obj_path = c_build(obj, obj.src_path)
|
2024-10-14 16:58:15 +00:00
|
|
|
elif file_is_asm(obj.src_path):
|
2024-09-24 01:17:42 +00:00
|
|
|
# Add assembler build rule
|
|
|
|
built_obj_path = asm_build(obj, obj.src_path, obj.src_obj_path)
|
|
|
|
else:
|
|
|
|
sys.exit(f"Unknown source file type {obj.src_path}")
|
|
|
|
else:
|
|
|
|
if config.warn_missing_source or obj.completed:
|
|
|
|
print(f"Missing source file {obj.src_path}")
|
|
|
|
link_built_obj = False
|
|
|
|
|
|
|
|
# Assembly overrides
|
|
|
|
if obj.asm_path is not None and obj.asm_path.exists():
|
|
|
|
link_built_obj = True
|
|
|
|
built_obj_path = asm_build(obj, obj.asm_path, obj.asm_obj_path)
|
|
|
|
|
|
|
|
if link_built_obj and built_obj_path is not None:
|
|
|
|
# Use the source-built object
|
|
|
|
link_step.add(built_obj_path)
|
|
|
|
elif obj_path is not None:
|
|
|
|
# Use the original (extracted) object
|
|
|
|
link_step.add(obj_path)
|
|
|
|
else:
|
|
|
|
lib_name = obj.options["lib"]
|
|
|
|
sys.exit(
|
|
|
|
f"Missing object for {obj_name}: {obj.src_path} {lib_name} {obj}"
|
|
|
|
)
|
|
|
|
|
|
|
|
# Add DOL link step
|
|
|
|
link_step = LinkStep(build_config)
|
|
|
|
for unit in build_config["units"]:
|
|
|
|
add_unit(unit, link_step)
|
|
|
|
link_steps.append(link_step)
|
|
|
|
|
|
|
|
if config.build_rels:
|
|
|
|
# Add REL link steps
|
|
|
|
for module in build_config["modules"]:
|
|
|
|
module_link_step = LinkStep(module)
|
|
|
|
for unit in module["units"]:
|
|
|
|
add_unit(unit, module_link_step)
|
|
|
|
# Add empty object to empty RELs
|
|
|
|
if len(module_link_step.inputs) == 0:
|
|
|
|
if config.rel_empty_file is None:
|
|
|
|
sys.exit("ProjectConfig.rel_empty_file missing")
|
|
|
|
add_unit(
|
|
|
|
{
|
|
|
|
"object": None,
|
|
|
|
"name": config.rel_empty_file,
|
|
|
|
"autogenerated": True,
|
|
|
|
},
|
|
|
|
module_link_step,
|
|
|
|
)
|
|
|
|
link_steps.append(module_link_step)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
# Check if all compiler versions exist
|
|
|
|
for mw_version in used_compiler_versions:
|
|
|
|
mw_path = compilers / mw_version / "mwcceppc.exe"
|
|
|
|
if config.compilers_path and not os.path.exists(mw_path):
|
|
|
|
sys.exit(f"Compiler {mw_path} does not exist")
|
|
|
|
|
|
|
|
# Check if linker exists
|
|
|
|
mw_path = compilers / str(config.linker_version) / "mwldeppc.exe"
|
|
|
|
if config.compilers_path and not os.path.exists(mw_path):
|
|
|
|
sys.exit(f"Linker {mw_path} does not exist")
|
|
|
|
|
|
|
|
# Add all build steps needed before we link and after compiling objects
|
|
|
|
postcompile_implicit = write_custom_step("post-compile")
|
|
|
|
|
|
|
|
###
|
|
|
|
# Link
|
|
|
|
###
|
|
|
|
for step in link_steps:
|
|
|
|
step.write(n)
|
|
|
|
link_outputs.append(step.output())
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
# Add all build steps needed after linking and before GC/Wii native format generation
|
|
|
|
postlink_implicit = write_custom_step("post-link")
|
|
|
|
|
|
|
|
###
|
|
|
|
# Generate DOL
|
|
|
|
###
|
|
|
|
n.build(
|
|
|
|
outputs=link_steps[0].output(),
|
|
|
|
rule="elf2dol",
|
|
|
|
inputs=link_steps[0].partial_output(),
|
|
|
|
implicit=[*postlink_implicit, dtk],
|
|
|
|
)
|
|
|
|
|
|
|
|
###
|
|
|
|
# Generate RELs
|
|
|
|
###
|
|
|
|
n.comment("Generate REL(s)")
|
|
|
|
flags = "-w"
|
|
|
|
if len(build_config["links"]) > 1:
|
|
|
|
flags += " -q"
|
|
|
|
n.rule(
|
|
|
|
name="makerel",
|
|
|
|
command=f"{dtk} rel make {flags} -c $config $names @$rspfile",
|
|
|
|
description="REL",
|
|
|
|
rspfile="$rspfile",
|
|
|
|
rspfile_content="$in_newline",
|
|
|
|
)
|
|
|
|
generated_rels: List[str] = []
|
|
|
|
for idx, link in enumerate(build_config["links"]):
|
|
|
|
# Map module names to link steps
|
|
|
|
link_steps_local = list(
|
|
|
|
filter(
|
|
|
|
lambda step: step.name in link["modules"],
|
|
|
|
link_steps,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
link_steps_local.sort(key=lambda step: step.module_id)
|
|
|
|
# RELs can be the output of multiple link steps,
|
|
|
|
# so we need to filter out duplicates
|
|
|
|
rels_to_generate = list(
|
|
|
|
filter(
|
|
|
|
lambda step: step.module_id != 0
|
|
|
|
and step.name not in generated_rels,
|
|
|
|
link_steps_local,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
if len(rels_to_generate) == 0:
|
|
|
|
continue
|
|
|
|
generated_rels.extend(map(lambda step: step.name, rels_to_generate))
|
|
|
|
rel_outputs = list(
|
|
|
|
map(
|
|
|
|
lambda step: step.output(),
|
|
|
|
rels_to_generate,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
rel_names = list(
|
|
|
|
map(
|
|
|
|
lambda step: step.name,
|
|
|
|
link_steps_local,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
rel_names_arg = " ".join(map(lambda name: f"-n {name}", rel_names))
|
|
|
|
n.build(
|
|
|
|
outputs=rel_outputs,
|
|
|
|
rule="makerel",
|
|
|
|
inputs=list(map(lambda step: step.partial_output(), link_steps_local)),
|
|
|
|
implicit=[dtk, config.config_path],
|
|
|
|
variables={
|
|
|
|
"config": config.config_path,
|
|
|
|
"rspfile": config.out_path() / f"rel{idx}.rsp",
|
|
|
|
"names": rel_names_arg,
|
|
|
|
},
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
# Add all build steps needed post-build (re-building archives and such)
|
|
|
|
postbuild_implicit = write_custom_step("post-build")
|
|
|
|
|
|
|
|
###
|
|
|
|
# Helper rule for building all source files
|
|
|
|
###
|
|
|
|
n.comment("Build all source files")
|
|
|
|
n.build(
|
|
|
|
outputs="all_source",
|
|
|
|
rule="phony",
|
|
|
|
inputs=source_inputs,
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Helper rule for building all source files, with a host compiler
|
|
|
|
###
|
|
|
|
n.comment("Build all source files with a host compiler")
|
|
|
|
n.build(
|
|
|
|
outputs="all_source_host",
|
|
|
|
rule="phony",
|
|
|
|
inputs=host_source_inputs,
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Check hash
|
|
|
|
###
|
|
|
|
n.comment("Check hash")
|
|
|
|
ok_path = build_path / "ok"
|
|
|
|
quiet = "-q " if len(link_steps) > 3 else ""
|
|
|
|
n.rule(
|
|
|
|
name="check",
|
|
|
|
command=f"{dtk} shasum {quiet} -c $in -o $out",
|
|
|
|
description="CHECK $in",
|
|
|
|
)
|
|
|
|
n.build(
|
|
|
|
outputs=ok_path,
|
|
|
|
rule="check",
|
|
|
|
inputs=config.check_sha_path,
|
|
|
|
implicit=[dtk, *link_outputs, *postbuild_implicit],
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Calculate progress
|
|
|
|
###
|
|
|
|
n.comment("Calculate progress")
|
|
|
|
n.rule(
|
|
|
|
name="progress",
|
|
|
|
command=f"$python {configure_script} $configure_args progress",
|
|
|
|
description="PROGRESS",
|
|
|
|
)
|
|
|
|
n.build(
|
|
|
|
outputs=progress_path,
|
|
|
|
rule="progress",
|
2024-10-14 16:58:15 +00:00
|
|
|
implicit=[
|
|
|
|
ok_path,
|
|
|
|
configure_script,
|
|
|
|
python_lib,
|
|
|
|
report_path,
|
|
|
|
],
|
2024-09-24 01:17:42 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
###
|
|
|
|
# Generate progress report
|
|
|
|
###
|
|
|
|
n.comment("Generate progress report")
|
|
|
|
n.rule(
|
|
|
|
name="report",
|
|
|
|
command=f"{objdiff} report generate -o $out",
|
|
|
|
description="REPORT",
|
|
|
|
)
|
|
|
|
report_implicit: List[str | Path] = [objdiff, "all_source"]
|
|
|
|
n.build(
|
|
|
|
outputs=report_path,
|
|
|
|
rule="report",
|
|
|
|
implicit=report_implicit,
|
|
|
|
)
|
|
|
|
|
|
|
|
###
|
|
|
|
# Helper tools
|
|
|
|
###
|
|
|
|
# TODO: make these rules work for RELs too
|
|
|
|
dol_link_step = link_steps[0]
|
|
|
|
dol_elf_path = dol_link_step.partial_output()
|
|
|
|
n.comment("Check for mismatching symbols")
|
|
|
|
n.rule(
|
|
|
|
name="dol_diff",
|
|
|
|
command=f"{dtk} -L error dol diff $in",
|
|
|
|
description=f"DIFF {dol_elf_path}",
|
|
|
|
)
|
|
|
|
n.build(
|
|
|
|
inputs=[config.config_path, dol_elf_path],
|
|
|
|
outputs="dol_diff",
|
|
|
|
rule="dol_diff",
|
|
|
|
)
|
|
|
|
n.build(
|
|
|
|
outputs="diff",
|
|
|
|
rule="phony",
|
|
|
|
inputs="dol_diff",
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
n.comment("Apply symbols from linked ELF")
|
|
|
|
n.rule(
|
|
|
|
name="dol_apply",
|
|
|
|
command=f"{dtk} dol apply $in",
|
|
|
|
description=f"APPLY {dol_elf_path}",
|
|
|
|
)
|
|
|
|
n.build(
|
|
|
|
inputs=[config.config_path, dol_elf_path],
|
|
|
|
outputs="dol_apply",
|
|
|
|
rule="dol_apply",
|
|
|
|
implicit=[ok_path],
|
|
|
|
)
|
|
|
|
n.build(
|
|
|
|
outputs="apply",
|
|
|
|
rule="phony",
|
|
|
|
inputs="dol_apply",
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Split DOL
|
|
|
|
###
|
|
|
|
build_config_path = build_path / "config.json"
|
|
|
|
n.comment("Split DOL into relocatable objects")
|
|
|
|
n.rule(
|
|
|
|
name="split",
|
|
|
|
command=f"{dtk} dol split $in $out_dir",
|
|
|
|
description="SPLIT $in",
|
|
|
|
depfile="$out_dir/dep",
|
|
|
|
deps="gcc",
|
|
|
|
)
|
|
|
|
n.build(
|
|
|
|
inputs=config.config_path,
|
|
|
|
outputs=build_config_path,
|
|
|
|
rule="split",
|
|
|
|
implicit=dtk,
|
|
|
|
variables={"out_dir": build_path},
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Regenerate on change
|
|
|
|
###
|
|
|
|
n.comment("Reconfigure on change")
|
|
|
|
n.rule(
|
|
|
|
name="configure",
|
|
|
|
command=f"$python {configure_script} $configure_args",
|
|
|
|
generator=True,
|
|
|
|
description=f"RUN {configure_script}",
|
|
|
|
)
|
|
|
|
n.build(
|
|
|
|
outputs="build.ninja",
|
|
|
|
rule="configure",
|
|
|
|
implicit=[
|
|
|
|
build_config_path,
|
|
|
|
configure_script,
|
|
|
|
python_lib,
|
|
|
|
python_lib_dir / "ninja_syntax.py",
|
|
|
|
*(config.reconfig_deps or []),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
n.newline()
|
|
|
|
|
|
|
|
###
|
|
|
|
# Default rule
|
|
|
|
###
|
|
|
|
n.comment("Default rule")
|
|
|
|
if build_config:
|
|
|
|
if config.non_matching:
|
|
|
|
n.default(link_outputs)
|
2024-10-14 16:58:15 +00:00
|
|
|
elif config.progress:
|
2024-09-24 01:17:42 +00:00
|
|
|
n.default(progress_path)
|
2024-10-14 16:58:15 +00:00
|
|
|
else:
|
|
|
|
n.default(ok_path)
|
2024-09-24 01:17:42 +00:00
|
|
|
else:
|
|
|
|
n.default(build_config_path)
|
|
|
|
|
|
|
|
# Write build.ninja
|
|
|
|
with open("build.ninja", "w", encoding="utf-8") as f:
|
|
|
|
f.write(out.getvalue())
|
|
|
|
out.close()
|
|
|
|
|
|
|
|
|
|
|
|
# Generate objdiff.json
|
|
|
|
def generate_objdiff_config(
|
|
|
|
config: ProjectConfig,
|
|
|
|
objects: Dict[str, Object],
|
|
|
|
build_config: Optional[Dict[str, Any]],
|
|
|
|
) -> None:
|
|
|
|
if build_config is None:
|
|
|
|
return
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Load existing objdiff.json
|
|
|
|
existing_units = {}
|
|
|
|
if Path("objdiff.json").is_file():
|
|
|
|
with open("objdiff.json", "r", encoding="utf-8") as r:
|
|
|
|
existing_config = json.load(r)
|
|
|
|
existing_units = {unit["name"]: unit for unit in existing_config["units"]}
|
|
|
|
|
2024-09-24 01:17:42 +00:00
|
|
|
objdiff_config: Dict[str, Any] = {
|
|
|
|
"min_version": "2.0.0-beta.5",
|
|
|
|
"custom_make": "ninja",
|
|
|
|
"build_target": False,
|
|
|
|
"watch_patterns": [
|
|
|
|
"*.c",
|
|
|
|
"*.cp",
|
|
|
|
"*.cpp",
|
|
|
|
"*.h",
|
|
|
|
"*.hpp",
|
|
|
|
"*.inc",
|
|
|
|
"*.py",
|
|
|
|
"*.yml",
|
|
|
|
"*.txt",
|
|
|
|
"*.json",
|
|
|
|
],
|
|
|
|
"units": [],
|
|
|
|
"progress_categories": [],
|
|
|
|
}
|
|
|
|
|
|
|
|
# decomp.me compiler name mapping
|
|
|
|
COMPILER_MAP = {
|
|
|
|
"GC/1.0": "mwcc_233_144",
|
|
|
|
"GC/1.1": "mwcc_233_159",
|
|
|
|
"GC/1.2.5": "mwcc_233_163",
|
|
|
|
"GC/1.2.5e": "mwcc_233_163e",
|
|
|
|
"GC/1.2.5n": "mwcc_233_163n",
|
|
|
|
"GC/1.3": "mwcc_242_53",
|
|
|
|
"GC/1.3.2": "mwcc_242_81",
|
|
|
|
"GC/1.3.2r": "mwcc_242_81r",
|
|
|
|
"GC/2.0": "mwcc_247_92",
|
|
|
|
"GC/2.5": "mwcc_247_105",
|
|
|
|
"GC/2.6": "mwcc_247_107",
|
|
|
|
"GC/2.7": "mwcc_247_108",
|
|
|
|
"GC/3.0a3": "mwcc_41_51213",
|
|
|
|
"GC/3.0a3.2": "mwcc_41_60126",
|
|
|
|
"GC/3.0a3.3": "mwcc_41_60209",
|
|
|
|
"GC/3.0a3.4": "mwcc_42_60308",
|
|
|
|
"GC/3.0a5": "mwcc_42_60422",
|
|
|
|
"GC/3.0a5.2": "mwcc_41_60831",
|
|
|
|
"GC/3.0": "mwcc_41_60831",
|
|
|
|
"Wii/1.0RC1": "mwcc_42_140",
|
|
|
|
"Wii/0x4201_127": "mwcc_42_142",
|
|
|
|
"Wii/1.0a": "mwcc_42_142",
|
|
|
|
"Wii/1.0": "mwcc_43_145",
|
|
|
|
"Wii/1.1": "mwcc_43_151",
|
|
|
|
"Wii/1.3": "mwcc_43_172",
|
|
|
|
"Wii/1.5": "mwcc_43_188",
|
|
|
|
"Wii/1.6": "mwcc_43_202",
|
|
|
|
"Wii/1.7": "mwcc_43_213",
|
|
|
|
}
|
|
|
|
|
|
|
|
def add_unit(
|
|
|
|
build_obj: Dict[str, Any], module_name: str, progress_categories: List[str]
|
|
|
|
) -> None:
|
|
|
|
obj_path, obj_name = build_obj["object"], build_obj["name"]
|
|
|
|
base_object = Path(obj_name).with_suffix("")
|
2024-10-14 16:58:15 +00:00
|
|
|
name = str(Path(module_name) / base_object).replace(os.sep, "/")
|
2024-09-24 01:17:42 +00:00
|
|
|
unit_config: Dict[str, Any] = {
|
2024-10-14 16:58:15 +00:00
|
|
|
"name": name,
|
2024-09-24 01:17:42 +00:00
|
|
|
"target_path": obj_path,
|
2024-10-14 16:58:15 +00:00
|
|
|
"base_path": None,
|
|
|
|
"scratch": None,
|
2024-09-24 01:17:42 +00:00
|
|
|
"metadata": {
|
2024-10-14 16:58:15 +00:00
|
|
|
"complete": None,
|
|
|
|
"reverse_fn_order": None,
|
|
|
|
"source_path": None,
|
2024-09-24 01:17:42 +00:00
|
|
|
"progress_categories": progress_categories,
|
2024-10-14 16:58:15 +00:00
|
|
|
"auto_generated": build_obj["autogenerated"],
|
2024-09-24 01:17:42 +00:00
|
|
|
},
|
2024-10-14 16:58:15 +00:00
|
|
|
"symbol_mappings": None,
|
2024-09-24 01:17:42 +00:00
|
|
|
}
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Preserve existing symbol mappings
|
|
|
|
existing_unit = existing_units.get(name)
|
|
|
|
if existing_unit is not None:
|
|
|
|
unit_config["symbol_mappings"] = existing_unit.get("symbol_mappings")
|
|
|
|
|
2024-09-24 01:17:42 +00:00
|
|
|
obj = objects.get(obj_name)
|
|
|
|
if obj is None:
|
|
|
|
objdiff_config["units"].append(unit_config)
|
|
|
|
return
|
|
|
|
|
|
|
|
src_exists = obj.src_path is not None and obj.src_path.exists()
|
|
|
|
if src_exists:
|
|
|
|
unit_config["base_path"] = obj.src_obj_path
|
2024-10-14 16:58:15 +00:00
|
|
|
unit_config["metadata"]["source_path"] = obj.src_path
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
cflags = obj.options["cflags"]
|
|
|
|
reverse_fn_order = False
|
2024-10-14 16:58:15 +00:00
|
|
|
for flag in cflags:
|
|
|
|
if not flag.startswith("-inline "):
|
|
|
|
continue
|
|
|
|
for value in flag.split(" ")[1].split(","):
|
|
|
|
if value == "deferred":
|
|
|
|
reverse_fn_order = True
|
|
|
|
elif value == "nodeferred":
|
|
|
|
reverse_fn_order = False
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Filter out include directories
|
|
|
|
def keep_flag(flag):
|
|
|
|
return not flag.startswith("-i ") and not flag.startswith("-I ")
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
cflags = list(filter(keep_flag, cflags))
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
compiler_version = COMPILER_MAP.get(obj.options["mw_version"])
|
|
|
|
if compiler_version is None:
|
|
|
|
print(f"Missing scratch compiler mapping for {obj.options['mw_version']}")
|
|
|
|
else:
|
|
|
|
cflags_str = make_flags_str(cflags)
|
|
|
|
if obj.options["extra_cflags"] is not None:
|
|
|
|
extra_cflags_str = make_flags_str(obj.options["extra_cflags"])
|
|
|
|
cflags_str += " " + extra_cflags_str
|
|
|
|
unit_config["scratch"] = {
|
|
|
|
"platform": "gc_wii",
|
|
|
|
"compiler": compiler_version,
|
|
|
|
"c_flags": cflags_str,
|
2024-11-21 01:58:20 +00:00
|
|
|
"preset_id": obj.options["scratch_preset_id"],
|
2024-09-24 01:17:42 +00:00
|
|
|
}
|
|
|
|
if src_exists:
|
|
|
|
unit_config["scratch"].update(
|
|
|
|
{
|
|
|
|
"ctx_path": obj.ctx_path,
|
|
|
|
"build_ctx": True,
|
|
|
|
}
|
|
|
|
)
|
|
|
|
category_opt: List[str] | str = obj.options["progress_category"]
|
|
|
|
if isinstance(category_opt, list):
|
|
|
|
progress_categories.extend(category_opt)
|
|
|
|
elif category_opt is not None:
|
|
|
|
progress_categories.append(category_opt)
|
|
|
|
unit_config["metadata"].update(
|
|
|
|
{
|
|
|
|
"complete": obj.completed,
|
|
|
|
"reverse_fn_order": reverse_fn_order,
|
|
|
|
"progress_categories": progress_categories,
|
|
|
|
}
|
|
|
|
)
|
|
|
|
objdiff_config["units"].append(unit_config)
|
|
|
|
|
|
|
|
# Add DOL units
|
|
|
|
for unit in build_config["units"]:
|
|
|
|
progress_categories = []
|
|
|
|
# Only include a "dol" category if there are any modules
|
|
|
|
# Otherwise it's redundant with the global report measures
|
|
|
|
if len(build_config["modules"]) > 0:
|
|
|
|
progress_categories.append("dol")
|
|
|
|
add_unit(unit, build_config["name"], progress_categories)
|
|
|
|
|
|
|
|
# Add REL units
|
|
|
|
for module in build_config["modules"]:
|
|
|
|
for unit in module["units"]:
|
|
|
|
progress_categories = []
|
|
|
|
if config.progress_modules:
|
|
|
|
progress_categories.append("modules")
|
|
|
|
if config.progress_each_module:
|
|
|
|
progress_categories.append(module["name"])
|
|
|
|
add_unit(unit, module["name"], progress_categories)
|
|
|
|
|
|
|
|
# Add progress categories
|
|
|
|
def add_category(id: str, name: str):
|
|
|
|
objdiff_config["progress_categories"].append(
|
|
|
|
{
|
|
|
|
"id": id,
|
|
|
|
"name": name,
|
|
|
|
}
|
|
|
|
)
|
|
|
|
|
|
|
|
if len(build_config["modules"]) > 0:
|
|
|
|
add_category("dol", "DOL")
|
|
|
|
if config.progress_modules:
|
|
|
|
add_category("modules", "Modules")
|
|
|
|
if config.progress_each_module:
|
|
|
|
for module in build_config["modules"]:
|
|
|
|
add_category(module["name"], module["name"])
|
|
|
|
for category in config.progress_categories:
|
|
|
|
add_category(category.id, category.name)
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
def cleandict(d):
|
|
|
|
if isinstance(d, dict):
|
|
|
|
return {k: cleandict(v) for k, v in d.items() if v is not None}
|
|
|
|
elif isinstance(d, list):
|
|
|
|
return [cleandict(v) for v in d]
|
|
|
|
else:
|
|
|
|
return d
|
|
|
|
|
2024-09-24 01:17:42 +00:00
|
|
|
# Write objdiff.json
|
|
|
|
with open("objdiff.json", "w", encoding="utf-8") as w:
|
|
|
|
|
|
|
|
def unix_path(input: Any) -> str:
|
|
|
|
return str(input).replace(os.sep, "/") if input else ""
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
json.dump(cleandict(objdiff_config), w, indent=2, default=unix_path)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
def generate_compile_commands(
|
|
|
|
config: ProjectConfig,
|
|
|
|
objects: Dict[str, Object],
|
|
|
|
build_config: Optional[Dict[str, Any]],
|
|
|
|
) -> None:
|
|
|
|
if build_config is None or not config.generate_compile_commands:
|
2024-09-24 01:17:42 +00:00
|
|
|
return
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# The following code attempts to convert mwcc flags to clang flags
|
|
|
|
# for use with clangd.
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Flags to ignore explicitly
|
|
|
|
CFLAG_IGNORE: Set[str] = {
|
|
|
|
# Search order modifier
|
|
|
|
# Has a different meaning to Clang, and would otherwise
|
|
|
|
# be picked up by the include passthrough prefix
|
|
|
|
"-I-",
|
|
|
|
"-i-",
|
|
|
|
}
|
|
|
|
CFLAG_IGNORE_PREFIX: Tuple[str, ...] = tuple()
|
|
|
|
|
|
|
|
# Flags to replace
|
|
|
|
CFLAG_REPLACE: Dict[str, str] = {}
|
|
|
|
CFLAG_REPLACE_PREFIX: Tuple[Tuple[str, str], ...] = (
|
|
|
|
# Includes
|
|
|
|
("-i ", "-I"),
|
|
|
|
("-I ", "-I"),
|
|
|
|
("-I+", "-I"),
|
|
|
|
# Defines
|
|
|
|
("-d ", "-D"),
|
|
|
|
("-D ", "-D"),
|
|
|
|
("-D+", "-D"),
|
|
|
|
)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Flags with a finite set of options
|
|
|
|
CFLAG_REPLACE_OPTIONS: Tuple[Tuple[str, Dict[str, Tuple[str, ...]]], ...] = (
|
|
|
|
# Exceptions
|
|
|
|
(
|
|
|
|
"-Cpp_exceptions",
|
|
|
|
{
|
|
|
|
"off": ("-fno-cxx-exceptions",),
|
|
|
|
"on": ("-fcxx-exceptions",),
|
|
|
|
},
|
|
|
|
),
|
|
|
|
# RTTI
|
|
|
|
(
|
|
|
|
"-RTTI",
|
|
|
|
{
|
|
|
|
"off": ("-fno-rtti",),
|
|
|
|
"on": ("-frtti",),
|
|
|
|
},
|
|
|
|
),
|
|
|
|
# Language configuration
|
|
|
|
(
|
|
|
|
"-lang",
|
|
|
|
{
|
|
|
|
"c": ("--language=c", "--std=c89"),
|
|
|
|
"c99": ("--language=c", "--std=c99"),
|
|
|
|
"c++": ("--language=c++", "--std=c++98"),
|
|
|
|
"cplus": ("--language=c++", "--std=c++98"),
|
|
|
|
},
|
|
|
|
),
|
|
|
|
)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Flags to pass through
|
|
|
|
CFLAG_PASSTHROUGH: Set[str] = set()
|
|
|
|
CFLAG_PASSTHROUGH_PREFIX: Tuple[str, ...] = (
|
|
|
|
"-I", # includes
|
|
|
|
"-D", # defines
|
|
|
|
)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
clangd_config = []
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
def add_unit(build_obj: Dict[str, Any]) -> None:
|
|
|
|
obj = objects.get(build_obj["name"])
|
|
|
|
if obj is None:
|
|
|
|
return
|
|
|
|
|
|
|
|
# Skip unresolved objects
|
|
|
|
if (
|
|
|
|
obj.src_path is None
|
|
|
|
or obj.src_obj_path is None
|
|
|
|
or not file_is_c_cpp(obj.src_path)
|
|
|
|
):
|
|
|
|
return
|
|
|
|
|
|
|
|
# Gather cflags for source file
|
|
|
|
cflags: list[str] = []
|
|
|
|
|
|
|
|
def append_cflags(flags: Iterable[str]) -> None:
|
|
|
|
# Match a flag against either a set of concrete flags, or a set of prefixes.
|
|
|
|
def flag_match(
|
|
|
|
flag: str, concrete: Set[str], prefixes: Tuple[str, ...]
|
|
|
|
) -> bool:
|
|
|
|
if flag in concrete:
|
|
|
|
return True
|
|
|
|
|
|
|
|
for prefix in prefixes:
|
|
|
|
if flag.startswith(prefix):
|
|
|
|
return True
|
|
|
|
|
|
|
|
return False
|
|
|
|
|
|
|
|
# Determine whether a flag should be ignored.
|
|
|
|
def should_ignore(flag: str) -> bool:
|
|
|
|
return flag_match(flag, CFLAG_IGNORE, CFLAG_IGNORE_PREFIX)
|
|
|
|
|
|
|
|
# Determine whether a flag should be passed through.
|
|
|
|
def should_passthrough(flag: str) -> bool:
|
|
|
|
return flag_match(flag, CFLAG_PASSTHROUGH, CFLAG_PASSTHROUGH_PREFIX)
|
|
|
|
|
|
|
|
# Attempts replacement for the given flag.
|
|
|
|
def try_replace(flag: str) -> bool:
|
|
|
|
replacement = CFLAG_REPLACE.get(flag)
|
|
|
|
if replacement is not None:
|
|
|
|
cflags.append(replacement)
|
|
|
|
return True
|
|
|
|
|
|
|
|
for prefix, replacement in CFLAG_REPLACE_PREFIX:
|
|
|
|
if flag.startswith(prefix):
|
|
|
|
cflags.append(flag.replace(prefix, replacement, 1))
|
|
|
|
return True
|
|
|
|
|
|
|
|
for prefix, options in CFLAG_REPLACE_OPTIONS:
|
|
|
|
if not flag.startswith(prefix):
|
|
|
|
continue
|
|
|
|
|
|
|
|
# "-lang c99" and "-lang=c99" are both generally valid option forms
|
|
|
|
option = flag.removeprefix(prefix).removeprefix("=").lstrip()
|
|
|
|
replacements = options.get(option)
|
|
|
|
if replacements is not None:
|
|
|
|
cflags.extend(replacements)
|
|
|
|
|
|
|
|
return True
|
|
|
|
|
|
|
|
return False
|
|
|
|
|
|
|
|
for flag in flags:
|
|
|
|
# Ignore flags first
|
|
|
|
if should_ignore(flag):
|
|
|
|
continue
|
|
|
|
|
|
|
|
# Then find replacements
|
|
|
|
if try_replace(flag):
|
|
|
|
continue
|
|
|
|
|
|
|
|
# Pass flags through last
|
|
|
|
if should_passthrough(flag):
|
|
|
|
cflags.append(flag)
|
|
|
|
continue
|
|
|
|
|
|
|
|
append_cflags(obj.options["cflags"])
|
|
|
|
if isinstance(obj.options["extra_cflags"], list):
|
|
|
|
append_cflags(obj.options["extra_cflags"])
|
|
|
|
|
|
|
|
unit_config = {
|
|
|
|
"directory": Path.cwd(),
|
|
|
|
"file": obj.src_path,
|
|
|
|
"output": obj.src_obj_path,
|
|
|
|
"arguments": [
|
|
|
|
"clang",
|
|
|
|
"-nostdinc",
|
|
|
|
"-fno-builtin",
|
|
|
|
"--target=powerpc-eabi",
|
|
|
|
*cflags,
|
|
|
|
"-c",
|
|
|
|
obj.src_path,
|
|
|
|
"-o",
|
|
|
|
obj.src_obj_path,
|
|
|
|
],
|
|
|
|
}
|
|
|
|
clangd_config.append(unit_config)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
# Add DOL units
|
|
|
|
for unit in build_config["units"]:
|
2024-10-14 16:58:15 +00:00
|
|
|
add_unit(unit)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
# Add REL units
|
|
|
|
for module in build_config["modules"]:
|
|
|
|
for unit in module["units"]:
|
2024-10-14 16:58:15 +00:00
|
|
|
add_unit(unit)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Write compile_commands.json
|
|
|
|
with open("compile_commands.json", "w", encoding="utf-8") as w:
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
def default_format(o):
|
|
|
|
if isinstance(o, Path):
|
|
|
|
return o.resolve().as_posix()
|
|
|
|
return str(o)
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
json.dump(clangd_config, w, indent=2, default=default_format)
|
|
|
|
|
|
|
|
|
|
|
|
# Calculate, print and write progress to progress.json
|
|
|
|
def calculate_progress(config: ProjectConfig) -> None:
|
|
|
|
config.validate()
|
|
|
|
out_path = config.out_path()
|
|
|
|
report_path = out_path / "report.json"
|
|
|
|
if not report_path.is_file():
|
|
|
|
sys.exit(f"Report file {report_path} does not exist")
|
|
|
|
|
|
|
|
report_data: Dict[str, Any] = {}
|
|
|
|
with open(report_path, "r", encoding="utf-8") as f:
|
|
|
|
report_data = json.load(f)
|
|
|
|
|
|
|
|
# Convert string numbers (u64) to int
|
|
|
|
def convert_numbers(data: Dict[str, Any]) -> None:
|
|
|
|
for key, value in data.items():
|
|
|
|
if isinstance(value, str) and value.isdigit():
|
|
|
|
data[key] = int(value)
|
|
|
|
|
|
|
|
convert_numbers(report_data["measures"])
|
|
|
|
for category in report_data["categories"]:
|
|
|
|
convert_numbers(category["measures"])
|
|
|
|
|
|
|
|
# Output to GitHub Actions job summary, if available
|
|
|
|
summary_path = os.getenv("GITHUB_STEP_SUMMARY")
|
|
|
|
summary_file: Optional[IO[str]] = None
|
|
|
|
if summary_path:
|
|
|
|
summary_file = open(summary_path, "a", encoding="utf-8")
|
|
|
|
summary_file.write("```\n")
|
|
|
|
|
|
|
|
def progress_print(s: str) -> None:
|
|
|
|
print(s)
|
|
|
|
if summary_file:
|
|
|
|
summary_file.write(s + "\n")
|
|
|
|
|
|
|
|
# Print human-readable progress
|
|
|
|
progress_print("Progress:")
|
|
|
|
|
|
|
|
def print_category(name: str, measures: Dict[str, Any]) -> None:
|
|
|
|
total_code = measures.get("total_code", 0)
|
|
|
|
matched_code = measures.get("matched_code", 0)
|
|
|
|
matched_code_percent = measures.get("matched_code_percent", 0)
|
|
|
|
total_data = measures.get("total_data", 0)
|
|
|
|
matched_data = measures.get("matched_data", 0)
|
|
|
|
matched_data_percent = measures.get("matched_data_percent", 0)
|
|
|
|
total_functions = measures.get("total_functions", 0)
|
|
|
|
matched_functions = measures.get("matched_functions", 0)
|
|
|
|
complete_code_percent = measures.get("complete_code_percent", 0)
|
|
|
|
total_units = measures.get("total_units", 0)
|
|
|
|
complete_units = measures.get("complete_units", 0)
|
|
|
|
|
|
|
|
progress_print(
|
|
|
|
f" {name}: {matched_code_percent:.2f}% matched, {complete_code_percent:.2f}% linked ({complete_units} / {total_units} files)"
|
|
|
|
)
|
|
|
|
progress_print(
|
|
|
|
f" Code: {matched_code} / {total_code} bytes ({matched_functions} / {total_functions} functions)"
|
|
|
|
)
|
|
|
|
progress_print(
|
|
|
|
f" Data: {matched_data} / {total_data} bytes ({matched_data_percent:.2f}%)"
|
2024-09-24 01:17:42 +00:00
|
|
|
)
|
2024-10-14 16:58:15 +00:00
|
|
|
|
|
|
|
print_category("All", report_data["measures"])
|
|
|
|
for category in report_data["categories"]:
|
|
|
|
if config.print_progress_categories is True or (
|
|
|
|
isinstance(config.print_progress_categories, list)
|
|
|
|
and category["id"] in config.print_progress_categories
|
|
|
|
):
|
|
|
|
print_category(category["name"], category["measures"])
|
2024-09-24 01:17:42 +00:00
|
|
|
|
|
|
|
if config.progress_use_fancy:
|
2024-10-14 16:58:15 +00:00
|
|
|
measures = report_data["measures"]
|
|
|
|
total_code = measures.get("total_code", 0)
|
|
|
|
total_data = measures.get("total_data", 0)
|
|
|
|
if total_code == 0 or total_data == 0:
|
2024-09-24 01:17:42 +00:00
|
|
|
return
|
2024-10-14 16:58:15 +00:00
|
|
|
code_frac = measures.get("complete_code", 0) / total_code
|
|
|
|
data_frac = measures.get("complete_data", 0) / total_data
|
2024-09-24 01:17:42 +00:00
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
progress_print(
|
2024-09-24 01:17:42 +00:00
|
|
|
"\nYou have {} out of {} {} and {} out of {} {}.".format(
|
|
|
|
math.floor(code_frac * config.progress_code_fancy_frac),
|
|
|
|
config.progress_code_fancy_frac,
|
|
|
|
config.progress_code_fancy_item,
|
|
|
|
math.floor(data_frac * config.progress_data_fancy_frac),
|
|
|
|
config.progress_data_fancy_frac,
|
|
|
|
config.progress_data_fancy_item,
|
|
|
|
)
|
|
|
|
)
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
# Finalize GitHub Actions job summary
|
|
|
|
if summary_file:
|
|
|
|
summary_file.write("```\n")
|
|
|
|
summary_file.close()
|
|
|
|
|
2024-09-24 01:17:42 +00:00
|
|
|
# Generate and write progress.json
|
|
|
|
progress_json: Dict[str, Any] = {}
|
2024-10-14 16:58:15 +00:00
|
|
|
|
|
|
|
def add_category(id: str, measures: Dict[str, Any]) -> None:
|
2024-09-24 01:17:42 +00:00
|
|
|
progress_json[id] = {
|
2024-10-14 16:58:15 +00:00
|
|
|
"code": measures.get("complete_code", 0),
|
|
|
|
"code/total": measures.get("total_code", 0),
|
|
|
|
"data": measures.get("complete_data", 0),
|
|
|
|
"data/total": measures.get("total_data", 0),
|
|
|
|
"matched_code": measures.get("matched_code", 0),
|
|
|
|
"matched_code/total": measures.get("total_code", 0),
|
|
|
|
"matched_data": measures.get("matched_data", 0),
|
|
|
|
"matched_data/total": measures.get("total_data", 0),
|
|
|
|
"matched_functions": measures.get("matched_functions", 0),
|
|
|
|
"matched_functions/total": measures.get("total_functions", 0),
|
|
|
|
"fuzzy_match": int(measures.get("fuzzy_match_percent", 0) * 100),
|
|
|
|
"fuzzy_match/total": 10000,
|
|
|
|
"units": measures.get("complete_units", 0),
|
|
|
|
"units/total": measures.get("total_units", 0),
|
2024-09-24 01:17:42 +00:00
|
|
|
}
|
|
|
|
|
2024-10-14 16:58:15 +00:00
|
|
|
if config.progress_all:
|
|
|
|
add_category("all", report_data["measures"])
|
|
|
|
else:
|
|
|
|
# Support for old behavior where "dol" was the main category
|
|
|
|
add_category("dol", report_data["measures"])
|
|
|
|
for category in report_data["categories"]:
|
|
|
|
add_category(category["id"], category["measures"])
|
|
|
|
|
2024-09-24 01:17:42 +00:00
|
|
|
with open(out_path / "progress.json", "w", encoding="utf-8") as w:
|
2024-10-14 16:58:15 +00:00
|
|
|
json.dump(progress_json, w, indent=2)
|