mirror of
https://github.com/doldecomp/melee.git
synced 2024-11-26 23:00:28 +00:00
Merge remote-tracking branch 'dtk-template/main' into wip/dtk-template
This commit is contained in:
commit
bbb527808f
30
.github/workflows/build.yml
vendored
30
.github/workflows/build.yml
vendored
@ -103,28 +103,27 @@ jobs:
|
||||
|
||||
mode='${{matrix.mode}}'
|
||||
config_args='--compilers /compilers --max-errors 0 --verbose --version ${{matrix.version}}'
|
||||
case "$mode" in
|
||||
'link')
|
||||
config_args="--map --require-protos $config_args"
|
||||
;;
|
||||
'diff') ;;
|
||||
*) exit 1 ;;
|
||||
esac
|
||||
|
||||
function run_configure {
|
||||
echo "$config_args" | xargs python configure.py
|
||||
}
|
||||
|
||||
case "$mode" in
|
||||
'link')
|
||||
config_args="--map --require-protos $config_args"
|
||||
run_configure
|
||||
ninja || ninja diff
|
||||
;;
|
||||
'diff')
|
||||
root='build/${{matrix.version}}'
|
||||
expected="$root/expected-report.json"
|
||||
actual="$root/report.json"
|
||||
current="$root/report.json"
|
||||
changes="$root/changes-report.json"
|
||||
head=$(git rev-parse HEAD)
|
||||
head="$(git rev-parse HEAD)"
|
||||
|
||||
ninja all_source
|
||||
objdiff-cli report generate -o "$actual"
|
||||
function run_ninja {
|
||||
ninja all_source "$current"
|
||||
}
|
||||
|
||||
case '${{github.event_name}}' in
|
||||
'pull_request')
|
||||
@ -137,11 +136,14 @@ jobs:
|
||||
*) exit 1 ;;
|
||||
esac
|
||||
|
||||
ninja all_source
|
||||
objdiff-cli report generate -o "$expected"
|
||||
run_configure
|
||||
run_ninja
|
||||
cp "$current" "$expected"
|
||||
|
||||
git checkout "$head"
|
||||
objdiff-cli report changes -o "$changes" "$expected" "$actual"
|
||||
run_ninja
|
||||
|
||||
build/tools/objdiff-cli report changes -o "$changes" "$expected" "$current"
|
||||
python tools/diff_changes.py "$changes"
|
||||
status="$?"
|
||||
exit "$status"
|
||||
|
1
.gitignore
vendored
1
.gitignore
vendored
@ -39,6 +39,7 @@ build/
|
||||
*.map
|
||||
*.dol
|
||||
*.pdb
|
||||
/result
|
||||
|
||||
# Temporary files
|
||||
*.swp
|
||||
|
@ -7,13 +7,13 @@
|
||||
|
||||
rustPlatform.buildRustPackage rec {
|
||||
pname = "decomp-toolkit";
|
||||
version = "0.9.3";
|
||||
version = "0.9.6";
|
||||
|
||||
src = fetchFromGitHub {
|
||||
owner = "encounter";
|
||||
repo = "decomp-toolkit";
|
||||
rev = "v${version}";
|
||||
hash = "sha256-5EWZwAQNso58WIWKtXNiDDBjMUYYArdRREtyD0bXurY=";
|
||||
hash = "sha256-mgdohsZ0ZkTSg/UsGVBPmtlJFa1rKMG6aa4XG1S9F4Y=";
|
||||
};
|
||||
|
||||
nativeBuildInputs = [
|
||||
@ -22,7 +22,6 @@ rustPlatform.buildRustPackage rec {
|
||||
|
||||
cargoLock.lockFile = "${src}/Cargo.lock";
|
||||
cargoLock.outputHashes."ar-0.8.0" = "sha256-OLyo+cRRWMsI1i8NsgsBKRJH1XsKW1CculQnJ/wcya0=";
|
||||
cargoLock.outputHashes."nod-1.2.0" = "sha256-M7jSBo1Dqrhy/F0osoUtTMNm2BkFnRy2MOmkTs1pvdM=";
|
||||
cargoLock.outputHashes."objdiff-core-2.0.0-alpha.3" = "sha256-E597zRlSpxrTGino7jqoQmyxWkLYXT1P6U2PRolm0Ek=";
|
||||
cargoLock.outputHashes."ppc750cl-0.3.0" = "sha256-nMJk+rgu7Ydi2VZfodJk0kBz9xLLVBVz0vEZPee8Q6M=";
|
||||
|
||||
|
64
configure.py
64
configure.py
@ -18,8 +18,9 @@ from pathlib import Path
|
||||
from typing import Iterator, List, Optional
|
||||
|
||||
from tools.project import (
|
||||
LibDict,
|
||||
Library,
|
||||
Object,
|
||||
ProgressCategory,
|
||||
ProjectConfig,
|
||||
calculate_progress,
|
||||
generate_build,
|
||||
@ -70,12 +71,6 @@ parser.add_argument(
|
||||
action="store_true",
|
||||
help="generate map file(s)",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--use-asm",
|
||||
dest="no_asm",
|
||||
action="store_false",
|
||||
help="incorporate .s files from asm directory",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--debug",
|
||||
action="store_true",
|
||||
@ -94,6 +89,12 @@ parser.add_argument(
|
||||
type=Path,
|
||||
help="path to decomp-toolkit binary or source (optional)",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--objdiff",
|
||||
metavar="BINARY | DIR",
|
||||
type=Path,
|
||||
help="path to objdiff-cli binary or source (optional)",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--sjiswrap",
|
||||
metavar="EXE",
|
||||
@ -150,21 +151,23 @@ version_num = VERSIONS.index(config.version)
|
||||
# Apply arguments
|
||||
config.build_dir = args.build_dir
|
||||
config.dtk_path = args.dtk
|
||||
config.objdiff_path = args.objdiff
|
||||
config.binutils_path = args.binutils
|
||||
config.compilers_path = args.compilers
|
||||
config.debug = args.debug
|
||||
config.generate_map = args.map
|
||||
config.non_matching = args.non_matching
|
||||
config.sjiswrap_path = args.sjiswrap
|
||||
if not is_windows():
|
||||
config.wrapper = args.wrapper
|
||||
if args.no_asm:
|
||||
# Don't build asm unless we're --non-matching
|
||||
if not config.non_matching:
|
||||
config.asm_dir = None
|
||||
|
||||
# Tool versions
|
||||
config.binutils_tag = "2.42-1"
|
||||
config.compilers_tag = "20231018"
|
||||
config.dtk_tag = "v0.9.3"
|
||||
config.compilers_tag = "20240706"
|
||||
config.dtk_tag = "v0.9.6"
|
||||
config.objdiff_tag = "v2.0.0-beta.6"
|
||||
config.sjiswrap_tag = "v1.1.1"
|
||||
config.wibo_tag = "0.6.11"
|
||||
|
||||
@ -183,6 +186,12 @@ config.ldflags = [
|
||||
"-nodefaults",
|
||||
"-warn off",
|
||||
]
|
||||
if args.debug:
|
||||
config.ldflags.append("-g") # Or -gdwarf-2 for Wii linkers
|
||||
if args.map:
|
||||
config.ldflags.append("-mapunused")
|
||||
# config.ldflags.append("-listclosure") # For Wii linkers
|
||||
|
||||
# Use for any additional files that should cause a re-configure when modified
|
||||
config.reconfig_deps = []
|
||||
|
||||
@ -214,7 +223,9 @@ cflags_base = [
|
||||
f"-DVERSION={version_num}",
|
||||
]
|
||||
|
||||
if config.debug:
|
||||
# Debug flags
|
||||
if args.debug:
|
||||
# Or -sym dwarf-2 for Wii compilers
|
||||
cflags_base.extend(["-sym on", "-DDEBUG=1"])
|
||||
else:
|
||||
cflags_base.append("-DNDEBUG=1")
|
||||
@ -268,7 +279,8 @@ def Lib(
|
||||
includes: List[str] = includes_base,
|
||||
system_includes: List[str] = system_includes_base,
|
||||
src_dir: Optional[str] = None,
|
||||
) -> LibDict:
|
||||
category: Optional[str] = None,
|
||||
) -> Library:
|
||||
def make_includes(includes: List[str]) -> Iterator[str]:
|
||||
return map(lambda s: f"-i {s}", includes)
|
||||
|
||||
@ -282,6 +294,7 @@ def Lib(
|
||||
*make_includes(system_includes),
|
||||
],
|
||||
"host": False,
|
||||
"progress_category": category,
|
||||
"objects": objects,
|
||||
}
|
||||
|
||||
@ -293,7 +306,7 @@ def Lib(
|
||||
|
||||
def DolphinLib(
|
||||
lib_name: str, objects: Objects, fix_epilogue=False, extern=False
|
||||
) -> LibDict:
|
||||
) -> Library:
|
||||
if extern:
|
||||
cflags = [
|
||||
"-c",
|
||||
@ -332,10 +345,11 @@ def DolphinLib(
|
||||
cflags=cflags,
|
||||
includes=includes,
|
||||
system_includes=system_includes,
|
||||
category="sdk",
|
||||
)
|
||||
|
||||
|
||||
def SysdolphinLib(lib_name: str, objects: Objects) -> LibDict:
|
||||
def SysdolphinLib(lib_name: str, objects: Objects) -> Library:
|
||||
return Lib(
|
||||
lib_name,
|
||||
objects,
|
||||
@ -347,10 +361,11 @@ def SysdolphinLib(lib_name: str, objects: Objects) -> LibDict:
|
||||
*system_includes_base,
|
||||
"src/dolphin",
|
||||
],
|
||||
category="hsd",
|
||||
)
|
||||
|
||||
|
||||
def MeleeLib(lib_name: str, objects: Objects) -> LibDict:
|
||||
def MeleeLib(lib_name: str, objects: Objects) -> Library:
|
||||
return Lib(
|
||||
lib_name,
|
||||
objects,
|
||||
@ -364,15 +379,17 @@ def MeleeLib(lib_name: str, objects: Objects) -> LibDict:
|
||||
"src/dolphin",
|
||||
"src/sysdolphin",
|
||||
],
|
||||
category="game",
|
||||
)
|
||||
|
||||
|
||||
def RuntimeLib(lib_name: str, objects: Objects) -> LibDict:
|
||||
def RuntimeLib(lib_name: str, objects: Objects) -> Library:
|
||||
return Lib(
|
||||
lib_name,
|
||||
objects,
|
||||
cflags=cflags_runtime,
|
||||
fix_epilogue=False,
|
||||
category="runtime",
|
||||
)
|
||||
|
||||
|
||||
@ -1472,13 +1489,22 @@ config.libs = [
|
||||
),
|
||||
]
|
||||
|
||||
# Optional extra categories for progress tracking
|
||||
# Adjust as desired for your project
|
||||
config.progress_categories = [
|
||||
ProgressCategory("game", "Game Code"),
|
||||
ProgressCategory("hsd", "HSD Code"),
|
||||
ProgressCategory("sdk", "Dolphin SDK Code"),
|
||||
ProgressCategory("runtime", "Gekko Runtime Code"),
|
||||
]
|
||||
config.progress_all = False
|
||||
config.progress_each_module = args.verbose
|
||||
|
||||
if args.mode == "configure":
|
||||
# Write build.ninja and objdiff.json
|
||||
generate_build(config)
|
||||
elif args.mode == "progress":
|
||||
# Print progress and write progress.json
|
||||
config.progress_each_module = args.verbose
|
||||
config.progress_all = False
|
||||
calculate_progress(config)
|
||||
else:
|
||||
sys.exit("Unknown mode: " + args.mode)
|
||||
|
@ -23,8 +23,8 @@ include_dirs = [
|
||||
# Add additional include directories here
|
||||
]
|
||||
|
||||
include_pattern = re.compile(r'^#include\s*[<"](.+?)[>"]$')
|
||||
guard_pattern = re.compile(r"^#ifndef\s+(.*)$")
|
||||
include_pattern = re.compile(r'^#\s*include\s*[<"](.+?)[>"]$')
|
||||
guard_pattern = re.compile(r"^#\s*ifndef\s+(.*)$")
|
||||
|
||||
defines = set()
|
||||
|
||||
|
@ -56,6 +56,22 @@ def dtk_url(tag: str) -> str:
|
||||
return f"{repo}/releases/download/{tag}/dtk-{system}-{arch}{suffix}"
|
||||
|
||||
|
||||
def objdiff_cli_url(tag: str) -> str:
|
||||
uname = platform.uname()
|
||||
suffix = ""
|
||||
system = uname.system.lower()
|
||||
if system == "darwin":
|
||||
system = "macos"
|
||||
elif system == "windows":
|
||||
suffix = ".exe"
|
||||
arch = uname.machine.lower()
|
||||
if arch == "amd64":
|
||||
arch = "x86_64"
|
||||
|
||||
repo = "https://github.com/encounter/objdiff"
|
||||
return f"{repo}/releases/download/{tag}/objdiff-cli-{system}-{arch}{suffix}"
|
||||
|
||||
|
||||
def sjiswrap_url(tag: str) -> str:
|
||||
repo = "https://github.com/encounter/sjiswrap"
|
||||
return f"{repo}/releases/download/{tag}/sjiswrap-windows-x86.exe"
|
||||
@ -70,6 +86,7 @@ TOOLS: Dict[str, Callable[[str], str]] = {
|
||||
"binutils": binutils_url,
|
||||
"compilers": compilers_url,
|
||||
"dtk": dtk_url,
|
||||
"objdiff-cli": objdiff_cli_url,
|
||||
"sjiswrap": sjiswrap_url,
|
||||
"wibo": wibo_url,
|
||||
}
|
||||
|
570
tools/project.py
570
tools/project.py
@ -17,7 +17,7 @@ import os
|
||||
import platform
|
||||
import sys
|
||||
from pathlib import Path
|
||||
from typing import Any, Dict, List, Optional, Set, Tuple, Union
|
||||
from typing import Any, Dict, List, Optional, Set, Tuple, Union, cast
|
||||
|
||||
from . import ninja_syntax
|
||||
from .ninja_syntax import serialize_path
|
||||
@ -29,26 +29,78 @@ if sys.platform == "cygwin":
|
||||
f"\n(Current path: {sys.executable})"
|
||||
)
|
||||
|
||||
LibDict = Dict[str, Any]
|
||||
Library = Dict[str, Any]
|
||||
|
||||
|
||||
class Object:
|
||||
def __init__(self, completed: bool, name: str, **options: Any) -> None:
|
||||
self.name = name
|
||||
self.base_name = Path(name).with_suffix("")
|
||||
self.completed = completed
|
||||
self.options: Dict[str, Any] = {
|
||||
"add_to_all": True,
|
||||
"add_to_all": None,
|
||||
"asflags": None,
|
||||
"extra_asflags": None,
|
||||
"asm_dir": None,
|
||||
"cflags": None,
|
||||
"extra_asflags": None,
|
||||
"extra_cflags": None,
|
||||
"host": None,
|
||||
"lib": None,
|
||||
"mw_version": None,
|
||||
"progress_category": None,
|
||||
"shift_jis": None,
|
||||
"source": name,
|
||||
"src_dir": None,
|
||||
}
|
||||
self.options.update(options)
|
||||
|
||||
# Internal
|
||||
self.src_path: Optional[Path] = None
|
||||
self.asm_path: Optional[Path] = None
|
||||
self.src_obj_path: Optional[Path] = None
|
||||
self.asm_obj_path: Optional[Path] = None
|
||||
self.host_obj_path: Optional[Path] = None
|
||||
self.ctx_path: Optional[Path] = None
|
||||
|
||||
def resolve(self, config: "ProjectConfig", lib: Library) -> "Object":
|
||||
# Use object options, then library options
|
||||
obj = Object(self.completed, self.name, **lib)
|
||||
for key, value in self.options.items():
|
||||
if value is not None or key not in obj.options:
|
||||
obj.options[key] = value
|
||||
|
||||
# Use default options from config
|
||||
def set_default(key: str, value: Any) -> None:
|
||||
if obj.options[key] is None:
|
||||
obj.options[key] = value
|
||||
|
||||
set_default("add_to_all", True)
|
||||
set_default("asflags", config.asflags)
|
||||
set_default("asm_dir", config.asm_dir)
|
||||
set_default("host", False)
|
||||
set_default("mw_version", config.linker_version)
|
||||
set_default("shift_jis", config.shift_jis)
|
||||
set_default("src_dir", config.src_dir)
|
||||
|
||||
# Resolve paths
|
||||
build_dir = config.out_path()
|
||||
obj.src_path = Path(obj.options["src_dir"]) / obj.options["source"]
|
||||
if obj.options["asm_dir"] is not None:
|
||||
obj.asm_path = (
|
||||
Path(obj.options["asm_dir"]) / obj.options["source"]
|
||||
).with_suffix(".s")
|
||||
base_name = Path(self.name).with_suffix("")
|
||||
obj.src_obj_path = build_dir / "src" / f"{base_name}.o"
|
||||
obj.asm_obj_path = build_dir / "mod" / f"{base_name}.o"
|
||||
obj.host_obj_path = build_dir / "host" / f"{base_name}.o"
|
||||
obj.ctx_path = build_dir / "src" / f"{base_name}.ctx"
|
||||
return obj
|
||||
|
||||
|
||||
class ProgressCategory:
|
||||
def __init__(self, id: str, name: str) -> None:
|
||||
self.id = id
|
||||
self.name = name
|
||||
|
||||
|
||||
class ProjectConfig:
|
||||
def __init__(self) -> None:
|
||||
@ -71,17 +123,18 @@ class ProjectConfig:
|
||||
self.wrapper: Optional[Path] = None # If None, download wibo on Linux
|
||||
self.sjiswrap_tag: Optional[str] = None # Git tag
|
||||
self.sjiswrap_path: Optional[Path] = None # If None, download
|
||||
self.objdiff_tag: Optional[str] = None # Git tag
|
||||
self.objdiff_path: Optional[Path] = None # If None, download
|
||||
|
||||
# Project config
|
||||
self.non_matching: bool = False
|
||||
self.build_rels: bool = True # Build REL files
|
||||
self.check_sha_path: Optional[Path] = None # Path to version.sha1
|
||||
self.config_path: Optional[Path] = None # Path to config.yml
|
||||
self.debug: bool = False # Build with debug info
|
||||
self.generate_map: bool = False # Generate map file(s)
|
||||
self.asflags: Optional[List[str]] = None # Assembler flags
|
||||
self.ldflags: Optional[List[str]] = None # Linker flags
|
||||
self.libs: Optional[List[Dict[str, Any]]] = None # List of libraries
|
||||
self.libs: Optional[List[Library]] = None # List of libraries
|
||||
self.linker_version: Optional[str] = None # mwld version
|
||||
self.version: Optional[str] = None # Version name
|
||||
self.warn_missing_config: bool = False # Warn on missing unit configuration
|
||||
@ -103,12 +156,13 @@ class ProjectConfig:
|
||||
None # Custom build steps, types are ["pre-compile", "post-compile", "post-link", "post-build"]
|
||||
)
|
||||
|
||||
# Progress output and progress.json config
|
||||
# Progress output, progress.json and report.json config
|
||||
self.progress_all: bool = True # Include combined "all" category
|
||||
self.progress_modules: bool = True # Include combined "modules" category
|
||||
self.progress_each_module: bool = (
|
||||
True # Include individual modules, disable for large numbers of modules
|
||||
False # Include individual modules, disable for large numbers of modules
|
||||
)
|
||||
self.progress_categories: List[ProgressCategory] = [] # Additional categories
|
||||
|
||||
# Progress fancy printing
|
||||
self.progress_use_fancy: bool = False
|
||||
@ -133,12 +187,17 @@ class ProjectConfig:
|
||||
if getattr(self, attr) is None:
|
||||
sys.exit(f"ProjectConfig.{attr} missing")
|
||||
|
||||
def find_object(self, name: str) -> Optional[Tuple[Dict[str, Any], Object]]:
|
||||
# Creates a map of object names to Object instances
|
||||
# Options are fully resolved from the library and object
|
||||
def objects(self) -> Dict[str, Object]:
|
||||
out = {}
|
||||
for lib in self.libs or {}:
|
||||
for obj in lib["objects"]:
|
||||
if obj.name == name:
|
||||
return lib, obj
|
||||
return None
|
||||
objects: List[Object] = lib["objects"]
|
||||
for obj in objects:
|
||||
if obj.name in out:
|
||||
sys.exit(f"Duplicate object name {obj.name}")
|
||||
out[obj.name] = obj.resolve(self, lib)
|
||||
return out
|
||||
|
||||
def out_path(self) -> Path:
|
||||
return self.build_dir / str(self.version)
|
||||
@ -174,7 +233,7 @@ def load_build_config(
|
||||
f = open(build_config_path, "r", encoding="utf-8")
|
||||
build_config: Dict[str, Any] = json.load(f)
|
||||
config_version = build_config.get("version")
|
||||
if not config_version:
|
||||
if config_version is None:
|
||||
# Invalid config.json
|
||||
f.close()
|
||||
os.remove(build_config_path)
|
||||
@ -193,17 +252,19 @@ def load_build_config(
|
||||
|
||||
# Generate build.ninja and objdiff.json
|
||||
def generate_build(config: ProjectConfig) -> None:
|
||||
config.validate()
|
||||
objects = config.objects()
|
||||
build_config = load_build_config(config, config.out_path() / "config.json")
|
||||
generate_build_ninja(config, build_config)
|
||||
generate_objdiff_config(config, build_config)
|
||||
generate_build_ninja(config, objects, build_config)
|
||||
generate_objdiff_config(config, objects, build_config)
|
||||
|
||||
|
||||
# Generate build.ninja
|
||||
def generate_build_ninja(
|
||||
config: ProjectConfig, build_config: Optional[Dict[str, Any]]
|
||||
config: ProjectConfig,
|
||||
objects: Dict[str, Object],
|
||||
build_config: Optional[Dict[str, Any]],
|
||||
) -> None:
|
||||
config.validate()
|
||||
|
||||
out = io.StringIO()
|
||||
n = ninja_syntax.Writer(out)
|
||||
n.variable("ninja_required_version", "1.3")
|
||||
@ -221,13 +282,8 @@ def generate_build_ninja(
|
||||
# Variables
|
||||
###
|
||||
n.comment("Variables")
|
||||
ldflags = " ".join(config.ldflags or [])
|
||||
if config.generate_map:
|
||||
ldflags += " -mapunused"
|
||||
if config.debug:
|
||||
ldflags += " -g"
|
||||
n.variable("ldflags", ldflags)
|
||||
if not config.linker_version:
|
||||
n.variable("ldflags", " ".join(config.ldflags or []))
|
||||
if config.linker_version is None:
|
||||
sys.exit("ProjectConfig.linker_version missing")
|
||||
n.variable("mw_version", Path(config.linker_version))
|
||||
n.newline()
|
||||
@ -239,6 +295,7 @@ def generate_build_ninja(
|
||||
|
||||
build_path = config.out_path()
|
||||
progress_path = build_path / "progress.json"
|
||||
report_path = build_path / "report.json"
|
||||
build_tools_path = config.build_dir / "tools"
|
||||
download_tool = config.tools_dir / "download_tool.py"
|
||||
n.rule(
|
||||
@ -256,17 +313,27 @@ def generate_build_ninja(
|
||||
deps="gcc",
|
||||
)
|
||||
|
||||
if config.dtk_path is not None and config.dtk_path.is_file():
|
||||
dtk = config.dtk_path
|
||||
elif config.dtk_path is not None:
|
||||
dtk = build_tools_path / "release" / f"dtk{EXE}"
|
||||
cargo_rule_written = False
|
||||
|
||||
def write_cargo_rule():
|
||||
nonlocal cargo_rule_written
|
||||
if not cargo_rule_written:
|
||||
n.pool("cargo", 1)
|
||||
n.rule(
|
||||
name="cargo",
|
||||
command="cargo build --release --manifest-path $in --bin $bin --target-dir $target",
|
||||
description="CARGO $bin",
|
||||
pool="cargo",
|
||||
depfile=Path("$target") / "release" / "$bin.d",
|
||||
deps="gcc",
|
||||
)
|
||||
cargo_rule_written = True
|
||||
|
||||
if config.dtk_path is not None and config.dtk_path.is_file():
|
||||
dtk = config.dtk_path
|
||||
elif config.dtk_path is not None:
|
||||
dtk = build_tools_path / "release" / f"dtk{EXE}"
|
||||
write_cargo_rule()
|
||||
n.build(
|
||||
outputs=dtk,
|
||||
rule="cargo",
|
||||
@ -291,6 +358,35 @@ def generate_build_ninja(
|
||||
else:
|
||||
sys.exit("ProjectConfig.dtk_tag missing")
|
||||
|
||||
if config.objdiff_path is not None and config.objdiff_path.is_file():
|
||||
objdiff = config.objdiff_path
|
||||
elif config.objdiff_path is not None:
|
||||
objdiff = build_tools_path / "release" / f"objdiff-cli{EXE}"
|
||||
write_cargo_rule()
|
||||
n.build(
|
||||
outputs=objdiff,
|
||||
rule="cargo",
|
||||
inputs=config.objdiff_path / "Cargo.toml",
|
||||
implicit=config.objdiff_path / "Cargo.lock",
|
||||
variables={
|
||||
"bin": "objdiff-cli",
|
||||
"target": build_tools_path,
|
||||
},
|
||||
)
|
||||
elif config.objdiff_tag:
|
||||
objdiff = build_tools_path / f"objdiff-cli{EXE}"
|
||||
n.build(
|
||||
outputs=objdiff,
|
||||
rule="download_tool",
|
||||
implicit=download_tool,
|
||||
variables={
|
||||
"tool": "objdiff-cli",
|
||||
"tag": config.objdiff_tag,
|
||||
},
|
||||
)
|
||||
else:
|
||||
sys.exit("ProjectConfig.objdiff_tag missing")
|
||||
|
||||
if config.sjiswrap_path:
|
||||
sjiswrap = config.sjiswrap_path
|
||||
elif config.sjiswrap_tag:
|
||||
@ -369,6 +465,17 @@ def generate_build_ninja(
|
||||
|
||||
n.newline()
|
||||
|
||||
###
|
||||
# Helper rule for downloading all tools
|
||||
###
|
||||
n.comment("Download all tools")
|
||||
n.build(
|
||||
outputs="tools",
|
||||
rule="phony",
|
||||
inputs=[dtk, sjiswrap, wrapper, compilers, binutils, objdiff],
|
||||
)
|
||||
n.newline()
|
||||
|
||||
###
|
||||
# Build rules
|
||||
###
|
||||
@ -451,11 +558,12 @@ def generate_build_ninja(
|
||||
)
|
||||
n.newline()
|
||||
|
||||
if len(config.custom_build_rules or {}) > 0:
|
||||
n.comment("Custom project build rules (pre/post-processing)")
|
||||
for rule in config.custom_build_rules or {}:
|
||||
n.rule(
|
||||
name=rule.get("name"),
|
||||
command=rule.get("command"),
|
||||
name=cast(str, rule.get("name")),
|
||||
command=cast(str, rule.get("command")),
|
||||
description=rule.get("description", None),
|
||||
depfile=rule.get("depfile", None),
|
||||
generator=rule.get("generator", False),
|
||||
@ -467,12 +575,12 @@ def generate_build_ninja(
|
||||
)
|
||||
n.newline()
|
||||
|
||||
def write_custom_step(step: str) -> List[str]:
|
||||
implicit = []
|
||||
def write_custom_step(step: str) -> List[str | Path]:
|
||||
implicit: List[str | Path] = []
|
||||
if config.custom_build_steps and step in config.custom_build_steps:
|
||||
n.comment(f"Custom build steps ({step})")
|
||||
for custom_step in config.custom_build_steps[step]:
|
||||
outputs = custom_step.get("outputs")
|
||||
outputs = cast(List[str | Path], custom_step.get("outputs"))
|
||||
|
||||
if isinstance(outputs, list):
|
||||
implicit.extend(outputs)
|
||||
@ -481,7 +589,7 @@ def generate_build_ninja(
|
||||
|
||||
n.build(
|
||||
outputs=outputs,
|
||||
rule=custom_step.get("rule"),
|
||||
rule=cast(str, custom_step.get("rule")),
|
||||
inputs=custom_step.get("inputs", None),
|
||||
implicit=custom_step.get("implicit", None),
|
||||
order_only=custom_step.get("order_only", None),
|
||||
@ -518,10 +626,6 @@ def generate_build_ninja(
|
||||
# Source files
|
||||
###
|
||||
n.comment("Source files")
|
||||
build_asm_path = build_path / "mod"
|
||||
build_src_path = build_path / "src"
|
||||
build_host_path = build_path / "host"
|
||||
build_config_path = build_path / "config.json"
|
||||
|
||||
def map_path(path: Path) -> Path:
|
||||
return path.parent / (path.name + ".MAP")
|
||||
@ -617,95 +721,85 @@ def generate_build_ninja(
|
||||
host_source_inputs: List[Path] = []
|
||||
source_added: Set[Path] = set()
|
||||
|
||||
def c_build(
|
||||
obj: Object, options: Dict[str, Any], lib_name: str, src_path: Path
|
||||
) -> Optional[Path]:
|
||||
cflags_str = make_flags_str(options["cflags"])
|
||||
if options["extra_cflags"] is not None:
|
||||
extra_cflags_str = make_flags_str(options["extra_cflags"])
|
||||
def c_build(obj: Object, src_path: Path) -> Optional[Path]:
|
||||
cflags_str = make_flags_str(obj.options["cflags"])
|
||||
if obj.options["extra_cflags"] is not None:
|
||||
extra_cflags_str = make_flags_str(obj.options["extra_cflags"])
|
||||
cflags_str += " " + extra_cflags_str
|
||||
used_compiler_versions.add(options["mw_version"])
|
||||
|
||||
src_obj_path = build_src_path / f"{obj.base_name}.o"
|
||||
src_base_path = build_src_path / obj.base_name
|
||||
used_compiler_versions.add(obj.options["mw_version"])
|
||||
|
||||
# Avoid creating duplicate build rules
|
||||
if src_obj_path in source_added:
|
||||
return src_obj_path
|
||||
source_added.add(src_obj_path)
|
||||
|
||||
shift_jis = options["shift_jis"]
|
||||
if shift_jis is None:
|
||||
shift_jis = config.shift_jis
|
||||
if obj.src_obj_path is None or obj.src_obj_path in source_added:
|
||||
return obj.src_obj_path
|
||||
source_added.add(obj.src_obj_path)
|
||||
|
||||
# Add MWCC build rule
|
||||
lib_name = obj.options["lib"]
|
||||
n.comment(f"{obj.name}: {lib_name} (linked {obj.completed})")
|
||||
n.build(
|
||||
outputs=src_obj_path,
|
||||
rule="mwcc_sjis" if shift_jis else "mwcc",
|
||||
outputs=obj.src_obj_path,
|
||||
rule="mwcc_sjis" if obj.options["shift_jis"] else "mwcc",
|
||||
inputs=src_path,
|
||||
variables={
|
||||
"mw_version": Path(options["mw_version"]),
|
||||
"mw_version": Path(obj.options["mw_version"]),
|
||||
"cflags": cflags_str,
|
||||
"basedir": os.path.dirname(src_base_path),
|
||||
"basefile": src_base_path,
|
||||
"basedir": os.path.dirname(obj.src_obj_path),
|
||||
"basefile": obj.src_obj_path.with_suffix(""),
|
||||
},
|
||||
implicit=mwcc_sjis_implicit if shift_jis else mwcc_implicit,
|
||||
implicit=(
|
||||
mwcc_sjis_implicit if obj.options["shift_jis"] else mwcc_implicit
|
||||
),
|
||||
)
|
||||
|
||||
# Add ctx build rule
|
||||
ctx_path = build_src_path / f"{obj.base_name}.ctx"
|
||||
if obj.ctx_path is not None:
|
||||
n.build(
|
||||
outputs=ctx_path,
|
||||
outputs=obj.ctx_path,
|
||||
rule="decompctx",
|
||||
inputs=src_path,
|
||||
implicit=decompctx,
|
||||
)
|
||||
|
||||
# Add host build rule
|
||||
if options.get("host", False):
|
||||
host_obj_path = build_host_path / f"{obj.base_name}.o"
|
||||
host_base_path = build_host_path / obj.base_name
|
||||
if obj.options["host"] and obj.host_obj_path is not None:
|
||||
n.build(
|
||||
outputs=host_obj_path,
|
||||
outputs=obj.host_obj_path,
|
||||
rule="host_cc" if src_path.suffix == ".c" else "host_cpp",
|
||||
inputs=src_path,
|
||||
variables={
|
||||
"basedir": os.path.dirname(host_base_path),
|
||||
"basefile": host_base_path,
|
||||
"basedir": os.path.dirname(obj.host_obj_path),
|
||||
"basefile": obj.host_obj_path.with_suffix(""),
|
||||
},
|
||||
)
|
||||
if options["add_to_all"]:
|
||||
host_source_inputs.append(host_obj_path)
|
||||
if obj.options["add_to_all"]:
|
||||
host_source_inputs.append(obj.host_obj_path)
|
||||
n.newline()
|
||||
|
||||
if options["add_to_all"]:
|
||||
source_inputs.append(src_obj_path)
|
||||
if obj.options["add_to_all"]:
|
||||
source_inputs.append(obj.src_obj_path)
|
||||
|
||||
return src_obj_path
|
||||
return obj.src_obj_path
|
||||
|
||||
def asm_build(
|
||||
obj: Object, options: Dict[str, Any], lib_name: str, src_path: Path
|
||||
obj: Object, src_path: Path, obj_path: Optional[Path]
|
||||
) -> Optional[Path]:
|
||||
asflags = options["asflags"] or config.asflags
|
||||
if asflags is None:
|
||||
if obj.options["asflags"] is None:
|
||||
sys.exit("ProjectConfig.asflags missing")
|
||||
asflags_str = make_flags_str(asflags)
|
||||
if options["extra_asflags"] is not None:
|
||||
extra_asflags_str = make_flags_str(options["extra_asflags"])
|
||||
asflags_str = make_flags_str(obj.options["asflags"])
|
||||
if obj.options["extra_asflags"] is not None:
|
||||
extra_asflags_str = make_flags_str(obj.options["extra_asflags"])
|
||||
asflags_str += " " + extra_asflags_str
|
||||
|
||||
asm_obj_path = build_asm_path / f"{obj.base_name}.o"
|
||||
|
||||
# Avoid creating duplicate build rules
|
||||
if asm_obj_path in source_added:
|
||||
return asm_obj_path
|
||||
source_added.add(asm_obj_path)
|
||||
if obj_path is None or obj_path in source_added:
|
||||
return obj_path
|
||||
source_added.add(obj_path)
|
||||
|
||||
# Add assembler build rule
|
||||
lib_name = obj.options["lib"]
|
||||
n.comment(f"{obj.name}: {lib_name} (linked {obj.completed})")
|
||||
n.build(
|
||||
outputs=asm_obj_path,
|
||||
outputs=obj_path,
|
||||
rule="as",
|
||||
inputs=src_path,
|
||||
variables={"asflags": asflags_str},
|
||||
@ -713,57 +807,40 @@ def generate_build_ninja(
|
||||
)
|
||||
n.newline()
|
||||
|
||||
if options["add_to_all"]:
|
||||
source_inputs.append(asm_obj_path)
|
||||
if obj.options["add_to_all"]:
|
||||
source_inputs.append(obj_path)
|
||||
|
||||
return asm_obj_path
|
||||
return obj_path
|
||||
|
||||
def add_unit(build_obj, link_step: LinkStep):
|
||||
obj_path, obj_name = build_obj["object"], build_obj["name"]
|
||||
result = config.find_object(obj_name)
|
||||
if not result:
|
||||
obj = objects.get(obj_name)
|
||||
if obj is None:
|
||||
if config.warn_missing_config and not build_obj["autogenerated"]:
|
||||
print(f"Missing configuration for {obj_name}")
|
||||
link_step.add(obj_path)
|
||||
return
|
||||
|
||||
lib, obj = result
|
||||
lib_name = lib["lib"]
|
||||
|
||||
# Use object options, then library options
|
||||
options = lib.copy()
|
||||
for key, value in obj.options.items():
|
||||
if value is not None or key not in options:
|
||||
options[key] = value
|
||||
|
||||
unit_src_path = Path(lib.get("src_dir", config.src_dir)) / options["source"]
|
||||
|
||||
unit_asm_path: Optional[Path] = None
|
||||
if config.asm_dir is not None:
|
||||
unit_asm_path = (
|
||||
Path(lib.get("asm_dir", config.asm_dir)) / options["source"]
|
||||
).with_suffix(".s")
|
||||
|
||||
link_built_obj = obj.completed
|
||||
built_obj_path: Optional[Path] = None
|
||||
if unit_src_path.exists():
|
||||
if unit_src_path.suffix in (".c", ".cp", ".cpp"):
|
||||
if obj.src_path is not None and obj.src_path.exists():
|
||||
if obj.src_path.suffix in (".c", ".cp", ".cpp"):
|
||||
# Add MWCC & host build rules
|
||||
built_obj_path = c_build(obj, options, lib_name, unit_src_path)
|
||||
elif unit_src_path.suffix == ".s":
|
||||
built_obj_path = c_build(obj, obj.src_path)
|
||||
elif obj.src_path.suffix == ".s":
|
||||
# Add assembler build rule
|
||||
built_obj_path = asm_build(obj, options, lib_name, unit_src_path)
|
||||
built_obj_path = asm_build(obj, obj.src_path, obj.src_obj_path)
|
||||
else:
|
||||
sys.exit(f"Unknown source file type {unit_src_path}")
|
||||
sys.exit(f"Unknown source file type {obj.src_path}")
|
||||
else:
|
||||
if config.warn_missing_source or obj.completed:
|
||||
print(f"Missing source file {unit_src_path}")
|
||||
print(f"Missing source file {obj.src_path}")
|
||||
link_built_obj = False
|
||||
|
||||
# Assembly overrides
|
||||
if unit_asm_path is not None and unit_asm_path.exists():
|
||||
if obj.asm_path is not None and obj.asm_path.exists():
|
||||
link_built_obj = True
|
||||
built_obj_path = asm_build(obj, options, lib_name, unit_asm_path)
|
||||
built_obj_path = asm_build(obj, obj.asm_path, obj.asm_obj_path)
|
||||
|
||||
if link_built_obj and built_obj_path is not None:
|
||||
# Use the source-built object
|
||||
@ -772,7 +849,10 @@ def generate_build_ninja(
|
||||
# Use the original (extracted) object
|
||||
link_step.add(obj_path)
|
||||
else:
|
||||
sys.exit(f"Missing object for {obj_name}: {unit_src_path} {lib} {obj}")
|
||||
lib_name = obj.options["lib"]
|
||||
sys.exit(
|
||||
f"Missing object for {obj_name}: {obj.src_path} {lib_name} {obj}"
|
||||
)
|
||||
|
||||
# Add DOL link step
|
||||
link_step = LinkStep(build_config)
|
||||
@ -788,7 +868,7 @@ def generate_build_ninja(
|
||||
add_unit(unit, module_link_step)
|
||||
# Add empty object to empty RELs
|
||||
if len(module_link_step.inputs) == 0:
|
||||
if not config.rel_empty_file:
|
||||
if config.rel_empty_file is None:
|
||||
sys.exit("ProjectConfig.rel_empty_file missing")
|
||||
add_unit(
|
||||
{
|
||||
@ -850,7 +930,7 @@ def generate_build_ninja(
|
||||
rspfile="$rspfile",
|
||||
rspfile_content="$in_newline",
|
||||
)
|
||||
generated_rels = []
|
||||
generated_rels: List[str] = []
|
||||
for idx, link in enumerate(build_config["links"]):
|
||||
# Map module names to link steps
|
||||
link_steps_local = list(
|
||||
@ -957,6 +1037,22 @@ def generate_build_ninja(
|
||||
implicit=[ok_path, configure_script, python_lib, config.config_path],
|
||||
)
|
||||
|
||||
###
|
||||
# Generate progress report
|
||||
###
|
||||
n.comment("Generate progress report")
|
||||
n.rule(
|
||||
name="report",
|
||||
command=f"{objdiff} report generate -o $out",
|
||||
description="REPORT",
|
||||
)
|
||||
report_implicit: List[str | Path] = [objdiff, "all_source"]
|
||||
n.build(
|
||||
outputs=report_path,
|
||||
rule="report",
|
||||
implicit=report_implicit,
|
||||
)
|
||||
|
||||
###
|
||||
# Helper tools
|
||||
###
|
||||
@ -1003,6 +1099,7 @@ def generate_build_ninja(
|
||||
###
|
||||
# Split DOL
|
||||
###
|
||||
build_config_path = build_path / "config.json"
|
||||
n.comment("Split DOL into relocatable objects")
|
||||
n.rule(
|
||||
name="split",
|
||||
@ -1063,13 +1160,15 @@ def generate_build_ninja(
|
||||
|
||||
# Generate objdiff.json
|
||||
def generate_objdiff_config(
|
||||
config: ProjectConfig, build_config: Optional[Dict[str, Any]]
|
||||
config: ProjectConfig,
|
||||
objects: Dict[str, Object],
|
||||
build_config: Optional[Dict[str, Any]],
|
||||
) -> None:
|
||||
if not build_config:
|
||||
if build_config is None:
|
||||
return
|
||||
|
||||
objdiff_config: Dict[str, Any] = {
|
||||
"min_version": "1.0.0",
|
||||
"min_version": "2.0.0-beta.5",
|
||||
"custom_make": "ninja",
|
||||
"build_target": False,
|
||||
"watch_patterns": [
|
||||
@ -1085,6 +1184,7 @@ def generate_objdiff_config(
|
||||
"*.json",
|
||||
],
|
||||
"units": [],
|
||||
"progress_categories": [],
|
||||
}
|
||||
|
||||
# decomp.me compiler name mapping
|
||||
@ -1095,6 +1195,7 @@ def generate_objdiff_config(
|
||||
"GC/1.2.5": "mwcc_233_163",
|
||||
"GC/1.2.5e": "mwcc_233_163e",
|
||||
"GC/1.2.5n": "mwcc_233_163n",
|
||||
"GC/1.3": "mwcc_242_53",
|
||||
"GC/1.3.2": "mwcc_242_81",
|
||||
"GC/1.3.2r": "mwcc_242_81r",
|
||||
"GC/2.0": "mwcc_247_92",
|
||||
@ -1119,44 +1220,30 @@ def generate_objdiff_config(
|
||||
"Wii/1.7": "mwcc_43_213",
|
||||
}
|
||||
|
||||
build_path = config.out_path()
|
||||
|
||||
def add_unit(build_obj: Dict[str, Any], module_name: str) -> None:
|
||||
if build_obj["autogenerated"]:
|
||||
# Skip autogenerated objects
|
||||
return
|
||||
|
||||
def add_unit(
|
||||
build_obj: Dict[str, Any], module_name: str, progress_categories: List[str]
|
||||
) -> None:
|
||||
obj_path, obj_name = build_obj["object"], build_obj["name"]
|
||||
base_object = Path(obj_name).with_suffix("")
|
||||
unit_config: Dict[str, Any] = {
|
||||
"name": Path(module_name) / base_object,
|
||||
"target_path": obj_path,
|
||||
"metadata": {
|
||||
"auto_generated": build_obj["autogenerated"],
|
||||
"progress_categories": progress_categories,
|
||||
},
|
||||
}
|
||||
|
||||
result = config.find_object(obj_name)
|
||||
if not result:
|
||||
obj = objects.get(obj_name)
|
||||
if obj is None:
|
||||
objdiff_config["units"].append(unit_config)
|
||||
return
|
||||
|
||||
lib, obj = result
|
||||
src_dir = Path(lib.get("src_dir", config.src_dir))
|
||||
|
||||
# Use object options, then library options
|
||||
options = lib.copy()
|
||||
for key, value in obj.options.items():
|
||||
if value is not None or key not in options:
|
||||
options[key] = value
|
||||
|
||||
unit_src_path = src_dir / str(options["source"])
|
||||
|
||||
if not unit_src_path.exists():
|
||||
objdiff_config["units"].append(unit_config)
|
||||
return
|
||||
|
||||
cflags = options["cflags"]
|
||||
src_obj_path = build_path / "src" / f"{obj.base_name}.o"
|
||||
src_ctx_path = build_path / "src" / f"{obj.base_name}.ctx"
|
||||
src_exists = obj.src_path is not None and obj.src_path.exists()
|
||||
if src_exists:
|
||||
unit_config["base_path"] = obj.src_obj_path
|
||||
|
||||
cflags = obj.options["cflags"]
|
||||
reverse_fn_order = False
|
||||
if type(cflags) is list:
|
||||
for flag in cflags:
|
||||
@ -1175,39 +1262,86 @@ def generate_objdiff_config(
|
||||
cflags = list(filter(keep_flag, cflags))
|
||||
|
||||
# Add appropriate lang flag
|
||||
if unit_src_path.suffix in (".cp", ".cpp"):
|
||||
if obj.src_path is not None and not any(
|
||||
flag.startswith("-lang") for flag in cflags
|
||||
):
|
||||
if obj.src_path.suffix in (".cp", ".cpp"):
|
||||
cflags.insert(0, "-lang=c++")
|
||||
else:
|
||||
cflags.insert(0, "-lang=c")
|
||||
|
||||
unit_config["base_path"] = src_obj_path
|
||||
unit_config["reverse_fn_order"] = reverse_fn_order
|
||||
unit_config["complete"] = obj.completed
|
||||
compiler_version = COMPILER_MAP.get(options["mw_version"])
|
||||
compiler_version = COMPILER_MAP.get(obj.options["mw_version"])
|
||||
if compiler_version is None:
|
||||
print(f"Missing scratch compiler mapping for {options['mw_version']}")
|
||||
print(f"Missing scratch compiler mapping for {obj.options['mw_version']}")
|
||||
else:
|
||||
cflags_str = make_flags_str(cflags)
|
||||
if options["extra_cflags"] is not None:
|
||||
extra_cflags_str = make_flags_str(options["extra_cflags"])
|
||||
if obj.options["extra_cflags"] is not None:
|
||||
extra_cflags_str = make_flags_str(obj.options["extra_cflags"])
|
||||
cflags_str += " " + extra_cflags_str
|
||||
unit_config["scratch"] = {
|
||||
"platform": "gc_wii",
|
||||
"compiler": compiler_version,
|
||||
"c_flags": cflags_str,
|
||||
"ctx_path": src_ctx_path,
|
||||
}
|
||||
if src_exists:
|
||||
unit_config["scratch"].update(
|
||||
{
|
||||
"ctx_path": obj.ctx_path,
|
||||
"build_ctx": True,
|
||||
}
|
||||
)
|
||||
category_opt: List[str] | str = obj.options["progress_category"]
|
||||
if isinstance(category_opt, list):
|
||||
progress_categories.extend(category_opt)
|
||||
elif category_opt is not None:
|
||||
progress_categories.append(category_opt)
|
||||
unit_config["metadata"].update(
|
||||
{
|
||||
"complete": obj.completed,
|
||||
"reverse_fn_order": reverse_fn_order,
|
||||
"source_path": obj.src_path,
|
||||
"progress_categories": progress_categories,
|
||||
}
|
||||
)
|
||||
objdiff_config["units"].append(unit_config)
|
||||
|
||||
# Add DOL units
|
||||
for unit in build_config["units"]:
|
||||
add_unit(unit, build_config["name"])
|
||||
progress_categories = []
|
||||
# Only include a "dol" category if there are any modules
|
||||
# Otherwise it's redundant with the global report measures
|
||||
if len(build_config["modules"]) > 0:
|
||||
progress_categories.append("dol")
|
||||
add_unit(unit, build_config["name"], progress_categories)
|
||||
|
||||
# Add REL units
|
||||
for module in build_config["modules"]:
|
||||
for unit in module["units"]:
|
||||
add_unit(unit, module["name"])
|
||||
progress_categories = []
|
||||
if config.progress_modules:
|
||||
progress_categories.append("modules")
|
||||
if config.progress_each_module:
|
||||
progress_categories.append(module["name"])
|
||||
add_unit(unit, module["name"], progress_categories)
|
||||
|
||||
# Add progress categories
|
||||
def add_category(id: str, name: str):
|
||||
objdiff_config["progress_categories"].append(
|
||||
{
|
||||
"id": id,
|
||||
"name": name,
|
||||
}
|
||||
)
|
||||
|
||||
if len(build_config["modules"]) > 0:
|
||||
add_category("dol", "DOL")
|
||||
if config.progress_modules:
|
||||
add_category("modules", "Modules")
|
||||
if config.progress_each_module:
|
||||
for module in build_config["modules"]:
|
||||
add_category(module["name"], module["name"])
|
||||
for category in config.progress_categories:
|
||||
add_category(category.id, category.name)
|
||||
|
||||
# Write objdiff.json
|
||||
with open("objdiff.json", "w", encoding="utf-8") as w:
|
||||
@ -1220,9 +1354,11 @@ def generate_objdiff_config(
|
||||
|
||||
# Calculate, print and write progress to progress.json
|
||||
def calculate_progress(config: ProjectConfig) -> None:
|
||||
config.validate()
|
||||
objects = config.objects()
|
||||
out_path = config.out_path()
|
||||
build_config = load_build_config(config, out_path / "config.json")
|
||||
if not build_config:
|
||||
if build_config is None:
|
||||
return
|
||||
|
||||
class ProgressUnit:
|
||||
@ -1254,12 +1390,8 @@ def calculate_progress(config: ProjectConfig) -> None:
|
||||
# Skip autogenerated objects
|
||||
return
|
||||
|
||||
result = config.find_object(build_obj["name"])
|
||||
if not result:
|
||||
return
|
||||
|
||||
_, obj = result
|
||||
if not obj.completed:
|
||||
obj = objects.get(build_obj["name"])
|
||||
if obj is None or not obj.completed:
|
||||
return
|
||||
|
||||
self.code_progress += build_obj["code_size"]
|
||||
@ -1268,38 +1400,68 @@ def calculate_progress(config: ProjectConfig) -> None:
|
||||
self.objects_progress += 1
|
||||
|
||||
def code_frac(self) -> float:
|
||||
if self.code_total == 0:
|
||||
return 1.0
|
||||
return self.code_progress / self.code_total
|
||||
|
||||
def data_frac(self) -> float:
|
||||
if self.data_total == 0:
|
||||
return 1.0
|
||||
return self.data_progress / self.data_total
|
||||
|
||||
progress_units: Dict[str, ProgressUnit] = {}
|
||||
if config.progress_all:
|
||||
progress_units["all"] = ProgressUnit("All")
|
||||
progress_units["dol"] = ProgressUnit("DOL")
|
||||
if len(build_config["modules"]) > 0:
|
||||
if config.progress_modules:
|
||||
progress_units["modules"] = ProgressUnit("Modules")
|
||||
if len(config.progress_categories) > 0:
|
||||
for category in config.progress_categories:
|
||||
progress_units[category.id] = ProgressUnit(category.name)
|
||||
if config.progress_each_module:
|
||||
for module in build_config["modules"]:
|
||||
progress_units[module["name"]] = ProgressUnit(module["name"])
|
||||
|
||||
def add_unit(id: str, unit: Dict[str, Any]) -> None:
|
||||
progress = progress_units.get(id)
|
||||
if progress is not None:
|
||||
progress.add(unit)
|
||||
|
||||
# Add DOL units
|
||||
all_progress = ProgressUnit("All") if config.progress_all else None
|
||||
dol_progress = ProgressUnit("DOL")
|
||||
for unit in build_config["units"]:
|
||||
if all_progress:
|
||||
all_progress.add(unit)
|
||||
dol_progress.add(unit)
|
||||
add_unit("all", unit)
|
||||
add_unit("dol", unit)
|
||||
obj = objects.get(unit["name"])
|
||||
if obj is not None:
|
||||
category_opt = obj.options["progress_category"]
|
||||
if isinstance(category_opt, list):
|
||||
for id in category_opt:
|
||||
add_unit(id, unit)
|
||||
elif category_opt is not None:
|
||||
add_unit(category_opt, unit)
|
||||
|
||||
# Add REL units
|
||||
rels_progress = ProgressUnit("Modules") if config.progress_modules else None
|
||||
modules_progress: List[ProgressUnit] = []
|
||||
for module in build_config["modules"]:
|
||||
progress = ProgressUnit(module["name"])
|
||||
modules_progress.append(progress)
|
||||
for unit in module["units"]:
|
||||
if all_progress:
|
||||
all_progress.add(unit)
|
||||
if rels_progress:
|
||||
rels_progress.add(unit)
|
||||
progress.add(unit)
|
||||
add_unit("all", unit)
|
||||
add_unit("modules", unit)
|
||||
add_unit(module["name"], unit)
|
||||
obj = objects.get(unit["name"])
|
||||
if obj is not None:
|
||||
category_opt = obj.options["progress_category"]
|
||||
if isinstance(category_opt, list):
|
||||
for id in category_opt:
|
||||
add_unit(id, unit)
|
||||
elif category_opt is not None:
|
||||
add_unit(category_opt, unit)
|
||||
|
||||
# Print human-readable progress
|
||||
print("Progress:")
|
||||
|
||||
def print_category(unit: Optional[ProgressUnit]) -> None:
|
||||
if unit is None:
|
||||
return
|
||||
for unit in progress_units.values():
|
||||
if unit.objects_total == 0:
|
||||
continue
|
||||
|
||||
code_frac = unit.code_frac()
|
||||
data_frac = unit.data_frac()
|
||||
@ -1321,41 +1483,17 @@ def calculate_progress(config: ProjectConfig) -> None:
|
||||
)
|
||||
)
|
||||
|
||||
bytes_per_frac = unit.code_total / unit.code_fancy_frac
|
||||
bytes_next = math.ceil((code_items + 1) * bytes_per_frac)
|
||||
bytes_to_go = bytes_next - unit.code_progress
|
||||
|
||||
print(f"Code bytes to go for next trophy: {bytes_to_go}")
|
||||
|
||||
if all_progress:
|
||||
print_category(all_progress)
|
||||
print_category(dol_progress)
|
||||
module_count = len(build_config["modules"])
|
||||
if module_count > 0:
|
||||
print_category(rels_progress)
|
||||
if config.progress_each_module:
|
||||
for progress in modules_progress:
|
||||
print_category(progress)
|
||||
|
||||
# Generate and write progress.json
|
||||
progress_json: Dict[str, Any] = {}
|
||||
|
||||
def add_category(category: str, unit: ProgressUnit) -> None:
|
||||
progress_json[category] = {
|
||||
for id, unit in progress_units.items():
|
||||
if unit.objects_total == 0:
|
||||
continue
|
||||
progress_json[id] = {
|
||||
"code": unit.code_progress,
|
||||
"code/total": unit.code_total,
|
||||
"data": unit.data_progress,
|
||||
"data/total": unit.data_total,
|
||||
}
|
||||
|
||||
if all_progress:
|
||||
add_category("all", all_progress)
|
||||
add_category("dol", dol_progress)
|
||||
if len(build_config["modules"]) > 0:
|
||||
if rels_progress:
|
||||
add_category("modules", rels_progress)
|
||||
if config.progress_each_module:
|
||||
for progress in modules_progress:
|
||||
add_category(progress.name, progress)
|
||||
with open(out_path / "progress.json", "w", encoding="utf-8") as w:
|
||||
json.dump(progress_json, w, indent=4)
|
||||
|
Loading…
Reference in New Issue
Block a user