### # decomp-toolkit project generator # Generates build.ninja and objdiff.json. # # This generator is intentionally project-agnostic # and shared between multiple projects. Any configuration # specific to a project should be added to `configure.py`. # # If changes are made, please submit a PR to # https://github.com/encounter/dtk-template ### import io import json import math import os import platform import sys from pathlib import Path from typing import Any, cast from . import ninja_syntax from .ninja_syntax import serialize_path if sys.platform == "cygwin": sys.exit( "Cygwin/MSYS2 is not supported." + f"\nPlease use native Windows Python instead." + f"\n(Current path: {sys.executable})" ) Library = dict[str, Any] class Object: def __init__(self, completed: bool, name: str, **options: Any) -> None: self.name = name self.completed = completed self.options: dict[str, Any] = { "add_to_all": None, "asflags": None, "asm_dir": None, "cflags": None, "extra_asflags": None, "extra_cflags": None, "host": None, "lib": None, "mw_version": None, "progress_category": None, "shift_jis": None, "source": name, "src_dir": None, } self.options.update(options) # Internal self.src_path: Path | None = None self.asm_path: Path | None = None self.src_obj_path: Path | None = None self.asm_obj_path: Path | None = None self.host_obj_path: Path | None = None self.ctx_path: Path | None = None def resolve(self, config: "ProjectConfig", lib: Library) -> "Object": # Use object options, then library options obj = Object(self.completed, self.name, **lib) for key, value in self.options.items(): if value is not None or key not in obj.options: obj.options[key] = value # Use default options from config def set_default(key: str, value: Any) -> None: if obj.options[key] is None: obj.options[key] = value set_default("add_to_all", True) set_default("asflags", config.asflags) set_default("asm_dir", config.asm_dir) set_default("host", False) set_default("mw_version", config.linker_version) set_default("shift_jis", config.shift_jis) set_default("src_dir", config.src_dir) # Resolve paths build_dir = config.out_path() obj.src_path = Path(obj.options["src_dir"]) / obj.options["source"] if obj.options["asm_dir"] is not None: obj.asm_path = ( Path(obj.options["asm_dir"]) / obj.options["source"] ).with_suffix(".s") base_name = Path(self.name).with_suffix("") obj.src_obj_path = build_dir / "src" / f"{base_name}.o" obj.asm_obj_path = build_dir / "mod" / f"{base_name}.o" obj.host_obj_path = build_dir / "host" / f"{base_name}.o" obj.ctx_path = build_dir / "src" / f"{base_name}.ctx" return obj class ProgressCategory: def __init__(self, id: str, name: str) -> None: self.id = id self.name = name class ProjectConfig: def __init__(self) -> None: # Paths self.build_dir: Path = Path("build") # Output build files self.src_dir: Path = Path("src") # C/C++/asm source files self.tools_dir: Path = Path("tools") # Python scripts self.asm_dir: Path | None = Path( "asm" ) # Override incomplete objects (for modding) # Tooling self.binutils_tag: str | None = None # Git tag self.binutils_path: Path | None = None # If None, download self.dtk_tag: str | None = None # Git tag self.dtk_path: Path | None = None # If None, download self.compilers_tag: str | None = None # 1 self.compilers_path: Path | None = None # If None, download self.wibo_tag: str | None = None # Git tag self.wrapper: Path | None = None # If None, download wibo on Linux self.sjiswrap_tag: str | None = None # Git tag self.sjiswrap_path: Path | None = None # If None, download self.objdiff_tag: str | None = None # Git tag self.objdiff_path: Path | None = None # If None, download # Project config self.non_matching: bool = False self.build_rels: bool = True # Build REL files self.check_sha_path: Path | None = None # Path to version.sha1 self.config_path: Path | None = None # Path to config.yml self.generate_map: bool = False # Generate map file(s) self.asflags: list[str | None] = None # Assembler flags self.ldflags: list[str | None] = None # Linker flags self.libs: list[Library | None] = None # list of libraries self.linker_version: str | None = None # mwld version self.version: str | None = None # Version name self.warn_missing_config: bool = False # Warn on missing unit configuration self.warn_missing_source: bool = False # Warn on missing source file self.rel_strip_partial: bool = True # Generate PLFs with -strip_partial self.rel_empty_file: str | None = None # Object name for generating empty RELs self.shift_jis = ( True # Convert source files from UTF-8 to Shift JIS automatically ) self.reconfig_deps: list[Path | None] = ( None # Additional re-configuration dependency files ) self.custom_build_rules: list[dict[str, Any | None]] = ( None # Custom ninja build rules ) self.custom_build_steps: dict[str, list[dict[str, Any | None]]] = ( None # Custom build steps, types are ["pre-compile", "post-compile", "post-link", "post-build"] ) # Progress output, progress.json and report.json config self.progress_all: bool = True # Include combined "all" category self.progress_modules: bool = True # Include combined "modules" category self.progress_each_module: bool = ( False # Include individual modules, disable for large numbers of modules ) self.progress_categories: list[ProgressCategory] = [] # Additional categories # Progress fancy printing self.progress_use_fancy: bool = False self.progress_code_fancy_frac: int = 0 self.progress_code_fancy_item: str = "" self.progress_data_fancy_frac: int = 0 self.progress_data_fancy_item: str = "" def validate(self) -> None: required_attrs = [ "build_dir", "src_dir", "tools_dir", "check_sha_path", "config_path", "ldflags", "linker_version", "libs", "version", ] for attr in required_attrs: if getattr(self, attr) is None: sys.exit(f"ProjectConfig.{attr} missing") # Creates a map of object names to Object instances # Options are fully resolved from the library and object def objects(self) -> dict[str, Object]: out = {} for lib in self.libs or {}: objects: list[Object] = lib["objects"] for obj in objects: if obj.name in out: sys.exit(f"Duplicate object name {obj.name}") out[obj.name] = obj.resolve(self, lib) return out def out_path(self) -> Path: return self.build_dir / str(self.version) def is_windows() -> bool: return os.name == "nt" # On Windows, we need this to use && in commands CHAIN = "cmd /c " if is_windows() else "" # Native executable extension EXE = ".exe" if is_windows() else "" def make_flags_str(flags: str | list[str] | None) -> str: if flags is None: return "" elif isinstance(flags, list): return " ".join(flags) else: return flags # Load decomp-toolkit generated config.json def load_build_config( config: ProjectConfig, build_config_path: Path ) -> dict[str, Any | None]: if not build_config_path.is_file(): return None def versiontuple(v: str) -> Tuple[int, ...]: return tuple(map(int, (v.split(".")))) f = open(build_config_path, "r", encoding="utf-8") build_config: dict[str, Any] = json.load(f) config_version = build_config.get("version") if config_version is None: print("Invalid config.json, regenerating...") f.close() os.remove(build_config_path) return None dtk_version = str(config.dtk_tag)[1:] # Strip v if versiontuple(config_version) < versiontuple(dtk_version): print("Outdated config.json, regenerating...") f.close() os.remove(build_config_path) return None f.close() return build_config # Generate build.ninja and objdiff.json def generate_build(config: ProjectConfig) -> None: config.validate() objects = config.objects() build_config = load_build_config(config, config.out_path() / "config.json") generate_build_ninja(config, objects, build_config) generate_objdiff_config(config, objects, build_config) # Generate build.ninja def generate_build_ninja( config: ProjectConfig, objects: dict[str, Object], build_config: dict[str, Any | None], ) -> None: out = io.StringIO() n = ninja_syntax.Writer(out) n.variable("ninja_required_version", "1.3") n.newline() configure_script = Path(os.path.relpath(os.path.abspath(sys.argv[0]))) python_lib = Path(os.path.relpath(__file__)) python_lib_dir = python_lib.parent n.comment("The arguments passed to configure.py, for rerunning it.") n.variable("configure_args", sys.argv[1:]) n.variable("python", f'"{sys.executable}"') n.newline() ### # Variables ### n.comment("Variables") n.variable("ldflags", make_flags_str(config.ldflags)) if config.linker_version is None: sys.exit("ProjectConfig.linker_version missing") n.variable("mw_version", Path(config.linker_version)) n.newline() ### # Tooling ### n.comment("Tooling") build_path = config.out_path() progress_path = build_path / "progress.json" report_path = build_path / "report.json" build_tools_path = config.build_dir / "tools" download_tool = config.tools_dir / "download_tool.py" n.rule( name="download_tool", command=f"$python {download_tool} $tool $out --tag $tag", description="TOOL $out", ) decompctx = config.tools_dir / "decompctx.py" n.rule( name="decompctx", command=f"$python {decompctx} $in -o $out -d $out.d", description="CTX $in", depfile="$out.d", deps="gcc", ) cargo_rule_written = False def write_cargo_rule(): nonlocal cargo_rule_written if not cargo_rule_written: n.pool("cargo", 1) n.rule( name="cargo", command="cargo build --release --manifest-path $in --bin $bin --target-dir $target", description="CARGO $bin", pool="cargo", depfile=Path("$target") / "release" / "$bin.d", deps="gcc", ) cargo_rule_written = True if config.dtk_path is not None and config.dtk_path.is_file(): dtk = config.dtk_path elif config.dtk_path is not None: dtk = build_tools_path / "release" / f"dtk{EXE}" write_cargo_rule() n.build( outputs=dtk, rule="cargo", inputs=config.dtk_path / "Cargo.toml", implicit=config.dtk_path / "Cargo.lock", variables={ "bin": "dtk", "target": build_tools_path, }, ) elif config.dtk_tag: dtk = build_tools_path / f"dtk{EXE}" n.build( outputs=dtk, rule="download_tool", implicit=download_tool, variables={ "tool": "dtk", "tag": config.dtk_tag, }, ) else: sys.exit("ProjectConfig.dtk_tag missing") if config.objdiff_path is not None and config.objdiff_path.is_file(): objdiff = config.objdiff_path elif config.objdiff_path is not None: objdiff = build_tools_path / "release" / f"objdiff-cli{EXE}" write_cargo_rule() n.build( outputs=objdiff, rule="cargo", inputs=config.objdiff_path / "Cargo.toml", implicit=config.objdiff_path / "Cargo.lock", variables={ "bin": "objdiff-cli", "target": build_tools_path, }, ) elif config.objdiff_tag: objdiff = build_tools_path / f"objdiff-cli{EXE}" n.build( outputs=objdiff, rule="download_tool", implicit=download_tool, variables={ "tool": "objdiff-cli", "tag": config.objdiff_tag, }, ) else: sys.exit("ProjectConfig.objdiff_tag missing") if config.sjiswrap_path: sjiswrap = config.sjiswrap_path elif config.sjiswrap_tag: sjiswrap = build_tools_path / "sjiswrap.exe" n.build( outputs=sjiswrap, rule="download_tool", implicit=download_tool, variables={ "tool": "sjiswrap", "tag": config.sjiswrap_tag, }, ) else: sys.exit("ProjectConfig.sjiswrap_tag missing") # Only add an implicit dependency on wibo if we download it wrapper = config.wrapper wrapper_implicit: Path | None = None if ( config.wibo_tag is not None and sys.platform == "linux" and platform.machine() in ("i386", "x86_64") and config.wrapper is None ): wrapper = build_tools_path / "wibo" wrapper_implicit = wrapper n.build( outputs=wrapper, rule="download_tool", implicit=download_tool, variables={ "tool": "wibo", "tag": config.wibo_tag, }, ) if not is_windows() and wrapper is None: wrapper = Path("wine") wrapper_cmd = f"{wrapper} " if wrapper else "" compilers_implicit: Path | None = None if config.compilers_path: compilers = config.compilers_path elif config.compilers_tag: compilers = config.build_dir / "compilers" compilers_implicit = compilers n.build( outputs=compilers, rule="download_tool", implicit=download_tool, variables={ "tool": "compilers", "tag": config.compilers_tag, }, ) else: sys.exit("ProjectConfig.compilers_tag missing") binutils_implicit = None if config.binutils_path: binutils = config.binutils_path elif config.binutils_tag: binutils = config.build_dir / "binutils" binutils_implicit = binutils n.build( outputs=binutils, rule="download_tool", implicit=download_tool, variables={ "tool": "binutils", "tag": config.binutils_tag, }, ) else: sys.exit("ProjectConfig.binutils_tag missing") n.newline() ### # Helper rule for downloading all tools ### n.comment("Download all tools") n.build( outputs="tools", rule="phony", inputs=[dtk, sjiswrap, wrapper, compilers, binutils, objdiff], ) n.newline() ### # Build rules ### compiler_path = compilers / "$mw_version" # MWCC mwcc = compiler_path / "mwcceppc.exe" mwcc_cmd = f"{wrapper_cmd}{mwcc} $cflags -MMD -c $in -o $basedir" mwcc_implicit: list[Path | None] = [compilers_implicit or mwcc, wrapper_implicit] # MWCC with UTF-8 to Shift JIS wrapper mwcc_sjis_cmd = f"{wrapper_cmd}{sjiswrap} {mwcc} $cflags -MMD -c $in -o $basedir" mwcc_sjis_implicit: list[Path | None] = [*mwcc_implicit, sjiswrap] # MWLD mwld = compiler_path / "mwldeppc.exe" mwld_cmd = f"{wrapper_cmd}{mwld} $ldflags -o $out @$out.rsp" mwld_implicit: list[Path | None] = [compilers_implicit or mwld, wrapper_implicit] # GNU as gnu_as = binutils / f"powerpc-eabi-as{EXE}" gnu_as_cmd = ( f"{CHAIN}{gnu_as} $asflags -o $out $in -MD $out.d" + f" && {dtk} elf fixup $out $out" ) gnu_as_implicit = [binutils_implicit or gnu_as, dtk] if os.name != "nt": transform_dep = config.tools_dir / "transform_dep.py" mwcc_cmd += f" && $python {transform_dep} $basefile.d $basefile.d" mwcc_sjis_cmd += f" && $python {transform_dep} $basefile.d $basefile.d" mwcc_implicit.append(transform_dep) mwcc_sjis_implicit.append(transform_dep) n.comment("Link ELF file") n.rule( name="link", command=mwld_cmd, description="LINK $out", rspfile="$out.rsp", rspfile_content="$in_newline", ) n.newline() n.comment("Generate DOL") n.rule( name="elf2dol", command=f"{dtk} elf2dol $in $out", description="DOL $out", ) n.newline() n.comment("MWCC build") n.rule( name="mwcc", command=mwcc_cmd, description="MWCC $out", depfile="$basefile.d", deps="gcc", ) n.newline() n.comment("MWCC build (with UTF-8 to Shift JIS wrapper)") n.rule( name="mwcc_sjis", command=mwcc_sjis_cmd, description="MWCC $out", depfile="$basefile.d", deps="gcc", ) n.newline() n.comment("Assemble asm") n.rule( name="as", command=gnu_as_cmd, description="AS $out", depfile="$out.d", deps="gcc", ) n.newline() if len(config.custom_build_rules or {}) > 0: n.comment("Custom project build rules (pre/post-processing)") for rule in config.custom_build_rules or {}: n.rule( name=cast(str, rule.get("name")), command=cast(str, rule.get("command")), description=rule.get("description", None), depfile=rule.get("depfile", None), generator=rule.get("generator", False), pool=rule.get("pool", None), restat=rule.get("restat", False), rspfile=rule.get("rspfile", None), rspfile_content=rule.get("rspfile_content", None), deps=rule.get("deps", None), ) n.newline() def write_custom_step(step: str) -> list[str | Path]: implicit: list[str | Path] = [] if config.custom_build_steps and step in config.custom_build_steps: n.comment(f"Custom build steps ({step})") for custom_step in config.custom_build_steps[step]: outputs = cast(list[str | Path], custom_step.get("outputs")) if isinstance(outputs, list): implicit.extend(outputs) else: implicit.append(outputs) n.build( outputs=outputs, rule=cast(str, custom_step.get("rule")), inputs=custom_step.get("inputs", None), implicit=custom_step.get("implicit", None), order_only=custom_step.get("order_only", None), variables=custom_step.get("variables", None), implicit_outputs=custom_step.get("implicit_outputs", None), pool=custom_step.get("pool", None), dyndep=custom_step.get("dyndep", None), ) n.newline() return implicit n.comment("Host build") n.variable("host_cflags", "-I include -Wno-trigraphs") n.variable( "host_cppflags", "-std=c++98 -I include -fno-exceptions -fno-rtti -D_CRT_SECURE_NO_WARNINGS -Wno-trigraphs -Wno-c++11-extensions", ) n.rule( name="host_cc", command="clang $host_cflags -c -o $out $in", description="CC $out", ) n.rule( name="host_cpp", command="clang++ $host_cppflags -c -o $out $in", description="CXX $out", ) n.newline() # Add all build steps needed before we compile (e.g. processing assets) precompile_implicit = write_custom_step("pre-compile") ### # Source files ### n.comment("Source files") def map_path(path: Path) -> Path: return path.parent / (path.name + ".MAP") class LinkStep: def __init__(self, config: dict[str, Any]) -> None: self.name: str = config["name"] self.module_id: int = config["module_id"] self.ldscript: Path | None = Path(config["ldscript"]) self.entry = config["entry"] self.inputs: list[str] = [] def add(self, obj: Path) -> None: self.inputs.append(serialize_path(obj)) def output(self) -> Path: if self.module_id == 0: return build_path / f"{self.name}.dol" else: return build_path / self.name / f"{self.name}.rel" def partial_output(self) -> Path: if self.module_id == 0: return build_path / f"{self.name}.elf" else: return build_path / self.name / f"{self.name}.plf" def write(self, n: ninja_syntax.Writer) -> None: n.comment(f"Link {self.name}") if self.module_id == 0: elf_path = build_path / f"{self.name}.elf" dol_path = build_path / f"{self.name}.dol" elf_ldflags = f"$ldflags -lcf {serialize_path(self.ldscript)}" if config.generate_map: elf_map = map_path(elf_path) elf_ldflags += f" -map {serialize_path(elf_map)}" else: elf_map = None n.build( outputs=elf_path, rule="link", inputs=self.inputs, implicit=[ *precompile_implicit, self.ldscript, *mwld_implicit, *postcompile_implicit, ], implicit_outputs=elf_map, variables={"ldflags": elf_ldflags}, ) else: preplf_path = build_path / self.name / f"{self.name}.preplf" plf_path = build_path / self.name / f"{self.name}.plf" preplf_ldflags = "$ldflags -sdata 0 -sdata2 0 -r" plf_ldflags = f"$ldflags -sdata 0 -sdata2 0 -r1 -lcf {serialize_path(self.ldscript)}" if self.entry: plf_ldflags += f" -m {self.entry}" # -strip_partial is only valid with -m if config.rel_strip_partial: plf_ldflags += " -strip_partial" if config.generate_map: preplf_map = map_path(preplf_path) preplf_ldflags += f" -map {serialize_path(preplf_map)}" plf_map = map_path(plf_path) plf_ldflags += f" -map {serialize_path(plf_map)}" else: preplf_map = None plf_map = None n.build( outputs=preplf_path, rule="link", inputs=self.inputs, implicit=mwld_implicit, implicit_outputs=preplf_map, variables={"ldflags": preplf_ldflags}, ) n.build( outputs=plf_path, rule="link", inputs=self.inputs, implicit=[self.ldscript, preplf_path, *mwld_implicit], implicit_outputs=plf_map, variables={"ldflags": plf_ldflags}, ) n.newline() link_outputs: list[Path] = [] if build_config: link_steps: list[LinkStep] = [] used_compiler_versions: Set[str] = set() source_inputs: list[Path] = [] host_source_inputs: list[Path] = [] source_added: Set[Path] = set() def c_build(obj: Object, src_path: Path) -> Path | None: cflags_str = make_flags_str(obj.options["cflags"]) if obj.options["extra_cflags"] is not None: extra_cflags_str = make_flags_str(obj.options["extra_cflags"]) cflags_str += " " + extra_cflags_str used_compiler_versions.add(obj.options["mw_version"]) # Avoid creating duplicate build rules if obj.src_obj_path is None or obj.src_obj_path in source_added: return obj.src_obj_path source_added.add(obj.src_obj_path) # Add MWCC build rule lib_name = obj.options["lib"] n.comment(f"{obj.name}: {lib_name} (linked {obj.completed})") n.build( outputs=obj.src_obj_path, rule="mwcc_sjis" if obj.options["shift_jis"] else "mwcc", inputs=src_path, variables={ "mw_version": Path(obj.options["mw_version"]), "cflags": cflags_str, "basedir": os.path.dirname(obj.src_obj_path), "basefile": obj.src_obj_path.with_suffix(""), }, implicit=( mwcc_sjis_implicit if obj.options["shift_jis"] else mwcc_implicit ), ) # Add ctx build rule if obj.ctx_path is not None: n.build( outputs=obj.ctx_path, rule="decompctx", inputs=src_path, implicit=decompctx, ) # Add host build rule if obj.options["host"] and obj.host_obj_path is not None: n.build( outputs=obj.host_obj_path, rule="host_cc" if src_path.suffix == ".c" else "host_cpp", inputs=src_path, variables={ "basedir": os.path.dirname(obj.host_obj_path), "basefile": obj.host_obj_path.with_suffix(""), }, ) if obj.options["add_to_all"]: host_source_inputs.append(obj.host_obj_path) n.newline() if obj.options["add_to_all"]: source_inputs.append(obj.src_obj_path) return obj.src_obj_path def asm_build( obj: Object, src_path: Path, obj_path: Path | None ) -> Path | None: if obj.options["asflags"] is None: sys.exit("ProjectConfig.asflags missing") asflags_str = make_flags_str(obj.options["asflags"]) if obj.options["extra_asflags"] is not None: extra_asflags_str = make_flags_str(obj.options["extra_asflags"]) asflags_str += " " + extra_asflags_str # Avoid creating duplicate build rules if obj_path is None or obj_path in source_added: return obj_path source_added.add(obj_path) # Add assembler build rule lib_name = obj.options["lib"] n.comment(f"{obj.name}: {lib_name} (linked {obj.completed})") n.build( outputs=obj_path, rule="as", inputs=src_path, variables={"asflags": asflags_str}, implicit=gnu_as_implicit, ) n.newline() if obj.options["add_to_all"]: source_inputs.append(obj_path) return obj_path def add_unit(build_obj, link_step: LinkStep): obj_path, obj_name = build_obj["object"], build_obj["name"] obj = objects.get(obj_name) if obj is None: if config.warn_missing_config and not build_obj["autogenerated"]: print(f"Missing configuration for {obj_name}") link_step.add(obj_path) return link_built_obj = obj.completed built_obj_path: Path | None = None if obj.src_path is not None and obj.src_path.exists(): if obj.src_path.suffix in (".c", ".cp", ".cpp"): # Add MWCC & host build rules built_obj_path = c_build(obj, obj.src_path) elif obj.src_path.suffix == ".s": # Add assembler build rule built_obj_path = asm_build(obj, obj.src_path, obj.src_obj_path) else: sys.exit(f"Unknown source file type {obj.src_path}") else: if config.warn_missing_source or obj.completed: print(f"Missing source file {obj.src_path}") link_built_obj = False # Assembly overrides if obj.asm_path is not None and obj.asm_path.exists(): link_built_obj = True built_obj_path = asm_build(obj, obj.asm_path, obj.asm_obj_path) if link_built_obj and built_obj_path is not None: # Use the source-built object link_step.add(built_obj_path) elif obj_path is not None: # Use the original (extracted) object link_step.add(obj_path) else: lib_name = obj.options["lib"] sys.exit( f"Missing object for {obj_name}: {obj.src_path} {lib_name} {obj}" ) # Add DOL link step link_step = LinkStep(build_config) for unit in build_config["units"]: add_unit(unit, link_step) link_steps.append(link_step) if config.build_rels: # Add REL link steps for module in build_config["modules"]: module_link_step = LinkStep(module) for unit in module["units"]: add_unit(unit, module_link_step) # Add empty object to empty RELs if len(module_link_step.inputs) == 0: if config.rel_empty_file is None: sys.exit("ProjectConfig.rel_empty_file missing") add_unit( { "object": None, "name": config.rel_empty_file, "autogenerated": True, }, module_link_step, ) link_steps.append(module_link_step) n.newline() # Check if all compiler versions exist for mw_version in used_compiler_versions: mw_path = compilers / mw_version / "mwcceppc.exe" if config.compilers_path and not os.path.exists(mw_path): sys.exit(f"Compiler {mw_path} does not exist") # Check if linker exists mw_path = compilers / str(config.linker_version) / "mwldeppc.exe" if config.compilers_path and not os.path.exists(mw_path): sys.exit(f"Linker {mw_path} does not exist") # Add all build steps needed before we link and after compiling objects postcompile_implicit = write_custom_step("post-compile") ### # Link ### for step in link_steps: step.write(n) link_outputs.append(step.output()) n.newline() # Add all build steps needed after linking and before GC/Wii native format generation postlink_implicit = write_custom_step("post-link") ### # Generate DOL ### n.build( outputs=link_steps[0].output(), rule="elf2dol", inputs=link_steps[0].partial_output(), implicit=[*postlink_implicit, dtk], ) ### # Generate RELs ### n.comment("Generate REL(s)") flags = "-w" if len(build_config["links"]) > 1: flags += " -q" n.rule( name="makerel", command=f"{dtk} rel make {flags} -c $config $names @$rspfile", description="REL", rspfile="$rspfile", rspfile_content="$in_newline", ) generated_rels: list[str] = [] for idx, link in enumerate(build_config["links"]): # Map module names to link steps link_steps_local = list( filter( lambda step: step.name in link["modules"], link_steps, ) ) link_steps_local.sort(key=lambda step: step.module_id) # RELs can be the output of multiple link steps, # so we need to filter out duplicates rels_to_generate = list( filter( lambda step: step.module_id != 0 and step.name not in generated_rels, link_steps_local, ) ) if len(rels_to_generate) == 0: continue generated_rels.extend(map(lambda step: step.name, rels_to_generate)) rel_outputs = list( map( lambda step: step.output(), rels_to_generate, ) ) rel_names = list( map( lambda step: step.name, link_steps_local, ) ) rel_names_arg = " ".join(map(lambda name: f"-n {name}", rel_names)) n.build( outputs=rel_outputs, rule="makerel", inputs=list(map(lambda step: step.partial_output(), link_steps_local)), implicit=[dtk, config.config_path], variables={ "config": config.config_path, "rspfile": config.out_path() / f"rel{idx}.rsp", "names": rel_names_arg, }, ) n.newline() # Add all build steps needed post-build (re-building archives and such) postbuild_implicit = write_custom_step("post-build") ### # Helper rule for building all source files ### n.comment("Build all source files") n.build( outputs="all_source", rule="phony", inputs=source_inputs, ) n.newline() ### # Helper rule for building all source files, with a host compiler ### n.comment("Build all source files with a host compiler") n.build( outputs="all_source_host", rule="phony", inputs=host_source_inputs, ) n.newline() ### # Check hash ### n.comment("Check hash") ok_path = build_path / "ok" quiet = "-q " if len(link_steps) > 3 else "" n.rule( name="check", command=f"{dtk} shasum {quiet} -c $in -o $out", description="CHECK $in", ) n.build( outputs=ok_path, rule="check", inputs=config.check_sha_path, implicit=[dtk, *link_outputs, *postbuild_implicit], ) n.newline() ### # Calculate progress ### n.comment("Calculate progress") n.rule( name="progress", command=f"$python {configure_script} $configure_args progress", description="PROGRESS", ) n.build( outputs=progress_path, rule="progress", implicit=[ok_path, configure_script, python_lib, config.config_path], ) ### # Generate progress report ### n.comment("Generate progress report") n.rule( name="report", command=f"{objdiff} report generate -o $out", description="REPORT", ) report_implicit: list[str | Path] = [objdiff, "all_source"] n.build( outputs=report_path, rule="report", implicit=report_implicit, ) ### # Helper tools ### # TODO: make these rules work for RELs too dol_link_step = link_steps[0] dol_elf_path = dol_link_step.partial_output() n.comment("Check for mismatching symbols") n.rule( name="dol_diff", command=f"{dtk} -L error dol diff $in", description=f"DIFF {dol_elf_path}", ) n.build( inputs=[config.config_path, dol_elf_path], outputs="dol_diff", rule="dol_diff", ) n.build( outputs="diff", rule="phony", inputs="dol_diff", ) n.newline() n.comment("Apply symbols from linked ELF") n.rule( name="dol_apply", command=f"{dtk} dol apply $in", description=f"APPLY {dol_elf_path}", ) n.build( inputs=[config.config_path, dol_elf_path], outputs="dol_apply", rule="dol_apply", implicit=[ok_path], ) n.build( outputs="apply", rule="phony", inputs="dol_apply", ) n.newline() ### # Split DOL ### build_config_path = build_path / "config.json" n.comment("Split DOL into relocatable objects") n.rule( name="split", command=f"{dtk} dol split $in $out_dir", description="SPLIT $in", depfile="$out_dir/dep", deps="gcc", ) n.build( inputs=config.config_path, outputs=build_config_path, rule="split", implicit=dtk, variables={"out_dir": build_path}, ) n.newline() ### # Regenerate on change ### n.comment("Reconfigure on change") n.rule( name="configure", command=f"$python {configure_script} $configure_args", generator=True, description=f"RUN {configure_script}", ) n.build( outputs="build.ninja", rule="configure", implicit=[ build_config_path, configure_script, python_lib, python_lib_dir / "ninja_syntax.py", *(config.reconfig_deps or []), ], ) n.newline() ### # Default rule ### n.comment("Default rule") if build_config: if config.non_matching: n.default(link_outputs) else: n.default(progress_path) else: n.default(build_config_path) # Write build.ninja with open("build.ninja", "w", encoding="utf-8") as f: f.write(out.getvalue()) out.close() # Generate objdiff.json def generate_objdiff_config( config: ProjectConfig, objects: dict[str, Object], build_config: dict[str, Any | None], ) -> None: if build_config is None: return objdiff_config: dict[str, Any] = { "min_version": "2.0.0-beta.5", "custom_make": "ninja", "build_target": False, "watch_patterns": [ "*.c", "*.cp", "*.cpp", "*.h", "*.hpp", "*.inc", "*.py", "*.yml", "*.txt", "*.json", ], "units": [], "progress_categories": [], } # decomp.me compiler name mapping COMPILER_MAP = { "GC/1.0": "mwcc_233_144", "GC/1.1": "mwcc_233_159", "GC/1.2.5": "mwcc_233_163", "GC/1.2.5e": "mwcc_233_163e", "GC/1.2.5n": "mwcc_233_163n", "GC/1.3": "mwcc_242_53", "GC/1.3.2": "mwcc_242_81", "GC/1.3.2r": "mwcc_242_81r", "GC/2.0": "mwcc_247_92", "GC/2.5": "mwcc_247_105", "GC/2.6": "mwcc_247_107", "GC/2.7": "mwcc_247_108", "GC/3.0a3": "mwcc_41_51213", "GC/3.0a3.2": "mwcc_41_60126", "GC/3.0a3.3": "mwcc_41_60209", "GC/3.0a3.4": "mwcc_42_60308", "GC/3.0a5": "mwcc_42_60422", "GC/3.0a5.2": "mwcc_41_60831", "GC/3.0": "mwcc_41_60831", "Wii/1.0RC1": "mwcc_42_140", "Wii/0x4201_127": "mwcc_42_142", "Wii/1.0a": "mwcc_42_142", "Wii/1.0": "mwcc_43_145", "Wii/1.1": "mwcc_43_151", "Wii/1.3": "mwcc_43_172", "Wii/1.5": "mwcc_43_188", "Wii/1.6": "mwcc_43_202", "Wii/1.7": "mwcc_43_213", } def add_unit( build_obj: dict[str, Any], module_name: str, progress_categories: list[str] ) -> None: obj_path, obj_name = build_obj["object"], build_obj["name"] base_object = Path(obj_name).with_suffix("") unit_config: dict[str, Any] = { "name": Path(module_name) / base_object, "target_path": obj_path, "metadata": { "auto_generated": build_obj["autogenerated"], "progress_categories": progress_categories, }, } obj = objects.get(obj_name) if obj is None: objdiff_config["units"].append(unit_config) return src_exists = obj.src_path is not None and obj.src_path.exists() if src_exists: unit_config["base_path"] = obj.src_obj_path cflags = obj.options["cflags"] reverse_fn_order = False if type(cflags) is list: for flag in cflags: if not flag.startswith("-inline "): continue for value in flag.split(" ")[1].split(","): if value == "deferred": reverse_fn_order = True elif value == "nodeferred": reverse_fn_order = False # Filter out include directories def keep_flag(flag): return not flag.startswith("-i ") and not flag.startswith("-I ") cflags = list(filter(keep_flag, cflags)) # Add appropriate lang flag if obj.src_path is not None and not any( flag.startswith("-lang") for flag in cflags ): if obj.src_path.suffix in (".cp", ".cpp"): cflags.insert(0, "-lang=c++") else: cflags.insert(0, "-lang=c") compiler_version = COMPILER_MAP.get(obj.options["mw_version"]) if compiler_version is None: print(f"Missing scratch compiler mapping for {obj.options['mw_version']}") else: cflags_str = make_flags_str(cflags) if obj.options["extra_cflags"] is not None: extra_cflags_str = make_flags_str(obj.options["extra_cflags"]) cflags_str += " " + extra_cflags_str unit_config["scratch"] = { "platform": "gc_wii", "compiler": compiler_version, "c_flags": cflags_str, } if src_exists: unit_config["scratch"].update( { "ctx_path": obj.ctx_path, "build_ctx": True, } ) category_opt: list[str] | str = obj.options["progress_category"] if isinstance(category_opt, list): progress_categories.extend(category_opt) elif category_opt is not None: progress_categories.append(category_opt) unit_config["metadata"].update( { "complete": obj.completed, "reverse_fn_order": reverse_fn_order, "source_path": obj.src_path, "progress_categories": progress_categories, } ) objdiff_config["units"].append(unit_config) # Add DOL units for unit in build_config["units"]: progress_categories = [] # Only include a "dol" category if there are any modules # Otherwise it's redundant with the global report measures if len(build_config["modules"]) > 0: progress_categories.append("dol") add_unit(unit, build_config["name"], progress_categories) # Add REL units for module in build_config["modules"]: for unit in module["units"]: progress_categories = [] if config.progress_modules: progress_categories.append("modules") if config.progress_each_module: progress_categories.append(module["name"]) add_unit(unit, module["name"], progress_categories) # Add progress categories def add_category(id: str, name: str): objdiff_config["progress_categories"].append( { "id": id, "name": name, } ) if len(build_config["modules"]) > 0: add_category("dol", "DOL") if config.progress_modules: add_category("modules", "Modules") if config.progress_each_module: for module in build_config["modules"]: add_category(module["name"], module["name"]) for category in config.progress_categories: add_category(category.id, category.name) # Write objdiff.json with open("objdiff.json", "w", encoding="utf-8") as w: def unix_path(input: Any) -> str: return str(input).replace(os.sep, "/") if input else "" json.dump(objdiff_config, w, indent=4, default=unix_path) # Calculate, print and write progress to progress.json def calculate_progress(config: ProjectConfig) -> None: config.validate() objects = config.objects() out_path = config.out_path() build_config = load_build_config(config, out_path / "config.json") if build_config is None: return class ProgressUnit: def __init__(self, name: str) -> None: self.name: str = name self.code_total: int = 0 self.code_progress: int = 0 self.data_total: int = 0 self.data_progress: int = 0 self.objects: Set[Object] = set() self.objects_progress: int = 0 def add(self, build_obj: dict[str, Any]) -> None: self.code_total += build_obj["code_size"] self.data_total += build_obj["data_size"] # Avoid counting the same object in different modules twice include_object = build_obj["name"] not in self.objects if include_object: self.objects.add(build_obj["name"]) if build_obj["autogenerated"]: # Skip autogenerated objects return obj = objects.get(build_obj["name"]) if obj is None or not obj.completed: return self.code_progress += build_obj["code_size"] self.data_progress += build_obj["data_size"] if include_object: self.objects_progress += 1 def code_frac(self) -> float: if self.code_total == 0: return 1.0 return self.code_progress / self.code_total def data_frac(self) -> float: if self.data_total == 0: return 1.0 return self.data_progress / self.data_total progress_units: dict[str, ProgressUnit] = {} if config.progress_all: progress_units["all"] = ProgressUnit("All") progress_units["dol"] = ProgressUnit("DOL") if len(build_config["modules"]) > 0: if config.progress_modules: progress_units["modules"] = ProgressUnit("Modules") if len(config.progress_categories) > 0: for category in config.progress_categories: progress_units[category.id] = ProgressUnit(category.name) if config.progress_each_module: for module in build_config["modules"]: progress_units[module["name"]] = ProgressUnit(module["name"]) def add_unit(id: str, unit: dict[str, Any]) -> None: progress = progress_units.get(id) if progress is not None: progress.add(unit) # Add DOL units for unit in build_config["units"]: add_unit("all", unit) add_unit("dol", unit) obj = objects.get(unit["name"]) if obj is not None: category_opt = obj.options["progress_category"] if isinstance(category_opt, list): for id in category_opt: add_unit(id, unit) elif category_opt is not None: add_unit(category_opt, unit) # Add REL units for module in build_config["modules"]: for unit in module["units"]: add_unit("all", unit) add_unit("modules", unit) add_unit(module["name"], unit) obj = objects.get(unit["name"]) if obj is not None: category_opt = obj.options["progress_category"] if isinstance(category_opt, list): for id in category_opt: add_unit(id, unit) elif category_opt is not None: add_unit(category_opt, unit) # Print human-readable progress print("Progress:") for unit in progress_units.values(): if len(unit.objects) == 0: continue code_frac = unit.code_frac() data_frac = unit.data_frac() print( f" {unit.name}: {code_frac:.2%} code, {data_frac:.2%} data ({unit.objects_progress} / {len(unit.objects)} files)" ) print(f" Code: {unit.code_progress} / {unit.code_total} bytes") print(f" Data: {unit.data_progress} / {unit.data_total} bytes") if config.progress_use_fancy: unit = progress_units.get("all") or progress_units.get("dol") if unit is None or len(unit.objects) == 0: return code_frac = unit.code_frac() data_frac = unit.data_frac() print( "\nYou have {} out of {} {} and {} out of {} {}.".format( math.floor(code_frac * config.progress_code_fancy_frac), config.progress_code_fancy_frac, config.progress_code_fancy_item, math.floor(data_frac * config.progress_data_fancy_frac), config.progress_data_fancy_frac, config.progress_data_fancy_item, ) ) # Generate and write progress.json progress_json: dict[str, Any] = {} for id, unit in progress_units.items(): if len(unit.objects) == 0: continue progress_json[id] = { "code": unit.code_progress, "code/total": unit.code_total, "data": unit.data_progress, "data/total": unit.data_total, } with open(out_path / "progress.json", "w", encoding="utf-8") as w: json.dump(progress_json, w, indent=4)