import atexit
import contextlib
import glob
import math
import os
import re
import subprocess
import sys
import textwrap
import zlib
from collections import OrderedDict
from io import StringIO, TextIOBase
from pathlib import Path
from typing import Generator, List, Optional, Union, cast
from misc.utility.color import print_error, print_info, print_warning
from platform_methods import detect_arch
base_folder = Path(__file__).resolve().parent
compiler_version_cache = None
_scu_folders = set()
def set_scu_folders(scu_folders):
global _scu_folders
_scu_folders = scu_folders
def add_source_files_orig(self, sources, files, allow_gen=False):
if isinstance(files, str):
skip_gen_cpp = "*" in files
files = self.Glob(files)
if skip_gen_cpp and not allow_gen:
files = [f for f in files if not str(f).endswith(".gen.cpp")]
for path in files:
obj = self.Object(path)
if obj in sources:
print_warning('Object "{}" already included in environment sources.'.format(obj))
continue
sources.append(obj)
def add_source_files_scu(self, sources, files, allow_gen=False):
if self["scu_build"] and isinstance(files, str):
if "*." not in files:
return False
subdir = os.path.dirname(files)
subdir = subdir if subdir == "" else subdir + "/"
section_name = self.Dir(subdir).tpath
section_name = section_name.replace("\\", "/")
global _scu_folders
if section_name not in (_scu_folders):
return False
add_source_files_orig(self, sources, subdir + ".scu/scu_*.gen.cpp", True)
return True
return False
def add_source_files(self, sources, files, allow_gen=False):
if not add_source_files_scu(self, sources, files, allow_gen):
add_source_files_orig(self, sources, files, allow_gen)
return False
return True
def redirect_emitter(target, source, env):
"""
Emitter to automatically redirect object/library build files to the `bin/obj` directory,
retaining subfolder structure. External build files will attempt to retain subfolder
structure relative to their environment's parent directory, sorted under `bin/obj/external`.
If `redirect_build_objects` is `False`, or an external build file isn't relative to the
passed environment, this emitter does nothing.
"""
if not env["redirect_build_objects"]:
return target, source
redirected_targets = []
for item in target:
if base_folder in (path := Path(item.get_abspath()).resolve()).parents:
item = env.File(f"#bin/obj/{path.relative_to(base_folder)}")
elif (alt_base := Path(env.Dir(".").get_abspath()).resolve().parent) in path.parents:
item = env.File(f"#bin/obj/external/{path.relative_to(alt_base)}")
else:
print_warning(f'Failed to redirect "{path}"')
redirected_targets.append(item)
return redirected_targets, source
def disable_warnings(self):
if self.msvc and not using_clang(self):
self["WARNLEVEL"] = "/w"
else:
self["WARNLEVEL"] = "-w"
def force_optimization_on_debug(self):
if self["target"] == "template_release":
return
elif self.msvc:
self["OPTIMIZELEVEL"] = "/O2"
else:
self["OPTIMIZELEVEL"] = "-O3"
def add_module_version_string(self, s):
self.module_version_string += "." + s
def get_version_info(module_version_string="", silent=False):
build_name = "custom_build"
if os.getenv("BUILD_NAME") is not None:
build_name = str(os.getenv("BUILD_NAME"))
if not silent:
print_info(f"Using custom build name: '{build_name}'.")
import version
version_info = {
"short_name": str(version.short_name),
"name": str(version.name),
"major": int(version.major),
"minor": int(version.minor),
"patch": int(version.patch),
"status": str(version.status),
"build": str(build_name),
"module_config": str(version.module_config) + module_version_string,
"website": str(version.website),
"docs_branch": str(version.docs),
}
if os.getenv("GODOT_VERSION_STATUS") is not None:
version_info["status"] = str(os.getenv("GODOT_VERSION_STATUS"))
if not silent:
print_info(f"Using version status '{version_info['status']}', overriding the original '{version.status}'.")
return version_info
def get_git_info():
os.chdir(base_folder)
git_hash = ""
git_folder = ".git"
if os.path.isfile(".git"):
with open(".git", "r", encoding="utf-8") as file:
module_folder = file.readline().strip()
if module_folder.startswith("gitdir: "):
git_folder = module_folder[8:]
if os.path.isfile(os.path.join(git_folder, "HEAD")):
with open(os.path.join(git_folder, "HEAD"), "r", encoding="utf8") as file:
head = file.readline().strip()
if head.startswith("ref: "):
ref = head[5:]
parts = git_folder.split("/")
if len(parts) > 2 and parts[-2] == "worktrees":
git_folder = "/".join(parts[0:-2])
head = os.path.join(git_folder, ref)
packedrefs = os.path.join(git_folder, "packed-refs")
if os.path.isfile(head):
with open(head, "r", encoding="utf-8") as file:
git_hash = file.readline().strip()
elif os.path.isfile(packedrefs):
for line in open(packedrefs, "r", encoding="utf-8").read().splitlines():
if line.startswith("#"):
continue
(line_hash, line_ref) = line.split(" ")
if ref == line_ref:
git_hash = line_hash
break
else:
git_hash = head
git_timestamp = 0
if os.path.exists(".git"):
try:
git_timestamp = subprocess.check_output(
["git", "log", "-1", "--pretty=format:%ct", "--no-show-signature", git_hash], encoding="utf-8"
)
except (subprocess.CalledProcessError, OSError):
pass
return {
"git_hash": git_hash,
"git_timestamp": git_timestamp,
}
def get_cmdline_bool(option, default):
"""We use `ARGUMENTS.get()` to check if options were manually overridden on the command line,
and SCons' _text2bool helper to convert them to booleans, otherwise they're handled as strings.
"""
from SCons.Script import ARGUMENTS
from SCons.Variables.BoolVariable import _text2bool
cmdline_val = ARGUMENTS.get(option)
if cmdline_val is not None:
return _text2bool(cmdline_val)
else:
return default
def detect_modules(search_path, recursive=False):
"""Detects and collects a list of C++ modules at specified path
`search_path` - a directory path containing modules. The path may point to
a single module, which may have other nested modules. A module must have
"register_types.h", "SCsub", "config.py" files created to be detected.
`recursive` - if `True`, then all subdirectories are searched for modules as
specified by the `search_path`, otherwise collects all modules under the
`search_path` directory. If the `search_path` is a module, it is collected
in all cases.
Returns an `OrderedDict` with module names as keys, and directory paths as
values. If a path is relative, then it is a built-in module. If a path is
absolute, then it is a custom module collected outside of the engine source.
"""
modules = OrderedDict()
def add_module(path):
module_name = os.path.basename(path)
module_path = path.replace("\\", "/")
modules[module_name] = module_path
def is_engine(path):
version_path = os.path.join(path, "version.py")
if os.path.exists(version_path):
with open(version_path, "r", encoding="utf-8") as f:
if 'short_name = "godot"' in f.read():
return True
return False
def get_files(path):
files = glob.glob(os.path.join(path, "*"))
files.sort()
return files
if not recursive:
if is_module(search_path):
add_module(search_path)
for path in get_files(search_path):
if is_engine(path):
continue
if is_module(path):
add_module(path)
else:
to_search = [search_path]
while to_search:
path = to_search.pop()
if is_module(path):
add_module(path)
for child in get_files(path):
if not os.path.isdir(child):
continue
if is_engine(child):
continue
to_search.insert(0, child)
return modules
def is_module(path):
if not os.path.isdir(path):
return False
must_exist = ["register_types.h", "SCsub", "config.py"]
for f in must_exist:
if not os.path.exists(os.path.join(path, f)):
return False
return True
def convert_custom_modules_path(path):
if not path:
return path
path = os.path.realpath(os.path.expanduser(os.path.expandvars(path)))
err_msg = "Build option 'custom_modules' must %s"
if not os.path.isdir(path):
raise ValueError(err_msg % "point to an existing directory.")
if path == os.path.realpath("modules"):
raise ValueError(err_msg % "be a directory other than built-in `modules` directory.")
return path
def module_add_dependencies(self, module, dependencies, optional=False):
"""
Adds dependencies for a given module.
Meant to be used in module `can_build` methods.
"""
if module not in self.module_dependencies:
self.module_dependencies[module] = [[], []]
if optional:
self.module_dependencies[module][1].extend(dependencies)
else:
self.module_dependencies[module][0].extend(dependencies)
def module_check_dependencies(self, module):
"""
Checks if module dependencies are enabled for a given module,
and prints a warning if they aren't.
Meant to be used in module `can_build` methods.
Returns a boolean (True if dependencies are satisfied).
"""
missing_deps = set()
required_deps = self.module_dependencies[module][0] if module in self.module_dependencies else []
for dep in required_deps:
opt = "module_{}_enabled".format(dep)
if opt not in self or not self[opt] or not module_check_dependencies(self, dep):
missing_deps.add(dep)
if missing_deps:
if module not in self.disabled_modules:
print_warning(
"Disabling '{}' module as the following dependencies are not satisfied: {}".format(
module, ", ".join(missing_deps)
)
)
self.disabled_modules.add(module)
return False
else:
return True
def sort_module_list(env):
deps = {k: v[0] + list(filter(lambda x: x in env.module_list, v[1])) for k, v in env.module_dependencies.items()}
frontier = list(env.module_list.keys())
explored = []
while len(frontier):
cur = frontier.pop()
deps_list = deps[cur] if cur in deps else []
if len(deps_list) and any([d not in explored for d in deps_list]):
frontier.insert(0, cur)
continue
explored.append(cur)
for k in explored:
env.module_list.move_to_end(k)
def use_windows_spawn_fix(self, platform=None):
if os.name != "nt":
return
def mySubProcess(cmdline, env):
startupinfo = subprocess.STARTUPINFO()
startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW
popen_args = {
"stdin": subprocess.PIPE,
"stdout": subprocess.PIPE,
"stderr": subprocess.PIPE,
"startupinfo": startupinfo,
"shell": False,
"env": env,
}
popen_args["text"] = True
proc = subprocess.Popen(cmdline, **popen_args)
_, err = proc.communicate()
rv = proc.wait()
if rv:
print_error(err)
elif len(err) > 0 and not err.isspace():
print(err)
return rv
def mySpawn(sh, escape, cmd, args, env):
if cmd == "del":
os.remove(args[1])
return 0
newargs = " ".join(args[1:])
cmdline = cmd + " " + newargs
rv = 0
env = {str(key): str(value) for key, value in iter(env.items())}
rv = mySubProcess(cmdline, env)
return rv
self["SPAWN"] = mySpawn
def no_verbose(env):
from misc.utility.color import Ansi, is_stdout_color
colors = [Ansi.BLUE, Ansi.BOLD, Ansi.REGULAR, Ansi.RESET] if is_stdout_color() else ["", "", "", ""]
compile_source_message = "{}Compiling {}$SOURCE{} ...{}".format(*colors)
java_compile_source_message = "{}Compiling {}$SOURCE{} ...{}".format(*colors)
compile_shared_source_message = "{}Compiling shared {}$SOURCE{} ...{}".format(*colors)
link_program_message = "{}Linking Program {}$TARGET{} ...{}".format(*colors)
link_library_message = "{}Linking Static Library {}$TARGET{} ...{}".format(*colors)
ranlib_library_message = "{}Ranlib Library {}$TARGET{} ...{}".format(*colors)
link_shared_library_message = "{}Linking Shared Library {}$TARGET{} ...{}".format(*colors)
java_library_message = "{}Creating Java Archive {}$TARGET{} ...{}".format(*colors)
compiled_resource_message = "{}Creating Compiled Resource {}$TARGET{} ...{}".format(*colors)
zip_archive_message = "{}Archiving {}$TARGET{} ...{}".format(*colors)
generated_file_message = "{}Generating {}$TARGET{} ...{}".format(*colors)
env["CXXCOMSTR"] = compile_source_message
env["CCCOMSTR"] = compile_source_message
env["SHCCCOMSTR"] = compile_shared_source_message
env["SHCXXCOMSTR"] = compile_shared_source_message
env["ARCOMSTR"] = link_library_message
env["RANLIBCOMSTR"] = ranlib_library_message
env["SHLINKCOMSTR"] = link_shared_library_message
env["LINKCOMSTR"] = link_program_message
env["JARCOMSTR"] = java_library_message
env["JAVACCOMSTR"] = java_compile_source_message
env["RCCOMSTR"] = compiled_resource_message
env["ZIPCOMSTR"] = zip_archive_message
env["GENCOMSTR"] = generated_file_message
def detect_visual_c_compiler_version(tools_env):
vc_chosen_compiler_index = -1
vc_chosen_compiler_str = ""
if "VCTOOLSINSTALLDIR" in tools_env:
vc_amd64_compiler_detection_index = (
tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX64\\X64;")
)
if vc_amd64_compiler_detection_index > -1:
vc_chosen_compiler_index = vc_amd64_compiler_detection_index
vc_chosen_compiler_str = "amd64"
vc_amd64_x86_compiler_detection_index = (
tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX64\\X86;")
)
if vc_amd64_x86_compiler_detection_index > -1 and (
vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_amd64_x86_compiler_detection_index
):
vc_chosen_compiler_index = vc_amd64_x86_compiler_detection_index
vc_chosen_compiler_str = "amd64_x86"
vc_x86_compiler_detection_index = (
tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX86\\X86;")
)
if vc_x86_compiler_detection_index > -1 and (
vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_compiler_detection_index
):
vc_chosen_compiler_index = vc_x86_compiler_detection_index
vc_chosen_compiler_str = "x86"
vc_x86_amd64_compiler_detection_index = (
tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX86\\X64;")
)
if vc_x86_amd64_compiler_detection_index > -1 and (
vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_amd64_compiler_detection_index
):
vc_chosen_compiler_str = "x86_amd64"
return vc_chosen_compiler_str
def find_visual_c_batch_file(env):
from SCons.Tool.MSCommon.vc import find_batch_file, find_vc_pdir, get_default_version, get_host_target
msvc_version = get_default_version(env)
if env.scons_version >= (4, 4, 0):
(host_platform, target_platform, _) = get_host_target(env, msvc_version)
else:
(host_platform, target_platform, _) = get_host_target(env)
if env.scons_version < (4, 6, 0):
return find_batch_file(env, msvc_version, host_platform, target_platform)[0]
product_dir = find_vc_pdir(msvc_version=msvc_version, env=env)
return find_batch_file(msvc_version, host_platform, target_platform, product_dir)[0]
def generate_cpp_hint_file(filename):
if os.path.isfile(filename):
pass
else:
try:
with open(filename, "w", encoding="utf-8", newline="\n") as fd:
fd.write("#define GDCLASS(m_class, m_inherits)\n")
for name in ["GDVIRTUAL", "EXBIND", "MODBIND"]:
for count in range(13):
for suffix in ["", "R", "C", "RC"]:
fd.write(f"#define {name}{count}{suffix}(")
if "R" in suffix:
fd.write("m_ret, ")
fd.write("m_name")
for idx in range(1, count + 1):
fd.write(f", type{idx}")
fd.write(")\n")
except OSError:
print_warning("Could not write cpp.hint file.")
def glob_recursive(pattern, node="."):
from SCons import Node
from SCons.Script import Glob
results = []
for f in Glob(str(node) + "/*", source=True):
if type(f) is Node.FS.Dir:
results += glob_recursive(pattern, f)
results += Glob(str(node) + "/" + pattern, source=True)
return results
def precious_program(env, program, sources, **args):
program = env.Program(program, sources, **args)
env.Precious(program)
return program
def add_shared_library(env, name, sources, **args):
library = env.SharedLibrary(name, sources, **args)
env.NoCache(library)
return library
def add_library(env, name, sources, **args):
library = env.Library(name, sources, **args)
env.NoCache(library)
return library
def add_program(env, name, sources, **args):
program = env.Program(name, sources, **args)
env.NoCache(program)
return program
def CommandNoCache(env, target, sources, command, **args):
result = env.Command(target, sources, command, **args)
env.NoCache(result)
return result
def Run(env, function, comstr="$GENCOMSTR"):
from SCons.Script import Action
return Action(function, comstr)
def detect_darwin_toolchain_path(env):
var_name = "APPLE_TOOLCHAIN_PATH"
if not env[var_name]:
try:
xcode_path = subprocess.check_output(["xcode-select", "-p"]).strip().decode("utf-8")
if xcode_path:
env[var_name] = xcode_path + "/Toolchains/XcodeDefault.xctoolchain"
except (subprocess.CalledProcessError, OSError):
print_error("Failed to find SDK path while running 'xcode-select -p'.")
raise
def detect_darwin_sdk_path(platform, env):
sdk_name = ""
if platform == "macos":
sdk_name = "macosx"
var_name = "MACOS_SDK_PATH"
elif platform == "ios":
sdk_name = "iphoneos"
var_name = "IOS_SDK_PATH"
elif platform == "iossimulator":
sdk_name = "iphonesimulator"
var_name = "IOS_SDK_PATH"
elif platform == "visionos":
sdk_name = "xros"
var_name = "VISIONOS_SDK_PATH"
elif platform == "visionossimulator":
sdk_name = "xrsimulator"
var_name = "VISIONOS_SDK_PATH"
else:
raise Exception("Invalid platform argument passed to detect_darwin_sdk_path")
if not env[var_name]:
try:
sdk_path = subprocess.check_output(["xcrun", "--sdk", sdk_name, "--show-sdk-path"]).strip().decode("utf-8")
if sdk_path:
env[var_name] = sdk_path
except (subprocess.CalledProcessError, OSError):
print_error("Failed to find SDK path while running 'xcrun --sdk {} --show-sdk-path'.".format(sdk_name))
raise
def is_apple_clang(env):
import shlex
if env["platform"] not in ["macos", "ios"]:
return False
if not using_clang(env):
return False
try:
version = (
subprocess.check_output(shlex.split(env.subst(env["CXX"]), posix=False) + ["--version"])
.strip()
.decode("utf-8")
)
except (subprocess.CalledProcessError, OSError):
print_warning("Couldn't parse CXX environment variable to infer compiler version.")
return False
return version.startswith("Apple")
def get_compiler_version(env):
"""
Returns a dictionary with various version information:
- major, minor, patch: Version following semantic versioning system
- metadata1, metadata2: Extra information
- date: Date of the build
"""
global compiler_version_cache
if compiler_version_cache is not None:
return compiler_version_cache
import shlex
ret = {
"major": -1,
"minor": -1,
"patch": -1,
"metadata1": "",
"metadata2": "",
"date": "",
"apple_major": -1,
"apple_minor": -1,
"apple_patch1": -1,
"apple_patch2": -1,
"apple_patch3": -1,
}
if env.msvc and not using_clang(env):
try:
args = [
env["VSWHERE"],
"-latest",
"-prerelease",
"-products",
"*",
"-requires",
"Microsoft.Component.MSBuild",
"-utf8",
]
version = subprocess.check_output(args, encoding="utf-8").strip()
for line in version.splitlines():
split = line.split(":", 1)
if split[0] == "catalog_productDisplayVersion":
sem_ver = split[1].split(".")
ret["major"] = int(sem_ver[0])
ret["minor"] = int(sem_ver[1])
ret["patch"] = int(sem_ver[2].split()[0])
if split[0] == "catalog_buildVersion":
ret["metadata1"] = split[1]
except (subprocess.CalledProcessError, OSError):
print_warning("Couldn't find vswhere to determine compiler version.")
return update_compiler_version_cache(ret)
try:
version = subprocess.check_output(
shlex.split(env.subst(env["CXX"]), posix=False) + ["--version"], shell=(os.name == "nt"), encoding="utf-8"
).strip()
except (subprocess.CalledProcessError, OSError):
print_warning("Couldn't parse CXX environment variable to infer compiler version.")
return update_compiler_version_cache(ret)
match = re.search(
r"(?:(?<=version )|(?<=\) )|(?<=^))"
r"(?P<major>\d+)"
r"(?:\.(?P<minor>\d*))?"
r"(?:\.(?P<patch>\d*))?"
r"(?:-(?P<metadata1>[0-9a-zA-Z-]*))?"
r"(?:\+(?P<metadata2>[0-9a-zA-Z-]*))?"
r"(?: (?P<date>[0-9]{8}|[0-9]{6})(?![0-9a-zA-Z]))?",
version,
)
if match is not None:
for key, value in match.groupdict().items():
if value is not None:
ret[key] = value
match_apple = re.search(
r"(?:(?<=clang-)|(?<=\) )|(?<=^))"
r"(?P<apple_major>\d+)"
r"(?:\.(?P<apple_minor>\d*))?"
r"(?:\.(?P<apple_patch1>\d*))?"
r"(?:\.(?P<apple_patch2>\d*))?"
r"(?:\.(?P<apple_patch3>\d*))?",
version,
)
if match_apple is not None:
for key, value in match_apple.groupdict().items():
if value is not None:
ret[key] = value
for key in [
"major",
"minor",
"patch",
"apple_major",
"apple_minor",
"apple_patch1",
"apple_patch2",
"apple_patch3",
]:
ret[key] = int(ret[key] or -1)
return update_compiler_version_cache(ret)
def update_compiler_version_cache(value):
global compiler_version_cache
compiler_version_cache = value
return value
def using_gcc(env):
return "gcc" in os.path.basename(env["CC"])
def using_clang(env):
return "clang" in os.path.basename(env["CC"])
def using_emcc(env):
return "emcc" in os.path.basename(env["CC"])
def show_progress(env):
if env["ninja"]:
return
NODE_COUNT_FILENAME = base_folder / ".scons_node_count"
class ShowProgress:
def __init__(self):
self.count = 0
self.max = 0
try:
with open(NODE_COUNT_FILENAME, "r", encoding="utf-8") as f:
self.max = int(f.readline())
except OSError:
pass
self.display = cast(bool, env["progress"] and sys.stdout.isatty())
if self.display and not self.max:
print_info("Performing initial build, progress percentage unavailable!")
self.display = False
def __call__(self, node, *args, **kw):
self.count += 1
if self.display:
percent = int(min(self.count * 100 / self.max, 100))
sys.stdout.write(f"\r[{percent:3d}%] ")
sys.stdout.flush()
from SCons.Script import Progress
from SCons.Script.Main import GetBuildFailures
progressor = ShowProgress()
Progress(progressor)
def progress_finish():
if GetBuildFailures() or not progressor.count:
return
try:
with open(NODE_COUNT_FILENAME, "w", encoding="utf-8", newline="\n") as f:
f.write(f"{progressor.count}\n")
except OSError:
pass
atexit.register(progress_finish)
def convert_size(size_bytes: int) -> str:
if size_bytes == 0:
return "0 bytes"
SIZE_NAMES = ["bytes", "KiB", "MiB", "GiB", "TiB", "PiB", "EiB", "ZiB", "YiB"]
index = math.floor(math.log(size_bytes, 1024))
power = math.pow(1024, index)
size = round(size_bytes / power, 2)
return f"{size} {SIZE_NAMES[index]}"
def get_size(start_path: str = ".") -> int:
total_size = 0
for dirpath, _, filenames in os.walk(start_path):
for file in filenames:
path = os.path.join(dirpath, file)
total_size += os.path.getsize(path)
return total_size
def clean_cache(cache_path: str, cache_limit: int, verbose: bool) -> None:
if not cache_limit:
return
files = glob.glob(os.path.join(cache_path, "*", "*"))
if not files:
return
stats = []
for file in files:
try:
stats.append((file, *os.stat(file)[6:8]))
except OSError:
print_error(f'Failed to access cache file "{file}"; skipping.')
stats.sort(key=lambda x: x[2], reverse=True)
sum = 0
for index, stat in enumerate(stats):
sum += stat[1]
if sum > cache_limit:
purge = [x[0] for x in stats[index:]]
count = len(purge)
for file in purge:
try:
os.remove(file)
except OSError:
print_error(f'Failed to remove cache file "{file}"; skipping.')
count -= 1
if verbose and count:
print_info(f"Purged {count} file{'s' if count else ''} from cache.")
break
def prepare_cache(env) -> None:
cache_path = ""
if env["cache_path"]:
cache_path = cast(str, env["cache_path"])
elif os.environ.get("SCONS_CACHE"):
print_warning("Environment variable `SCONS_CACHE` is deprecated; use `cache_path` argument instead.")
cache_path = cast(str, os.environ.get("SCONS_CACHE"))
if not cache_path:
return
env.CacheDir(cache_path)
print(f'SCons cache enabled... (path: "{cache_path}")')
if env["cache_limit"]:
cache_limit = float(env["cache_limit"])
elif os.environ.get("SCONS_CACHE_LIMIT"):
print_warning("Environment variable `SCONS_CACHE_LIMIT` is deprecated; use `cache_limit` argument instead.")
cache_limit = float(os.getenv("SCONS_CACHE_LIMIT", "0")) / 1024
cache_limit = max(0, int(cache_limit * 1024 * 1024 * 1024))
if env["verbose"]:
print_info(
f"Current cache size is {convert_size(get_size(cache_path))}"
+ (f" (limit: {convert_size(cache_limit)})" if cache_limit else "")
)
atexit.register(clean_cache, cache_path, cache_limit, env["verbose"])
def prepare_purge(env):
from SCons.Script.Main import GetBuildFailures
def purge_flaky_files():
paths_to_keep = [env["ninja_file"]]
for build_failure in GetBuildFailures():
path = build_failure.node.path
if os.path.isfile(path) and path not in paths_to_keep:
os.remove(path)
atexit.register(purge_flaky_files)
def prepare_timer():
import time
def print_elapsed_time(time_at_start: float):
time_elapsed = time.time() - time_at_start
time_formatted = time.strftime("%H:%M:%S", time.gmtime(time_elapsed))
time_centiseconds = (time_elapsed % 1) * 100
print_info(f"Time elapsed: {time_formatted}.{time_centiseconds:02.0f}")
atexit.register(print_elapsed_time, time.time())
def dump(env):
"""
Dumps latest build information for debugging purposes and external tools.
"""
with open(".scons_env.json", "w", encoding="utf-8", newline="\n") as file:
file.write(env.Dump(format="json"))
def generate_vs_project(env, original_args, project_name="godot"):
def glob_recursive_2(pattern, dirs, node="."):
from SCons import Node
from SCons.Script import Glob
results = []
for f in Glob(str(node) + "/*", source=True):
if type(f) is Node.FS.Dir:
results += glob_recursive_2(pattern, dirs, f)
r = Glob(str(node) + "/" + pattern, source=True)
if len(r) > 0 and str(node) not in dirs:
d = ""
for part in str(node).split("\\"):
d += part
if d not in dirs:
dirs.append(d)
d += "\\"
results += r
return results
def get_bool(args, option, default):
from SCons.Variables.BoolVariable import _text2bool
val = args.get(option, default)
if val is not None:
try:
return _text2bool(val)
except (ValueError, AttributeError):
return default
else:
return default
def format_key_value(v):
if type(v) in [tuple, list]:
return v[0] if len(v) == 1 else f"{v[0]}={v[1]}"
return v
def get_dependencies(file, env, exts, headers, sources, others):
for child in file.children():
if isinstance(child, str):
child = env.File(x)
fname = ""
try:
fname = child.path
except AttributeError:
pass
if fname:
parts = os.path.splitext(fname)
if len(parts) > 1:
ext = parts[1].lower()
if ext in exts["sources"]:
sources += [fname]
elif ext in exts["headers"]:
headers += [fname]
elif ext in exts["others"]:
others += [fname]
get_dependencies(child, env, exts, headers, sources, others)
filtered_args = original_args.copy()
filtered_args.pop("vsproj", None)
filtered_args.pop("vsproj_gen_only", None)
filtered_args.pop("vsproj_props_only", None)
filtered_args.pop("progress", None)
filtered_args.pop("platform", None)
filtered_args.pop("target", None)
filtered_args.pop("arch", None)
platform = env["platform"]
target = env["target"]
arch = env["arch"]
host_arch = detect_arch()
host_platform = "windows"
if (
sys.platform.startswith("linux")
or sys.platform.startswith("dragonfly")
or sys.platform.startswith("freebsd")
or sys.platform.startswith("netbsd")
or sys.platform.startswith("openbsd")
):
host_platform = "linuxbsd"
elif sys.platform == "darwin":
host_platform = "macos"
vs_configuration = {}
host_vs_configuration = {}
common_build_prefix = []
confs = []
for x in sorted(glob.glob("platform/*")):
if not os.path.isdir(x) or not os.path.exists(x + "/msvs.py"):
continue
tmppath = "./" + x
sys.path.insert(0, tmppath)
import msvs
vs_plats = []
vs_confs = []
try:
platform_name = x[9:]
vs_plats = msvs.get_platforms()
vs_confs = msvs.get_configurations()
val = []
for plat in vs_plats:
val += [{"platform": plat[0], "architecture": plat[1]}]
vsconf = {"platform": platform_name, "targets": vs_confs, "arches": val}
confs += [vsconf]
if platform == platform_name:
common_build_prefix = msvs.get_build_prefix(env)
vs_configuration = vsconf
if platform_name == host_platform:
host_vs_configuration = vsconf
for a in vsconf["arches"]:
if host_arch == a["architecture"]:
host_arch = a["platform"]
break
except Exception:
pass
sys.path.remove(tmppath)
sys.modules.pop("msvs")
extensions = {}
extensions["headers"] = [".h", ".hh", ".hpp", ".hxx", ".inc"]
extensions["sources"] = [".c", ".cc", ".cpp", ".cxx", ".m", ".mm", ".java"]
extensions["others"] = [".natvis", ".glsl", ".rc"]
headers = []
headers_dirs = []
for ext in extensions["headers"]:
for file in glob_recursive_2("*" + ext, headers_dirs):
headers.append(str(file).replace("/", "\\"))
sources = []
sources_dirs = []
for ext in extensions["sources"]:
for file in glob_recursive_2("*" + ext, sources_dirs):
sources.append(str(file).replace("/", "\\"))
others = []
others_dirs = []
for ext in extensions["others"]:
for file in glob_recursive_2("*" + ext, others_dirs):
others.append(str(file).replace("/", "\\"))
skip_filters = False
import hashlib
import json
md5 = hashlib.md5(
json.dumps(sorted(headers + headers_dirs + sources + sources_dirs + others + others_dirs)).encode("utf-8")
).hexdigest()
if os.path.exists(f"{project_name}.vcxproj.filters"):
with open(f"{project_name}.vcxproj.filters", "r", encoding="utf-8") as file:
existing_filters = file.read()
match = re.search(r"(?ms)^<!-- CHECKSUM$.([0-9a-f]{32})", existing_filters)
if match is not None and md5 == match.group(1):
skip_filters = True
import uuid
if not skip_filters:
print(f"Regenerating {project_name}.vcxproj.filters")
with open("misc/msvs/vcxproj.filters.template", "r", encoding="utf-8") as file:
filters_template = file.read()
for i in range(1, 10):
filters_template = filters_template.replace(f"%%UUID{i}%%", str(uuid.uuid4()))
filters = ""
for d in headers_dirs:
filters += f'<Filter Include="Header Files\\{d}"><UniqueIdentifier>{{{str(uuid.uuid4())}}}</UniqueIdentifier></Filter>\n'
for d in sources_dirs:
filters += f'<Filter Include="Source Files\\{d}"><UniqueIdentifier>{{{str(uuid.uuid4())}}}</UniqueIdentifier></Filter>\n'
for d in others_dirs:
filters += f'<Filter Include="Other Files\\{d}"><UniqueIdentifier>{{{str(uuid.uuid4())}}}</UniqueIdentifier></Filter>\n'
filters_template = filters_template.replace("%%FILTERS%%", filters)
filters = ""
for file in headers:
filters += (
f'<ClInclude Include="{file}"><Filter>Header Files\\{os.path.dirname(file)}</Filter></ClInclude>\n'
)
filters_template = filters_template.replace("%%INCLUDES%%", filters)
filters = ""
for file in sources:
filters += (
f'<ClCompile Include="{file}"><Filter>Source Files\\{os.path.dirname(file)}</Filter></ClCompile>\n'
)
filters_template = filters_template.replace("%%COMPILES%%", filters)
filters = ""
for file in others:
filters += f'<None Include="{file}"><Filter>Other Files\\{os.path.dirname(file)}</Filter></None>\n'
filters_template = filters_template.replace("%%OTHERS%%", filters)
filters_template = filters_template.replace("%%HASH%%", md5)
with open(f"{project_name}.vcxproj.filters", "w", encoding="utf-8", newline="\r\n") as f:
f.write(filters_template)
headers_active = []
sources_active = []
others_active = []
get_dependencies(
env.File(f"#bin/godot{env['PROGSUFFIX']}"), env, extensions, headers_active, sources_active, others_active
)
all_items = []
properties = []
activeItems = []
extraItems = []
set_headers = set(headers_active)
set_sources = set(sources_active)
set_others = set(others_active)
for file in headers:
base_path = os.path.dirname(file).replace("\\", "_")
all_items.append(f'<ClInclude Include="{file}">')
all_items.append(
f" <ExcludedFromBuild Condition=\"!$(ActiveProjectItemList_{base_path}.Contains(';{file};'))\">true</ExcludedFromBuild>"
)
all_items.append("</ClInclude>")
if file in set_headers:
activeItems.append(file)
for file in sources:
base_path = os.path.dirname(file).replace("\\", "_")
all_items.append(f'<ClCompile Include="{file}">')
all_items.append(
f" <ExcludedFromBuild Condition=\"!$(ActiveProjectItemList_{base_path}.Contains(';{file};'))\">true</ExcludedFromBuild>"
)
all_items.append("</ClCompile>")
if file in set_sources:
activeItems.append(file)
for file in others:
base_path = os.path.dirname(file).replace("\\", "_")
all_items.append(f'<None Include="{file}">')
all_items.append(
f" <ExcludedFromBuild Condition=\"!$(ActiveProjectItemList_{base_path}.Contains(';{file};'))\">true</ExcludedFromBuild>"
)
all_items.append("</None>")
if file in set_others:
activeItems.append(file)
if vs_configuration:
vsconf = ""
for a in vs_configuration["arches"]:
if arch == a["architecture"]:
vsconf = f"{target}|{a['platform']}"
break
condition = "'$(GodotConfiguration)|$(GodotPlatform)'=='" + vsconf + "'"
itemlist = {}
for item in activeItems:
key = os.path.dirname(item).replace("\\", "_")
if key not in itemlist:
itemlist[key] = [item]
else:
itemlist[key] += [item]
for x in itemlist.keys():
properties.append(
"<ActiveProjectItemList_%s>;%s;</ActiveProjectItemList_%s>" % (x, ";".join(itemlist[x]), x)
)
output = os.path.join("bin", f"godot{env['PROGSUFFIX']}")
env.Append(VSHINT_DEFINES=[f"MODULE_{module.upper()}_ENABLED" for module in env.module_list])
with open("misc/msvs/props.template", "r", encoding="utf-8") as file:
props_template = file.read()
props_template = props_template.replace("%%CONDITION%%", condition)
props_template = props_template.replace("%%PROPERTIES%%", "\n ".join(properties))
props_template = props_template.replace("%%EXTRA_ITEMS%%", "\n ".join(extraItems))
props_template = props_template.replace("%%OUTPUT%%", output)
proplist = [format_key_value(j) for j in list(env["CPPDEFINES"])]
proplist += [format_key_value(j) for j in env.get("VSHINT_DEFINES", [])]
props_template = props_template.replace("%%DEFINES%%", ";".join(proplist))
proplist = [str(j) for j in env["CPPPATH"]]
proplist += [str(j) for j in env.get("VSHINT_INCLUDES", [])]
proplist += [str(j) for j in get_default_include_paths(env)]
props_template = props_template.replace("%%INCLUDES%%", ";".join(proplist))
proplist = [env.subst("$CCFLAGS")]
proplist += [env.subst("$CXXFLAGS")]
proplist += [env.subst("$VSHINT_OPTIONS")]
props_template = props_template.replace("%%OPTIONS%%", " ".join(proplist))
common_build_postfix = [
"--directory="$(ProjectDir.TrimEnd('\\'))"",
"progress=no",
f"platform={platform}",
f"target={target}",
f"arch={arch}",
]
for arg, value in filtered_args.items():
common_build_postfix.append(f"{arg}={value}")
cmd_rebuild = [
"vsproj=yes",
"vsproj_props_only=yes",
"vsproj_gen_only=no",
f"vsproj_name={project_name}",
] + common_build_postfix
cmd_clean = [
"--clean",
] + common_build_postfix
commands = "scons"
if len(common_build_prefix) == 0:
commands = "echo Starting SCons & " + commands
else:
common_build_prefix[0] = "echo Starting SCons & " + common_build_prefix[0]
cmd = " ".join(common_build_prefix + [" ".join([commands] + common_build_postfix)])
props_template = props_template.replace("%%BUILD%%", cmd)
cmd = " ".join(common_build_prefix + [" ".join([commands] + cmd_rebuild)])
props_template = props_template.replace("%%REBUILD%%", cmd)
cmd = " ".join(common_build_prefix + [" ".join([commands] + cmd_clean)])
props_template = props_template.replace("%%CLEAN%%", cmd)
with open(
f"{project_name}.{platform}.{target}.{arch}.generated.props", "w", encoding="utf-8", newline="\r\n"
) as f:
f.write(props_template)
proj_uuid = str(uuid.uuid4())
sln_uuid = str(uuid.uuid4())
if os.path.exists(f"{project_name}.sln"):
for line in open(f"{project_name}.sln", "r", encoding="utf-8").read().splitlines():
if line.startswith('Project("{8BC9CEB8-8B4A-11D0-8D11-00A0C91BC942}")'):
proj_uuid = re.search(
r"\"{(\b[0-9a-fA-F]{8}\b-[0-9a-fA-F]{4}-[0-9a-fA-F]{4}-[0-9a-fA-F]{4}-\b[0-9a-fA-F]{12}\b)}\"$",
line,
).group(1)
elif line.strip().startswith("SolutionGuid ="):
sln_uuid = re.search(
r"{(\b[0-9a-fA-F]{8}\b-[0-9a-fA-F]{4}-[0-9a-fA-F]{4}-[0-9a-fA-F]{4}-\b[0-9a-fA-F]{12}\b)}", line
).group(1)
break
configurations = []
imports = []
properties = []
section1 = []
section2 = []
for conf in confs:
godot_platform = conf["platform"]
has_editor = "editor" in conf["targets"]
if has_editor and godot_platform != host_vs_configuration["platform"]:
continue
for p in conf["arches"]:
sln_plat = p["platform"]
proj_plat = sln_plat
godot_arch = p["architecture"]
if not has_editor:
section1 += [f"editor|{sln_plat} = editor|{proj_plat}"]
section2 += [f"{{{proj_uuid}}}.editor|{proj_plat}.ActiveCfg = editor|{host_arch}"]
configurations += [
f'<ProjectConfiguration Include="editor|{proj_plat}">',
" <Configuration>editor</Configuration>",
f" <Platform>{proj_plat}</Platform>",
"</ProjectConfiguration>",
]
properties += [
f"<PropertyGroup Condition=\"'$(Configuration)|$(Platform)'=='editor|{proj_plat}'\">",
" <GodotConfiguration>editor</GodotConfiguration>",
f" <GodotPlatform>{proj_plat}</GodotPlatform>",
"</PropertyGroup>",
]
for t in conf["targets"]:
godot_target = t
if godot_platform == "windows" and godot_target == "editor" and godot_arch == "x86_32":
sln_plat = "x86"
configurations += [
f'<ProjectConfiguration Include="{godot_target}|{proj_plat}">',
f" <Configuration>{godot_target}</Configuration>",
f" <Platform>{proj_plat}</Platform>",
"</ProjectConfiguration>",
]
properties += [
f"<PropertyGroup Condition=\"'$(Configuration)|$(Platform)'=='{godot_target}|{proj_plat}'\">",
f" <GodotConfiguration>{godot_target}</GodotConfiguration>",
f" <GodotPlatform>{proj_plat}</GodotPlatform>",
"</PropertyGroup>",
]
p = f"{project_name}.{godot_platform}.{godot_target}.{godot_arch}.generated.props"
imports += [
f'<Import Project="$(MSBuildProjectDirectory)\\{p}" Condition="Exists(\'$(MSBuildProjectDirectory)\\{p}\')"/>'
]
section1 += [f"{godot_target}|{sln_plat} = {godot_target}|{sln_plat}"]
section2 += [
f"{{{proj_uuid}}}.{godot_target}|{sln_plat}.ActiveCfg = {godot_target}|{proj_plat}",
f"{{{proj_uuid}}}.{godot_target}|{sln_plat}.Build.0 = {godot_target}|{proj_plat}",
]
imports += [
f'<Import Project="$(MSBuildProjectDirectory)\\{project_name}.vs.user.props" Condition="Exists(\'$(MSBuildProjectDirectory)\\{project_name}.vs.user.props\')"/>'
]
section1 = sorted(section1)
section2 = sorted(section2)
if not get_bool(original_args, "vsproj_props_only", False):
with open("misc/msvs/vcxproj.template", "r", encoding="utf-8") as file:
proj_template = file.read()
proj_template = proj_template.replace("%%UUID%%", proj_uuid)
proj_template = proj_template.replace("%%CONFS%%", "\n ".join(configurations))
proj_template = proj_template.replace("%%IMPORTS%%", "\n ".join(imports))
proj_template = proj_template.replace("%%DEFAULT_ITEMS%%", "\n ".join(all_items))
proj_template = proj_template.replace("%%PROPERTIES%%", "\n ".join(properties))
with open(f"{project_name}.vcxproj", "w", encoding="utf-8", newline="\r\n") as f:
f.write(proj_template)
if not get_bool(original_args, "vsproj_props_only", False):
with open("misc/msvs/sln.template", "r", encoding="utf-8") as file:
sln_template = file.read()
sln_template = sln_template.replace("%%NAME%%", project_name)
sln_template = sln_template.replace("%%UUID%%", proj_uuid)
sln_template = sln_template.replace("%%SLNUUID%%", sln_uuid)
sln_template = sln_template.replace("%%SECTION1%%", "\n\t\t".join(section1))
sln_template = sln_template.replace("%%SECTION2%%", "\n\t\t".join(section2))
with open(f"{project_name}.sln", "w", encoding="utf-8", newline="\r\n") as f:
f.write(sln_template)
if get_bool(original_args, "vsproj_gen_only", True):
sys.exit()
def generate_copyright_header(filename: str) -> str:
MARGIN = 70
TEMPLATE = """\
/**************************************************************************/
/* %s*/
/**************************************************************************/
/* This file is part of: */
/* GODOT ENGINE */
/* https://godotengine.org */
/**************************************************************************/
/* Copyright (c) 2014-present Godot Engine contributors (see AUTHORS.md). */
/* Copyright (c) 2007-2014 Juan Linietsky, Ariel Manzur. */
/* */
/* Permission is hereby granted, free of charge, to any person obtaining */
/* a copy of this software and associated documentation files (the */
/* "Software"), to deal in the Software without restriction, including */
/* without limitation the rights to use, copy, modify, merge, publish, */
/* distribute, sublicense, and/or sell copies of the Software, and to */
/* permit persons to whom the Software is furnished to do so, subject to */
/* the following conditions: */
/* */
/* The above copyright notice and this permission notice shall be */
/* included in all copies or substantial portions of the Software. */
/* */
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. */
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
/**************************************************************************/
"""
if len(filename := os.path.basename(filename).ljust(MARGIN)) > MARGIN:
print_warning(f'Filename "{filename}" too large for copyright header.')
return TEMPLATE % filename
@contextlib.contextmanager
def generated_wrapper(
path: str,
guard: Optional[bool] = None,
) -> Generator[TextIOBase, None, None]:
"""
Wrapper class to automatically handle copyright headers and header guards
for generated scripts. Meant to be invoked via `with` statement similar to
creating a file.
- `path`: The path of the file to be created.
- `guard`: Optional bool to determine if `#pragma once` should be added. If
unassigned, the value is determined by file extension.
"""
with open(path, "wt", encoding="utf-8", newline="\n") as file:
if not path.endswith(".out"):
file.write(generate_copyright_header(path))
file.write("\n/* THIS FILE IS GENERATED. EDITS WILL BE LOST. */\n\n")
if guard is None:
guard = path.endswith((".h", ".hh", ".hpp", ".hxx", ".inc"))
if guard:
file.write("#pragma once\n\n")
with StringIO(newline="\n") as str_io:
yield str_io
file.write(str_io.getvalue().strip() or "/* NO CONTENT */")
file.write("\n")
def get_buffer(path: str) -> bytes:
with open(path, "rb") as file:
return file.read()
def compress_buffer(buffer: bytes) -> bytes:
return zlib.compress(buffer, zlib.Z_BEST_COMPRESSION)
def format_buffer(buffer: bytes, indent: int = 0, width: int = 120, initial_indent: bool = False) -> str:
return textwrap.fill(
", ".join(str(byte) for byte in buffer),
width=width,
initial_indent="\t" * indent if initial_indent else "",
subsequent_indent="\t" * indent,
tabsize=4,
)
C_ESCAPABLES = [
("\\", "\\\\"),
("\a", "\\a"),
("\b", "\\b"),
("\f", "\\f"),
("\n", "\\n"),
("\r", "\\r"),
("\t", "\\t"),
("\v", "\\v"),
('"', '\\"'),
]
C_ESCAPE_TABLE = str.maketrans(dict((x, y) for x, y in C_ESCAPABLES))
def to_escaped_cstring(value: str) -> str:
return value.translate(C_ESCAPE_TABLE)
def to_raw_cstring(value: Union[str, List[str]]) -> str:
MAX_LITERAL = 16 * 1024
if isinstance(value, list):
value = "\n".join(value) + "\n"
split: List[bytes] = []
offset = 0
encoded = value.encode()
while offset <= len(encoded):
segment = encoded[offset : offset + MAX_LITERAL]
offset += MAX_LITERAL
if len(segment) == MAX_LITERAL:
pretty_break = segment.rfind(b"\n\n")
if pretty_break != -1:
segment = segment[: pretty_break + 1]
offset -= MAX_LITERAL - pretty_break - 1
elif segment[-1] & 0b10000000:
last_11xxxxxx_index = [i for i in range(-1, -5, -1) if segment[i] & 0b11000000 == 0b11000000][0]
last_11xxxxxx = segment[last_11xxxxxx_index]
if not last_11xxxxxx & 0b00100000:
last_char_length = 2
elif not last_11xxxxxx & 0b0010000:
last_char_length = 3
elif not last_11xxxxxx & 0b0001000:
last_char_length = 4
if last_char_length > -last_11xxxxxx_index:
segment = segment[:last_11xxxxxx_index]
offset += last_11xxxxxx_index
split += [segment]
if len(split) == 1:
return f'R"<!>({split[0].decode()})<!>"'
else:
return "({})".format(" ".join(f'R"<!>({segment.decode()})<!>"' for segment in split))
def get_default_include_paths(env):
if env.msvc:
return []
compiler = env.subst("$CXX")
target = os.path.join(env.Dir("#main").abspath, "main.cpp")
args = [compiler, target, "-x", "c++", "-v"]
ret = subprocess.run(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, text=True)
output = ret.stdout
match = re.search(r"#include <\.\.\.> search starts here:([\S\s]*)End of search list.", output)
if not match:
print_warning("Failed to find the include paths in the compiler output.")
return []
return [x.strip() for x in match[1].strip().splitlines()]