import os import sys import re import glob import subprocess from collections import OrderedDict from collections.abc import Mapping from typing import Iterator def add_source_files(self, sources, files): # Convert string to list of absolute paths (including expanding wildcard) if isinstance(files, (str, bytes)): # Keep SCons project-absolute path as they are (no wildcard support) if files.startswith("#"): if "*" in files: print("ERROR: Wildcards can't be expanded in SCons project-absolute path: '{}'".format(files)) return files = [files] else: # Exclude .gen.cpp files from globbing, to avoid including obsolete ones. # They should instead be added manually. skip_gen_cpp = "*" in files dir_path = self.Dir(".").abspath files = sorted(glob.glob(dir_path + "/" + files)) if skip_gen_cpp: files = [f for f in files if not f.endswith(".gen.cpp")] # Add each path as compiled Object following environment (self) configuration for path in files: obj = self.Object(path) if obj in sources: print('WARNING: Object "{}" already included in environment sources.'.format(obj)) continue sources.append(obj) def disable_warnings(self): # 'self' is the environment if self.msvc: # We have to remove existing warning level defines before appending /w, # otherwise we get: "warning D9025 : overriding '/W3' with '/w'" self["CCFLAGS"] = [x for x in self["CCFLAGS"] if not (x.startswith("/W") or x.startswith("/w"))] self["CFLAGS"] = [x for x in self["CFLAGS"] if not (x.startswith("/W") or x.startswith("/w"))] self["CXXFLAGS"] = [x for x in self["CXXFLAGS"] if not (x.startswith("/W") or x.startswith("/w"))] self.AppendUnique(CCFLAGS=["/w"]) else: self.AppendUnique(CCFLAGS=["-w"]) def force_optimization_on_debug(self): # 'self' is the environment if self["target"] != "template-release": return if self.msvc: # We have to remove existing optimization level defines before appending /O2, # otherwise we get: "warning D9025 : overriding '/0d' with '/02'" self["CCFLAGS"] = [x for x in self["CCFLAGS"] if not x.startswith("/O")] self["CFLAGS"] = [x for x in self["CFLAGS"] if not x.startswith("/O")] self["CXXFLAGS"] = [x for x in self["CXXFLAGS"] if not x.startswith("/O")] self.AppendUnique(CCFLAGS=["/O2"]) else: self.AppendUnique(CCFLAGS=["-O3"]) def add_module_version_string(self, s): self.module_version_string += "." + s def get_version_info(module_version_string="", silent=False): build_name = "custom_build" if os.getenv("BUILD_NAME") != None: build_name = str(os.getenv("BUILD_NAME")) if not silent: print(f"Using custom build name: '{build_name}'.") import version version_info = { "short_name": str(version.short_name), "name": str(version.name), "major": int(version.major), "minor": int(version.minor), "patch": int(version.patch), "status": str(version.status), "build": str(build_name), "module_config": str(version.module_config) + module_version_string, "year": int(version.year), "website": str(version.website), "docs_branch": str(version.docs), } # For dev snapshots (alpha, beta, RC, etc.) we do not commit status change to Git, # so this define provides a way to override it without having to modify the source. if os.getenv("GODOT_VERSION_STATUS") != None: version_info["status"] = str(os.getenv("GODOT_VERSION_STATUS")) if not silent: print(f"Using version status '{version_info['status']}', overriding the original '{version.status}'.") # Parse Git hash if we're in a Git repo. githash = "" gitfolder = ".git" if os.path.isfile(".git"): module_folder = open(".git", "r").readline().strip() if module_folder.startswith("gitdir: "): gitfolder = module_folder[8:] if os.path.isfile(os.path.join(gitfolder, "HEAD")): head = open(os.path.join(gitfolder, "HEAD"), "r", encoding="utf8").readline().strip() if head.startswith("ref: "): ref = head[5:] # If this directory is a Git worktree instead of a root clone. parts = gitfolder.split("/") if len(parts) > 2 and parts[-2] == "worktrees": gitfolder = "/".join(parts[0:-2]) head = os.path.join(gitfolder, ref) packedrefs = os.path.join(gitfolder, "packed-refs") if os.path.isfile(head): githash = open(head, "r").readline().strip() elif os.path.isfile(packedrefs): # Git may pack refs into a single file. This code searches .git/packed-refs file for the current ref's hash. # https://mirrors.edge.kernel.org/pub/software/scm/git/docs/git-pack-refs.html for line in open(packedrefs, "r").read().splitlines(): if line.startswith("#"): continue (line_hash, line_ref) = line.split(" ") if ref == line_ref: githash = line_hash break else: githash = head version_info["git_hash"] = githash return version_info def generate_version_header(module_version_string=""): version_info = get_version_info(module_version_string) # NOTE: It is safe to generate these files here, since this is still executed serially. f = open("core/version_generated.gen.h", "w") f.write( """/* THIS FILE IS GENERATED DO NOT EDIT */ #ifndef VERSION_GENERATED_GEN_H #define VERSION_GENERATED_GEN_H #define VERSION_SHORT_NAME "{short_name}" #define VERSION_NAME "{name}" #define VERSION_MAJOR {major} #define VERSION_MINOR {minor} #define VERSION_PATCH {patch} #define VERSION_STATUS "{status}" #define VERSION_BUILD "{build}" #define VERSION_MODULE_CONFIG "{module_config}" #define VERSION_YEAR {year} #define VERSION_WEBSITE "{website}" #define VERSION_DOCS_BRANCH "{docs_branch}" #define VERSION_DOCS_URL "https://docs.godotengine.org/en/" VERSION_DOCS_BRANCH #endif // VERSION_GENERATED_GEN_H """.format( **version_info ) ) f.close() fhash = open("core/version_hash.gen.cpp", "w") fhash.write( """/* THIS FILE IS GENERATED DO NOT EDIT */ #include "core/version.h" const char *const VERSION_HASH = "{git_hash}"; """.format( **version_info ) ) fhash.close() def parse_cg_file(fname, uniforms, sizes, conditionals): fs = open(fname, "r") line = fs.readline() while line: if re.match(r"^\s*uniform", line): res = re.match(r"uniform ([\d\w]*) ([\d\w]*)") type = res.groups(1) name = res.groups(2) uniforms.append(name) if type.find("texobj") != -1: sizes.append(1) else: t = re.match(r"float(\d)x(\d)", type) if t: sizes.append(int(t.groups(1)) * int(t.groups(2))) else: t = re.match(r"float(\d)", type) sizes.append(int(t.groups(1))) if line.find("[branch]") != -1: conditionals.append(name) line = fs.readline() fs.close() def get_cmdline_bool(option, default): """We use `ARGUMENTS.get()` to check if options were manually overridden on the command line, and SCons' _text2bool helper to convert them to booleans, otherwise they're handled as strings. """ from SCons.Script import ARGUMENTS from SCons.Variables.BoolVariable import _text2bool cmdline_val = ARGUMENTS.get(option) if cmdline_val is not None: return _text2bool(cmdline_val) else: return default def detect_modules(search_path, recursive=False): """Detects and collects a list of C++ modules at specified path `search_path` - a directory path containing modules. The path may point to a single module, which may have other nested modules. A module must have "register_types.h", "SCsub", "config.py" files created to be detected. `recursive` - if `True`, then all subdirectories are searched for modules as specified by the `search_path`, otherwise collects all modules under the `search_path` directory. If the `search_path` is a module, it is collected in all cases. Returns an `OrderedDict` with module names as keys, and directory paths as values. If a path is relative, then it is a built-in module. If a path is absolute, then it is a custom module collected outside of the engine source. """ modules = OrderedDict() def add_module(path): module_name = os.path.basename(path) module_path = path.replace("\\", "/") # win32 modules[module_name] = module_path def is_engine(path): # Prevent recursively detecting modules in self and other # Godot sources when using `custom_modules` build option. version_path = os.path.join(path, "version.py") if os.path.exists(version_path): with open(version_path) as f: if 'short_name = "godot"' in f.read(): return True return False def get_files(path): files = glob.glob(os.path.join(path, "*")) # Sort so that `register_module_types` does not change that often, # and plugins are registered in alphabetic order as well. files.sort() return files if not recursive: if is_module(search_path): add_module(search_path) for path in get_files(search_path): if is_engine(path): continue if is_module(path): add_module(path) else: to_search = [search_path] while to_search: path = to_search.pop() if is_module(path): add_module(path) for child in get_files(path): if not os.path.isdir(child): continue if is_engine(child): continue to_search.insert(0, child) return modules def is_module(path): if not os.path.isdir(path): return False must_exist = ["register_types.h", "SCsub", "config.py"] for f in must_exist: if not os.path.exists(os.path.join(path, f)): return False return True def write_disabled_classes(class_list): f = open("core/disabled_classes.gen.h", "w") f.write("/* THIS FILE IS GENERATED DO NOT EDIT */\n") f.write("#ifndef DISABLED_CLASSES_GEN_H\n") f.write("#define DISABLED_CLASSES_GEN_H\n\n") for c in class_list: cs = c.strip() if cs != "": f.write("#define ClassDB_Disable_" + cs + " 1\n") f.write("\n#endif\n") def write_modules(modules): includes_cpp = "" initialize_cpp = "" uninitialize_cpp = "" for name, path in modules.items(): try: with open(os.path.join(path, "register_types.h")): includes_cpp += '#include "' + path + '/register_types.h"\n' initialize_cpp += "#ifdef MODULE_" + name.upper() + "_ENABLED\n" initialize_cpp += "\tinitialize_" + name + "_module(p_level);\n" initialize_cpp += "#endif\n" uninitialize_cpp += "#ifdef MODULE_" + name.upper() + "_ENABLED\n" uninitialize_cpp += "\tuninitialize_" + name + "_module(p_level);\n" uninitialize_cpp += "#endif\n" except OSError: pass modules_cpp = """// register_module_types.gen.cpp /* THIS FILE IS GENERATED DO NOT EDIT */ #include "register_module_types.h" #include "modules/modules_enabled.gen.h" %s void initialize_modules(ModuleInitializationLevel p_level) { %s } void uninitialize_modules(ModuleInitializationLevel p_level) { %s } """ % ( includes_cpp, initialize_cpp, uninitialize_cpp, ) # NOTE: It is safe to generate this file here, since this is still executed serially with open("modules/register_module_types.gen.cpp", "w") as f: f.write(modules_cpp) def convert_custom_modules_path(path): if not path: return path path = os.path.realpath(os.path.expanduser(os.path.expandvars(path))) err_msg = "Build option 'custom_modules' must %s" if not os.path.isdir(path): raise ValueError(err_msg % "point to an existing directory.") if path == os.path.realpath("modules"): raise ValueError(err_msg % "be a directory other than built-in `modules` directory.") return path def disable_module(self): self.disabled_modules.append(self.current_module) def module_add_dependencies(self, module, dependencies, optional=False): """ Adds dependencies for a given module. Meant to be used in module `can_build` methods. """ if module not in self.module_dependencies: self.module_dependencies[module] = [[], []] if optional: self.module_dependencies[module][1].extend(dependencies) else: self.module_dependencies[module][0].extend(dependencies) def module_check_dependencies(self, module): """ Checks if module dependencies are enabled for a given module, and prints a warning if they aren't. Meant to be used in module `can_build` methods. Returns a boolean (True if dependencies are satisfied). """ missing_deps = [] required_deps = self.module_dependencies[module][0] if module in self.module_dependencies else [] for dep in required_deps: opt = "module_{}_enabled".format(dep) if not opt in self or not self[opt]: missing_deps.append(dep) if missing_deps != []: print( "Disabling '{}' module as the following dependencies are not satisfied: {}".format( module, ", ".join(missing_deps) ) ) return False else: return True def sort_module_list(env): out = OrderedDict() deps = {k: v[0] + list(filter(lambda x: x in env.module_list, v[1])) for k, v in env.module_dependencies.items()} frontier = list(env.module_list.keys()) explored = [] while len(frontier): cur = frontier.pop() deps_list = deps[cur] if cur in deps else [] if len(deps_list) and any([d not in explored for d in deps_list]): # Will explore later, after its dependencies frontier.insert(0, cur) continue explored.append(cur) for k in explored: env.module_list.move_to_end(k) def use_windows_spawn_fix(self, platform=None): if os.name != "nt": return # not needed, only for windows # On Windows, due to the limited command line length, when creating a static library # from a very high number of objects SCons will invoke "ar" once per object file; # that makes object files with same names to be overwritten so the last wins and # the library loses symbols defined by overwritten objects. # By enabling quick append instead of the default mode (replacing), libraries will # got built correctly regardless the invocation strategy. # Furthermore, since SCons will rebuild the library from scratch when an object file # changes, no multiple versions of the same object file will be present. self.Replace(ARFLAGS="q") def mySubProcess(cmdline, env): startupinfo = subprocess.STARTUPINFO() startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW popen_args = { "stdin": subprocess.PIPE, "stdout": subprocess.PIPE, "stderr": subprocess.PIPE, "startupinfo": startupinfo, "shell": False, "env": env, } if sys.version_info >= (3, 7, 0): popen_args["text"] = True proc = subprocess.Popen(cmdline, **popen_args) _, err = proc.communicate() rv = proc.wait() if rv: print("=====") print(err) print("=====") return rv def mySpawn(sh, escape, cmd, args, env): newargs = " ".join(args[1:]) cmdline = cmd + " " + newargs rv = 0 env = {str(key): str(value) for key, value in iter(env.items())} if len(cmdline) > 32000 and cmd.endswith("ar"): cmdline = cmd + " " + args[1] + " " + args[2] + " " for i in range(3, len(args)): rv = mySubProcess(cmdline + args[i], env) if rv: break else: rv = mySubProcess(cmdline, env) return rv self["SPAWN"] = mySpawn def save_active_platforms(apnames, ap): for x in ap: svg_names = [] if os.path.isfile(x + "/logo.svg"): svg_names.append("logo") if os.path.isfile(x + "/run_icon.svg"): svg_names.append("run_icon") for name in svg_names: svgf = open(x + "/" + name + ".svg", "rb") b = svgf.read(1) svg_str = " /* AUTOGENERATED FILE, DO NOT EDIT */ \n" svg_str += " static const char *_" + x[9:] + "_" + name + '_svg = "' while len(b) == 1: svg_str += "\\" + hex(ord(b))[1:] b = svgf.read(1) svg_str += '";\n' svgf.close() # NOTE: It is safe to generate this file here, since this is still executed serially wf = x + "/" + name + "_svg.gen.h" with open(wf, "w") as svgw: svgw.write(svg_str) def no_verbose(sys, env): colors = {} # Colors are disabled in non-TTY environments such as pipes. This means # that if output is redirected to a file, it will not contain color codes if sys.stdout.isatty(): colors["blue"] = "\033[0;94m" colors["bold_blue"] = "\033[1;94m" colors["reset"] = "\033[0m" else: colors["blue"] = "" colors["bold_blue"] = "" colors["reset"] = "" # There is a space before "..." to ensure that source file names can be # Ctrl + clicked in the VS Code terminal. compile_source_message = "{}Compiling {}$SOURCE{} ...{}".format( colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"] ) java_compile_source_message = "{}Compiling {}$SOURCE{} ...{}".format( colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"] ) compile_shared_source_message = "{}Compiling shared {}$SOURCE{} ...{}".format( colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"] ) link_program_message = "{}Linking Program {}$TARGET{} ...{}".format( colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"] ) link_library_message = "{}Linking Static Library {}$TARGET{} ...{}".format( colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"] ) ranlib_library_message = "{}Ranlib Library {}$TARGET{} ...{}".format( colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"] ) link_shared_library_message = "{}Linking Shared Library {}$TARGET{} ...{}".format( colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"] ) java_library_message = "{}Creating Java Archive {}$TARGET{} ...{}".format( colors["blue"], colors["bold_blue"], colors["blue"], colors["reset"] ) env.Append(CXXCOMSTR=[compile_source_message]) env.Append(CCCOMSTR=[compile_source_message]) env.Append(SHCCCOMSTR=[compile_shared_source_message]) env.Append(SHCXXCOMSTR=[compile_shared_source_message]) env.Append(ARCOMSTR=[link_library_message]) env.Append(RANLIBCOMSTR=[ranlib_library_message]) env.Append(SHLINKCOMSTR=[link_shared_library_message]) env.Append(LINKCOMSTR=[link_program_message]) env.Append(JARCOMSTR=[java_library_message]) env.Append(JAVACCOMSTR=[java_compile_source_message]) def detect_visual_c_compiler_version(tools_env): # tools_env is the variable scons uses to call tools that execute tasks, SCons's env['ENV'] that executes tasks... # (see the SCons documentation for more information on what it does)... # in order for this function to be well encapsulated i choose to force it to receive SCons's TOOLS env (env['ENV'] # and not scons setup environment (env)... so make sure you call the right environment on it or it will fail to detect # the proper vc version that will be called # There is no flag to give to visual c compilers to set the architecture, i.e. scons arch argument (x86_32, x86_64, arm64, etc.). # There are many different cl.exe files that are run, and each one compiles & links to a different architecture # As far as I know, the only way to figure out what compiler will be run when Scons calls cl.exe via Program() # is to check the PATH variable and figure out which one will be called first. Code below does that and returns: # the following string values: # "" Compiler not detected # "amd64" Native 64 bit compiler # "amd64_x86" 64 bit Cross Compiler for 32 bit # "x86" Native 32 bit compiler # "x86_amd64" 32 bit Cross Compiler for 64 bit # There are other architectures, but Godot does not support them currently, so this function does not detect arm/amd64_arm # and similar architectures/compilers # Set chosen compiler to "not detected" vc_chosen_compiler_index = -1 vc_chosen_compiler_str = "" # Start with Pre VS 2017 checks which uses VCINSTALLDIR: if "VCINSTALLDIR" in tools_env: # print("Checking VCINSTALLDIR") # find() works with -1 so big ifs below are needed... the simplest solution, in fact # First test if amd64 and amd64_x86 compilers are present in the path vc_amd64_compiler_detection_index = tools_env["PATH"].find(tools_env["VCINSTALLDIR"] + "BIN\\amd64;") if vc_amd64_compiler_detection_index > -1: vc_chosen_compiler_index = vc_amd64_compiler_detection_index vc_chosen_compiler_str = "amd64" vc_amd64_x86_compiler_detection_index = tools_env["PATH"].find(tools_env["VCINSTALLDIR"] + "BIN\\amd64_x86;") if vc_amd64_x86_compiler_detection_index > -1 and ( vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_amd64_x86_compiler_detection_index ): vc_chosen_compiler_index = vc_amd64_x86_compiler_detection_index vc_chosen_compiler_str = "amd64_x86" # Now check the 32 bit compilers vc_x86_compiler_detection_index = tools_env["PATH"].find(tools_env["VCINSTALLDIR"] + "BIN;") if vc_x86_compiler_detection_index > -1 and ( vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_compiler_detection_index ): vc_chosen_compiler_index = vc_x86_compiler_detection_index vc_chosen_compiler_str = "x86" vc_x86_amd64_compiler_detection_index = tools_env["PATH"].find(tools_env["VCINSTALLDIR"] + "BIN\\x86_amd64;") if vc_x86_amd64_compiler_detection_index > -1 and ( vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_amd64_compiler_detection_index ): vc_chosen_compiler_index = vc_x86_amd64_compiler_detection_index vc_chosen_compiler_str = "x86_amd64" # and for VS 2017 and newer we check VCTOOLSINSTALLDIR: if "VCTOOLSINSTALLDIR" in tools_env: # Newer versions have a different path available vc_amd64_compiler_detection_index = ( tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX64\\X64;") ) if vc_amd64_compiler_detection_index > -1: vc_chosen_compiler_index = vc_amd64_compiler_detection_index vc_chosen_compiler_str = "amd64" vc_amd64_x86_compiler_detection_index = ( tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX64\\X86;") ) if vc_amd64_x86_compiler_detection_index > -1 and ( vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_amd64_x86_compiler_detection_index ): vc_chosen_compiler_index = vc_amd64_x86_compiler_detection_index vc_chosen_compiler_str = "amd64_x86" vc_x86_compiler_detection_index = ( tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX86\\X86;") ) if vc_x86_compiler_detection_index > -1 and ( vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_compiler_detection_index ): vc_chosen_compiler_index = vc_x86_compiler_detection_index vc_chosen_compiler_str = "x86" vc_x86_amd64_compiler_detection_index = ( tools_env["PATH"].upper().find(tools_env["VCTOOLSINSTALLDIR"].upper() + "BIN\\HOSTX86\\X64;") ) if vc_x86_amd64_compiler_detection_index > -1 and ( vc_chosen_compiler_index == -1 or vc_chosen_compiler_index > vc_x86_amd64_compiler_detection_index ): vc_chosen_compiler_str = "x86_amd64" return vc_chosen_compiler_str def find_visual_c_batch_file(env): from SCons.Tool.MSCommon.vc import ( get_default_version, get_host_target, find_batch_file, ) # Syntax changed in SCons 4.4.0. from SCons import __version__ as scons_raw_version scons_ver = env._get_major_minor_revision(scons_raw_version) version = get_default_version(env) if scons_ver >= (4, 4, 0): (host_platform, target_platform, _) = get_host_target(env, version) else: (host_platform, target_platform, _) = get_host_target(env) return find_batch_file(env, version, host_platform, target_platform)[0] def generate_cpp_hint_file(filename): if os.path.isfile(filename): # Don't overwrite an existing hint file since the user may have customized it. pass else: try: with open(filename, "w") as fd: fd.write("#define GDCLASS(m_class, m_inherits)\n") except OSError: print("Could not write cpp.hint file.") def glob_recursive(pattern, node="."): from SCons import Node from SCons.Script import Glob results = [] for f in Glob(str(node) + "/*", source=True): if type(f) is Node.FS.Dir: results += glob_recursive(pattern, f) results += Glob(str(node) + "/" + pattern, source=True) return results def add_to_vs_project(env, sources): for x in sources: if type(x) == type(""): fname = env.File(x).path else: fname = env.File(x)[0].path pieces = fname.split(".") if len(pieces) > 0: basename = pieces[0] basename = basename.replace("\\\\", "/") if os.path.isfile(basename + ".h"): env.vs_incs += [basename + ".h"] elif os.path.isfile(basename + ".hpp"): env.vs_incs += [basename + ".hpp"] if os.path.isfile(basename + ".c"): env.vs_srcs += [basename + ".c"] elif os.path.isfile(basename + ".cpp"): env.vs_srcs += [basename + ".cpp"] def generate_vs_project(env, num_jobs): batch_file = find_visual_c_batch_file(env) if batch_file: class ModuleConfigs(Mapping): # This version information (Win32, x64, Debug, Release) seems to be # required for Visual Studio to understand that it needs to generate an NMAKE # project. Do not modify without knowing what you are doing. PLATFORMS = ["Win32", "x64"] PLATFORM_IDS = ["x86_32", "x86_64"] CONFIGURATIONS = ["editor", "template_release", "template_debug"] DEV_SUFFIX = ".dev" if env["dev_build"] else "" @staticmethod def for_every_variant(value): return [value for _ in range(len(ModuleConfigs.CONFIGURATIONS) * len(ModuleConfigs.PLATFORMS))] def __init__(self): shared_targets_array = [] self.names = [] self.arg_dict = { "variant": [], "runfile": shared_targets_array, "buildtarget": shared_targets_array, "cpppaths": [], "cppdefines": [], "cmdargs": [], } self.add_mode() # default def add_mode( self, name: str = "", includes: str = "", cli_args: str = "", defines=None, ): if defines is None: defines = [] self.names.append(name) self.arg_dict["variant"] += [ f'{config}{f"_[{name}]" if name else ""}|{platform}' for config in ModuleConfigs.CONFIGURATIONS for platform in ModuleConfigs.PLATFORMS ] self.arg_dict["runfile"] += [ f'bin\\godot.windows.{config}{ModuleConfigs.DEV_SUFFIX}{".double" if env["precision"] == "double" else ""}.{plat_id}{f".{name}" if name else ""}.exe' for config in ModuleConfigs.CONFIGURATIONS for plat_id in ModuleConfigs.PLATFORM_IDS ] self.arg_dict["cpppaths"] += ModuleConfigs.for_every_variant(env["CPPPATH"] + [includes]) self.arg_dict["cppdefines"] += ModuleConfigs.for_every_variant(env["CPPDEFINES"] + defines) self.arg_dict["cmdargs"] += ModuleConfigs.for_every_variant(cli_args) def build_commandline(self, commands): configuration_getter = ( "$(Configuration" + "".join([f'.Replace("{name}", "")' for name in self.names[1:]]) + '.Replace("_[]", "")' + ")" ) common_build_prefix = [ 'cmd /V /C set "plat=$(PlatformTarget)"', '(if "$(PlatformTarget)"=="x64" (set "plat=x86_amd64"))', 'call "' + batch_file + '" !plat!', ] # Windows allows us to have spaces in paths, so we need # to double quote off the directory. However, the path ends # in a backslash, so we need to remove this, lest it escape the # last double quote off, confusing MSBuild common_build_postfix = [ "--directory=\"$(ProjectDir.TrimEnd('\\'))\"", "platform=windows", f"target={configuration_getter}", "progress=no", "-j%s" % num_jobs, ] if env["dev_build"]: common_build_postfix.append("dev_build=yes") if env["dev_mode"]: common_build_postfix.append("dev_mode=yes") elif env["tests"]: common_build_postfix.append("tests=yes") if env["custom_modules"]: common_build_postfix.append("custom_modules=%s" % env["custom_modules"]) if env["precision"] == "double": common_build_postfix.append("precision=double") result = " ^& ".join(common_build_prefix + [" ".join([commands] + common_build_postfix)]) return result # Mappings interface definitions def __iter__(self) -> Iterator[str]: for x in self.arg_dict: yield x def __len__(self) -> int: return len(self.names) def __getitem__(self, k: str): return self.arg_dict[k] add_to_vs_project(env, env.core_sources) add_to_vs_project(env, env.drivers_sources) add_to_vs_project(env, env.main_sources) add_to_vs_project(env, env.modules_sources) add_to_vs_project(env, env.scene_sources) add_to_vs_project(env, env.servers_sources) if env["tests"]: add_to_vs_project(env, env.tests_sources) if env.editor_build: add_to_vs_project(env, env.editor_sources) for header in glob_recursive("**/*.h"): env.vs_incs.append(str(header)) module_configs = ModuleConfigs() if env.get("module_mono_enabled"): mono_defines = [("GD_MONO_HOT_RELOAD",)] if env.editor_build else [] module_configs.add_mode( "mono", cli_args="module_mono_enabled=yes", defines=mono_defines, ) env["MSVSBUILDCOM"] = module_configs.build_commandline("scons") env["MSVSREBUILDCOM"] = module_configs.build_commandline("scons vsproj=yes") env["MSVSCLEANCOM"] = module_configs.build_commandline("scons --clean") if not env.get("MSVS"): env["MSVS"]["PROJECTSUFFIX"] = ".vcxproj" env["MSVS"]["SOLUTIONSUFFIX"] = ".sln" env.MSVSProject( target=["#godot" + env["MSVSPROJECTSUFFIX"]], incs=env.vs_incs, srcs=env.vs_srcs, auto_build_solution=1, **module_configs, ) else: print("Could not locate Visual Studio batch file to set up the build environment. Not generating VS project.") def precious_program(env, program, sources, **args): program = env.ProgramOriginal(program, sources, **args) env.Precious(program) return program def add_shared_library(env, name, sources, **args): library = env.SharedLibrary(name, sources, **args) env.NoCache(library) return library def add_library(env, name, sources, **args): library = env.Library(name, sources, **args) env.NoCache(library) return library def add_program(env, name, sources, **args): program = env.Program(name, sources, **args) env.NoCache(program) return program def CommandNoCache(env, target, sources, command, **args): result = env.Command(target, sources, command, **args) env.NoCache(result) return result def Run(env, function, short_message, subprocess=True): from SCons.Script import Action from platform_methods import run_in_subprocess output_print = short_message if not env["verbose"] else "" if not subprocess: return Action(function, output_print) else: return Action(run_in_subprocess(function), output_print) def detect_darwin_sdk_path(platform, env): sdk_name = "" if platform == "macos": sdk_name = "macosx" var_name = "MACOS_SDK_PATH" elif platform == "ios": sdk_name = "iphoneos" var_name = "IOS_SDK_PATH" elif platform == "iossimulator": sdk_name = "iphonesimulator" var_name = "IOS_SDK_PATH" else: raise Exception("Invalid platform argument passed to detect_darwin_sdk_path") if not env[var_name]: try: sdk_path = subprocess.check_output(["xcrun", "--sdk", sdk_name, "--show-sdk-path"]).strip().decode("utf-8") if sdk_path: env[var_name] = sdk_path except (subprocess.CalledProcessError, OSError): print("Failed to find SDK path while running xcrun --sdk {} --show-sdk-path.".format(sdk_name)) raise def is_vanilla_clang(env): if not using_clang(env): return False try: version = subprocess.check_output([env.subst(env["CXX"]), "--version"]).strip().decode("utf-8") except (subprocess.CalledProcessError, OSError): print("Couldn't parse CXX environment variable to infer compiler version.") return False return not version.startswith("Apple") def get_compiler_version(env): """ Returns an array of version numbers as ints: [major, minor, patch]. The return array should have at least two values (major, minor). """ if not env.msvc: # Not using -dumpversion as some GCC distros only return major, and # Clang used to return hardcoded 4.2.1: # https://reviews.llvm.org/D56803 try: version = subprocess.check_output([env.subst(env["CXX"]), "--version"]).strip().decode("utf-8") except (subprocess.CalledProcessError, OSError): print("Couldn't parse CXX environment variable to infer compiler version.") return None else: # TODO: Implement for MSVC return None match = re.search( "(?:(?<=version )|(?<=\) )|(?<=^))" "(?P<major>\d+)" "(?:\.(?P<minor>\d*))?" "(?:\.(?P<patch>\d*))?" "(?:-(?P<metadata1>[0-9a-zA-Z-]*))?" "(?:\+(?P<metadata2>[0-9a-zA-Z-]*))?" "(?: (?P<date>[0-9]{8}|[0-9]{6})(?![0-9a-zA-Z]))?", version, ) if match is not None: return match.groupdict() else: return None def using_gcc(env): return "gcc" in os.path.basename(env["CC"]) def using_clang(env): return "clang" in os.path.basename(env["CC"]) def using_emcc(env): return "emcc" in os.path.basename(env["CC"]) def show_progress(env): import sys from SCons.Script import Progress, Command, AlwaysBuild screen = sys.stdout # Progress reporting is not available in non-TTY environments since it # messes with the output (for example, when writing to a file) show_progress = env["progress"] and sys.stdout.isatty() node_count = 0 node_count_max = 0 node_count_interval = 1 node_count_fname = str(env.Dir("#")) + "/.scons_node_count" import time, math class cache_progress: # The default is 1 GB cache and 12 hours half life def __init__(self, path=None, limit=1073741824, half_life=43200): self.path = path self.limit = limit self.exponent_scale = math.log(2) / half_life if env["verbose"] and path != None: screen.write( "Current cache limit is {} (used: {})\n".format( self.convert_size(limit), self.convert_size(self.get_size(path)) ) ) self.delete(self.file_list()) def __call__(self, node, *args, **kw): nonlocal node_count, node_count_max, node_count_interval, node_count_fname, show_progress if show_progress: # Print the progress percentage node_count += node_count_interval if node_count_max > 0 and node_count <= node_count_max: screen.write("\r[%3d%%] " % (node_count * 100 / node_count_max)) screen.flush() elif node_count_max > 0 and node_count > node_count_max: screen.write("\r[100%] ") screen.flush() else: screen.write("\r[Initial build] ") screen.flush() def delete(self, files): if len(files) == 0: return if env["verbose"]: # Utter something screen.write("\rPurging %d %s from cache...\n" % (len(files), len(files) > 1 and "files" or "file")) [os.remove(f) for f in files] def file_list(self): if self.path is None: # Nothing to do return [] # Gather a list of (filename, (size, atime)) within the # cache directory file_stat = [(x, os.stat(x)[6:8]) for x in glob.glob(os.path.join(self.path, "*", "*"))] if file_stat == []: # Nothing to do return [] # Weight the cache files by size (assumed to be roughly # proportional to the recompilation time) times an exponential # decay since the ctime, and return a list with the entries # (filename, size, weight). current_time = time.time() file_stat = [(x[0], x[1][0], (current_time - x[1][1])) for x in file_stat] # Sort by the most recently accessed files (most sensible to keep) first file_stat.sort(key=lambda x: x[2]) # Search for the first entry where the storage limit is # reached sum, mark = 0, None for i, x in enumerate(file_stat): sum += x[1] if sum > self.limit: mark = i break if mark is None: return [] else: return [x[0] for x in file_stat[mark:]] def convert_size(self, size_bytes): if size_bytes == 0: return "0 bytes" size_name = ("bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB") i = int(math.floor(math.log(size_bytes, 1024))) p = math.pow(1024, i) s = round(size_bytes / p, 2) return "%s %s" % (int(s) if i == 0 else s, size_name[i]) def get_size(self, start_path="."): total_size = 0 for dirpath, dirnames, filenames in os.walk(start_path): for f in filenames: fp = os.path.join(dirpath, f) total_size += os.path.getsize(fp) return total_size def progress_finish(target, source, env): nonlocal node_count, progressor try: with open(node_count_fname, "w") as f: f.write("%d\n" % node_count) progressor.delete(progressor.file_list()) except Exception: pass try: with open(node_count_fname) as f: node_count_max = int(f.readline()) except Exception: pass cache_directory = os.environ.get("SCONS_CACHE") # Simple cache pruning, attached to SCons' progress callback. Trim the # cache directory to a size not larger than cache_limit. cache_limit = float(os.getenv("SCONS_CACHE_LIMIT", 1024)) * 1024 * 1024 progressor = cache_progress(cache_directory, cache_limit) Progress(progressor, interval=node_count_interval) progress_finish_command = Command("progress_finish", [], progress_finish) AlwaysBuild(progress_finish_command) def dump(env): # Dumps latest build information for debugging purposes and external tools. from json import dump def non_serializable(obj): return "<<non-serializable: %s>>" % (type(obj).__qualname__) with open(".scons_env.json", "w") as f: dump(env.Dictionary(), f, indent=4, default=non_serializable)