diff --git a/.github/workflows/javascript_builds.yml b/.github/workflows/javascript_builds.yml index 8739a65307ee..895869063b67 100644 --- a/.github/workflows/javascript_builds.yml +++ b/.github/workflows/javascript_builds.yml @@ -8,7 +8,6 @@ env: GODOT_BASE_BRANCH: 3.x SCONSFLAGS: verbose=yes warnings=all werror=yes debug_symbols=no EM_VERSION: 3.1.39 - EM_CACHE_FOLDER: "emsdk-cache" concurrency: group: ci-${{github.actor}}-${{github.head_ref || github.run_number}}-${{github.ref}}-javascript @@ -26,8 +25,7 @@ jobs: uses: mymindstorm/setup-emsdk@v14 with: version: ${{env.EM_VERSION}} - actions-cache-folder: ${{env.EM_CACHE_FOLDER}} - cache-key: emsdk-${{ matrix.cache-name }}-${{env.GODOT_BASE_BRANCH}}-${{github.ref}}-${{github.sha}} + no-cache: true - name: Verify Emscripten setup run: | diff --git a/SConstruct b/SConstruct index d0ffcde3e35b..48632cbaa8e3 100644 --- a/SConstruct +++ b/SConstruct @@ -796,12 +796,13 @@ elif selected_platform != "": # The following only makes sense when the 'env' is defined, and assumes it is. if "env" in locals(): - # FIXME: This method mixes both cosmetic progress stuff and cache handling... methods.show_progress(env) # TODO: replace this with `env.Dump(format="json")` # once we start requiring SCons 4.0 as min version. methods.dump(env) + methods.clean_cache(env) + def print_elapsed_time(): elapsed_time_sec = round(time.time() - time_at_start, 3) diff --git a/doc/classes/Timer.xml b/doc/classes/Timer.xml index 10b6e2e22ff4..f4f9e3c0d332 100644 --- a/doc/classes/Timer.xml +++ b/doc/classes/Timer.xml @@ -37,6 +37,7 @@ If [code]true[/code], the timer will automatically start when entering the scene tree. [b]Note:[/b] This property is automatically set to [code]false[/code] after the timer enters the scene tree and starts. + [b]Note:[/b] This property does nothing when the timer is running in the editor. If [code]true[/code], the timer will stop when reaching 0. If [code]false[/code], it will restart. diff --git a/methods.py b/methods.py index 07469f60be1d..d125c119e545 100644 --- a/methods.py +++ b/methods.py @@ -1126,21 +1126,19 @@ def show_progress(env): "fname": str(env.Dir("#")) + "/.scons_node_count", } - import time, math + import math class cache_progress: - # The default is 1 GB cache and 12 hours half life - def __init__(self, path=None, limit=1073741824, half_life=43200): + # The default is 1 GB cache + def __init__(self, path=None, limit=pow(1024, 3)): self.path = path self.limit = limit - self.exponent_scale = math.log(2) / half_life if env["verbose"] and path != None: screen.write( "Current cache limit is {} (used: {})\n".format( self.convert_size(limit), self.convert_size(self.get_size(path)) ) ) - self.delete(self.file_list()) def __call__(self, node, *args, **kw): if show_progress: @@ -1158,12 +1156,65 @@ def __call__(self, node, *args, **kw): screen.write("\r[Initial build] ") screen.flush() + def convert_size(self, size_bytes): + if size_bytes == 0: + return "0 bytes" + size_name = ("bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB") + i = int(math.floor(math.log(size_bytes, 1024))) + p = math.pow(1024, i) + s = round(size_bytes / p, 2) + return "%s %s" % (int(s) if i == 0 else s, size_name[i]) + + def get_size(self, start_path="."): + total_size = 0 + for dirpath, dirnames, filenames in os.walk(start_path): + for f in filenames: + fp = os.path.join(dirpath, f) + total_size += os.path.getsize(fp) + return total_size + + def progress_finish(target, source, env): + try: + with open(node_count_data["fname"], "w") as f: + f.write("%d\n" % node_count_data["count"]) + except Exception: + pass + + try: + with open(node_count_data["fname"]) as f: + node_count_data["max"] = int(f.readline()) + except Exception: + pass + + cache_directory = os.environ.get("SCONS_CACHE") + # Simple cache pruning, attached to SCons' progress callback. Trim the + # cache directory to a size not larger than cache_limit. + cache_limit = float(os.getenv("SCONS_CACHE_LIMIT", 1024)) * 1024 * 1024 + progressor = cache_progress(cache_directory, cache_limit) + Progress(progressor, interval=node_count_data["interval"]) + + progress_finish_command = Command("progress_finish", [], progress_finish) + AlwaysBuild(progress_finish_command) + + +def clean_cache(env): + import atexit + import time + + class cache_clean: + def __init__(self, path=None, limit=pow(1024, 3)): + self.path = path + self.limit = limit + + def clean(self): + self.delete(self.file_list()) + def delete(self, files): if len(files) == 0: return if env["verbose"]: # Utter something - screen.write("\rPurging %d %s from cache...\n" % (len(files), len(files) > 1 and "files" or "file")) + print("Purging %d %s from cache..." % (len(files), "files" if len(files) > 1 else "file")) [os.remove(f) for f in files] def file_list(self): @@ -1197,46 +1248,20 @@ def file_list(self): else: return [x[0] for x in file_stat[mark:]] - def convert_size(self, size_bytes): - if size_bytes == 0: - return "0 bytes" - size_name = ("bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB") - i = int(math.floor(math.log(size_bytes, 1024))) - p = math.pow(1024, i) - s = round(size_bytes / p, 2) - return "%s %s" % (int(s) if i == 0 else s, size_name[i]) - - def get_size(self, start_path="."): - total_size = 0 - for dirpath, dirnames, filenames in os.walk(start_path): - for f in filenames: - fp = os.path.join(dirpath, f) - total_size += os.path.getsize(fp) - return total_size - - def progress_finish(target, source, env): + def cache_finally(): + nonlocal cleaner try: - with open(node_count_data["fname"], "w") as f: - f.write("%d\n" % node_count_data["count"]) - progressor.delete(progressor.file_list()) + cleaner.clean() except Exception: pass - try: - with open(node_count_data["fname"]) as f: - node_count_data["max"] = int(f.readline()) - except Exception: - pass - cache_directory = os.environ.get("SCONS_CACHE") # Simple cache pruning, attached to SCons' progress callback. Trim the # cache directory to a size not larger than cache_limit. cache_limit = float(os.getenv("SCONS_CACHE_LIMIT", 1024)) * 1024 * 1024 - progressor = cache_progress(cache_directory, cache_limit) - Progress(progressor, interval=node_count_data["interval"]) + cleaner = cache_clean(cache_directory, cache_limit) - progress_finish_command = Command("progress_finish", [], progress_finish) - AlwaysBuild(progress_finish_command) + atexit.register(cache_finally) def dump(env):