Commit 876a96f0 authored by AUTOMATIC's avatar AUTOMATIC
Browse files

remove erroneous dir in the extension directory

remove loading .js files from scripts dir (they go into javascript)
load scripts after models, for scripts that depend on loaded models
parent 999929be
Loading
Loading
Loading
Loading

stable-diffusion-webui-inspiration @ a0b96664

Original line number Diff line number Diff line
Subproject commit a0b96664d2524b87916ae463fbb65411b13a569b
+1 −1
Original line number Diff line number Diff line
@@ -1885,7 +1885,7 @@ def load_javascript(raw_response):
        javascript = f'<script>{jsfile.read()}</script>'

    scripts_list = modules.scripts.list_scripts("javascript", ".js")
    scripts_list += modules.scripts.list_scripts("scripts", ".js")
    
    for basedir, filename, path in scripts_list:
        with open(path, "r", encoding="utf8") as jsfile:
            javascript += f"\n<!-- {filename} --><script>{jsfile.read()}</script>"
+6 −5
Original line number Diff line number Diff line
@@ -9,7 +9,7 @@ from fastapi.middleware.gzip import GZipMiddleware

from modules.paths import script_path

from modules import devices, sd_samplers
from modules import devices, sd_samplers, upscaler
import modules.codeformer_model as codeformer
import modules.extras
import modules.face_restoration
@@ -73,12 +73,11 @@ def wrap_gradio_gpu_call(func, extra_outputs=None):


def initialize():
    modules.scripts.load_scripts()
    if cmd_opts.ui_debug_mode:
        class enmpty():
            name = None
        shared.sd_upscalers = [enmpty()]
        shared.sd_upscalers = upscaler.UpscalerLanczos().scalers
        modules.scripts.load_scripts()
        return

    modelloader.cleanup_models()
    modules.sd_models.setup_model()
    codeformer.setup_model(cmd_opts.codeformer_models_path)
@@ -86,6 +85,8 @@ def initialize():
    shared.face_restorers.append(modules.face_restoration.FaceRestoration())
    modelloader.load_upscalers()

    modules.scripts.load_scripts()

    modules.sd_models.load_model()
    shared.opts.onchange("sd_model_checkpoint", wrap_queued_call(lambda: modules.sd_models.reload_model_weights(shared.sd_model)))
    shared.opts.onchange("sd_hypernetwork", wrap_queued_call(lambda: modules.hypernetworks.hypernetwork.load_hypernetwork(shared.opts.sd_hypernetwork)))