Commit 51864790 authored by Aarni Koskela's avatar Aarni Koskela
Browse files

Simplify a bunch of `len(x) > 0`/`len(x) == 0` style expressions

parent 3e995778
Loading
Loading
Loading
Loading
+2 −1
Original line number Original line Diff line number Diff line
@@ -91,8 +91,9 @@ class VQModel(pl.LightningModule):
                    del sd[k]
                    del sd[k]
        missing, unexpected = self.load_state_dict(sd, strict=False)
        missing, unexpected = self.load_state_dict(sd, strict=False)
        print(f"Restored from {path} with {len(missing)} missing and {len(unexpected)} unexpected keys")
        print(f"Restored from {path} with {len(missing)} missing and {len(unexpected)} unexpected keys")
        if len(missing) > 0:
        if missing:
            print(f"Missing Keys: {missing}")
            print(f"Missing Keys: {missing}")
        if unexpected:
            print(f"Unexpected Keys: {unexpected}")
            print(f"Unexpected Keys: {unexpected}")


    def on_train_batch_end(self, *args, **kwargs):
    def on_train_batch_end(self, *args, **kwargs):
+2 −2
Original line number Original line Diff line number Diff line
@@ -195,9 +195,9 @@ class DDPMV1(pl.LightningModule):
        missing, unexpected = self.load_state_dict(sd, strict=False) if not only_model else self.model.load_state_dict(
        missing, unexpected = self.load_state_dict(sd, strict=False) if not only_model else self.model.load_state_dict(
            sd, strict=False)
            sd, strict=False)
        print(f"Restored from {path} with {len(missing)} missing and {len(unexpected)} unexpected keys")
        print(f"Restored from {path} with {len(missing)} missing and {len(unexpected)} unexpected keys")
        if len(missing) > 0:
        if missing:
            print(f"Missing Keys: {missing}")
            print(f"Missing Keys: {missing}")
        if len(unexpected) > 0:
        if unexpected:
            print(f"Unexpected Keys: {unexpected}")
            print(f"Unexpected Keys: {unexpected}")


    def q_mean_variance(self, x_start, t):
    def q_mean_variance(self, x_start, t):
+2 −2
Original line number Original line Diff line number Diff line
@@ -9,14 +9,14 @@ class ExtraNetworkLora(extra_networks.ExtraNetwork):
    def activate(self, p, params_list):
    def activate(self, p, params_list):
        additional = shared.opts.sd_lora
        additional = shared.opts.sd_lora


        if additional != "None" and additional in lora.available_loras and len([x for x in params_list if x.items[0] == additional]) == 0:
        if additional != "None" and additional in lora.available_loras and not any(x for x in params_list if x.items[0] == additional):
            p.all_prompts = [x + f"<lora:{additional}:{shared.opts.extra_networks_default_multiplier}>" for x in p.all_prompts]
            p.all_prompts = [x + f"<lora:{additional}:{shared.opts.extra_networks_default_multiplier}>" for x in p.all_prompts]
            params_list.append(extra_networks.ExtraNetworkParams(items=[additional, shared.opts.extra_networks_default_multiplier]))
            params_list.append(extra_networks.ExtraNetworkParams(items=[additional, shared.opts.extra_networks_default_multiplier]))


        names = []
        names = []
        multipliers = []
        multipliers = []
        for params in params_list:
        for params in params_list:
            assert len(params.items) > 0
            assert params.items


            names.append(params.items[0])
            names.append(params.items[0])
            multipliers.append(float(params.items[1]) if len(params.items) > 1 else 1.0)
            multipliers.append(float(params.items[1]) if len(params.items) > 1 else 1.0)
+2 −2
Original line number Original line Diff line number Diff line
@@ -219,7 +219,7 @@ def load_lora(name, lora_on_disk):
        else:
        else:
            raise AssertionError(f"Bad Lora layer name: {key_diffusers} - must end in lora_up.weight, lora_down.weight or alpha")
            raise AssertionError(f"Bad Lora layer name: {key_diffusers} - must end in lora_up.weight, lora_down.weight or alpha")


    if len(keys_failed_to_match) > 0:
    if keys_failed_to_match:
        print(f"Failed to match keys when loading Lora {lora_on_disk.filename}: {keys_failed_to_match}")
        print(f"Failed to match keys when loading Lora {lora_on_disk.filename}: {keys_failed_to_match}")


    return lora
    return lora
@@ -267,7 +267,7 @@ def load_loras(names, multipliers=None):
        lora.multiplier = multipliers[i] if multipliers else 1.0
        lora.multiplier = multipliers[i] if multipliers else 1.0
        loaded_loras.append(lora)
        loaded_loras.append(lora)


    if len(failed_to_load_loras) > 0:
    if failed_to_load_loras:
        sd_hijack.model_hijack.comments.append("Failed to find Loras: " + ", ".join(failed_to_load_loras))
        sd_hijack.model_hijack.comments.append("Failed to find Loras: " + ", ".join(failed_to_load_loras))




+1 −1
Original line number Original line Diff line number Diff line
@@ -21,7 +21,7 @@ class ExtraOptionsSection(scripts.Script):
        self.setting_names = []
        self.setting_names = []


        with gr.Blocks() as interface:
        with gr.Blocks() as interface:
            with gr.Accordion("Options", open=False) if shared.opts.extra_options_accordion and len(shared.opts.extra_options) > 0 else gr.Group(), gr.Row():
            with gr.Accordion("Options", open=False) if shared.opts.extra_options_accordion and shared.opts.extra_options else gr.Group(), gr.Row():
                for setting_name in shared.opts.extra_options:
                for setting_name in shared.opts.extra_options:
                    with FormColumn():
                    with FormColumn():
                        comp = ui_settings.create_setting_component(setting_name)
                        comp = ui_settings.create_setting_component(setting_name)
Loading