Commit 505ec7e4 authored by AUTOMATIC's avatar AUTOMATIC
Browse files

cleanup some unneeded imports for hijack files

parent 7dbfd8a7
Loading
Loading
Loading
Loading
+2 −8
Original line number Original line Diff line number Diff line
import math
import os
import sys
import traceback
import torch
import torch
import numpy as np
from torch import einsum
from torch.nn.functional import silu
from torch.nn.functional import silu


import modules.textual_inversion.textual_inversion
import modules.textual_inversion.textual_inversion
from modules import prompt_parser, devices, sd_hijack_optimizations, shared, sd_hijack_checkpoint
from modules import devices, sd_hijack_optimizations, shared, sd_hijack_checkpoint
from modules.hypernetworks import hypernetwork
from modules.hypernetworks import hypernetwork
from modules.shared import opts, device, cmd_opts
from modules.shared import cmd_opts
from modules import sd_hijack_clip, sd_hijack_open_clip, sd_hijack_unet
from modules import sd_hijack_clip, sd_hijack_open_clip, sd_hijack_unet


from modules.sd_hijack_optimizations import invokeAI_mps_available
from modules.sd_hijack_optimizations import invokeAI_mps_available
+0 −3
Original line number Original line Diff line number Diff line
@@ -5,7 +5,6 @@ import importlib


import torch
import torch
from torch import einsum
from torch import einsum
import torch.nn.functional as F


from ldm.util import default
from ldm.util import default
from einops import rearrange
from einops import rearrange
@@ -13,8 +12,6 @@ from einops import rearrange
from modules import shared
from modules import shared
from modules.hypernetworks import hypernetwork
from modules.hypernetworks import hypernetwork


from ldm.modules.diffusionmodules.util import timestep_embedding



if shared.cmd_opts.xformers or shared.cmd_opts.force_enable_xformers:
if shared.cmd_opts.xformers or shared.cmd_opts.force_enable_xformers:
    try:
    try: