Unverified Commit 85232a5b authored by AUTOMATIC1111's avatar AUTOMATIC1111 Committed by GitHub
Browse files

Merge branch 'dev' into taesd-a

parents 56a26728 4b07f2f5
Loading
Loading
Loading
Loading
+12 −0
Original line number Original line Diff line number Diff line
@@ -59,6 +59,18 @@ body:
        - iOS
        - iOS
        - Android
        - Android
        - Other/Cloud
        - Other/Cloud
  - type: dropdown
    id: device
    attributes:
        label: What device are you running WebUI on?
        multiple: true
        options:
        - Nvidia GPUs (RTX 20 above)
        - Nvidia GPUs (GTX 16 below)
        - AMD GPUs (RX 6000 above)
        - AMD GPUs (RX 5000 below)
        - CPU
        - Other GPUs
  - type: dropdown
  - type: dropdown
    id: browsers
    id: browsers
    attributes:
    attributes:
+25 −18
Original line number Original line Diff line number Diff line
@@ -18,22 +18,29 @@ jobs:
    steps:
    steps:
      - name: Checkout Code
      - name: Checkout Code
        uses: actions/checkout@v3
        uses: actions/checkout@v3
      - name: Set up Python 3.10
      - uses: actions/setup-python@v4
        uses: actions/setup-python@v4
        with:
        with:
          python-version: 3.10.6
          python-version: 3.11
          cache: pip
          # NB: there's no cache: pip here since we're not installing anything
          cache-dependency-path: |
          #     from the requirements.txt file(s) in the repository; it's faster
            **/requirements*txt
          #     not to have GHA download an (at the time of writing) 4 GB cache
      - name: Install PyLint
          #     of PyTorch and other dependencies.
        run: | 
      - name: Install Ruff
          python -m pip install --upgrade pip
        run: pip install ruff==0.0.265
          pip install pylint
      - name: Run Ruff
      # This lets PyLint check to see if it can resolve imports
        run: ruff .
      - name: Install dependencies

        run: |
# The rest are currently disabled pending fixing of e.g. installing the torch dependency.
          export COMMANDLINE_ARGS="--skip-torch-cuda-test --exit"

          python launch.py
#      - name: Install PyLint
      - name: Analysing the code with pylint
#        run: |
        run: |
#          python -m pip install --upgrade pip
          pylint $(git ls-files '*.py')
#          pip install pylint
#      # This lets PyLint check to see if it can resolve imports
#      - name: Install dependencies
#        run: |
#          export COMMANDLINE_ARGS="--skip-torch-cuda-test --exit"
#          python launch.py
#      - name: Analysing the code with pylint
#        run: |
#          pylint $(git ls-files '*.py')
+6 −0
Original line number Original line Diff line number Diff line
@@ -17,8 +17,14 @@ jobs:
          cache: pip
          cache: pip
          cache-dependency-path: |
          cache-dependency-path: |
            **/requirements*txt
            **/requirements*txt
            launch.py
      - name: Run tests
      - name: Run tests
        run: python launch.py --tests test --no-half --disable-opt-split-attention --use-cpu all --skip-torch-cuda-test
        run: python launch.py --tests test --no-half --disable-opt-split-attention --use-cpu all --skip-torch-cuda-test
        env:
          PIP_DISABLE_PIP_VERSION_CHECK: "1"
          PIP_PROGRESS_BAR: "off"
          TORCH_INDEX_URL: https://download.pytorch.org/whl/cpu
          WEBUI_LAUNCH_LIVE_OUTPUT: "1"
      - name: Upload main app stdout-stderr
      - name: Upload main app stdout-stderr
        uses: actions/upload-artifact@v3
        uses: actions/upload-artifact@v3
        if: always()
        if: always()
+13 −0
Original line number Original line Diff line number Diff line
## 1.2.1

### Features:
 * add an option to always refer to lora by filenames

### Bug Fixes:
 * never refer to lora by an alias if multiple loras have same alias or the alias is called none
 * fix upscalers disappearing after the user reloads UI
 * allow bf16 in safe unpickler (resolves problems with loading some loras)
 * allow web UI to be ran fully offline
 * fix localizations not working
 * fix error for loras: 'LatentDiffusion' object has no attribute 'lora_layer_mapping'

## 1.2.0
## 1.2.0


### Features:
### Features:
+6 −7
Original line number Original line Diff line number Diff line
@@ -88,7 +88,7 @@ class LDSR:


        x_t = None
        x_t = None
        logs = None
        logs = None
        for n in range(n_runs):
        for _ in range(n_runs):
            if custom_shape is not None:
            if custom_shape is not None:
                x_t = torch.randn(1, custom_shape[1], custom_shape[2], custom_shape[3]).to(model.device)
                x_t = torch.randn(1, custom_shape[1], custom_shape[2], custom_shape[3]).to(model.device)
                x_t = repeat(x_t, '1 c h w -> b c h w', b=custom_shape[0])
                x_t = repeat(x_t, '1 c h w -> b c h w', b=custom_shape[0])
@@ -110,7 +110,6 @@ class LDSR:
        diffusion_steps = int(steps)
        diffusion_steps = int(steps)
        eta = 1.0
        eta = 1.0


        down_sample_method = 'Lanczos'


        gc.collect()
        gc.collect()
        if torch.cuda.is_available:
        if torch.cuda.is_available:
@@ -158,7 +157,7 @@ class LDSR:




def get_cond(selected_path):
def get_cond(selected_path):
    example = dict()
    example = {}
    up_f = 4
    up_f = 4
    c = selected_path.convert('RGB')
    c = selected_path.convert('RGB')
    c = torch.unsqueeze(torchvision.transforms.ToTensor()(c), 0)
    c = torch.unsqueeze(torchvision.transforms.ToTensor()(c), 0)
@@ -196,7 +195,7 @@ def convsample_ddim(model, cond, steps, shape, eta=1.0, callback=None, normals_s
@torch.no_grad()
@torch.no_grad()
def make_convolutional_sample(batch, model, custom_steps=None, eta=1.0, quantize_x0=False, custom_shape=None, temperature=1., noise_dropout=0., corrector=None,
def make_convolutional_sample(batch, model, custom_steps=None, eta=1.0, quantize_x0=False, custom_shape=None, temperature=1., noise_dropout=0., corrector=None,
                              corrector_kwargs=None, x_T=None, ddim_use_x0_pred=False):
                              corrector_kwargs=None, x_T=None, ddim_use_x0_pred=False):
    log = dict()
    log = {}


    z, c, x, xrec, xc = model.get_input(batch, model.first_stage_key,
    z, c, x, xrec, xc = model.get_input(batch, model.first_stage_key,
                                        return_first_stage_outputs=True,
                                        return_first_stage_outputs=True,
@@ -244,7 +243,7 @@ def make_convolutional_sample(batch, model, custom_steps=None, eta=1.0, quantize
        x_sample_noquant = model.decode_first_stage(sample, force_not_quantize=True)
        x_sample_noquant = model.decode_first_stage(sample, force_not_quantize=True)
        log["sample_noquant"] = x_sample_noquant
        log["sample_noquant"] = x_sample_noquant
        log["sample_diff"] = torch.abs(x_sample_noquant - x_sample)
        log["sample_diff"] = torch.abs(x_sample_noquant - x_sample)
    except:
    except Exception:
        pass
        pass


    log["sample"] = x_sample
    log["sample"] = x_sample
Loading