Comments (31)
line 796, in main
I see the issue. Should be fixed with 1ddfcfc
So far its fixed, yhea, it generated till 100 steps as a test and it successfully made an image and saved a ckpt file. However....it got stuck on a loop saying training complete? and it kept saving a ckpt file every 1 step.
Wheeee!
Allright, I'll se what I did wrong now. :P
from sd_dreambooth_extension.
just got the same error.
from sd_dreambooth_extension.
supposedly Nerdy Rodent said it had to do with export LD_LIBRARY_PATH=/usr/lib/wsl/lib:$LD_LIBRARY_PATH
.
Something that it had to be in the pastebin for it to work or something. But i haven't figured it out.
bitsandbytes-foundation/bitsandbytes#52
from sd_dreambooth_extension.
supposedly Nerdy Rodent said it had to do with
export LD_LIBRARY_PATH=/usr/lib/wsl/lib:$LD_LIBRARY_PATH
. Something that it had to be in the pastebin for it to work or something. But i haven't figured it out.
I don't think that's the issue here. That's for WSL, we're trying to do this thing without WSL.
I think the issue is just that we need to do 'conda install cudatoolkit=11.3', but getting an issue running from my install script.
from sd_dreambooth_extension.
supposedly Nerdy Rodent said it had to do with
export LD_LIBRARY_PATH=/usr/lib/wsl/lib:$LD_LIBRARY_PATH
. Something that it had to be in the pastebin for it to work or something. But i haven't figured it out.
TimDettmers/bitsandbytes#52I don't think that's the issue here. That's for WSL, we're trying to do this thing without WSL.
I think the issue is just that we need to do 'conda install cudatoolkit=11.3', but getting an issue running from my install script.
Is there another way to install it, if it doesn't work?
from sd_dreambooth_extension.
Is this because we don't use conda for the virtual environment? Basically you can't install cudatoolkit using pip which what the issue is. PyTorch doesn't install everything that this needs to run under Windows. I think.
Not an expert on this. I've tried all kinds of instructions from various sources and had no luck getting 8bit Adam to work in Windows with A111 venv.
from sd_dreambooth_extension.
Is this because we don't use conda for the virtual environment? Basically you can't install cudatoolkit using pip which what the issue is. PyTorch doesn't install everything that this needs to run under Windows. I think.
Not an expert on this. I've tried all kinds of instructions from various sources and had no luck getting 8bit Adam to work in Windows with A111 venv.
Not 100%, but that's what I'm leaning towards as the issue? Seems like we might need to manually use conda to install cuda-toolkit to the venv, or figure out why it's failing from within the install.py script.
from sd_dreambooth_extension.
I fixed this by changing the line in cextension.py:
binary_name = evaluate_cuda_setup()
to hardcode the cuda DLL path to libbitsandbytes_cuda116.dll
from sd_dreambooth_extension.
I think there is already a solution for this in Windows here: bitsandbytes-foundation/bitsandbytes#30 (comment)
Basically, you download 2 dll files from this repo: https://github.com/DeXtmL/bitsandbytes-win-prebuilt
and drop them both into webui's sitepackage bitandbytes folder, should be following: stable-diffusion-webui\venv\lib\site-packages\bitsandbytes
After that, edit and replace cextension.py file in the same folder with this: https://pastebin.com/jjgxuh8V
and go to cuda_setup folder, replace the main.py file with this: https://pastebin.com/BsEzpdpw
That is all, problem solved, I tried and no more CUDA_SETUP: WARNING!
from sd_dreambooth_extension.
Found the problem, as LaikaSA pointed out.
I was copying the main.py to the wrong spot.
Should be fixed and working now on windoze with c42b4f3
from sd_dreambooth_extension.
Found the problem, as LaikaSA pointed out.
I was copying the main.py to the wrong spot.
Should be fixed and working now on windoze with c42b4f3
Yhea it worked, but now i got CUDA out of memory on my 3080TI, however disabling Text encoder
fixed it and now its training. Running at 11.7 GB /12. Getting about 3it/s
from sd_dreambooth_extension.
Now i got this.
Traceback (most recent call last):
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\modules\ui.py", line 185, in f
res = list(func(*args, **kwargs))
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\webui.py", line 54, in f
res = func(*args, **kwargs)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\extensions\sd_dreambooth_extension\dreambooth\dreambooth.py", line 256, in start_training
trained_steps = main(config)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\extensions\sd_dreambooth_extension\dreambooth\train_dreambooth.py", line 796, in main
text_enc_model = CLIPTextModel.from_pretrained(args.model_name,
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\transformers\modeling_utils.py", line 1966, in from_pretrained
config, model_kwargs = cls.config_class.from_pretrained(
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\transformers\models\clip\configuration_clip.py", line 133, in from_pretrained
config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\transformers\configuration_utils.py", line 559, in get_config_dict
config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\transformers\configuration_utils.py", line 614, in _get_config_dict
resolved_config_file = cached_file(
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\transformers\utils\hub.py", line 424, in cached_file
raise EnvironmentError(
OSError: Luvelia is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models'
If this is a private repository, make sure to pass a token having permission to this repo with use_auth_token
or log in with huggingface-cli login
and pass use_auth_token=True
.
from sd_dreambooth_extension.
line 796, in main
I see the issue. Should be fixed with 1ddfcfc
from sd_dreambooth_extension.
line 796, in main
I see the issue. Should be fixed with 1ddfcfc
So far its fixed, yhea, it generated till 100 steps as a test and it successfully made an image and saved a ckpt file.
However....it got stuck on a loop saying training complete? and it kept saving a ckpt file every 1 step.
from sd_dreambooth_extension.
line 796, in main
I see the issue. Should be fixed with 1ddfcfc
So far its fixed, yhea, it generated till 100 steps as a test and it successfully made an image and saved a ckpt file. However....it got stuck on a loop saying training complete? and it kept saving a ckpt file every 1 step.
Wheeee!
Allright, I'll se what I did wrong now. :P
Forgot to mention it only happened clicking the blue button to restore after hitting cancel and then train again from 100
from sd_dreambooth_extension.
line 796, in main
I see the issue. Should be fixed with 1ddfcfc
So far its fixed, yhea, it generated till 100 steps as a test and it successfully made an image and saved a ckpt file. However....it got stuck on a loop saying training complete? and it kept saving a ckpt file every 1 step.
Wheeee!
Allright, I'll se what I did wrong now. :PForgot to mention it only happened clicking the blue button to restore after hitting cancel and then train again from 100
Can you post the parameters it spits out in the console when you start training after doing this?
from sd_dreambooth_extension.
let me check if i can replicate it...
Trying to parse: [{'instance_prompt': 'Luvelia', 'class_prompt': 'person', 'instance_data_dir': 'C:\Users\Hector\Downloads\New folder\New folder', 'class_data_dir': ''}]
Unable to load concepts as JSON, trying as file.
Loaded model.
Allocated: 0.0GB
Reserved: 0.0GB
The config attributes {'set_alpha_to_one': False, 'skip_prk_steps': True, 'steps_offset': 1} were passed to DDPMScheduler, but are not expected and will be ignored. Please verify your scheduler_config.json configuration file.
Scheduler Loaded
Allocated: 0.2GB
Reserved: 0.2GB
Total target lifetime optimization steps = 2000
CPU: False Adam: True, Prec: fp16, Prior: False, Grad: True, TextTr: True
Allocated: 3.8GB
Reserved: 3.9GB
Steps: 0%| | 1/1000 [00:00<10:28, 1.59it/s, loss=0.00983, lr=5e-6]Error completing request
Arguments: ('Luvelia', 'C:\Users\Hector\Downloads\New folder\New folder', '', 'Luvelia', 'person', '', '', 1.0, 7.5, 40.0, 0, 512, False, True, 1, 1, 10, 1000, 1, True, 5e-06, False, 'constant', 0, True, 0.9, 0.999, 0.01, 1e-08, 1, 100, 1000, 'fp16', True, "[{'instance_prompt': 'Luvelia', 'class_prompt': 'person', 'instance_data_dir': 'C:\\Users\\Hector\\Downloads\\New folder\\New folder', 'class_data_dir': ''}]", False) {}
Traceback (most recent call last):
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\modules\ui.py", line 185, in f
res = list(func(*args, **kwargs))
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\webui.py", line 54, in f
res = func(*args, **kwargs)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\extensions\sd_dreambooth_extension\dreambooth\dreambooth.py", line 256, in start_training
trained_steps = main(config)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\extensions\sd_dreambooth_extension\dreambooth\train_dreambooth.py", line 766, in main
accelerator.backward(loss)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\accelerate\accelerator.py", line 882, in backward
self.scaler.scale(loss).backward(**kwargs)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch_tensor.py", line 396, in backward
torch.autograd.backward(self, gradient, retain_graph, create_graph, inputs=inputs)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch\autograd_init_.py", line 173, in backward
Variable.execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch\autograd\function.py", line 253, in apply
return user_fn(self, *args)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch\utils\checkpoint.py", line 146, in backward
torch.autograd.backward(outputs_with_grad, args_with_grad)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch\autograd_init.py", line 173, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
RuntimeError: CUDA out of memory. Tried to allocate 512.00 MiB (GPU 0; 12.00 GiB total capacity; 10.19 GiB already allocated; 0 bytes free; 10.74 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
Steps: 0%| | 1/1000 [00:02<45:06, 2.71s/it, loss=0.00983, lr=5e-6]
Starting Dreambooth training...
VRAM cleared.
Allocated: 0.0GB
Reserved: 8.2GB
Trying to parse: [{'instance_prompt': 'Luvelia', 'class_prompt': 'person', 'instance_data_dir': 'C:\Users\Hector\Downloads\New folder\New folder', 'class_data_dir': ''}]
Unable to load concepts as JSON, trying as file.
Loaded model.
Allocated: 0.0GB
Reserved: 8.2GB
The config attributes {'set_alpha_to_one': False, 'skip_prk_steps': True, 'steps_offset': 1} were passed to DDPMScheduler, but are not expected and will be ignored. Please verify your scheduler_config.json configuration file.
Scheduler Loaded
Allocated: 0.2GB
Reserved: 8.2GB
Total target lifetime optimization steps = 2000
CPU: False Adam: True, Prec: fp16, Prior: False, Grad: True, TextTr: True
Allocated: 3.8GB
Reserved: 8.2GB
Steps: 2%|▊ | 17/1000 [00:06<06:08, 2.67it/s, loss=0.0705, lr=5e-6] Loaded pipeline for preview...
Allocated: 7.1GB
Reserved: 11.2GB
Saving checkpoint at step 1017.
Successfully trained model for a total of 1017 steps, converting to ckpt.
Pipeline cleared...
Allocated: 5.7GB
Reserved: 6.7GB
Training complete??
Error completing request
Arguments: ('Luvelia', 'C:\Users\Hector\Downloads\New folder\New folder', '', 'Luvelia', 'person', '', '', 1.0, 7.5, 40.0, 0, 512, False, True, 1, 1, 10, 1000, 1, True, 5e-06, False, 'constant', 0, True, 0.9, 0.999, 0.01, 1e-08, 1, 100, 1000, 'fp16', True, "[{'instance_prompt': 'Luvelia', 'class_prompt': 'person', 'instance_data_dir': 'C:\\Users\\Hector\\Downloads\\New folder\\New folder', 'class_data_dir': ''}]", False) {}
Traceback (most recent call last):
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\modules\ui.py", line 185, in f
res = list(func(*args, **kwargs))
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\webui.py", line 54, in f
res = func(*args, **kwargs)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\extensions\sd_dreambooth_extension\dreambooth\dreambooth.py", line 256, in start_training
trained_steps = main(config)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\extensions\sd_dreambooth_extension\dreambooth\train_dreambooth.py", line 766, in main
accelerator.backward(loss)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\accelerate\accelerator.py", line 882, in backward
self.scaler.scale(loss).backward(**kwargs)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch_tensor.py", line 396, in backward
torch.autograd.backward(self, gradient, retain_graph, create_graph, inputs=inputs)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch\autograd_init_.py", line 173, in backward
Variable.execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch\autograd\function.py", line 253, in apply
return user_fn(self, *args)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch\utils\checkpoint.py", line 146, in backward
torch.autograd.backward(outputs_with_grad, args_with_grad)
File "C:\SUPER_SD_2.0\stable-diffusion-webui-master\venv\lib\site-packages\torch\autograd_init.py", line 173, in backward
Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass
RuntimeError: CUDA out of memory. Tried to allocate 512.00 MiB (GPU 0; 12.00 GiB total capacity; 10.19 GiB already allocated; 0 bytes free; 10.75 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
Steps: 2%|▊ | 17/1000 [00:21<20:17, 1.24s/it, loss=0.0705, lr=5e-6]
Starting Dreambooth training...
VRAM cleared.
Allocated: 0.0GB
Reserved: 8.2GB
Trying to parse: [{'instance_prompt': 'Luvelia', 'class_prompt': 'person', 'instance_data_dir': 'C:\Users\Hector\Downloads\New folder\New folder', 'class_data_dir': ''}]
Unable to load concepts as JSON, trying as file.
Loaded model.
Allocated: 0.0GB
Reserved: 8.2GB
The config attributes {'set_alpha_to_one': False, 'skip_prk_steps': True, 'steps_offset': 1} were passed to DDPMScheduler, but are not expected and will be ignored. Please verify your scheduler_config.json configuration file.
Scheduler Loaded
Allocated: 0.2GB
Reserved: 8.2GB
Total target lifetime optimization steps = 2000
CPU: False Adam: True, Prec: fp16, Prior: False, Grad: True, TextTr: True
Allocated: 3.8GB
Reserved: 8.2GB
Steps: 5%|██▋ | 52/1000 [00:20<06:26, 2.45it/s, loss=0.0851, lr=5e-6] Loaded pipeline for preview...
Allocated: 7.1GB
Reserved: 11.2GB
Saving checkpoint at step 1052.
Successfully trained model for a total of 1052 steps, converting to ckpt.
Pipeline cleared...
Allocated: 5.7GB
Reserved: 7.1GB
Training complete??
Steps: 5%|██▌ | 53/1000 [00:36<1:18:38, 4.98s/it, loss=0.0851, lr=5e-6] Loaded pipeline for preview...
Allocated: 7.1GB
Reserved: 11.2GB
Saving checkpoint at step 1053.
Successfully trained model for a total of 1053 steps, converting to ckpt.
Pipeline cleared...
Allocated: 5.7GB
Reserved: 7.1GB
Training complete??
Steps: 5%|██▋ | 54/1000 [01:03<3:04:55, 11.73s/it, loss=0.0851, lr=5e-6] Loaded pipeline for preview...
Allocated: 7.1GB
Reserved: 11.2GB
Saving checkpoint at step 1054.
from sd_dreambooth_extension.
after pressing cancel it gets stuck on a loop, i had to close the cmd to stop it
from sd_dreambooth_extension.
Doesn't work.
Exception importing 8bit adam: cannot import name 'CUDASetup' from partially initialized module 'bitsandbytes.cextension' (most likely due to a circular import)
However, the same DLLs work with geocine's version because that force installs cu116
So I went to webui's main folder and did this:
.\venv\Scripts\activate
pip install torch==1.12.1+cu116 torchvision==0.13.1+cu116 --extra-index-url https://download.pytorch.org/whl/cu116
It then uninstalled cu113
Found existing installation: torchvision 0.13.1+cu113
Uninstalling torchvision-0.13.1+cu113:
Successfully uninstalled torchvision-0.13.1+cu113
Successfully installed torch-1.12.1+cu116 torchvision-0.13.1+cu116
and now it works.
from sd_dreambooth_extension.
Just added this same bit in install.py via 9e7ec85
from sd_dreambooth_extension.
Great. Now this works on windows fine for me.
from sd_dreambooth_extension.
Now we need a good training guide, i'm getting confused and sometimes bad results, mostly because i don't know how to train this well.
from sd_dreambooth_extension.
from sd_dreambooth_extension.
I'm going to mark this as closed. It may be necessary to specify a custom torch cmd in webui-user.bat, which I will document briefly in the readme. But, I can get bnb to load and run fine now, so we should be good here.
from sd_dreambooth_extension.
Readme has been updated with installation instructions, but the trick to update CUDA (if needed) is to add this to your webui-user.bat file:
set TORCH_COMMAND="pip install torch==1.12.1+cu116 torchvision==0.13.1+cu116 --extra-index-url https://download.pytorch.org/whl/cu116"
from sd_dreambooth_extension.
I'm a bit confused.
Automatic1111 suggest this in its webui-user.sh
(I'm a linux user):
export TORCH_COMMAND="pip install torch==1.12.1+cu113 --extra-index-url https://download.pytorch.org/whl/cu113"
But you suggests this other one (for Windows users at least):
TORCH_COMMAND="pip install torch==1.12.1+cu116 torchvision==0.13.1+cu116 --extra-index-url https://download.pytorch.org/whl/cu116"
Is Automatic1111 in the wrong? Shoud he/she use cu116 too?
Also, it seems a bit arbitrary to install one cuda version or another one without checking our own systems. So shouldn't the CUDA version be the one our own system has (i.e. nvcc --version)? Or am I missing something? Thanks in advance!
EDIT:
Disregard all that. I noticed bitsandbytes has some pretty stringent system requirements which I wasn't aware:
- Hardware requirements:
- LLM.int8(): NVIDIA Turing (RTX 20xx; T4) or Ampere GPU (RTX 30xx; A4-A100); (a GPU from 2018 or older).
- 8-bit optimizers and quantization: NVIDIA Maxwell GPU or newer (>=GTX 9XX).
from sd_dreambooth_extension.
8-bit optimizers and quantization: NVIDIA Maxwell GPU or newer (>=GTX 9XX).
So if I'm not mistaken choosing 8-bit adam will fail if you don't have those NVIDIA specs.
from sd_dreambooth_extension.
Similar issues taking place with newer versions now.
Anyone has a solution for CUDA 11.7 + ? Talking about bitsandbytes on windows (11) and cuda accel.
Thanks
from sd_dreambooth_extension.
Possible solution for Windows users:
- Go to ...\stable-diffusion-webui\venv\Lib\site-packages
- Delete the folder "bitsandbytes"
- Use the bitsandbytes version with the following link:
https://pypi.org/project/bitsandbytes-windows/ - Install the version and ensure it was added to site-packages of stable diffusion
- If you use xformers ensure you use version 0.0.22 in requirements.txt (xformers==0.0.22)
from sd_dreambooth_extension.
Bitsandbytes was not supported windows before, but my method can support windows.(yuhuang)
1 open folder J:\StableDiffusion\sdwebui,Click the address bar of the folder and enter CMD
or WIN+R, CMD 。enter,cd /d J:\StableDiffusion\sdwebui
2 J:\StableDiffusion\sdwebui\py310\python.exe -m pip uninstall bitsandbytes
3 J:\StableDiffusion\sdwebui\py310\python.exe -m pip uninstall bitsandbytes-windows
4 J:\StableDiffusion\sdwebui\py310\python.exe -m pip install https://github.com/jllllll/bitsandbytes-windows-webui/releases/download/wheels/bitsandbytes-0.41.1-py3-none-win_amd64.whl
Replace your SD venv directory file(python.exe Folder) here(J:\StableDiffusion\sdwebui\py310)
from sd_dreambooth_extension.
OR you are Linux distribution (Ubuntu, MacOS, etc.)system ,AND CUDA Version: 11.X.
Bitsandbytes can support ubuntu.(yuhuang)
1 open folder J:\StableDiffusion\sdwebui,Click the address bar of the folder and enter CMD
or WIN+R, CMD 。enter,cd /d J:\StableDiffusion\sdwebui
2 J:\StableDiffusion\sdwebui\py310\python.exe -m pip uninstall bitsandbytes
3 J:\StableDiffusion\sdwebui\py310\python.exe -m pip uninstall bitsandbytes-windows
4 J:\StableDiffusion\sdwebui\py310\python.exe -m pip install https://github.com/TimDettmers/bitsandbytes/releases/download/0.41.0/bitsandbytes-0.41.0-py3-none-any.whl
Replace your SD venv directory file(python.exe Folder) here(J:\StableDiffusion\sdwebui\py310)
from sd_dreambooth_extension.
Related Issues (20)
- [Bug]: Can't create model using trained & generated model using dreambooth. HOT 3
- [Bug]: Exception training model: 'Cannot copy out of meta tensor; no data!'. HOT 2
- [Bug]: HOT 1
- [Bug]: The deprecation tuple ('LoRAAttnProcessor2_0', '0.26.0', 'Make sure use AttnProcessor2_0 instead by settingLoRA layers to `self.{to_q,to_k,to_v,to_out[0]}.lora_layer` respectively. This will be done automatically when using `LoraLoaderMixin.load_lora_weights`') should be removed since diffusers' version 0.26.1 is >= 0.26.0 HOT 4
- [Bug]: Completely unable to train any LORA with CUDA out of memory error HOT 2
- RuntimeError: Expected query, key, and value to have the same dtype, but got query.dtype: c10::Half key.dtype: float and value.dtype: float instead. HOT 9
- [Bug]: HOT 1
- [Bug]: OSError: Can't load tokenizer for 'laion/CLIP-ViT-bigG-14-laion2B-39B-b160k'. HOT 2
- [Bug]: TypeError: intercept_args() got an unexpected keyword argument 'multiprocessing_context' HOT 1
- Error al cargar sd_dreambooth_extension en Windows 10: 'LoRAAttnProcessor2_0' no definido HOT 1
- [Bug]: Dreambooth (input tab) not showing correctly HOT 10
- [Bug]: Exception training model: 'type object 'LoraLoaderMixin' has no attribute '_modify_text_encoder''. HOT 4
- [Bug]: AttributeError: 'NoneType' object has no attribute 'unscale_grads' HOT 1
- [Bug]: Unable to further train using previously trained ckpt in dreambooth. HOT 1
- [Bug]: Dreambooth can not start training HOT 3
- AttributeError: module 'jax.random' has no attribute 'KeyArray'[Bug]: HOT 1
- [Bug]: Fast api of concepts don't work HOT 2
- [Bug]: AttributeError: 'NoneType' object has no attribute 'keys' HOT 1
- [Bug]: Unable to do training on sdxl model HOT 4
- [Bug]: Memory Attention default try to use xformers if Class Images Per Instance Image is greather that zero and need to generate images HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from sd_dreambooth_extension.