Code Monkey home page Code Monkey logo

kohya-config-webui's Introduction

kohya-config-webui

A WebUI for making config files used by kohya_sd_script 9@`$94}$U~DCAQCEA }20AX

现在我们有什么?

目前,我为kohya-lora-dreambooth训练中较为常用、实用的训练参数,使用gradio和toml库,编写了一个交互式的WebUI生成工具,可以在带有python环境的windows和Colab环境中快速部署。

使用本项目,你可以快速指定训练参数,并生成config_file.toml和sample_prompt.txt。

如果你觉得此项目有用,可以给我一颗小星星,非常感谢你⭐


Notebook Name Description Link Old-Version
Colab_Lora_train 基于Akegarasu/lora-scripts的定制化Colab notebook
kohya_train_webui NEW 基于WSH032/kohya-config-webui的WebUI版Colab notebook

使用方法

(一)Colab版本(带有完整训练环境):

(二)AUTOMATIC1111/stable-diffusion-webui插件:

https://github.com/WSH032/kohya-config-webui.git

将这个仓库连接复制到SD-WebUi的扩展->从网址安装界面下载即可

(三)直接下载:

运行以下代码,或直接从github上下载zip并解压(直接下载将无法使用update.bat),

git clone https://github.com/WSH032/kohya-config-webui.git

运行目录下的install_webui.ps1在虚拟环境中安装gradio和toml,或者

pip install  gradio>=3.24.1
pip install  toml>=0.10.2

运行run_webui.ps1,或者

python .\module\kohya_config_webui.py

Todo

  • 增加上一次参数保存功能
  • 增加dataset_config.toml生成功能

Credit

Based on the work of kohya-ss , Linaqruf. Thanks to them.

  • 这个项目使用的的训练脚本来自kohya-ss
  • notebook中部分代码(如下载模块)来自Linaqruf

Attention: It's called kohya-config-webui, but I don't have a license for kohya. It just creates config files for kohya-ss.

  • 这个插件叫kohya-config-webui,但是我没有kohya的授权,我只是为了说明它的作用是生成一个用于kohya-ss训练的config文件

上述两位作者和我目前采取的是Apache-2.0 license

如果你基于此项目进行了修改、引用等用途,请注意原作者的协议。

请在你使用的部分标明代码来源。

其他部分展示

0S2_@5B6638VJ 4%Y@TQDXK

7$K@5833YHY(8T1 `3RE03L

kohya-config-webui's People

Contributors

wsh032 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar

kohya-config-webui's Issues

开始训练时的报错

(B站过来的)
我今天又跑了一下,做了以下事情,没有重现之前的报错,但是有新的问题。
1 下载1.5和2.1的模型 2 修改生成的名称 3 修改底模为1.5 4 修改采样的prompt.其他参数默认
报错我贴在下面,你看看。

Traceback (most recent call last):
  File "/content/sd-scripts/train_network.py", line 773, in <module>
    train(args)
  File "/content/sd-scripts/train_network.py", line 146, in train
    text_encoder, vae, unet, _ = train_util.load_target_model(args, weight_dtype, accelerator)
  File "/content/sd-scripts/library/train_util.py", line 2923, in load_target_model
    text_encoder, vae, unet, load_stable_diffusion_format = _load_target_model(
  File "/content/sd-scripts/library/train_util.py", line 2889, in _load_target_model
    text_encoder, vae, unet = model_util.load_models_from_stable_diffusion_checkpoint(args.v2, name_or_path, device)
  File "/content/sd-scripts/library/model_util.py", line 856, in load_models_from_stable_diffusion_checkpoint
    _, state_dict = load_checkpoint_with_text_encoder_conversion(ckpt_path, device)
  File "/content/sd-scripts/library/model_util.py", line 833, in load_checkpoint_with_text_encoder_conversion
    checkpoint = torch.load(ckpt_path, map_location=device)
  File "/usr/local/lib/python3.10/dist-packages/torch/serialization.py", line 815, in load
    return _legacy_load(opened_file, map_location, pickle_module, **pickle_load_args)
  File "/usr/local/lib/python3.10/dist-packages/torch/serialization.py", line 1033, in _legacy_load
    magic_number = pickle_module.load(f, **pickle_load_args)
_pickle.UnpicklingError: invalid load key, '\x09'.

我看上面的\x09很像是中文的原因,但实际情况是没有输入任何中文的

python setup.py egg_info did not run successfully 和 Encountered error while generating package metadata.

Cloning into '/content/sd-scripts'...
remote: Enumerating objects: 2808, done.
remote: Counting objects: 100% (1021/1021), done.
remote: Compressing objects: 100% (276/276), done.
remote: Total 2808 (delta 842), reused 847 (delta 745), pack-reused 1787
Receiving objects: 100% (2808/2808), 3.28 MiB | 4.97 MiB/s, done.
Resolving deltas: 100% (1922/1922), done.
Cloning into '/content/kohya-config-webui'...
remote: Enumerating objects: 118, done.
remote: Counting objects: 100% (118/118), done.
remote: Compressing objects: 100% (91/91), done.
remote: Total 118 (delta 58), reused 53 (delta 22), pack-reused 0
Receiving objects: 100% (118/118), 67.27 KiB | 13.45 MiB/s, done.
Resolving deltas: 100% (58/58), done.
torch安装中
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 108.2/108.2 MB 8.2 MB/s eta 0:00:00
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 123.8/123.8 MB 8.0 MB/s eta 0:00:00
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 123.6/123.6 MB 7.9 MB/s eta 0:00:00
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 50.9/50.9 MB 12.9 MB/s eta 0:00:00
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 292.5/292.5 kB 33.1 MB/s eta 0:00:00
error: subprocess-exited-with-error

× python setup.py egg_info did not run successfully.
│ exit code: 1
╰─> See above for output.

note: This error originates from a subprocess, and is likely not a problem with pip.
Preparing metadata (setup.py) ... error
error: metadata-generation-failed

× Encountered error while generating package metadata.
╰─> See above for output.

note: This is an issue with the package mentioned above, not pip.
hint: See above for details.
torch安装完成
kohya依赖安装中

使用safetensor是不是会报错?

──────────────────── Traceback (most recent call last) ──────────────────────╮
│ /content/sd-scripts/train_network.py:873 in │
│ │
│ 870 │ args = parser.parse_args() │
│ 871 │ args = train_util.read_config_from_file(args, parser) │
│ 872 │ │
│ ❱ 873 │ train(args) │
│ 874 │
│ │
│ /content/sd-scripts/train_network.py:168 in train │
│ │
│ 165 │ weight_dtype, save_dtype = train_util.prepare_dtype(args) │
│ 166 │ │
│ 167 │ # モデルを読み込む │
│ ❱ 168 │ text_encoder, vae, unet, _ = train_util.load_target_model(args, we │
│ 169 │ │
│ 170 │ # モデルに xformers とか memory efficient attention を組み込む │
│ 171 │ train_util.replace_unet_modules(unet, args.mem_eff_attn, args.xfor │
│ │
│ /content/sd-scripts/library/train_util.py:3149 in load_target_model │
│ │
│ 3146 │ │ if pi == accelerator.state.local_process_index: │
│ 3147 │ │ │ print(f"loading model for process {accelerator.state.loca │
│ 3148 │ │ │ │
│ ❱ 3149 │ │ │ text_encoder, vae, unet, load_stable_diffusion_format = _ │
│ 3150 │ │ │ │ args, weight_dtype, accelerator.device if args.lowram │
│ 3151 │ │ │ ) │
│ 3152 │
│ │
│ /content/sd-scripts/library/train_util.py:3115 in load_target_model │
│ │
│ 3112 │ load_stable_diffusion_format = os.path.isfile(name_or_path) # de │
│ 3113 │ if load_stable_diffusion_format: │
│ 3114 │ │ print(f"load StableDiffusion checkpoint: {name_or_path}") │
│ ❱ 3115 │ │ text_encoder, vae, unet = model_util.load_models_from_stable

│ 3116 │ else: │
│ 3117 │ │ # Diffusers model is loaded to CPU │
│ 3118 │ │ print(f"load Diffusers pretrained models: {name_or_path}") │
│ │
│ /content/sd-scripts/library/model_util.py:856 in │
│ load_models_from_stable_diffusion_checkpoint │
│ │
│ 853 │
│ 854 # TODO dtype指定の動作が怪しいので確認する text_encoderを指定形式で作 │
│ 855 def load_models_from_stable_diffusion_checkpoint(v2, ckpt_path, devic │
│ ❱ 856 │ _, state_dict = load_checkpoint_with_text_encoder_conversion(ckpt │
│ 857 │ │
│ 858 │ # Convert the UNet2DConditionModel model. │
│ 859 │ unet_config = create_unet_diffusers_config(v2, unet_use_linear_pr │
│ │
│ /content/sd-scripts/library/model_util.py:833 in │
│ load_checkpoint_with_text_encoder_conversion │
│ │
│ 830 │ │ checkpoint = None │
│ 831 │ │ state_dict = load_file(ckpt_path) # , device) # may causes e │
│ 832 │ else: │
│ ❱ 833 │ │ checkpoint = torch.load(ckpt_path, map_location=device) │
│ 834 │ │ if "state_dict" in checkpoint: │
│ 835 │ │ │ state_dict = checkpoint["state_dict"] │
│ 836 │ │ else: │
│ │
│ /usr/local/lib/python3.10/dist-packages/torch/serialization.py:815 in load │
│ │
│ 812 │ │ │ │ return _legacy_load(opened_file, map_location, _weigh │
│ 813 │ │ │ except RuntimeError as e: │
│ 814 │ │ │ │ raise pickle.UnpicklingError(UNSAFE_MESSAGE + str(e)) │
│ ❱ 815 │ │ return _legacy_load(opened_file, map_location, pickle_module, │
│ 816 │
│ 817 │
│ 818 # Register pickling support for layout instances such as │
│ │
│ /usr/local/lib/python3.10/dist-packages/torch/serialization.py:1033 in │
│ _legacy_load │
│ │
│ 1030 │ │ │ f"Received object of type "{type(f)}". Please update to │
│ 1031 │ │ │ "functionality.") │
│ 1032 │ │
│ ❱ 1033 │ magic_number = pickle_module.load(f, **pickle_load_args) │
│ 1034 │ if magic_number != MAGIC_NUMBER: │
│ 1035 │ │ raise RuntimeError("Invalid magic number; corrupt file?") │
│ 1036 │ protocol_version = pickle_module.load(f, **pickle_load_args) │
╰──────────────────────────────────────────────────────────────────────────────╯
UnpicklingError: invalid load key, '^'.
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /usr/local/bin/accelerate:8 in │
│ │
│ 5 from accelerate.commands.accelerate_cli import main │
│ 6 if name == 'main': │
│ 7 │ sys.argv[0] = re.sub(r'(-script.pyw|.exe)?$', '', sys.argv[0]) │
│ ❱ 8 │ sys.exit(main()) │
│ 9 │
│ │
│ /usr/local/lib/python3.10/dist-packages/accelerate/commands/accelerate_cli.p │
│ y:45 in main │
│ │
│ 42 │ │ exit(1) │
│ 43 │ │
│ 44 │ # Run │
│ ❱ 45 │ args.func(args) │
│ 46 │
│ 47 │
│ 48 if name == "main": │
│ │
│ /usr/local/lib/python3.10/dist-packages/accelerate/commands/launch.py:1104 │
│ in launch_command │
│ │
│ 1101 │ elif defaults is not None and defaults.compute_environment == Com │
│ 1102 │ │ sagemaker_launcher(defaults, args) │
│ 1103 │ else: │
│ ❱ 1104 │ │ simple_launcher(args) │
│ 1105 │
│ 1106 │
│ 1107 def main(): │
│ │
│ /usr/local/lib/python3.10/dist-packages/accelerate/commands/launch.py:567 in │
│ simple_launcher │
│ │
│ 564 │ process = subprocess.Popen(cmd, env=current_env) │
│ 565 │ process.wait() │
│ 566 │ if process.returncode != 0: │
│ ❱ 567 │ │ raise subprocess.CalledProcessError(returncode=process.return │
│ 568 │
│ 569 │
│ 570 def multi_gpu_launcher(args): │
╰──────────────────────────────────────────────────────────────────────────────╯
CalledProcessError: Command '['/usr/bin/python3', 'train_network.py',
'--config_file=config_file.toml', '--sample_prompts=sample_prompts.txt']'
returned non-zero exit status 1.

求助AssertionError: resolution is required / resolution

Loading settings from /content/drive/MyDrive/Lora/kohya_config_webui_save/kohya_config_webui_save.toml...
/content/drive/MyDrive/Lora/kohya_config_webui_save/kohya_config_webui_save
prepare tokenizer
update token length: 225
Using DreamBooth method.
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /content/sd-scripts/train_network.py:873 in │
│ │
│ 870 │ args = parser.parse_args() │
│ 871 │ args = train_util.read_config_from_file(args, parser) │
│ 872 │ │
│ ❱ 873 │ train(args) │
│ 874 │
│ │
│ /content/sd-scripts/train_network.py:135 in train │
│ │
│ 132 │ │ │ │ } │
│ 133 │ │ │
│ 134 │ │ blueprint = blueprint_generator.generate(user_config, args, to │
│ ❱ 135 │ │ train_dataset_group = config_util.generate_dataset_group_by_bl │
│ 136 │ else: │
│ 137 │ │ # use arbitrary dataset class │
│ 138 │ │ train_dataset_group = train_util.load_arbitrary_dataset(args, │
│ │
│ /content/sd-scripts/library/config_util.py:375 in │
│ generate_dataset_group_by_blueprint │
│ │
│ 372 │ dataset_klass = FineTuningDataset │
│ 373 │ │
│ 374 │ subsets = [subset_klass(**asdict(subset_blueprint.params)) for sub │
│ ❱ 375 │ dataset = dataset_klass(subsets=subsets, **asdict(dataset_blueprin │
│ 376 │ datasets.append(dataset) │
│ 377 │
│ 378 # print info │
│ │
│ /content/sd-scripts/library/train_util.py:1028 in init
│ │
│ 1025 │ ) -> None: │
│ 1026 │ │ super().init(tokenizer, max_token_length, resolution, deb │
│ 1027 │ │ │
│ ❱ 1028 │ │ assert resolution is not None, f"resolution is required / res │
│ 1029 │ │ │
│ 1030 │ │ self.batch_size = batch_size │
│ 1031 │ │ self.size = min(self.width, self.height) # 短いほう │
╰──────────────────────────────────────────────────────────────────────────────╯
AssertionError: resolution is required / resolution(解像度)指定は必須です
╭───────────────────── Traceback (most recent call last) ──────────────────────╮
│ /usr/local/bin/accelerate:8 in │
│ │
│ 5 from accelerate.commands.accelerate_cli import main │
│ 6 if name == 'main': │
│ 7 │ sys.argv[0] = re.sub(r'(-script.pyw|.exe)?$', '', sys.argv[0]) │
│ ❱ 8 │ sys.exit(main()) │
│ 9 │
│ │
│ /usr/local/lib/python3.10/dist-packages/accelerate/commands/accelerate_cli.p │
│ y:45 in main │
│ │
│ 42 │ │ exit(1) │
│ 43 │ │
│ 44 │ # Run │
│ ❱ 45 │ args.func(args) │
│ 46 │
│ 47 │
│ 48 if name == "main": │
│ │
│ /usr/local/lib/python3.10/dist-packages/accelerate/commands/launch.py:1104 │
│ in launch_command │
│ │
│ 1101 │ elif defaults is not None and defaults.compute_environment == Com │
│ 1102 │ │ sagemaker_launcher(defaults, args) │
│ 1103 │ else: │
│ ❱ 1104 │ │ simple_launcher(args) │
│ 1105 │
│ 1106 │
│ 1107 def main(): │
│ │
│ /usr/local/lib/python3.10/dist-packages/accelerate/commands/launch.py:567 in │
│ simple_launcher │
│ │
│ 564 │ process = subprocess.Popen(cmd, env=current_env) │
│ 565 │ process.wait() │
│ 566 │ if process.returncode != 0: │
│ ❱ 567 │ │ raise subprocess.CalledProcessError(returncode=process.return │
│ 568 │
│ 569 │
│ 570 def multi_gpu_launcher(args): │
╰──────────────────────────────────────────────────────────────────────────────╯
CalledProcessError: Command '['/usr/bin/python3', 'train_network.py',
'--config_file=/content/drive/MyDrive/Lora/kohya_config_webui_save/kohya_config_
webui_save.toml',
'--sample_prompts=/content/drive/MyDrive/Lora/kohya_config_webui_save/kohya_conf
ig_webui_save.toml']' returned non-zero exit status 1.

求助:开始训练时出现的报错

/usr/local/lib/python3.10/dist-packages/transformers/utils/generic.py:441: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
_torch_pytree._register_pytree_node(
/usr/local/lib/python3.10/dist-packages/transformers/utils/generic.py:309: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
_torch_pytree._register_pytree_node(
/usr/local/lib/python3.10/dist-packages/transformers/utils/generic.py:441: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
_torch_pytree._register_pytree_node(
/usr/local/lib/python3.10/dist-packages/transformers/utils/generic.py:309: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
_torch_pytree._register_pytree_node(
/usr/local/lib/python3.10/dist-packages/diffusers/utils/outputs.py:63: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
torch.utils._pytree._register_pytree_node(
/usr/local/lib/python3.10/dist-packages/diffusers/utils/outputs.py:63: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
torch.utils._pytree._register_pytree_node(
/usr/local/lib/python3.10/dist-packages/diffusers/utils/outputs.py:63: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
torch.utils._pytree._register_pytree_node(
Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1382, in _get_module
return importlib.import_module("." + module_name, self.name)
File "/usr/lib/python3.10/importlib/init.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "", line 1050, in _gcd_import
File "", line 1027, in _find_and_load
File "", line 1006, in _find_and_load_unlocked
File "", line 688, in _load_unlocked
File "", line 883, in exec_module
File "", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/transformers/models/clip/image_processing_clip.py", line 21, in
from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict
File "/usr/local/lib/python3.10/dist-packages/transformers/image_processing_utils.py", line 28, in
from .image_transforms import center_crop, normalize, rescale
File "/usr/local/lib/python3.10/dist-packages/transformers/image_transforms.py", line 47, in
import tensorflow as tf
File "/usr/local/lib/python3.10/dist-packages/tensorflow/init.py", line 48, in
from tensorflow._api.v2 import internal
File "/usr/local/lib/python3.10/dist-packages/tensorflow/_api/v2/internal/init.py", line 8, in
from tensorflow._api.v2.internal import autograph
File "/usr/local/lib/python3.10/dist-packages/tensorflow/_api/v2/internal/autograph/init.py", line 8, in
from tensorflow.python.autograph.core.ag_ctx import control_status_ctx # line: 34
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/autograph/core/ag_ctx.py", line 21, in
from tensorflow.python.autograph.utils import ag_logging
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/autograph/utils/init.py", line 17, in
from tensorflow.python.autograph.utils.context_managers import control_dependency_on_returns
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/autograph/utils/context_managers.py", line 19, in
from tensorflow.python.framework import ops
File "/usr/local/lib/python3.10/dist-packages/tensorflow/python/framework/ops.py", line 29, in
from tensorflow.core.framework import attr_value_pb2
File "/usr/local/lib/python3.10/dist-packages/tensorflow/core/framework/attr_value_pb2.py", line 5, in
from google.protobuf.internal import builder as _builder
ImportError: cannot import name 'builder' from 'google.protobuf.internal' (/usr/local/lib/python3.10/dist-packages/google/protobuf/internal/init.py)

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
File "/usr/local/lib/python3.10/dist-packages/diffusers/utils/import_utils.py", line 710, in _get_module
return importlib.import_module("." + module_name, self.name)
File "/usr/lib/python3.10/importlib/init.py", line 126, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "", line 1050, in _gcd_import
File "", line 1027, in _find_and_load
File "", line 1006, in _find_and_load_unlocked
File "", line 688, in _load_unlocked
File "", line 883, in exec_module
File "", line 241, in _call_with_frames_removed
File "/usr/local/lib/python3.10/dist-packages/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion.py", line 20, in
from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection
File "", line 1075, in _handle_fromlist
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1373, in getattr
value = getattr(module, name)
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1372, in getattr
module = self._get_module(self._class_to_module[name])
File "/usr/local/lib/python3.10/dist-packages/transformers/utils/import_utils.py", line 1384, in _get_module
raise RuntimeError(
RuntimeError: Failed to import transformers.models.clip.image_processing_clip because of the following error (look up to see its traceback):
cannot import name 'builder' from 'google.protobuf.internal' (/usr/local/lib/python3.10/dist-packages/google/protobuf/internal/init.py)

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
File "/content/sd-scripts/train_network.py", line 22, in
from library import model_util
File "/content/sd-scripts/library/model_util.py", line 13, in
from diffusers import AutoencoderKL, DDIMScheduler, StableDiffusionPipeline # , UNet2DConditionModel
File "", line 1075, in _handle_fromlist
File "/usr/local/lib/python3.10/dist-packages/diffusers/utils/import_utils.py", line 701, in getattr
value = getattr(module, name)
File "/usr/local/lib/python3.10/dist-packages/diffusers/utils/import_utils.py", line 701, in getattr
value = getattr(module, name)
File "/usr/local/lib/python3.10/dist-packages/diffusers/utils/import_utils.py", line 700, in getattr
module = self._get_module(self._class_to_module[name])
File "/usr/local/lib/python3.10/dist-packages/diffusers/utils/import_utils.py", line 712, in _get_module
raise RuntimeError(
RuntimeError: Failed to import diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion because of the following error (look up to see its traceback):
Failed to import transformers.models.clip.image_processing_clip because of the following error (look up to see its traceback):
cannot import name 'builder' from 'google.protobuf.internal' (/usr/local/lib/python3.10/dist-packages/google/protobuf/internal/init.py)
Traceback (most recent call last):
File "/usr/local/bin/accelerate", line 8, in
sys.exit(main())
File "/usr/local/lib/python3.10/dist-packages/accelerate/commands/accelerate_cli.py", line 47, in main
args.func(args)
File "/usr/local/lib/python3.10/dist-packages/accelerate/commands/launch.py", line 1017, in launch_command
simple_launcher(args)
File "/usr/local/lib/python3.10/dist-packages/accelerate/commands/launch.py", line 637, in simple_launcher
raise subprocess.CalledProcessError(returncode=process.returncode, cmd=cmd)
subprocess.CalledProcessError: Command '['/usr/bin/python3', 'train_network.py', '--config_file=config_file.toml', '--sample_prompts=sample_prompts.txt']' returned non-zero exit status 1.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.