alfredplpl's picture
Update README.md
aec65d2 verified
---
license: other
license_name: stabilityai-ai-community
license_link: >-
https://huggingface.co/stabilityai/stable-diffusion-3.5-large/blob/main/LICENSE.md
language:
- en
base_model:
- stabilityai/stable-diffusion-3.5-large
---
![eyecatch](eyecatch.jpg)
# SD 3.5 Large Modern Anime Full Model Card
This is an experimental model. I full-finetune SD 3.5 Large by Quality Tuning only.
# Usage
- ComfyUI
1. Download the [model](sd3_5_large_modern_anime_full.safetensors). [fp8](sd3_5_large_modern_anime_full_fp8.safetensors)
2. Enjoy! (trigger word: modern anime style, )
- diffusers
1. Run the code:
```python
import torch
from diffusers import SD3Transformer2DModel, StableDiffusion3Pipeline
transformer = SD3Transformer2DModel.from_single_file(
"https://huggingface.co/alfredplpl/sd3-5-large-modern-anime-full/blob/main/sd3_5_large_modern_anime_full.safetensors",
torch_dtype=torch.bfloat16,
)
pipe = StableDiffusion3Pipeline.from_pretrained(
"stabilityai/stable-diffusion-3.5-large",
transformer=transformer,
torch_dtype=torch.bfloat16,
)
pipe.enable_model_cpu_offload()
image = pipe("modern anime style, A close-up shot of a girl's face in the center, looking directly at the viewer. Autumn maple trees with red leaves frame both the left and right sides of the background, with the sky visible in the middle.").images[0]
image.save("sd35.png")
```
# How to Make
## Prerequisites
- A6000x1 (48GB)
- Private dataset: 3000 images (collected manually)
## Procedure
I used sd-scripts. The dataset config as is follows:
```toml
[general]
enable_bucket = true # Aspect Ratio Bucketingを使うか否か
[[datasets]]
resolution = 1024 # 学習解像度
batch_size = 4 # バッチサイズ
[[datasets.subsets]]
image_dir = '/mnt/NVM2/manual_now' # 学習用画像を入れたフォルダを指定
metadata_file = 'manual_dcap2.json' # メタデータファイル名
```
I ran the command:
```bash
accelerate launch --num_cpu_threads_per_process 1 sd3_train.py --pretrained_model_name_or_path='/mnt/NVM2/sd3_5/sd3.5_large.safetensors' --output_dir='/mnt/NVM2/sd3_5' --output_name=modern_anime --dataset_config=manual_dcap2.toml --save_model_as=safetensors --learning_rate=5e-6 --sdpa --gradient_checkpointing --mixed_precision=bf16 --full_bf16 --max_train_epochs=10 --min_bucket_reso=512 --max_bucket_reso=2048 --clip_l='/mnt/NVM2/sd3_5/clip_l.safetensors' --clip_g='/mnt/NVM2/sd3_5/clip_g.safetensors' --gradient_accumulation_steps=1 --t5xxl='/mnt/NVM2/sd3_5/t5xxl_fp16.safetensors' --optimizer_type=Lion8bit --cache_text_encoder_outputs_to_disk --cache_text_encoder_outputs --cache_latents --cache_latents_to_disk --save_every_n_epochs=1
```