Upload 4 files
Browse files- README.md +16 -0
- config.json +58 -0
- model_index.json +38 -0
- pytorch_lora_weights.safetensors +3 -0
README.md
ADDED
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- trl
|
5 |
+
- ddpo
|
6 |
+
- diffusers
|
7 |
+
- reinforcement-learning
|
8 |
+
- text-to-image
|
9 |
+
- stable-diffusion
|
10 |
+
---
|
11 |
+
|
12 |
+
# TRL DDPO Model
|
13 |
+
|
14 |
+
This is a diffusion model that has been fine-tuned with reinforcement learning to
|
15 |
+
guide the model outputs according to a value, function, or human feedback. The model can be used for image generation conditioned with text.
|
16 |
+
|
config.json
ADDED
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accelerator_kwargs": {},
|
3 |
+
"allow_tf32": true,
|
4 |
+
"artistic_log_on": true,
|
5 |
+
"async_reward_computation": false,
|
6 |
+
"exp_name": "o2o",
|
7 |
+
"global_step": 0,
|
8 |
+
"high_reward": 100.0,
|
9 |
+
"huggingface_note": "RePic250_2on",
|
10 |
+
"log_with": "wandb",
|
11 |
+
"logdir": "ll",
|
12 |
+
"low_reward": 50.0,
|
13 |
+
"max_workers": 2,
|
14 |
+
"mixed_precision": "fp16",
|
15 |
+
"negative_prompts": "",
|
16 |
+
"num_checkpoint_limit": 10,
|
17 |
+
"num_epochs": 250,
|
18 |
+
"offpolicy_sample_batch_size": 1,
|
19 |
+
"pass_images": 0,
|
20 |
+
"per_prompt_stat_tracking": true,
|
21 |
+
"per_prompt_stat_tracking_buffer_size": 32,
|
22 |
+
"per_prompt_stat_tracking_min_count": 16,
|
23 |
+
"project_kwargs": {
|
24 |
+
"automatic_checkpoint_naming": true,
|
25 |
+
"logging_dir": "./outputs/logs",
|
26 |
+
"project_dir": "./outputs/",
|
27 |
+
"total_limit": 5
|
28 |
+
},
|
29 |
+
"resolution": 768,
|
30 |
+
"resume_from": "",
|
31 |
+
"reward_function_usage": false,
|
32 |
+
"run_name": "",
|
33 |
+
"sample_batch_size": 3,
|
34 |
+
"sample_eta": 1.0,
|
35 |
+
"sample_guidance_scale": 5.0,
|
36 |
+
"sample_num_batches_per_epoch": 1,
|
37 |
+
"sample_num_steps": 50,
|
38 |
+
"save_freq": 1000,
|
39 |
+
"seed": 1,
|
40 |
+
"tracker_kwargs": {},
|
41 |
+
"tracker_project_name": "Dev",
|
42 |
+
"train_adam_beta1": 0.9,
|
43 |
+
"train_adam_beta2": 0.999,
|
44 |
+
"train_adam_epsilon": 1e-08,
|
45 |
+
"train_adam_weight_decay": 0.0001,
|
46 |
+
"train_adv_clip_max": 5,
|
47 |
+
"train_batch_size": 3,
|
48 |
+
"train_cfg": true,
|
49 |
+
"train_clip_range": 0.0001,
|
50 |
+
"train_gradient_accumulation_steps": 1,
|
51 |
+
"train_learning_rate": 0.0003,
|
52 |
+
"train_max_grad_norm": 1.0,
|
53 |
+
"train_num_inner_epochs": 1,
|
54 |
+
"train_timestep_fraction": 1.0,
|
55 |
+
"train_use_8bit_adam": false,
|
56 |
+
"valid_batch_size": 1,
|
57 |
+
"valid_size": 10
|
58 |
+
}
|
model_index.json
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "StableDiffusionPipeline",
|
3 |
+
"_diffusers_version": "0.32.2",
|
4 |
+
"_name_or_path": "stabilityai/stable-diffusion-2-1",
|
5 |
+
"feature_extractor": [
|
6 |
+
"transformers",
|
7 |
+
"CLIPImageProcessor"
|
8 |
+
],
|
9 |
+
"image_encoder": [
|
10 |
+
null,
|
11 |
+
null
|
12 |
+
],
|
13 |
+
"requires_safety_checker": false,
|
14 |
+
"safety_checker": [
|
15 |
+
null,
|
16 |
+
null
|
17 |
+
],
|
18 |
+
"scheduler": [
|
19 |
+
"diffusers",
|
20 |
+
"DDIMScheduler"
|
21 |
+
],
|
22 |
+
"text_encoder": [
|
23 |
+
"transformers",
|
24 |
+
"CLIPTextModel"
|
25 |
+
],
|
26 |
+
"tokenizer": [
|
27 |
+
"transformers",
|
28 |
+
"CLIPTokenizer"
|
29 |
+
],
|
30 |
+
"unet": [
|
31 |
+
"diffusers",
|
32 |
+
"UNet2DConditionModel"
|
33 |
+
],
|
34 |
+
"vae": [
|
35 |
+
"diffusers",
|
36 |
+
"AutoencoderKL"
|
37 |
+
]
|
38 |
+
}
|
pytorch_lora_weights.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9f25a035e4921c5a044a34caffe4f48276476fda6079c7c235ff71f7262b2e6b
|
3 |
+
size 3357296
|