jburtoft commited on
Commit
5730505
1 Parent(s): dd4e4ec

a2ee0201bf7701e699196a6a787b08a7a1b9d890dec9e0299541689f0b740139

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ text_encoder/model.neuron filter=lfs diff=lfs merge=lfs -text
model_index.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "NeuronStableDiffusionXLPipeline",
3
+ "_diffusers_version": "0.23.0",
4
+ "force_zeros_for_empty_prompt": true,
5
+ "scheduler": [
6
+ "diffusers",
7
+ "EulerDiscreteScheduler"
8
+ ],
9
+ "text_encoder": [
10
+ "optimum",
11
+ "NeuronModelTextEncoder"
12
+ ],
13
+ "text_encoder_2": [
14
+ "optimum",
15
+ "NeuronModelTextEncoder"
16
+ ],
17
+ "tokenizer": [
18
+ "transformers",
19
+ "CLIPTokenizer"
20
+ ],
21
+ "tokenizer_2": [
22
+ "transformers",
23
+ "CLIPTokenizer"
24
+ ],
25
+ "unet": [
26
+ "optimum",
27
+ "NeuronModelUnet"
28
+ ],
29
+ "vae_decoder": [
30
+ "optimum",
31
+ "NeuronModelVaeDecoder"
32
+ ],
33
+ "vae_encoder": [
34
+ "optimum",
35
+ "NeuronModelVaeEncoder"
36
+ ]
37
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "EulerDiscreteScheduler",
3
+ "_diffusers_version": "0.23.0",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "interpolation_type": "linear",
9
+ "num_train_timesteps": 1000,
10
+ "prediction_type": "epsilon",
11
+ "sample_max_value": 1.0,
12
+ "set_alpha_to_one": false,
13
+ "skip_prk_steps": true,
14
+ "steps_offset": 1,
15
+ "timestep_spacing": "leading",
16
+ "trained_betas": null,
17
+ "use_karras_sigmas": false
18
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/home/ubuntu/.cache/huggingface/hub/models--ThinkDiffusion--ThinkDiffusionXL/snapshots/e31c4693b7f87128fee02fa6b16c4dc2ada13721/text_encoder",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "quick_gelu",
11
+ "hidden_size": 768,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 3072,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "neuron": {
19
+ "auto_cast": "all",
20
+ "auto_cast_type": "bf16",
21
+ "compiler_type": "neuronx-cc",
22
+ "compiler_version": "2.12.54.0+f631c2365",
23
+ "disable_fallback": false,
24
+ "disable_fast_relayout": false,
25
+ "dynamic_batch_size": false,
26
+ "input_names": [
27
+ "input_ids"
28
+ ],
29
+ "model_type": "clip-text-model",
30
+ "optlevel": "2",
31
+ "output_attentions": false,
32
+ "output_hidden_states": false,
33
+ "output_names": [
34
+ "last_hidden_state",
35
+ "pooler_output",
36
+ "hidden_states"
37
+ ],
38
+ "static_batch_size": 1,
39
+ "static_num_beams": 1,
40
+ "static_sequence_length": 77
41
+ },
42
+ "num_attention_heads": 12,
43
+ "num_hidden_layers": 12,
44
+ "output_hidden_states": true,
45
+ "pad_token_id": 1,
46
+ "projection_dim": 768,
47
+ "task": "feature-extraction",
48
+ "torch_dtype": "float16",
49
+ "torchscript": true,
50
+ "transformers_version": "4.36.2",
51
+ "vocab_size": 49408
52
+ }
text_encoder/model.neuron ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35f676fb689abc2fe7068c0617655ab6c5dcbf3c8929ea80db6fc6fc75398d27
3
+ size 375884546
text_encoder_2/config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/home/ubuntu/.cache/huggingface/hub/models--ThinkDiffusion--ThinkDiffusionXL/snapshots/e31c4693b7f87128fee02fa6b16c4dc2ada13721/text_encoder_2",
3
+ "architectures": [
4
+ "CLIPTextModelWithProjection"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_size": 1280,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 5120,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "neuron": {
19
+ "auto_cast": "all",
20
+ "auto_cast_type": "bf16",
21
+ "compiler_type": "neuronx-cc",
22
+ "compiler_version": "2.12.54.0+f631c2365",
23
+ "disable_fallback": false,
24
+ "disable_fast_relayout": false,
25
+ "dynamic_batch_size": false,
26
+ "input_names": [
27
+ "input_ids"
28
+ ],
29
+ "model_type": "clip-text-model",
30
+ "optlevel": "2",
31
+ "output_attentions": false,
32
+ "output_hidden_states": false,
33
+ "output_names": [
34
+ "text_embeds",
35
+ "last_hidden_state",
36
+ "hidden_states"
37
+ ],
38
+ "static_batch_size": 1,
39
+ "static_num_beams": 1,
40
+ "static_sequence_length": 77
41
+ },
42
+ "num_attention_heads": 20,
43
+ "num_hidden_layers": 32,
44
+ "output_hidden_states": true,
45
+ "pad_token_id": 1,
46
+ "projection_dim": 1280,
47
+ "task": "feature-extraction",
48
+ "torch_dtype": "float16",
49
+ "torchscript": true,
50
+ "transformers_version": "4.36.2",
51
+ "vocab_size": 49408
52
+ }