zhangduo4610 commited on
Commit
7c5e6dc
1 Parent(s): bd5c9b8

Model save

Browse files
README.md CHANGED
@@ -1,57 +1,64 @@
1
  ---
2
  base_model: facebook/opt-350m
3
- library_name: transformers
4
- model_name: opt350
 
 
5
  tags:
6
- - generated_from_trainer
7
  - trl
8
  - sft
9
- licence: license
 
 
 
10
  ---
11
 
12
- # Model Card for opt350
 
13
 
14
- This model is a fine-tuned version of [facebook/opt-350m](https://huggingface.co/facebook/opt-350m).
15
- It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
- ## Quick start
18
 
19
- ```python
20
- from transformers import pipeline
21
 
22
- question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
- generator = pipeline("text-generation", model="zhangduo4610/opt350", device="cuda")
24
- output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
- print(output["generated_text"])
26
- ```
27
 
28
- ## Training procedure
29
 
 
30
 
 
31
 
32
- This model was trained with SFT.
33
 
34
- ### Framework versions
 
 
35
 
36
- - TRL: 0.12.0
37
- - Transformers: 4.46.1
38
- - Pytorch: 2.1.2
39
- - Datasets: 3.1.0
40
- - Tokenizers: 0.20.2
 
 
 
 
 
 
 
 
 
41
 
42
- ## Citations
43
 
44
 
45
 
46
- Cite TRL as:
47
-
48
- ```bibtex
49
- @misc{vonwerra2022trl,
50
- title = {{TRL: Transformer Reinforcement Learning}},
51
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
52
- year = 2020,
53
- journal = {GitHub repository},
54
- publisher = {GitHub},
55
- howpublished = {\url{https://github.com/huggingface/trl}}
56
- }
57
- ```
 
1
  ---
2
  base_model: facebook/opt-350m
3
+ datasets:
4
+ - generator
5
+ library_name: peft
6
+ license: other
7
  tags:
 
8
  - trl
9
  - sft
10
+ - generated_from_trainer
11
+ model-index:
12
+ - name: opt350
13
+ results: []
14
  ---
15
 
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
 
19
+ # opt350
 
20
 
21
+ This model is a fine-tuned version of [facebook/opt-350m](https://huggingface.co/facebook/opt-350m) on the generator dataset.
22
 
23
+ ## Model description
 
24
 
25
+ More information needed
 
 
 
 
26
 
27
+ ## Intended uses & limitations
28
 
29
+ More information needed
30
 
31
+ ## Training and evaluation data
32
 
33
+ More information needed
34
 
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
 
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 0.0002
41
+ - train_batch_size: 4
42
+ - eval_batch_size: 8
43
+ - seed: 42
44
+ - distributed_type: multi-GPU
45
+ - num_devices: 2
46
+ - gradient_accumulation_steps: 2
47
+ - total_train_batch_size: 16
48
+ - total_eval_batch_size: 16
49
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
+ - lr_scheduler_type: cosine
51
+ - lr_scheduler_warmup_ratio: 0.1
52
+ - num_epochs: 1
53
 
54
+ ### Training results
55
 
56
 
57
 
58
+ ### Framework versions
59
+
60
+ - PEFT 0.13.2
61
+ - Transformers 4.45.0
62
+ - Pytorch 2.1.2
63
+ - Datasets 3.1.0
64
+ - Tokenizers 0.20.2
 
 
 
 
 
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
- "q_proj",
25
  "k_proj",
26
- "gate_proj",
 
27
  "o_proj",
28
- "down_proj",
29
- "up_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "k_proj",
24
+ "v_proj",
25
+ "up_proj",
26
  "o_proj",
27
+ "gate_proj",
28
+ "q_proj",
29
+ "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
all_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 0.9999380306128772,
3
  "total_flos": 4248659495485440.0,
4
- "train_loss": 0.39469828721492317,
5
- "train_runtime": 1041.1882,
6
  "train_samples": 207864,
7
- "train_samples_per_second": 123.988,
8
- "train_steps_per_second": 7.749
9
  }
 
1
  {
2
  "epoch": 0.9999380306128772,
3
  "total_flos": 4248659495485440.0,
4
+ "train_loss": 0.0,
5
+ "train_runtime": 0.0126,
6
  "train_samples": 207864,
7
+ "train_samples_per_second": 10236188.723,
8
+ "train_steps_per_second": 639727.105
9
  }
runs/Nov05_10-32-53_gnode007.cluster/events.out.tfevents.1730832694.gnode007.cluster.1851533.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4e10a62d6bf762c2dc341c44ec2155e7d91b27293b2fc50ab064f4039655a64
3
+ size 5968
runs/Nov05_10-54-00_gnode007.cluster/events.out.tfevents.1730832849.gnode007.cluster.1852419.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3a39a1bf8021df3f3214325d182c14dde90e0867dcca30d2cef787bbb319172
3
+ size 5968
tokenizer_config.json CHANGED
@@ -21,7 +21,7 @@
21
  },
22
  "bos_token": "</s>",
23
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
24
- "clean_up_tokenization_spaces": false,
25
  "eos_token": "</s>",
26
  "errors": "replace",
27
  "model_max_length": 2048,
 
21
  },
22
  "bos_token": "</s>",
23
  "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
24
+ "clean_up_tokenization_spaces": true,
25
  "eos_token": "</s>",
26
  "errors": "replace",
27
  "model_max_length": 2048,
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 0.9999380306128772,
3
  "total_flos": 4248659495485440.0,
4
- "train_loss": 0.39469828721492317,
5
- "train_runtime": 1041.1882,
6
  "train_samples": 207864,
7
- "train_samples_per_second": 123.988,
8
- "train_steps_per_second": 7.749
9
  }
 
1
  {
2
  "epoch": 0.9999380306128772,
3
  "total_flos": 4248659495485440.0,
4
+ "train_loss": 0.0,
5
+ "train_runtime": 0.0126,
6
  "train_samples": 207864,
7
+ "train_samples_per_second": 10236188.723,
8
+ "train_steps_per_second": 639727.105
9
  }
trainer_state.json CHANGED
@@ -11306,22 +11306,14 @@
11306
  "loss": 3.5602,
11307
  "step": 8065
11308
  },
11309
- {
11310
- "epoch": 0.9999380306128772,
11311
- "eval_loss": 1.7876207828521729,
11312
- "eval_runtime": 155.4034,
11313
- "eval_samples_per_second": 91.883,
11314
- "eval_steps_per_second": 5.746,
11315
- "step": 8068
11316
- },
11317
  {
11318
  "epoch": 0.9999380306128772,
11319
  "step": 8068,
11320
  "total_flos": 4248659495485440.0,
11321
- "train_loss": 0.39469828721492317,
11322
- "train_runtime": 1041.1882,
11323
- "train_samples_per_second": 123.988,
11324
- "train_steps_per_second": 7.749
11325
  }
11326
  ],
11327
  "logging_steps": 5,
 
11306
  "loss": 3.5602,
11307
  "step": 8065
11308
  },
 
 
 
 
 
 
 
 
11309
  {
11310
  "epoch": 0.9999380306128772,
11311
  "step": 8068,
11312
  "total_flos": 4248659495485440.0,
11313
+ "train_loss": 0.0,
11314
+ "train_runtime": 0.0126,
11315
+ "train_samples_per_second": 10236188.723,
11316
+ "train_steps_per_second": 639727.105
11317
  }
11318
  ],
11319
  "logging_steps": 5,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b40232365cdac60e63e903629fe55e11c030e9274cbe1a3765db331b581f516c
3
- size 7096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:258e43470debd52839f53189bb0668179eb156a1bcda2b08fa180bb781f216cc
3
+ size 7032