superbigtree commited on
Commit
57f8d44
1 Parent(s): 967d807

Model save

Browse files
README.md CHANGED
@@ -2,7 +2,7 @@
2
  tags:
3
  - generated_from_trainer
4
  datasets:
5
- - ydshieh/coco_dataset_script
6
  model-index:
7
  - name: clip-roberta-finetuned
8
  results: []
@@ -13,9 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # clip-roberta-finetuned
15
 
16
- This model was trained from scratch on the ydshieh/coco_dataset_script 2017 dataset.
17
- It achieves the following results on the evaluation set:
18
- - Loss: 1.5825
19
 
20
  ## Model description
21
 
@@ -39,9 +37,9 @@ The following hyperparameters were used during training:
39
  - eval_batch_size: 64
40
  - seed: 42
41
  - distributed_type: multi-GPU
42
- - num_devices: 4
43
- - total_train_batch_size: 256
44
- - total_eval_batch_size: 256
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - num_epochs: 3.0
@@ -54,5 +52,5 @@ The following hyperparameters were used during training:
54
 
55
  - Transformers 4.38.0.dev0
56
  - Pytorch 2.0.1+gita61a294
57
- - Datasets 2.16.1
58
- - Tokenizers 0.15.1
 
2
  tags:
3
  - generated_from_trainer
4
  datasets:
5
+ - coco_dataset_script
6
  model-index:
7
  - name: clip-roberta-finetuned
8
  results: []
 
13
 
14
  # clip-roberta-finetuned
15
 
16
+ This model was trained from scratch on the coco_dataset_script dataset.
 
 
17
 
18
  ## Model description
19
 
 
37
  - eval_batch_size: 64
38
  - seed: 42
39
  - distributed_type: multi-GPU
40
+ - num_devices: 8
41
+ - total_train_batch_size: 512
42
+ - total_eval_batch_size: 512
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
  - num_epochs: 3.0
 
52
 
53
  - Transformers 4.38.0.dev0
54
  - Pytorch 2.0.1+gita61a294
55
+ - Datasets 2.17.0
56
+ - Tokenizers 0.15.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed169b2b85e2fc51c20d896f702cd424619ff37dac9c14cbcd0bbc46151b1a60
3
  size 851603588
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46a42031095858a91a0025b5efa14043b3ef5dff9416e941f182af1634102de3
3
  size 851603588
runs/Feb12_18-25-46_28fc6ffbaa77/events.out.tfevents.1707762445.28fc6ffbaa77.4275.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63cb880de7313979f59d051a4398c2034d49b8d1a48dd0be66848ba1d218f041
3
- size 9285
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84891b0a8ee1cc5bf03c7421cd1c30ab452094d87d358e66b728e3984a608c36
3
+ size 9639
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 128,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 128
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 1,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<pad>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,