NeverLearn commited on
Commit
7031cdc
·
verified ·
1 Parent(s): 57c6722

Training in progress, step 500

Browse files
Files changed (3) hide show
  1. config.json +3 -16
  2. model.safetensors +2 -2
  3. training_args.bin +1 -1
config.json CHANGED
@@ -1,34 +1,22 @@
1
  {
2
- "_name_or_path": "NeverLearn/gpt2-wikitext2",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
- "BertForMaskedLM"
6
  ],
7
- "attention_probs_dropout_prob": 0.1,
8
  "attn_pdrop": 0.1,
9
  "bos_token_id": 50256,
10
- "classifier_dropout": null,
11
  "embd_pdrop": 0.1,
12
  "eos_token_id": 50256,
13
- "hidden_act": "gelu",
14
- "hidden_dropout_prob": 0.1,
15
- "hidden_size": 768,
16
  "initializer_range": 0.02,
17
- "intermediate_size": 3072,
18
- "layer_norm_eps": 1e-12,
19
  "layer_norm_epsilon": 1e-05,
20
- "max_position_embeddings": 512,
21
- "model_type": "bert",
22
  "n_ctx": 1024,
23
  "n_embd": 768,
24
  "n_head": 12,
25
  "n_inner": null,
26
  "n_layer": 12,
27
  "n_positions": 1024,
28
- "num_attention_heads": 12,
29
- "num_hidden_layers": 12,
30
- "pad_token_id": 0,
31
- "position_embedding_type": "absolute",
32
  "reorder_and_upcast_attn": false,
33
  "resid_pdrop": 0.1,
34
  "scale_attn_by_inverse_layer_idx": false,
@@ -46,7 +34,6 @@
46
  },
47
  "torch_dtype": "float32",
48
  "transformers_version": "4.40.0",
49
- "type_vocab_size": 2,
50
  "use_cache": true,
51
  "vocab_size": 50257
52
  }
 
1
  {
2
+ "_name_or_path": "openai-community/gpt2",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
+ "GPT2LMHeadModel"
6
  ],
 
7
  "attn_pdrop": 0.1,
8
  "bos_token_id": 50256,
 
9
  "embd_pdrop": 0.1,
10
  "eos_token_id": 50256,
 
 
 
11
  "initializer_range": 0.02,
 
 
12
  "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
 
14
  "n_ctx": 1024,
15
  "n_embd": 768,
16
  "n_head": 12,
17
  "n_inner": null,
18
  "n_layer": 12,
19
  "n_positions": 1024,
 
 
 
 
20
  "reorder_and_upcast_attn": false,
21
  "resid_pdrop": 0.1,
22
  "scale_attn_by_inverse_layer_idx": false,
 
34
  },
35
  "torch_dtype": "float32",
36
  "transformers_version": "4.40.0",
 
37
  "use_cache": true,
38
  "vocab_size": 50257
39
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c829845c9837772684bddeca7d32d8b0a4172cb0b8382c4ce9a8cedfd971583f
3
- size 498785764
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:781696faa4767e8e3c880daef349ff82f4817eb73b77f1f6c2cbe6d9c6af75ae
3
+ size 497774208
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8016abe50aa5f911d6aa79bfce87a115b93cf85fe553553bdf84e6aa86490df
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1084752d82de5d3be258a37d82b895e4a218fc40d4156f7b87e8cd7c2794ddc
3
  size 4984