Training in progress, step 1600
Browse files- .gitignore +1 -0
- added_tokens.json +67 -0
- merges.txt +0 -0
- pytorch_model.bin +3 -0
- special_tokens_map.json +73 -0
- tokenizer.json +0 -0
- tokenizer_config.json +10 -0
- training_args.bin +3 -0
- vocab.json +0 -0
.gitignore
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
checkpoint-*/
|
added_tokens.json
ADDED
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"[Add]": 50311,
|
3 |
+
"[And]": 50291,
|
4 |
+
"[Assign]": 50286,
|
5 |
+
"[Attribute]": 50280,
|
6 |
+
"[AugAssign]": 50310,
|
7 |
+
"[BinOp]": 50299,
|
8 |
+
"[BoolOp]": 50290,
|
9 |
+
"[Break]": 50316,
|
10 |
+
"[CLS0]": 50261,
|
11 |
+
"[CLS1]": 50267,
|
12 |
+
"[CLS2]": 50258,
|
13 |
+
"[Call]": 50279,
|
14 |
+
"[ClassDef]": 50271,
|
15 |
+
"[Compare]": 50292,
|
16 |
+
"[Constant]": 50281,
|
17 |
+
"[Eq]": 50296,
|
18 |
+
"[Expr]": 50278,
|
19 |
+
"[For]": 50294,
|
20 |
+
"[FunctionDef]": 50275,
|
21 |
+
"[GtE]": 50297,
|
22 |
+
"[Gt]": 50303,
|
23 |
+
"[If]": 50302,
|
24 |
+
"[ImportFrom]": 50317,
|
25 |
+
"[In]": 50319,
|
26 |
+
"[Index]": 50307,
|
27 |
+
"[IsNot]": 50305,
|
28 |
+
"[Is]": 50301,
|
29 |
+
"[LST]": 50268,
|
30 |
+
"[Lambda]": 50295,
|
31 |
+
"[ListComp]": 50313,
|
32 |
+
"[List]": 50289,
|
33 |
+
"[Load]": 50274,
|
34 |
+
"[LtE]": 50304,
|
35 |
+
"[Lt]": 50293,
|
36 |
+
"[Module]": 50270,
|
37 |
+
"[NEND]": 50269,
|
38 |
+
"[Name]": 50273,
|
39 |
+
"[NoneType]": 50282,
|
40 |
+
"[NotEq]": 50298,
|
41 |
+
"[Not]": 50320,
|
42 |
+
"[Or]": 50312,
|
43 |
+
"[Return]": 50285,
|
44 |
+
"[Slice]": 50315,
|
45 |
+
"[Store]": 50287,
|
46 |
+
"[Sub]": 50300,
|
47 |
+
"[Subscript]": 50306,
|
48 |
+
"[Tuple]": 50321,
|
49 |
+
"[USub]": 50309,
|
50 |
+
"[UnaryOp]": 50308,
|
51 |
+
"[alias]": 50318,
|
52 |
+
"[arg]": 50277,
|
53 |
+
"[arguments]": 50276,
|
54 |
+
"[bool]": 50288,
|
55 |
+
"[comprehension]": 50314,
|
56 |
+
"[int]": 50283,
|
57 |
+
"[keyword]": 50284,
|
58 |
+
"[str]": 50272,
|
59 |
+
"[v0]": 50263,
|
60 |
+
"[v1]": 50262,
|
61 |
+
"[v2]": 50257,
|
62 |
+
"[v3]": 50266,
|
63 |
+
"[v4]": 50265,
|
64 |
+
"[v5]": 50259,
|
65 |
+
"[v6]": 50260,
|
66 |
+
"[v7]": 50264
|
67 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c52b3ca9df91ee2005299bc02340471e6f4ed7f69975e7c1bf37c613e6209bb8
|
3 |
+
size 334176057
|
special_tokens_map.json
ADDED
@@ -0,0 +1,73 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"[v2]",
|
4 |
+
"[CLS2]",
|
5 |
+
"[v5]",
|
6 |
+
"[v6]",
|
7 |
+
"[CLS0]",
|
8 |
+
"[v1]",
|
9 |
+
"[v0]",
|
10 |
+
"[v7]",
|
11 |
+
"[v4]",
|
12 |
+
"[v3]",
|
13 |
+
"[CLS1]",
|
14 |
+
"[LST]",
|
15 |
+
"[NEND]",
|
16 |
+
"[Module]",
|
17 |
+
"[ClassDef]",
|
18 |
+
"[str]",
|
19 |
+
"[Name]",
|
20 |
+
"[Load]",
|
21 |
+
"[FunctionDef]",
|
22 |
+
"[arguments]",
|
23 |
+
"[arg]",
|
24 |
+
"[Expr]",
|
25 |
+
"[Call]",
|
26 |
+
"[Attribute]",
|
27 |
+
"[Constant]",
|
28 |
+
"[NoneType]",
|
29 |
+
"[int]",
|
30 |
+
"[keyword]",
|
31 |
+
"[Return]",
|
32 |
+
"[Assign]",
|
33 |
+
"[Store]",
|
34 |
+
"[bool]",
|
35 |
+
"[List]",
|
36 |
+
"[BoolOp]",
|
37 |
+
"[And]",
|
38 |
+
"[Compare]",
|
39 |
+
"[Lt]",
|
40 |
+
"[For]",
|
41 |
+
"[Lambda]",
|
42 |
+
"[Eq]",
|
43 |
+
"[GtE]",
|
44 |
+
"[NotEq]",
|
45 |
+
"[BinOp]",
|
46 |
+
"[Sub]",
|
47 |
+
"[Is]",
|
48 |
+
"[If]",
|
49 |
+
"[Gt]",
|
50 |
+
"[LtE]",
|
51 |
+
"[IsNot]",
|
52 |
+
"[Subscript]",
|
53 |
+
"[Index]",
|
54 |
+
"[UnaryOp]",
|
55 |
+
"[USub]",
|
56 |
+
"[AugAssign]",
|
57 |
+
"[Add]",
|
58 |
+
"[Or]",
|
59 |
+
"[ListComp]",
|
60 |
+
"[comprehension]",
|
61 |
+
"[Slice]",
|
62 |
+
"[Break]",
|
63 |
+
"[ImportFrom]",
|
64 |
+
"[alias]",
|
65 |
+
"[In]",
|
66 |
+
"[Not]",
|
67 |
+
"[Tuple]"
|
68 |
+
],
|
69 |
+
"bos_token": "<|endoftext|>",
|
70 |
+
"eos_token": "<|endoftext|>",
|
71 |
+
"pad_token": "<|endoftext|>",
|
72 |
+
"unk_token": "<|endoftext|>"
|
73 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"bos_token": "<|endoftext|>",
|
4 |
+
"eos_token": "<|endoftext|>",
|
5 |
+
"model_max_length": 1024,
|
6 |
+
"name_or_path": "distilgpt2",
|
7 |
+
"special_tokens_map_file": null,
|
8 |
+
"tokenizer_class": "GPT2Tokenizer",
|
9 |
+
"unk_token": "<|endoftext|>"
|
10 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9f6569c0370c367e160d794b569f3babcb0081ebedb312726c05295765ce7644
|
3 |
+
size 3387
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|