dvitel commited on
Commit
6ce4a45
1 Parent(s): b7d70d1

Training in progress, step 1600

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
added_tokens.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "[Add]": 50311,
3
+ "[And]": 50291,
4
+ "[Assign]": 50286,
5
+ "[Attribute]": 50280,
6
+ "[AugAssign]": 50310,
7
+ "[BinOp]": 50299,
8
+ "[BoolOp]": 50290,
9
+ "[Break]": 50316,
10
+ "[CLS0]": 50262,
11
+ "[CLS1]": 50261,
12
+ "[CLS2]": 50260,
13
+ "[Call]": 50279,
14
+ "[ClassDef]": 50271,
15
+ "[Compare]": 50292,
16
+ "[Constant]": 50281,
17
+ "[Eq]": 50296,
18
+ "[Expr]": 50278,
19
+ "[For]": 50294,
20
+ "[FunctionDef]": 50275,
21
+ "[GtE]": 50297,
22
+ "[Gt]": 50303,
23
+ "[If]": 50302,
24
+ "[ImportFrom]": 50317,
25
+ "[In]": 50319,
26
+ "[Index]": 50307,
27
+ "[IsNot]": 50305,
28
+ "[Is]": 50301,
29
+ "[LST]": 50268,
30
+ "[Lambda]": 50295,
31
+ "[ListComp]": 50313,
32
+ "[List]": 50289,
33
+ "[Load]": 50274,
34
+ "[LtE]": 50304,
35
+ "[Lt]": 50293,
36
+ "[Module]": 50270,
37
+ "[NEND]": 50269,
38
+ "[Name]": 50273,
39
+ "[NoneType]": 50282,
40
+ "[NotEq]": 50298,
41
+ "[Not]": 50320,
42
+ "[Or]": 50312,
43
+ "[Return]": 50285,
44
+ "[Slice]": 50315,
45
+ "[Store]": 50287,
46
+ "[Sub]": 50300,
47
+ "[Subscript]": 50306,
48
+ "[Tuple]": 50321,
49
+ "[USub]": 50309,
50
+ "[UnaryOp]": 50308,
51
+ "[alias]": 50318,
52
+ "[arg]": 50277,
53
+ "[arguments]": 50276,
54
+ "[bool]": 50288,
55
+ "[comprehension]": 50314,
56
+ "[int]": 50283,
57
+ "[keyword]": 50284,
58
+ "[str]": 50272,
59
+ "[v0]": 50257,
60
+ "[v1]": 50265,
61
+ "[v2]": 50263,
62
+ "[v3]": 50259,
63
+ "[v4]": 50258,
64
+ "[v5]": 50264,
65
+ "[v6]": 50267,
66
+ "[v7]": 50266
67
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d18f67cc8e9089e9c2279e42b1d666bbdb21848ed103cd4bf2b5ec59b7754af8
3
+ size 334176057
special_tokens_map.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "[v0]",
4
+ "[v4]",
5
+ "[v3]",
6
+ "[CLS2]",
7
+ "[CLS1]",
8
+ "[CLS0]",
9
+ "[v2]",
10
+ "[v5]",
11
+ "[v1]",
12
+ "[v7]",
13
+ "[v6]",
14
+ "[LST]",
15
+ "[NEND]",
16
+ "[Module]",
17
+ "[ClassDef]",
18
+ "[str]",
19
+ "[Name]",
20
+ "[Load]",
21
+ "[FunctionDef]",
22
+ "[arguments]",
23
+ "[arg]",
24
+ "[Expr]",
25
+ "[Call]",
26
+ "[Attribute]",
27
+ "[Constant]",
28
+ "[NoneType]",
29
+ "[int]",
30
+ "[keyword]",
31
+ "[Return]",
32
+ "[Assign]",
33
+ "[Store]",
34
+ "[bool]",
35
+ "[List]",
36
+ "[BoolOp]",
37
+ "[And]",
38
+ "[Compare]",
39
+ "[Lt]",
40
+ "[For]",
41
+ "[Lambda]",
42
+ "[Eq]",
43
+ "[GtE]",
44
+ "[NotEq]",
45
+ "[BinOp]",
46
+ "[Sub]",
47
+ "[Is]",
48
+ "[If]",
49
+ "[Gt]",
50
+ "[LtE]",
51
+ "[IsNot]",
52
+ "[Subscript]",
53
+ "[Index]",
54
+ "[UnaryOp]",
55
+ "[USub]",
56
+ "[AugAssign]",
57
+ "[Add]",
58
+ "[Or]",
59
+ "[ListComp]",
60
+ "[comprehension]",
61
+ "[Slice]",
62
+ "[Break]",
63
+ "[ImportFrom]",
64
+ "[alias]",
65
+ "[In]",
66
+ "[Not]",
67
+ "[Tuple]"
68
+ ],
69
+ "bos_token": "<|endoftext|>",
70
+ "eos_token": "<|endoftext|>",
71
+ "pad_token": "<|endoftext|>",
72
+ "unk_token": "<|endoftext|>"
73
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<|endoftext|>",
4
+ "eos_token": "<|endoftext|>",
5
+ "model_max_length": 1024,
6
+ "name_or_path": "distilgpt2",
7
+ "special_tokens_map_file": null,
8
+ "tokenizer_class": "GPT2Tokenizer",
9
+ "unk_token": "<|endoftext|>"
10
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3548deabb41f9d347d2883ea2f138a03a7e95c5131447d9739b3d6eea946d32
3
+ size 3387
vocab.json ADDED
The diff for this file is too large to render. See raw diff