Upload folder using huggingface_hub
Browse files- checkpoints/checkpoint-51000/config.json +29 -0
- checkpoints/checkpoint-51000/model.safetensors +3 -0
- checkpoints/checkpoint-51000/optimizer.pt +3 -0
- checkpoints/checkpoint-51000/rng_state.pth +3 -0
- checkpoints/checkpoint-51000/scheduler.pt +3 -0
- checkpoints/checkpoint-51000/trainer_state.json +0 -0
- checkpoints/checkpoint-51000/training_args.bin +3 -0
- checkpoints/checkpoint-52000/config.json +29 -0
- checkpoints/checkpoint-52000/model.safetensors +3 -0
- checkpoints/checkpoint-52000/optimizer.pt +3 -0
- checkpoints/checkpoint-52000/rng_state.pth +3 -0
- checkpoints/checkpoint-52000/scheduler.pt +3 -0
- checkpoints/checkpoint-52000/trainer_state.json +0 -0
- checkpoints/checkpoint-52000/training_args.bin +3 -0
- checkpoints/checkpoint-53000/config.json +29 -0
- checkpoints/checkpoint-53000/model.safetensors +3 -0
- checkpoints/checkpoint-53000/optimizer.pt +3 -0
- checkpoints/checkpoint-53000/rng_state.pth +3 -0
- checkpoints/checkpoint-53000/scheduler.pt +3 -0
- checkpoints/checkpoint-53000/trainer_state.json +0 -0
- checkpoints/checkpoint-53000/training_args.bin +3 -0
- checkpoints/checkpoint-54000/config.json +29 -0
- checkpoints/checkpoint-54000/model.safetensors +3 -0
- checkpoints/checkpoint-54000/optimizer.pt +3 -0
- checkpoints/checkpoint-54000/rng_state.pth +3 -0
- checkpoints/checkpoint-54000/scheduler.pt +3 -0
- checkpoints/checkpoint-54000/trainer_state.json +0 -0
- checkpoints/checkpoint-54000/training_args.bin +3 -0
- checkpoints/config.json +1 -1
- checkpoints/model.safetensors +1 -1
- checkpoints/training_args.bin +1 -1
checkpoints/checkpoint-51000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-51000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:06e658cc46bdd8a72cb52f403eef5d565dae45616f78bd3c57533755f9cb6041
|
3 |
+
size 44890256
|
checkpoints/checkpoint-51000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46dde2ed75d759c8e38c1336599c03c653f1b57bca5188d3ff869e9c9bef71fc
|
3 |
+
size 89797322
|
checkpoints/checkpoint-51000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d80778c8232421fd236e39309a51aaaa60f2e382c3a48206f6dbc992f76c370b
|
3 |
+
size 14244
|
checkpoints/checkpoint-51000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba7d5590af05df6e51ad422f658f48d4b533eb50367380ec9c9eff7012fe42eb
|
3 |
+
size 1064
|
checkpoints/checkpoint-51000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-51000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
|
3 |
+
size 5112
|
checkpoints/checkpoint-52000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-52000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4fa518a02969a2a8f6fdf740701e61a39bf4ac87eb9e5c5e17200dc0bb43f990
|
3 |
+
size 44890256
|
checkpoints/checkpoint-52000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c087906600c9662b09a8ff962630a01cc747ba7cf6e7fa2eff7c901618e2c19c
|
3 |
+
size 89797322
|
checkpoints/checkpoint-52000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8f0d6128daece7ee138488d0405ba20683554aebe56f4e3774203b1ff5e43d32
|
3 |
+
size 14244
|
checkpoints/checkpoint-52000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:50ee8f4ed8a0fd1b13d34b4eda8cf9645e38e4339d6fa5ca42244af1d707bfba
|
3 |
+
size 1064
|
checkpoints/checkpoint-52000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-52000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
|
3 |
+
size 5112
|
checkpoints/checkpoint-53000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-53000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:17183bf2f2d1e3e15f79ae798c9ccc9dcae0bb2add266d7b63cd3c880a971844
|
3 |
+
size 44890256
|
checkpoints/checkpoint-53000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9b32bf7b4072067bce6267388c3c9ab5bc404502e425eb13afa2bf854016f7aa
|
3 |
+
size 89797322
|
checkpoints/checkpoint-53000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5121d7ea8e1c63607830973f6a99e34ee3b048bf1d8313fd391e150255e789be
|
3 |
+
size 14244
|
checkpoints/checkpoint-53000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a151fdab51c034cd3fc8b0b65fee7659b99e5753fa0dff3edb44e998ba4aba41
|
3 |
+
size 1064
|
checkpoints/checkpoint-53000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-53000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
|
3 |
+
size 5112
|
checkpoints/checkpoint-54000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-54000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a906631ee85a758dafc7f566772038ce5c928a568c5a705ae27a2e9909adb9c1
|
3 |
+
size 44890256
|
checkpoints/checkpoint-54000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:07d4d67af68d2e7f47482742124a750511f13e4de0acc548a578349dbdd3da24
|
3 |
+
size 89797322
|
checkpoints/checkpoint-54000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:719dcc21f701523b0563a2506cce2906abe179e40c26e43d9781cc2652dc8f6c
|
3 |
+
size 14244
|
checkpoints/checkpoint-54000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:276adbe1735ddeb94c55b460a02fb15f06f275511fd3941aed6564ac8b3e883a
|
3 |
+
size 1064
|
checkpoints/checkpoint-54000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-54000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
|
3 |
+
size 5112
|
checkpoints/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-
|
3 |
"architectures": [
|
4 |
"AlbertForMaskedLM"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-14000",
|
3 |
"architectures": [
|
4 |
"AlbertForMaskedLM"
|
5 |
],
|
checkpoints/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 44890256
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a906631ee85a758dafc7f566772038ce5c928a568c5a705ae27a2e9909adb9c1
|
3 |
size 44890256
|
checkpoints/training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5112
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c74bde39c0eb85c341f3d86db723addaddac33096478be1fb933890f3bd3d68
|
3 |
size 5112
|