Upload folder using huggingface_hub
Browse files- checkpoints/checkpoint-34000/config.json +29 -0
- checkpoints/checkpoint-34000/model.safetensors +3 -0
- checkpoints/checkpoint-34000/optimizer.pt +3 -0
- checkpoints/checkpoint-34000/rng_state.pth +3 -0
- checkpoints/checkpoint-34000/scheduler.pt +3 -0
- checkpoints/checkpoint-34000/trainer_state.json +0 -0
- checkpoints/checkpoint-34000/training_args.bin +3 -0
- checkpoints/checkpoint-35000/config.json +29 -0
- checkpoints/checkpoint-35000/model.safetensors +3 -0
- checkpoints/checkpoint-35000/optimizer.pt +3 -0
- checkpoints/checkpoint-35000/rng_state.pth +3 -0
- checkpoints/checkpoint-35000/scheduler.pt +3 -0
- checkpoints/checkpoint-35000/trainer_state.json +0 -0
- checkpoints/checkpoint-35000/training_args.bin +3 -0
- checkpoints/checkpoint-36000/config.json +29 -0
- checkpoints/checkpoint-36000/model.safetensors +3 -0
- checkpoints/checkpoint-36000/optimizer.pt +3 -0
- checkpoints/checkpoint-36000/rng_state.pth +3 -0
- checkpoints/checkpoint-36000/scheduler.pt +3 -0
- checkpoints/checkpoint-36000/trainer_state.json +0 -0
- checkpoints/checkpoint-36000/training_args.bin +3 -0
- checkpoints/checkpoint-37000/config.json +29 -0
- checkpoints/checkpoint-37000/model.safetensors +3 -0
- checkpoints/checkpoint-37000/optimizer.pt +3 -0
- checkpoints/checkpoint-37000/rng_state.pth +3 -0
- checkpoints/checkpoint-37000/scheduler.pt +3 -0
- checkpoints/checkpoint-37000/trainer_state.json +0 -0
- checkpoints/checkpoint-37000/training_args.bin +3 -0
- checkpoints/checkpoint-38000/config.json +29 -0
- checkpoints/checkpoint-38000/model.safetensors +3 -0
- checkpoints/checkpoint-38000/optimizer.pt +3 -0
- checkpoints/checkpoint-38000/rng_state.pth +3 -0
- checkpoints/checkpoint-38000/scheduler.pt +3 -0
- checkpoints/checkpoint-38000/trainer_state.json +0 -0
- checkpoints/checkpoint-38000/training_args.bin +3 -0
- checkpoints/config.json +1 -1
- checkpoints/model.safetensors +1 -1
- checkpoints/training_args.bin +1 -1
checkpoints/checkpoint-34000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-25000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-34000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c45db85a8f1459282bcbbce3da1c71aa579e5b7c97efbc653dc13a7410ce77a1
|
3 |
+
size 44890256
|
checkpoints/checkpoint-34000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b6c3da189aa61c05703c10d589be03a1fb6ec6fc8eff31f395fbf128c402c009
|
3 |
+
size 89797322
|
checkpoints/checkpoint-34000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:294b92ecaaa21feb76133de4b6aa79252c9aa7a0d85c2df0f28e44645ab14e7d
|
3 |
+
size 14244
|
checkpoints/checkpoint-34000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae8fd1aff388b68cb8d83c1e5e0843b77b08f18010572e0571a6f2812eb41e28
|
3 |
+
size 1064
|
checkpoints/checkpoint-34000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-34000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2082f22c6e15e3580ffafef7d6546679ce8b0b31d9030c785ab63457d6d33de7
|
3 |
+
size 5112
|
checkpoints/checkpoint-35000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-25000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-35000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d3693881e9c8dce5107b1ba12a482e72367d61fd4604eb8e41f637bae33897b3
|
3 |
+
size 44890256
|
checkpoints/checkpoint-35000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee5ee6c846d7488949e0d6a7318dda4d244d2b2fbc51580d0ab4b329559c71b0
|
3 |
+
size 89797322
|
checkpoints/checkpoint-35000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a73953dd2539490d4aafcfd99c4b6409140fa071f758e25b0bb14c37034c0541
|
3 |
+
size 14244
|
checkpoints/checkpoint-35000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74faf87ce1ff0679419cbdc997cf31ec7b2458533f1e41453fe2e77a99b9a07d
|
3 |
+
size 1064
|
checkpoints/checkpoint-35000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-35000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2082f22c6e15e3580ffafef7d6546679ce8b0b31d9030c785ab63457d6d33de7
|
3 |
+
size 5112
|
checkpoints/checkpoint-36000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-25000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-36000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d7bc3560b550fcb0fb5ffeb54cc1601ce130e476a5cd6ff3f65cd7355dd78c8c
|
3 |
+
size 44890256
|
checkpoints/checkpoint-36000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f622c0025e933badcd56343f5cd9e90f238249631a9d1dccfd63cf13feae9dee
|
3 |
+
size 89797322
|
checkpoints/checkpoint-36000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96623784616bfb0dd7bff9072916c768d00a474b19169a85e1ea13f46cdf80c4
|
3 |
+
size 14244
|
checkpoints/checkpoint-36000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1dcf3acfecd82418c01a04c5820ad613f26ad20caf6b867682cc1d26af1bb676
|
3 |
+
size 1064
|
checkpoints/checkpoint-36000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-36000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2082f22c6e15e3580ffafef7d6546679ce8b0b31d9030c785ab63457d6d33de7
|
3 |
+
size 5112
|
checkpoints/checkpoint-37000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-25000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-37000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f48bbd8710f42d76bd25a8621f968b2c1c045163100da66292423effd0208beb
|
3 |
+
size 44890256
|
checkpoints/checkpoint-37000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e5b707a07c9583dcf8156150add31066ccadb7a54e6fc7c573b67dd74d4b869f
|
3 |
+
size 89797322
|
checkpoints/checkpoint-37000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3355245dfb82f5ebecc364907bbcc5345ae79a2bc103a3e1dcadc453c63437e2
|
3 |
+
size 14244
|
checkpoints/checkpoint-37000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:887f63dbc55e0735bef0668fb5eea33a071f5b5196a3473ebeaa954450dc84ae
|
3 |
+
size 1064
|
checkpoints/checkpoint-37000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-37000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2082f22c6e15e3580ffafef7d6546679ce8b0b31d9030c785ab63457d6d33de7
|
3 |
+
size 5112
|
checkpoints/checkpoint-38000/config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-25000",
|
3 |
+
"architectures": [
|
4 |
+
"AlbertForMaskedLM"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0,
|
7 |
+
"bos_token_id": 2,
|
8 |
+
"classifier_dropout_prob": 0.1,
|
9 |
+
"embedding_size": 128,
|
10 |
+
"eos_token_id": 3,
|
11 |
+
"hidden_act": "gelu_new",
|
12 |
+
"hidden_dropout_prob": 0,
|
13 |
+
"hidden_size": 768,
|
14 |
+
"initializer_range": 0.02,
|
15 |
+
"inner_group_num": 1,
|
16 |
+
"intermediate_size": 3072,
|
17 |
+
"layer_norm_eps": 1e-12,
|
18 |
+
"max_position_embeddings": 512,
|
19 |
+
"model_type": "albert",
|
20 |
+
"num_attention_heads": 12,
|
21 |
+
"num_hidden_groups": 1,
|
22 |
+
"num_hidden_layers": 12,
|
23 |
+
"pad_token_id": 0,
|
24 |
+
"position_embedding_type": "absolute",
|
25 |
+
"torch_dtype": "float32",
|
26 |
+
"transformers_version": "4.38.2",
|
27 |
+
"type_vocab_size": 2,
|
28 |
+
"vocab_size": 30000
|
29 |
+
}
|
checkpoints/checkpoint-38000/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:830cb24badbac18edfa59a7ad9bf307055bb91761ef656a235335d0f4bf22a10
|
3 |
+
size 44890256
|
checkpoints/checkpoint-38000/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5df45215e71e76fced7701d73ebd04ed7d2cba48533d596dd0199e16f6387598
|
3 |
+
size 89797322
|
checkpoints/checkpoint-38000/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11738d65197dcfeb3f50f6203e3719029ff39110abc36935ef0dd5274d5032d9
|
3 |
+
size 14244
|
checkpoints/checkpoint-38000/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a5068f32b6ad184892d872c68bd67c2aeda0b8333776681ca7444b62494ce6b2
|
3 |
+
size 1064
|
checkpoints/checkpoint-38000/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoints/checkpoint-38000/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2082f22c6e15e3580ffafef7d6546679ce8b0b31d9030c785ab63457d6d33de7
|
3 |
+
size 5112
|
checkpoints/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "/content/
|
3 |
"architectures": [
|
4 |
"AlbertForMaskedLM"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/content/drive/MyDrive/Colab Notebooks/nusa-albert/checkpoint-25000",
|
3 |
"architectures": [
|
4 |
"AlbertForMaskedLM"
|
5 |
],
|
checkpoints/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 44890256
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:830cb24badbac18edfa59a7ad9bf307055bb91761ef656a235335d0f4bf22a10
|
3 |
size 44890256
|
checkpoints/training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5112
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2082f22c6e15e3580ffafef7d6546679ce8b0b31d9030c785ab63457d6d33de7
|
3 |
size 5112
|