tarekxpc commited on
Commit
1f9a639
1 Parent(s): 0f6c108

Training complete

Browse files
Files changed (3) hide show
  1. README.md +16 -16
  2. pytorch_model.bin +1 -1
  3. tokenizer.json +1 -6
README.md CHANGED
@@ -13,10 +13,10 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # mamba_text_classification
15
 
16
- This model was trained from scratch on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.8755
19
- - Accuracy: 0.7778
20
 
21
  ## Model description
22
 
@@ -35,28 +35,28 @@ More information needed
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
- - learning_rate: 5e-05
39
- - train_batch_size: 2
40
- - eval_batch_size: 4
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.01
45
- - num_epochs: 1
46
 
47
  ### Training results
48
 
49
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
50
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
51
- | 2.0279 | 0.1 | 371 | 2.5090 | 0.1111 |
52
- | 1.3557 | 0.2 | 742 | 1.0745 | 0.6667 |
53
- | 1.5255 | 0.3 | 1113 | 1.1685 | 0.6667 |
54
- | 0.7389 | 0.4 | 1484 | 1.3240 | 0.5556 |
55
- | 0.9114 | 0.5 | 1855 | 1.2930 | 0.6667 |
56
- | 0.0422 | 0.6 | 2226 | 1.1987 | 0.6667 |
57
- | 1.5648 | 0.7 | 2597 | 0.5782 | 0.7778 |
58
- | 1.7356 | 0.8 | 2968 | 0.7707 | 0.6667 |
59
- | 0.0145 | 0.9 | 3339 | 0.8755 | 0.7778 |
60
 
61
 
62
  ### Framework versions
 
13
 
14
  # mamba_text_classification
15
 
16
+ This model was trained from scratch on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.6274
19
+ - Accuracy: 1.0
20
 
21
  ## Model description
22
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - learning_rate: 1e-06
39
+ - train_batch_size: 4
40
+ - eval_batch_size: 8
41
  - seed: 42
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.01
45
+ - num_epochs: 4
46
 
47
  ### Training results
48
 
49
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
50
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
51
+ | 0.6169 | 0.4 | 52 | 0.8997 | 0.0 |
52
+ | 0.698 | 0.81 | 104 | 0.7669 | 0.0 |
53
+ | 0.5953 | 1.21 | 156 | 0.6956 | 0.0 |
54
+ | 0.5979 | 1.61 | 208 | 0.6580 | 1.0 |
55
+ | 0.5949 | 2.02 | 260 | 0.6465 | 1.0 |
56
+ | 0.6608 | 2.42 | 312 | 0.6321 | 1.0 |
57
+ | 0.5082 | 2.82 | 364 | 0.6339 | 1.0 |
58
+ | 0.578 | 3.22 | 416 | 0.6302 | 1.0 |
59
+ | 0.6325 | 3.63 | 468 | 0.6274 | 1.0 |
60
 
61
 
62
  ### Framework versions
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:80c2f54576acbd5a4b54a1cf2fffa8bcf687f5286f04a290b9a50d1bafd27b07
3
  size 516640282
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afad85180d6c7b577c15d78afba6909e76b495afa89e4b6f361ff91ca68fca65
3
  size 516640282
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {