File size: 3,606 Bytes
a1262ea ebd2cf7 a1262ea c3a3ff3 24de895 44d8322 e41d963 746df02 17fbbfe a46da16 17d35ff 5cc4342 c426dc1 ce45537 c06e7f1 56414db 77882bf a479fc3 b223860 9dbf8ea 53741d0 36e4b16 2a236b6 c2bc997 8e1cddf c66c625 d50a39b dbe0a29 323919c d54825e a2ddf1c 36d33df bf7ebcc 71624f2 0c202df 7494c13 c88d031 7696e5d 225c798 5b83697 bc21343 4e31b47 27e3420 5940fac 6e3ce3d ef992b5 4f56fd5 d5ae141 891f130 fc1d24a 2ae7b9e 8afdd1f 98400dd 9239ffe d0d0b5a 2757e58 be0bf81 cc89ac5 49403ac ebd2cf7 a1262ea |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 |
---
license: apache-2.0
base_model: t5-small
tags:
- generated_from_keras_callback
model-index:
- name: tarsssss/eng-jagoy-t5-001
results: []
---
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# tarsssss/eng-jagoy-t5-001
This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 5.4795
- Validation Loss: 5.4816
- Epoch: 56
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32
### Training results
| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 7.8603 | 7.4105 | 0 |
| 7.3775 | 7.1273 | 1 |
| 7.1632 | 6.9598 | 2 |
| 7.0228 | 6.8372 | 3 |
| 6.9085 | 6.7335 | 4 |
| 6.8226 | 6.6458 | 5 |
| 6.7451 | 6.5671 | 6 |
| 6.6785 | 6.5022 | 7 |
| 6.6254 | 6.4409 | 8 |
| 6.5606 | 6.3842 | 9 |
| 6.5163 | 6.3361 | 10 |
| 6.4682 | 6.2908 | 11 |
| 6.4250 | 6.2436 | 12 |
| 6.3749 | 6.1907 | 13 |
| 6.3293 | 6.1494 | 14 |
| 6.2822 | 6.1098 | 15 |
| 6.2560 | 6.0750 | 16 |
| 6.2078 | 6.0508 | 17 |
| 6.1839 | 6.0229 | 18 |
| 6.1561 | 5.9944 | 19 |
| 6.1146 | 5.9732 | 20 |
| 6.0885 | 5.9490 | 21 |
| 6.0587 | 5.9243 | 22 |
| 6.0366 | 5.9064 | 23 |
| 6.0135 | 5.8857 | 24 |
| 5.9904 | 5.8675 | 25 |
| 5.9681 | 5.8482 | 26 |
| 5.9473 | 5.8262 | 27 |
| 5.9263 | 5.8127 | 28 |
| 5.9031 | 5.7896 | 29 |
| 5.8827 | 5.7721 | 30 |
| 5.8566 | 5.7482 | 31 |
| 5.8406 | 5.7355 | 32 |
| 5.8285 | 5.7231 | 33 |
| 5.7944 | 5.7049 | 34 |
| 5.7822 | 5.6968 | 35 |
| 5.7567 | 5.6813 | 36 |
| 5.7526 | 5.6650 | 37 |
| 5.7363 | 5.6614 | 38 |
| 5.7132 | 5.6398 | 39 |
| 5.6945 | 5.6383 | 40 |
| 5.6786 | 5.6243 | 41 |
| 5.6636 | 5.6071 | 42 |
| 5.6527 | 5.5955 | 43 |
| 5.6390 | 5.5876 | 44 |
| 5.6198 | 5.5754 | 45 |
| 5.6082 | 5.5663 | 46 |
| 5.6070 | 5.5572 | 47 |
| 5.5782 | 5.5493 | 48 |
| 5.5679 | 5.5487 | 49 |
| 5.5520 | 5.5301 | 50 |
| 5.5307 | 5.5261 | 51 |
| 5.5284 | 5.5089 | 52 |
| 5.5160 | 5.5003 | 53 |
| 5.4976 | 5.4981 | 54 |
| 5.4864 | 5.4860 | 55 |
| 5.4795 | 5.4816 | 56 |
### Framework versions
- Transformers 4.33.2
- TensorFlow 2.10.0
- Datasets 2.15.0
- Tokenizers 0.13.3
|