File size: 5,739 Bytes
a1262ea
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2360fbf
 
 
a1262ea
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c3a3ff3
24de895
44d8322
e41d963
746df02
17fbbfe
a46da16
17d35ff
5cc4342
c426dc1
ce45537
c06e7f1
56414db
77882bf
a479fc3
b223860
9dbf8ea
53741d0
36e4b16
2a236b6
c2bc997
8e1cddf
c66c625
d50a39b
dbe0a29
323919c
d54825e
a2ddf1c
36d33df
bf7ebcc
71624f2
0c202df
7494c13
c88d031
7696e5d
225c798
5b83697
bc21343
4e31b47
27e3420
5940fac
6e3ce3d
ef992b5
4f56fd5
d5ae141
891f130
fc1d24a
2ae7b9e
8afdd1f
98400dd
9239ffe
d0d0b5a
2757e58
be0bf81
cc89ac5
49403ac
ebd2cf7
7ad48b1
cc0aa93
c5aa30c
7e30e48
4d6feaf
50dcdda
84b7ee7
753e7e7
85e6817
b2cbd09
f5772ed
9df911c
3aa315b
5504f13
386e427
90ab3ce
a219f25
7314821
ffaf1f0
c541471
9979285
7446ee0
b89d71e
438681d
187ee4e
9b43f49
899d3a8
6abdc8b
965b304
78bb999
69e4af9
a2ec60c
123fca7
9893729
d05721c
5e6f7dd
08ea790
6a435cd
bb9a75b
59dd5cf
f594f09
0f1b37c
3fd9c58
4a693e1
55c7ed9
b5b6ed9
06ac3e7
ce27527
dbba31a
563b8e8
f47eb6e
2360fbf
a1262ea
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
---
license: apache-2.0
base_model: t5-small
tags:
- generated_from_keras_callback
model-index:
- name: tarsssss/eng-jagoy-t5-001
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# tarsssss/eng-jagoy-t5-001

This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 4.9612
- Validation Loss: 5.2003
- Epoch: 108

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 7.8603     | 7.4105          | 0     |
| 7.3775     | 7.1273          | 1     |
| 7.1632     | 6.9598          | 2     |
| 7.0228     | 6.8372          | 3     |
| 6.9085     | 6.7335          | 4     |
| 6.8226     | 6.6458          | 5     |
| 6.7451     | 6.5671          | 6     |
| 6.6785     | 6.5022          | 7     |
| 6.6254     | 6.4409          | 8     |
| 6.5606     | 6.3842          | 9     |
| 6.5163     | 6.3361          | 10    |
| 6.4682     | 6.2908          | 11    |
| 6.4250     | 6.2436          | 12    |
| 6.3749     | 6.1907          | 13    |
| 6.3293     | 6.1494          | 14    |
| 6.2822     | 6.1098          | 15    |
| 6.2560     | 6.0750          | 16    |
| 6.2078     | 6.0508          | 17    |
| 6.1839     | 6.0229          | 18    |
| 6.1561     | 5.9944          | 19    |
| 6.1146     | 5.9732          | 20    |
| 6.0885     | 5.9490          | 21    |
| 6.0587     | 5.9243          | 22    |
| 6.0366     | 5.9064          | 23    |
| 6.0135     | 5.8857          | 24    |
| 5.9904     | 5.8675          | 25    |
| 5.9681     | 5.8482          | 26    |
| 5.9473     | 5.8262          | 27    |
| 5.9263     | 5.8127          | 28    |
| 5.9031     | 5.7896          | 29    |
| 5.8827     | 5.7721          | 30    |
| 5.8566     | 5.7482          | 31    |
| 5.8406     | 5.7355          | 32    |
| 5.8285     | 5.7231          | 33    |
| 5.7944     | 5.7049          | 34    |
| 5.7822     | 5.6968          | 35    |
| 5.7567     | 5.6813          | 36    |
| 5.7526     | 5.6650          | 37    |
| 5.7363     | 5.6614          | 38    |
| 5.7132     | 5.6398          | 39    |
| 5.6945     | 5.6383          | 40    |
| 5.6786     | 5.6243          | 41    |
| 5.6636     | 5.6071          | 42    |
| 5.6527     | 5.5955          | 43    |
| 5.6390     | 5.5876          | 44    |
| 5.6198     | 5.5754          | 45    |
| 5.6082     | 5.5663          | 46    |
| 5.6070     | 5.5572          | 47    |
| 5.5782     | 5.5493          | 48    |
| 5.5679     | 5.5487          | 49    |
| 5.5520     | 5.5301          | 50    |
| 5.5307     | 5.5261          | 51    |
| 5.5284     | 5.5089          | 52    |
| 5.5160     | 5.5003          | 53    |
| 5.4976     | 5.4981          | 54    |
| 5.4864     | 5.4860          | 55    |
| 5.4795     | 5.4816          | 56    |
| 5.4653     | 5.4652          | 57    |
| 5.4484     | 5.4639          | 58    |
| 5.4335     | 5.4580          | 59    |
| 5.4231     | 5.4454          | 60    |
| 5.4132     | 5.4358          | 61    |
| 5.4064     | 5.4349          | 62    |
| 5.3886     | 5.4261          | 63    |
| 5.3913     | 5.4193          | 64    |
| 5.3692     | 5.4138          | 65    |
| 5.3556     | 5.4028          | 66    |
| 5.3469     | 5.4001          | 67    |
| 5.3421     | 5.3942          | 68    |
| 5.3194     | 5.3826          | 69    |
| 5.3243     | 5.3799          | 70    |
| 5.3081     | 5.3713          | 71    |
| 5.2921     | 5.3737          | 72    |
| 5.2845     | 5.3681          | 73    |
| 5.2754     | 5.3601          | 74    |
| 5.2594     | 5.3524          | 75    |
| 5.2527     | 5.3420          | 76    |
| 5.2496     | 5.3367          | 77    |
| 5.2360     | 5.3320          | 78    |
| 5.2193     | 5.3253          | 79    |
| 5.2141     | 5.3178          | 80    |
| 5.1993     | 5.3150          | 81    |
| 5.1923     | 5.3157          | 82    |
| 5.1875     | 5.3097          | 83    |
| 5.1776     | 5.3051          | 84    |
| 5.1693     | 5.3050          | 85    |
| 5.1533     | 5.3115          | 86    |
| 5.1567     | 5.2943          | 87    |
| 5.1348     | 5.2757          | 88    |
| 5.1317     | 5.2849          | 89    |
| 5.1191     | 5.2846          | 90    |
| 5.1102     | 5.2742          | 91    |
| 5.1054     | 5.2725          | 92    |
| 5.0944     | 5.2624          | 93    |
| 5.0906     | 5.2560          | 94    |
| 5.0712     | 5.2502          | 95    |
| 5.0719     | 5.2495          | 96    |
| 5.0628     | 5.2498          | 97    |
| 5.0597     | 5.2454          | 98    |
| 5.0402     | 5.2420          | 99    |
| 5.0308     | 5.2441          | 100   |
| 5.0193     | 5.2379          | 101   |
| 5.0198     | 5.2298          | 102   |
| 5.0110     | 5.2315          | 103   |
| 5.0087     | 5.2304          | 104   |
| 4.9906     | 5.2261          | 105   |
| 4.9883     | 5.2288          | 106   |
| 4.9818     | 5.2069          | 107   |
| 4.9612     | 5.2003          | 108   |


### Framework versions

- Transformers 4.33.2
- TensorFlow 2.10.0
- Datasets 2.15.0
- Tokenizers 0.13.3