Update README.md
Browse files
README.md
CHANGED
@@ -34,7 +34,7 @@ Notes and observations:
|
|
34 |
11. Using set_transform function to transform the samples on-the-fly was a good idea as it didn't require to save the transformed dataset.
|
35 |
12. Streaming dataset might be another good option if the dataset size were to increase any further.
|
36 |
13. Free GPU on colab seem not enough for this experiment, as keeping two models in GPU and training forces to keep batch size small and also the free GPUs (T4) are not fast enough.
|
37 |
-
14. A very important data cleaning step was to just check if the sample image and text can be converted to the input format expected by the model, the text should be non-empty value when converted back from the input IDs to text (some characters are not identified by the tokenizer and get converted to special token and we usually skip the special tokens when converting to text) as it is required to be non-empty while doing the CER calculation.
|
38 |
|
39 |
|
40 |
|
|
|
34 |
11. Using set_transform function to transform the samples on-the-fly was a good idea as it didn't require to save the transformed dataset.
|
35 |
12. Streaming dataset might be another good option if the dataset size were to increase any further.
|
36 |
13. Free GPU on colab seem not enough for this experiment, as keeping two models in GPU and training forces to keep batch size small and also the free GPUs (T4) are not fast enough.
|
37 |
+
14. A very important data cleaning step was to just check if the sample image and text can be converted to the input format expected by the model, the text should be non-empty value when converted back from the input IDs to text (some characters are not identified by the tokenizer and get converted to special token and we usually skip the special tokens when converting input IDs back to text) as it is required to be non-empty while doing the CER calculation.
|
38 |
|
39 |
|
40 |
|