Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
A small french language model for french text generation (and possibly more NLP tasks...)
|
2 |
+
|
3 |
+
**Introduction**
|
4 |
+
|
5 |
+
This french gpt2 model is based on openai GPT-2 small model.
|
6 |
+
|
7 |
+
It was trained on a <b>very small (190Mb) dataset </b> from french wikipedia using Transfer Learning and Fine-tuning techniques in just over a day, on one Colab pro with 1GPU 16GB.
|
8 |
+
|
9 |
+
It was created applying the recept of <b>Pierre Guillou</b>
|
10 |
+
|
11 |
+
See https://medium.com/@pierre_guillou/faster-than-training-from-scratch-fine-tuning-the-english-gpt-2-in-any-language-with-hugging-f2ec05c98787
|
12 |
+
|
13 |
+
It is a proof-of-concept that makes possible to get a language model in any language with low ressources.
|
14 |
+
|
15 |
+
It was fine-tuned from the English pre-trained GPT-2 small using the Hugging Face libraries (Transformers and Tokenizers) wrapped into the fastai v2 Deep Learning framework. All the fine-tuning fastai v2 techniques were used.
|
16 |
+
|
17 |
+
It is now available on Hugging Face. For further information or requests, please go to "Faster than training from scratch — Fine-tuning the English GPT-2 in any language with Hugging Face and fastai v2 (practical case with Portuguese)".
|
18 |
+
|
19 |
+
Model migth be improved by using larger dataset under larger powerful training infrastructure. At least this one can be used for small finetuning experimentation (i.e with aitextgen).
|
20 |
+
|
21 |
+
PS : I've lost the metrics but it speaks french with some minor grammar issues, coherence of text is somehow limited.
|