|
--- |
|
language: en |
|
widget: |
|
- text: "Arts" |
|
example_title: "Arts" |
|
- text: "Space" |
|
example_title: "Space" |
|
- text: "Time" |
|
example_title: "Time" |
|
|
|
tags: |
|
- text-generation |
|
- creative-writing |
|
- essay-writing |
|
|
|
inference: |
|
parameters: |
|
max_length: 400 |
|
num_beams: 10 |
|
early_stopping: true |
|
temperature: 0.3 |
|
no_repeat_ngram_size: 2 |
|
num_return_sequences: 2 |
|
|
|
|
|
--- |
|
Introduction: |
|
This repository contains a finetuned DistilGPT2 model for generating diverse essays on topics spanning Arts, Science, and Culture. |
|
|
|
Dataset: |
|
The training dataset comprises 2000+ essays covering diverse topics in Arts, Science, and Culture. These essays are written by human experts and contain a diverse set of opinions and knowledge, ensuring that the model learns from high-quality and diverse content. |
|
|
|
Model Training: |
|
- epoch: 50 |
|
- training_loss: 2.473200 |
|
- validation_loss: 4.569556 |
|
- perplexities: [517.4149169921875, 924.535888671875, 704.73291015625, 465.9677429199219, 577.629150390625, 443.994140625, 770.1861572265625, 683.028076171875, 1017.7510375976562, 880.795166015625] |
|
- mean_perplexity: 698.603519 |
|
|
|
Description: |
|
The model achieved a mean perplexity of 698.603519 on the validation set, indicating its ability to generate diverse and high-quality essays on the given topics. |
|
|
|
During Text Generation, the following parameters are used: |
|
|
|
- `max_length`: The maximum length of the generated text, set to 400 tokens. |
|
- `num_beams`: The number of beams for beam search, set to 10. A higher value will increase the diversity of the generated text but may also increase the inference time. |
|
- `early_stopping`: If set to True, the generation will stop as soon as the end-of-sequence token is generated. |
|
- `temperature`: The sampling temperature, is set to 0.3. |
|
- `no_repeat_ngram_size`: The size of the n-gram window to avoid repetitions, set to 2. |
|
|
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64fec5de57ccb8f1bdfbec54/ac89INQ8czj1u6WApI20J.png) |
|
|
|
|
|
Find the kaggle notebook for this project at |
|
|
|
[Kaggle Notebook](https://www.kaggle.com/code/vignesharjunraj/finetuned-distilgpt2-llm-for-essays-400-words/) |