nenad1002 commited on
Commit
6cdc647
1 Parent(s): 5eff434

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -127,7 +127,7 @@ The final evaluation cross-entropy ended around 0.4.
127
  Since the fine-tuned model is designed to explain, and if possible, summarize newly learned data, ROUGE and BERTScore metrics were measured on a sample of 50 manually crafted questions. The reference answers were constructed during the creation of the training and evaluation sets.
128
  Given that GPT-4-turbo was already used in this context for the reference questions generation, I did not compare my model against it. Instead, I chose to compare it against the following models:
129
 
130
- | Metric | quantum-research-bot-v1.0 | Meta-Llama-3.1-8B-Instruct | gemini-1.5-pro |
131
  |:------------------|:---------------------------|:--------------------|:------------------|
132
  | **BERTScore F1** | 0.5821 | 0.3305 | 0.4982 |
133
  | **ROUGE-1** | 0.6045 | 0.3152 |0.5029 |
 
127
  Since the fine-tuned model is designed to explain, and if possible, summarize newly learned data, ROUGE and BERTScore metrics were measured on a sample of 50 manually crafted questions. The reference answers were constructed during the creation of the training and evaluation sets.
128
  Given that GPT-4-turbo was already used in this context for the reference questions generation, I did not compare my model against it. Instead, I chose to compare it against the following models:
129
 
130
+ | Metric (mean) | quantum-research-bot-v1.0 | Meta-Llama-3.1-8B-Instruct | gemini-1.5-pro |
131
  |:------------------|:---------------------------|:--------------------|:------------------|
132
  | **BERTScore F1** | 0.5821 | 0.3305 | 0.4982 |
133
  | **ROUGE-1** | 0.6045 | 0.3152 |0.5029 |