Update README.md
Browse files
README.md
CHANGED
@@ -4,12 +4,13 @@ datasets:
|
|
4 |
language:
|
5 |
- en
|
6 |
pipeline_tag: summarization
|
|
|
7 |
---
|
8 |
# Model Card for roemmele/ablit-bart-base
|
9 |
|
10 |
<!-- Provide a quick summary of what the model is/does. -->
|
11 |
|
12 |
-
This model is initialized from facebook/bart-base. It has been fine-tuned on the AbLit dataset, which consists of abridged versions of books aligned with their original versions at the passage level. Given a text, the model generates an abridgement of the text based on what it has observed in AbLit. See the cited
|
13 |
|
14 |
## Model Details
|
15 |
|
@@ -20,15 +21,15 @@ This model is initialized from facebook/bart-base. It has been fine-tuned on the
|
|
20 |
- **Developed by:** Language Weaver (Melissa Roemmele, Kyle Shaffer, Katrina Olsen, Yiyi Wang, and Steve DeNeefe)
|
21 |
- **Model type:** Seq2SeqLM
|
22 |
- **Language(s) (NLP):** English
|
23 |
-
- **License:**
|
24 |
-
- **Finetuned from model [
|
25 |
|
26 |
-
### Model Sources
|
27 |
|
28 |
<!-- Provide the basic links for the model. -->
|
29 |
|
30 |
- **Repository:** [github.com/roemmele/AbLit](https://github.com/roemmele/AbLit)
|
31 |
-
- **Paper
|
32 |
|
33 |
## Uses
|
34 |
|
@@ -81,12 +82,9 @@ data = load_dataset("roemmele/ablit", "chunks-10-sentences")
|
|
81 |
|
82 |
We used the training script [here](https://github.com/huggingface/transformers/blob/main/examples/pytorch/summarization/run_summarization.py).
|
83 |
|
84 |
-
Hyperparameter settings: We specified maximum length of 1024 for both the source (original passage) and target (abridged passage), and truncated all tokens beyond this limit. We evaluated each model on the AbLit development set after each epoch and concluded training when cross-entropy loss stopped decreasing. We used a batch size of 4. For all other hyperparameters we used the default values set by this script.
|
85 |
-
|
86 |
-
|
87 |
#### Training Hyperparameters
|
88 |
|
89 |
-
|
90 |
|
91 |
#### Speeds, Sizes, Times
|
92 |
|
@@ -109,7 +107,7 @@ The model obtained a ROUGE-L score of 0.78 on the AbLit test set. See the paper
|
|
109 |
|
110 |
Our analysis shows that in comparison with human-authored abridgements, the model-generated abridgements tend to preserve more of the original text, suggesting it is challenging to learn what text can be removed while maintaining loyalty to the important parts of the original text.
|
111 |
|
112 |
-
## Citation
|
113 |
|
114 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
115 |
|
@@ -133,4 +131,4 @@ Melissa Roemmele
|
|
133 |
|
134 |
## Model Card Contact
|
135 |
|
136 |
|
|
4 |
language:
|
5 |
- en
|
6 |
pipeline_tag: summarization
|
7 |
+
license: mit
|
8 |
---
|
9 |
# Model Card for roemmele/ablit-bart-base
|
10 |
|
11 |
<!-- Provide a quick summary of what the model is/does. -->
|
12 |
|
13 |
+
This model is initialized from facebook/bart-base. It has been fine-tuned on the AbLit dataset, which consists of abridged versions of books aligned with their original versions at the passage level. Given a text, the model generates an abridgement of the text based on what it has observed in AbLit. See the paper cited below for more details.
|
14 |
|
15 |
## Model Details
|
16 |
|
|
|
21 |
- **Developed by:** Language Weaver (Melissa Roemmele, Kyle Shaffer, Katrina Olsen, Yiyi Wang, and Steve DeNeefe)
|
22 |
- **Model type:** Seq2SeqLM
|
23 |
- **Language(s) (NLP):** English
|
24 |
+
- **License:** mit
|
25 |
+
- **Finetuned from model:** [facebook/bart-base](https://huggingface.co/facebook/bart-base)
|
26 |
|
27 |
+
### Model Sources
|
28 |
|
29 |
<!-- Provide the basic links for the model. -->
|
30 |
|
31 |
- **Repository:** [github.com/roemmele/AbLit](https://github.com/roemmele/AbLit)
|
32 |
+
- **Paper:** [AbLit: A Resource for Analyzing and Generating Abridged Versions of English Literature](https://arxiv.org/pdf/2302.06579.pdf)
|
33 |
|
34 |
## Uses
|
35 |
|
|
|
82 |
|
83 |
We used the training script [here](https://github.com/huggingface/transformers/blob/main/examples/pytorch/summarization/run_summarization.py).
|
84 |
|
|
|
|
|
|
|
85 |
#### Training Hyperparameters
|
86 |
|
87 |
+
We specified maximum length of 1024 for both the source (original passage) and target (abridged passage), and truncated all tokens beyond this limit. We evaluated each model on the AbLit development set after each epoch and concluded training when cross-entropy loss stopped decreasing. We used a batch size of 4. For all other hyperparameters we used the default values set by this script.
|
88 |
|
89 |
#### Speeds, Sizes, Times
|
90 |
|
|
|
107 |
|
108 |
Our analysis shows that in comparison with human-authored abridgements, the model-generated abridgements tend to preserve more of the original text, suggesting it is challenging to learn what text can be removed while maintaining loyalty to the important parts of the original text.
|
109 |
|
110 |
+
## Citation
|
111 |
|
112 |
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
113 |
|
|
|
131 |
|
132 |
## Model Card Contact
|
133 |
|
134 |