jeffbritts commited on
Commit
a426a93
1 Parent(s): fdc70e0

Pushed by DataDreamer

Browse files
Files changed (1) hide show
  1. README.md +32 -191
README.md CHANGED
@@ -1,199 +1,40 @@
 
1
  ---
2
- library_name: transformers
3
- tags: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  ---
 
5
 
6
- # Model Card for Model ID
7
-
8
- <!-- Provide a quick summary of what the model is/does. -->
9
-
10
-
11
-
12
- ## Model Details
13
-
14
- ### Model Description
15
-
16
- <!-- Provide a longer summary of what this model is. -->
17
-
18
- This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
-
20
- - **Developed by:** [More Information Needed]
21
- - **Funded by [optional]:** [More Information Needed]
22
- - **Shared by [optional]:** [More Information Needed]
23
- - **Model type:** [More Information Needed]
24
- - **Language(s) (NLP):** [More Information Needed]
25
- - **License:** [More Information Needed]
26
- - **Finetuned from model [optional]:** [More Information Needed]
27
-
28
- ### Model Sources [optional]
29
-
30
- <!-- Provide the basic links for the model. -->
31
-
32
- - **Repository:** [More Information Needed]
33
- - **Paper [optional]:** [More Information Needed]
34
- - **Demo [optional]:** [More Information Needed]
35
-
36
- ## Uses
37
-
38
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
-
40
- ### Direct Use
41
-
42
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
-
44
- [More Information Needed]
45
-
46
- ### Downstream Use [optional]
47
-
48
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
-
50
- [More Information Needed]
51
-
52
- ### Out-of-Scope Use
53
-
54
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
-
56
- [More Information Needed]
57
-
58
- ## Bias, Risks, and Limitations
59
-
60
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
-
62
- [More Information Needed]
63
-
64
- ### Recommendations
65
-
66
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
-
68
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
-
70
- ## How to Get Started with the Model
71
-
72
- Use the code below to get started with the model.
73
-
74
- [More Information Needed]
75
-
76
- ## Training Details
77
-
78
- ### Training Data
79
-
80
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
-
82
- [More Information Needed]
83
-
84
- ### Training Procedure
85
-
86
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
-
88
- #### Preprocessing [optional]
89
-
90
- [More Information Needed]
91
-
92
-
93
- #### Training Hyperparameters
94
-
95
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
-
97
- #### Speeds, Sizes, Times [optional]
98
-
99
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
 
101
- [More Information Needed]
102
 
103
- ## Evaluation
 
104
 
105
- <!-- This section describes the evaluation protocols and provides the results. -->
 
 
106
 
107
- ### Testing Data, Factors & Metrics
 
 
108
 
109
- #### Testing Data
110
-
111
- <!-- This should link to a Dataset Card if possible. -->
112
-
113
- [More Information Needed]
114
-
115
- #### Factors
116
-
117
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
-
119
- [More Information Needed]
120
-
121
- #### Metrics
122
-
123
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
-
125
- [More Information Needed]
126
-
127
- ### Results
128
-
129
- [More Information Needed]
130
-
131
- #### Summary
132
-
133
-
134
-
135
- ## Model Examination [optional]
136
-
137
- <!-- Relevant interpretability work for the model goes here -->
138
-
139
- [More Information Needed]
140
-
141
- ## Environmental Impact
142
-
143
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
-
145
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
-
147
- - **Hardware Type:** [More Information Needed]
148
- - **Hours used:** [More Information Needed]
149
- - **Cloud Provider:** [More Information Needed]
150
- - **Compute Region:** [More Information Needed]
151
- - **Carbon Emitted:** [More Information Needed]
152
-
153
- ## Technical Specifications [optional]
154
-
155
- ### Model Architecture and Objective
156
-
157
- [More Information Needed]
158
-
159
- ### Compute Infrastructure
160
-
161
- [More Information Needed]
162
-
163
- #### Hardware
164
-
165
- [More Information Needed]
166
-
167
- #### Software
168
-
169
- [More Information Needed]
170
-
171
- ## Citation [optional]
172
-
173
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
-
175
- **BibTeX:**
176
-
177
- [More Information Needed]
178
-
179
- **APA:**
180
-
181
- [More Information Needed]
182
-
183
- ## Glossary [optional]
184
-
185
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
-
187
- [More Information Needed]
188
-
189
- ## More Information [optional]
190
-
191
- [More Information Needed]
192
-
193
- ## Model Card Authors [optional]
194
-
195
- [More Information Needed]
196
-
197
- ## Model Card Contact
198
-
199
- [More Information Needed]
 
1
+
2
  ---
3
+ base_model: google/t5-v1_1-base
4
+
5
+ tags:
6
+ - datadreamer
7
+ - datadreamer-0.28.0
8
+ - synthetic
9
+ - openai-community/gpt2
10
+ - openai-community/gpt2
11
+ - text2text-generation
12
+
13
+ widget:
14
+ - text: "Acknowledgments\n\nThank you to all our donors. Their input was invaluable, and many of them have kept this program active. I really appreciate some privacy concerns with these papers and the paper itself. However, thank you to my research team for helping get the entire research protocol up and running since 2010. It's been absolutely stunning for me to be a part of such a small organization, but when something like this happens, it is such a huge deal. It means it's hard not to get involved.\n\nYou will also get a new Open Science Foundation letter if you donate and support NLP. I know I am more than qualified to help you in any way you get involved. Thank you in advance.\n\nAs an additional thanks-good-ness, at the risk of repeating some of a large list, I will do an accompanying Google Hangout. The Hangout is where you can send an email at nlp-doc@umass-edu. In my time as a speaker, we'll do an ongoing Hangout video series and maybe even a live talk. The original YouTube channel is hosted here.\n\nIf you have any questions or concerns or would like to talk to a team member, write to my Open Science Committee through this website below or send your comments directly to me. Thanks."
15
+ example_title: "Example 1"
16
+ - text: "Findings:\n\nFor those of you whose work relies on an annotation, it may be helpful to check out my previous work by Fotus, Dandridgea, and Chudley on generative models. All were very effective! If there could be the opportunity for more work like that, it's great to have more!\n\nThis work was adapted verbatim into a textbook paper as a guest post for Kornbirch's \"I, The Computer Science Professorship\": The Importance of an Annotations \" for an excellent article.\n\nThis paper provides three new tools built into each, namely an online parser, two independent dictionaries, and a simple formulae for a variety of information types.\n\nIt was suggested that one or two authors could contribute their work to make the paper worthwhile by bringing a few new ideas or topics into the field. After learning how to extract such things, we now need to make it possible for authors to share what they want to talk about, even if this entails writing less text.\n\nIn addition to publishing annotations, annotations represent a new tool. A user's ability in their everyday life to learn with their own hands is vital to make good annotation tools. Annotations have the additional benefit of being simple to use and fast to produce. And these annotations can, given the nature of their usage, provide the motivation needed to improve, too.\n\nWith support for this paper from The University of Chicago, we now consider creating my own annotated text corpus. After a cursory search on the abstracts, a search of the articles in this article, one of my other annotated texts on annotation, had already been selected. And as long as this corpus was already available to help my other annotated articles in progress, we can go ahead and create our own custom annotations of my own work.\n\nNotes"
17
+ example_title: "Example 2"
18
+ - text: "This program will create some data, like this chart:\n\n\nThat's much nicer.\n\n\nThe author, I have the following project information (details, contact info, etc)::\n\nName\n\nDate posted last year\n\nTitle\n\nSubject\n\nComments on Papers\n\nNotes\n\nPseudoscience & Information\n\nIn my first presentation of a dataset, I said I had to send people out on my own for three days/event hours (to \"keep the record warm\"). I took this data, which you can buy and download on my web site, and found that the participants had been involved (3 days to 2 days plus 4 hours in \"special situations\" between 4, 5 and 8 a.m., etc). I was going to use \"research\" as my main interest group; hence I wrote that research report in a simple \"research summary sheet\" called Acknowledgements. This time, my focus was on making this research report available for the public, because many researchers will have seen it or read it and would be pleased to be able to send out paper in their papers or give a grant for this project to try to help others. I also want to provide people useful information that could help them improve their abilities to participate in important scientific research.\n\n\nIf you find your data useful, please share this information using the mailing list and Twitter or any other means where you can email me: [email protected]\n\n\nThank You.\n\nRobert\n\nJ.D. Wimp\n\nUniversity of Chicago Libraries\n\nChicago\n\nEmail\n\nMessage To\n\nTo\n\nTo\n\n\nDear J.O.,\n\nI would like to welcome each and every person of great interest within the U.S. of B.S. B.S. to you and I will keep an e-mail list at http://usb.beyond.ucla.ac.uk for this purpose. One last note about this program. During my presentation I referenced this program, but I also discussed the nature, length (days) and size of submissions. I would add this program in a follow up question I did at that address to people who think the use of these programs is wrong, i.e. they know of the lack of research using \"Research\" for their \"inform\" (rather than \"science\") (hereinafter referred to as \"the web page\").\n\n\nI would like to clarify: I was only talking to people who were interested and interested in improving their ability at their research work to send out \"research\" for their peer-reviewed journals. That means I was talking to some non-scientific professional, that did have a field interest, to ask questions on this program as well. As of this writing, I can confirm that there have been NO references to using \"Research\" or this data. Any other question that needs to be addressed by researchers for the \"experience\" of being a part of future research is welcome.\n\n\nSincerely...\n\nJ.E.W.\n\nUniversity of Wisconsin --\n\nMichaela\n\n\nYour research papers should not be used without an active or even active peer (not in direct relation to the actual papers/events or \"situations\"). Research must be presented with open, serious scholarly discussions of the main research findings.\n\nYou also shouldn't use \"Tribality\" as a proxy to describe the nature of certain types of study, since this type doesn't offer that useful understanding of many important issues of the study or of the subject. As an example, I've seen that there are researchers in psychology, where \"tribality\" refers to how they describe research from areas that could provide useful benefits. However, to compare this research with \"tribality\" you need to look at those cases with fewer \"tribalities\". Because I'm dealing with this research, I'm not actually talking about \"dementia\", here or elsewhere. Instead, I'm writing about the treatment for \"dementia\" or the cause of people's problem when they don't interact with the subject. There may be some \"tribalities\".\n\n\nBecause the people who do all this work are usually from marginalized populations I've done research projects with, those people are more likely to engage actively in research (whether by looking to create new or expanded scientific papers, because this is not being done; that is usually where resources are sparse), making it harder for others.\n\n\nThese are the people whose lives make them less productive (in terms Of being physically active and experiencing life more often and feeling healthier (I think); most of these studies use \"tragic illness\", in the sense that people with these conditions are most productive and thus less able to make improvements.\n\n\nThe people who take advantage of this have nothing to do with our culture and can participate"
19
+ example_title: "Example 3"
20
+ pipeline_tag: text2text-generation
21
  ---
22
+ # Model Card
23
 
24
+ [Add more information here](https://huggingface.co/templates/model-card-example)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
 
26
+ ## Example Usage
27
 
28
+ ```python3
29
+ from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, pipeline
30
 
31
+ tokenizer = AutoTokenizer.from_pretrained('jeffbritts/abstracts_to_post_model', revision=None) # Load tokenizer
32
+ model = AutoModelForSeq2SeqLM.from_pretrained('jeffbritts/abstracts_to_post_model', revision=None) # Load model
33
+ pipe = pipeline('text2text-generation', model=model, tokenizer=tokenizer, pad_token_id=tokenizer.pad_token_id)
34
 
35
+ inputs = ["Acknowledgments\n\nThank you to all our donors. Their input was invaluable, and many of them have kept this program active. I really appreciate some privacy concerns with these papers and the paper itself. However, thank you to my research team for helping get the entire research protocol up and running since 2010. It's been absolutely stunning for me to be a part of such a small organization, but when something like this happens, it is such a huge deal. It means it's hard not to get involved.\n\nYou will also get a new Open Science Foundation letter if you donate and support NLP. I know I am more than qualified to help you in any way you get involved. Thank you in advance.\n\nAs an additional thanks-good-ness, at the risk of repeating some of a large list, I will do an accompanying Google Hangout. The Hangout is where you can send an email at nlp-doc@umass-edu. In my time as a speaker, we'll do an ongoing Hangout video series and maybe even a live talk. The original YouTube channel is hosted here.\n\nIf you have any questions or concerns or would like to talk to a team member, write to my Open Science Committee through this website below or send your comments directly to me. Thanks."]
36
+ print(pipe(inputs, max_length=512, do_sample=False))
37
+ ```
38
 
39
+ ---
40
+ This model was trained with a synthetic dataset with [DataDreamer 🤖💤](https://datadreamer.dev). The synthetic dataset card and model card can be found [here](datadreamer.json). The training arguments can be found [here](training_args.json).