Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
---
|
2 |
-
license:
|
3 |
datasets:
|
4 |
- irds/codesearchnet
|
5 |
- giganticode/java-cmpx-v1
|
@@ -88,6 +88,34 @@ datasets:
|
|
88 |
- rombodawg/LosslessMegaCodeTrainingV3_MINI
|
89 |
- BelleGroup/multiturn_chat_0.8M
|
90 |
- smangrul/code-chat-assistant-v1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
91 |
language:
|
92 |
- en
|
93 |
- it
|
@@ -98,8 +126,11 @@ language:
|
|
98 |
- ro
|
99 |
- el
|
100 |
- ja
|
101 |
-
- ch
|
102 |
- zh
|
|
|
|
|
|
|
|
|
103 |
metrics:
|
104 |
- accuracy
|
105 |
- bertscore
|
@@ -109,9 +140,11 @@ metrics:
|
|
109 |
- brier_score
|
110 |
- cer
|
111 |
- chrf
|
|
|
|
|
112 |
tags:
|
113 |
- code
|
114 |
-
-
|
115 |
library_name: transformers
|
116 |
pipeline_tag: conversational
|
117 |
---
|
@@ -122,7 +155,7 @@ pipeline_tag: conversational
|
|
122 |
|
123 |
**Model type:** Large language model
|
124 |
|
125 |
-
**Model size:**
|
126 |
|
127 |
**Intended use:** Aiden T5 is a large language model that can be used for a variety of tasks, including text generation, translation, summarization, and question answering. It is still under development, but it has learned to perform many kinds of tasks surprisingly well.
|
128 |
|
@@ -137,11 +170,10 @@ pipeline_tag: conversational
|
|
137 |
**How to use Aiden T5:** Aiden T5 can be used through the Hugging Face Hub. To use Aiden T5, simply create a new project and select the Aiden T5 model. You can then use Aiden T5 to generate text, translate languages, summarize text, and answer questions.
|
138 |
|
139 |
|
140 |
-
The number of parameters in a machine learning model is a measure of its complexity. Aiden T5 has
|
141 |
|
142 |
The number of parameters is important because it affects the model's ability to learn from data. A model with more parameters can learn more complex relationships between the input and output data. However, a model with too many parameters can be overfitting, which means that it learns the training data too well and does not generalize well to new data.
|
143 |
|
144 |
The developers of Aiden T5 have carefully tuned the number of parameters to achieve a good balance between learning and generalization. As a result, Aiden T5 is able to learn complex relationships from the training data and generalize well to new data.
|
145 |
|
146 |
-
This is why Aiden T5 is able to perform many kinds of tasks surprisingly well, even though it is still under development.
|
147 |
-
|
|
|
1 |
---
|
2 |
+
license: openrail
|
3 |
datasets:
|
4 |
- irds/codesearchnet
|
5 |
- giganticode/java-cmpx-v1
|
|
|
88 |
- rombodawg/LosslessMegaCodeTrainingV3_MINI
|
89 |
- BelleGroup/multiturn_chat_0.8M
|
90 |
- smangrul/code-chat-assistant-v1
|
91 |
+
- goendalf666/sales-textbook_for_convincing_and_selling
|
92 |
+
- readerbench/ConversationalAgent-Ro
|
93 |
+
- beurkinger/autotrain-data-human-action-recognition
|
94 |
+
- jpwahle/autoencoder-paraphrase-dataset
|
95 |
+
- jpwahle/autoregressive-paraphrase-dataset
|
96 |
+
- teknium/GPT4-LLM-Cleaned
|
97 |
+
- Anthropic/model-written-evals
|
98 |
+
- openai_humaneval
|
99 |
+
- kye/all-google-ai-python-code
|
100 |
+
- kye/all-openai-github-code
|
101 |
+
- EleutherAI/lambada_openai
|
102 |
+
- CShorten/ML-ArXiv-Papers
|
103 |
+
- WaltonFuture/InstructionGPT-4
|
104 |
+
- open-llm-leaderboard/details_AIDC-ai-business__Marcoroni-70B
|
105 |
+
- seansullivan/INT-Business-Syllabus
|
106 |
+
- theoldmandthesea/17k_business_book
|
107 |
+
- SunRise228/business-doc
|
108 |
+
- gauravshrm211/VC-startup-evaluation-for-investment
|
109 |
+
- TuningAI/Startups_V1
|
110 |
+
- TuningAI/Startups_V2
|
111 |
+
- AdiOO7/llama-2-finance
|
112 |
+
- scillm/scientific_papers
|
113 |
+
- gokuls/wiki_book_corpus_complete_processed_bert_dataset
|
114 |
+
- the_pile_books3
|
115 |
+
- go_emotions
|
116 |
+
- yizhongw/self_instruct
|
117 |
+
- codeparrot/self-instruct-starcoder
|
118 |
+
- Amani27/massive_translation_dataset
|
119 |
language:
|
120 |
- en
|
121 |
- it
|
|
|
126 |
- ro
|
127 |
- el
|
128 |
- ja
|
|
|
129 |
- zh
|
130 |
+
- ga
|
131 |
+
- cy
|
132 |
+
- gd
|
133 |
+
- de
|
134 |
metrics:
|
135 |
- accuracy
|
136 |
- bertscore
|
|
|
140 |
- brier_score
|
141 |
- cer
|
142 |
- chrf
|
143 |
+
- charcut_mt
|
144 |
+
- bleurt
|
145 |
tags:
|
146 |
- code
|
147 |
+
- conversational
|
148 |
library_name: transformers
|
149 |
pipeline_tag: conversational
|
150 |
---
|
|
|
155 |
|
156 |
**Model type:** Large language model
|
157 |
|
158 |
+
**Model size:** 248B parameters
|
159 |
|
160 |
**Intended use:** Aiden T5 is a large language model that can be used for a variety of tasks, including text generation, translation, summarization, and question answering. It is still under development, but it has learned to perform many kinds of tasks surprisingly well.
|
161 |
|
|
|
170 |
**How to use Aiden T5:** Aiden T5 can be used through the Hugging Face Hub. To use Aiden T5, simply create a new project and select the Aiden T5 model. You can then use Aiden T5 to generate text, translate languages, summarize text, and answer questions.
|
171 |
|
172 |
|
173 |
+
The number of parameters in a machine learning model is a measure of its complexity. Aiden T5 has 248B parameters, which makes it one of the largest and most complex language models ever created.
|
174 |
|
175 |
The number of parameters is important because it affects the model's ability to learn from data. A model with more parameters can learn more complex relationships between the input and output data. However, a model with too many parameters can be overfitting, which means that it learns the training data too well and does not generalize well to new data.
|
176 |
|
177 |
The developers of Aiden T5 have carefully tuned the number of parameters to achieve a good balance between learning and generalization. As a result, Aiden T5 is able to learn complex relationships from the training data and generalize well to new data.
|
178 |
|
179 |
+
This is why Aiden T5 is able to perform many kinds of tasks surprisingly well, even though it is still under development.
|
|