or4cl3ai commited on
Commit
c08713c
1 Parent(s): 275a1f6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +39 -7
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- license: mit
3
  datasets:
4
  - irds/codesearchnet
5
  - giganticode/java-cmpx-v1
@@ -88,6 +88,34 @@ datasets:
88
  - rombodawg/LosslessMegaCodeTrainingV3_MINI
89
  - BelleGroup/multiturn_chat_0.8M
90
  - smangrul/code-chat-assistant-v1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
91
  language:
92
  - en
93
  - it
@@ -98,8 +126,11 @@ language:
98
  - ro
99
  - el
100
  - ja
101
- - ch
102
  - zh
 
 
 
 
103
  metrics:
104
  - accuracy
105
  - bertscore
@@ -109,9 +140,11 @@ metrics:
109
  - brier_score
110
  - cer
111
  - chrf
 
 
112
  tags:
113
  - code
114
- - text-generation-inference
115
  library_name: transformers
116
  pipeline_tag: conversational
117
  ---
@@ -122,7 +155,7 @@ pipeline_tag: conversational
122
 
123
  **Model type:** Large language model
124
 
125
- **Model size:** 175B parameters
126
 
127
  **Intended use:** Aiden T5 is a large language model that can be used for a variety of tasks, including text generation, translation, summarization, and question answering. It is still under development, but it has learned to perform many kinds of tasks surprisingly well.
128
 
@@ -137,11 +170,10 @@ pipeline_tag: conversational
137
  **How to use Aiden T5:** Aiden T5 can be used through the Hugging Face Hub. To use Aiden T5, simply create a new project and select the Aiden T5 model. You can then use Aiden T5 to generate text, translate languages, summarize text, and answer questions.
138
 
139
 
140
- The number of parameters in a machine learning model is a measure of its complexity. Aiden T5 has 175B parameters, which makes it one of the largest and most complex language models ever created.
141
 
142
  The number of parameters is important because it affects the model's ability to learn from data. A model with more parameters can learn more complex relationships between the input and output data. However, a model with too many parameters can be overfitting, which means that it learns the training data too well and does not generalize well to new data.
143
 
144
  The developers of Aiden T5 have carefully tuned the number of parameters to achieve a good balance between learning and generalization. As a result, Aiden T5 is able to learn complex relationships from the training data and generalize well to new data.
145
 
146
- This is why Aiden T5 is able to perform many kinds of tasks surprisingly well, even though it is still under development.
147
-
 
1
  ---
2
+ license: openrail
3
  datasets:
4
  - irds/codesearchnet
5
  - giganticode/java-cmpx-v1
 
88
  - rombodawg/LosslessMegaCodeTrainingV3_MINI
89
  - BelleGroup/multiturn_chat_0.8M
90
  - smangrul/code-chat-assistant-v1
91
+ - goendalf666/sales-textbook_for_convincing_and_selling
92
+ - readerbench/ConversationalAgent-Ro
93
+ - beurkinger/autotrain-data-human-action-recognition
94
+ - jpwahle/autoencoder-paraphrase-dataset
95
+ - jpwahle/autoregressive-paraphrase-dataset
96
+ - teknium/GPT4-LLM-Cleaned
97
+ - Anthropic/model-written-evals
98
+ - openai_humaneval
99
+ - kye/all-google-ai-python-code
100
+ - kye/all-openai-github-code
101
+ - EleutherAI/lambada_openai
102
+ - CShorten/ML-ArXiv-Papers
103
+ - WaltonFuture/InstructionGPT-4
104
+ - open-llm-leaderboard/details_AIDC-ai-business__Marcoroni-70B
105
+ - seansullivan/INT-Business-Syllabus
106
+ - theoldmandthesea/17k_business_book
107
+ - SunRise228/business-doc
108
+ - gauravshrm211/VC-startup-evaluation-for-investment
109
+ - TuningAI/Startups_V1
110
+ - TuningAI/Startups_V2
111
+ - AdiOO7/llama-2-finance
112
+ - scillm/scientific_papers
113
+ - gokuls/wiki_book_corpus_complete_processed_bert_dataset
114
+ - the_pile_books3
115
+ - go_emotions
116
+ - yizhongw/self_instruct
117
+ - codeparrot/self-instruct-starcoder
118
+ - Amani27/massive_translation_dataset
119
  language:
120
  - en
121
  - it
 
126
  - ro
127
  - el
128
  - ja
 
129
  - zh
130
+ - ga
131
+ - cy
132
+ - gd
133
+ - de
134
  metrics:
135
  - accuracy
136
  - bertscore
 
140
  - brier_score
141
  - cer
142
  - chrf
143
+ - charcut_mt
144
+ - bleurt
145
  tags:
146
  - code
147
+ - conversational
148
  library_name: transformers
149
  pipeline_tag: conversational
150
  ---
 
155
 
156
  **Model type:** Large language model
157
 
158
+ **Model size:** 248B parameters
159
 
160
  **Intended use:** Aiden T5 is a large language model that can be used for a variety of tasks, including text generation, translation, summarization, and question answering. It is still under development, but it has learned to perform many kinds of tasks surprisingly well.
161
 
 
170
  **How to use Aiden T5:** Aiden T5 can be used through the Hugging Face Hub. To use Aiden T5, simply create a new project and select the Aiden T5 model. You can then use Aiden T5 to generate text, translate languages, summarize text, and answer questions.
171
 
172
 
173
+ The number of parameters in a machine learning model is a measure of its complexity. Aiden T5 has 248B parameters, which makes it one of the largest and most complex language models ever created.
174
 
175
  The number of parameters is important because it affects the model's ability to learn from data. A model with more parameters can learn more complex relationships between the input and output data. However, a model with too many parameters can be overfitting, which means that it learns the training data too well and does not generalize well to new data.
176
 
177
  The developers of Aiden T5 have carefully tuned the number of parameters to achieve a good balance between learning and generalization. As a result, Aiden T5 is able to learn complex relationships from the training data and generalize well to new data.
178
 
179
+ This is why Aiden T5 is able to perform many kinds of tasks surprisingly well, even though it is still under development.