amezasor commited on
Commit
2b7658c
1 Parent(s): 3bb4b6a

developers update

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -208,7 +208,7 @@ model-index:
208
  **Model Summary:**
209
  Granite-3.0-8B-Base is a decoder-only language model to support a variety of text-to-text generation tasks. It is trained from scratch following a two-stage training strategy. In the first stage, it is trained on 10 trillion tokens sourced from diverse domains. During the second stage, it is further trained on 2 trillion tokens using a carefully curated mix of high-quality data, aiming to enhance its performance on specific tasks.
210
 
211
- - **Developers:** IBM Research
212
  - **GitHub Repository:** [ibm-granite/granite-3.0-language-models](https://github.com/ibm-granite/granite-3.0-language-models)
213
  - **Website**: [Granite Docs](https://www.ibm.com/granite/docs/)
214
  - **Paper:** [Granite 3.0 Language Models](https://github.com/ibm-granite/granite-3.0-language-models/blob/main/paper.pdf)
@@ -217,6 +217,7 @@ Granite-3.0-8B-Base is a decoder-only language model to support a variety of tex
217
 
218
  **Supported Languages:**
219
  English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, and Chinese. Users may finetune Granite 3.0 models for languages beyond these 12 languages.
 
220
  **Intended use:**
221
  Prominent use cases of LLMs in text-to-text generation include summarization, text classification, extraction, question-answering, and more. All Granite Base models are able to handle these tasks as they were trained on a large amount of data from various domains. Moreover, they can serve as baseline to create specialized models for specific application scenarios.
222
 
 
208
  **Model Summary:**
209
  Granite-3.0-8B-Base is a decoder-only language model to support a variety of text-to-text generation tasks. It is trained from scratch following a two-stage training strategy. In the first stage, it is trained on 10 trillion tokens sourced from diverse domains. During the second stage, it is further trained on 2 trillion tokens using a carefully curated mix of high-quality data, aiming to enhance its performance on specific tasks.
210
 
211
+ - **Developers:** IBM Granite Team
212
  - **GitHub Repository:** [ibm-granite/granite-3.0-language-models](https://github.com/ibm-granite/granite-3.0-language-models)
213
  - **Website**: [Granite Docs](https://www.ibm.com/granite/docs/)
214
  - **Paper:** [Granite 3.0 Language Models](https://github.com/ibm-granite/granite-3.0-language-models/blob/main/paper.pdf)
 
217
 
218
  **Supported Languages:**
219
  English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, and Chinese. Users may finetune Granite 3.0 models for languages beyond these 12 languages.
220
+
221
  **Intended use:**
222
  Prominent use cases of LLMs in text-to-text generation include summarization, text classification, extraction, question-answering, and more. All Granite Base models are able to handle these tasks as they were trained on a large amount of data from various domains. Moreover, they can serve as baseline to create specialized models for specific application scenarios.
223