Update README.md
Browse files
README.md
CHANGED
@@ -21,7 +21,7 @@ We introduce Bamba-9B, a decoder-only language model based on the [Mamba-2](http
|
|
21 |
The current release includes the following models:
|
22 |
| **Stage** | **Bamba 9B** | **Quantized** | **Note** |
|
23 |
|----------------------|----------------------------------------------------------------------|-------------------------------------------------------------------------|-------------------------------------------------------------------|
|
24 |
-
| **Base Model** | [ibm-fms/Bamba-9B](https://huggingface.co/ibm-fms/Bamba-9B)
|
25 |
| **Base Model** | [ibm-fms/Bamba-9B-2T](https://huggingface.co/ibm-fms/Bamba-9B-2T) | [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Stage 1 pretraining |
|
26 |
| **Base Model** | [ibm-fms/Bamba-9B-1.8T](https://huggingface.co/ibm-fms/Bamba-9B-1.8T)| [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Intermediate checkpoints during Stage 1, more to come |
|
27 |
| **SFT** | coming soon | coming soon | to be released in the next drop |
|
@@ -64,8 +64,8 @@ You can utilize our newly contributed HF integration to run inference on our Bam
|
|
64 |
```python
|
65 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
66 |
|
67 |
-
model = AutoModelForCausalLM.from_pretrained("ibm-fms/Bamba-9B")
|
68 |
-
tokenizer = AutoTokenizer.from_pretrained("ibm-fms/Bamba-9B")
|
69 |
|
70 |
message = ["Mamba is a snake with following properties "]
|
71 |
inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False)
|
|
|
21 |
The current release includes the following models:
|
22 |
| **Stage** | **Bamba 9B** | **Quantized** | **Note** |
|
23 |
|----------------------|----------------------------------------------------------------------|-------------------------------------------------------------------------|-------------------------------------------------------------------|
|
24 |
+
| **Base Model** | [ibm-fms/Bamba-9B-v1](https://huggingface.co/ibm-fms/Bamba-9B-v1) | [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Stage 2 pretraining |
|
25 |
| **Base Model** | [ibm-fms/Bamba-9B-2T](https://huggingface.co/ibm-fms/Bamba-9B-2T) | [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Stage 1 pretraining |
|
26 |
| **Base Model** | [ibm-fms/Bamba-9B-1.8T](https://huggingface.co/ibm-fms/Bamba-9B-1.8T)| [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Intermediate checkpoints during Stage 1, more to come |
|
27 |
| **SFT** | coming soon | coming soon | to be released in the next drop |
|
|
|
64 |
```python
|
65 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
66 |
|
67 |
+
model = AutoModelForCausalLM.from_pretrained("ibm-fms/Bamba-9B-v1")
|
68 |
+
tokenizer = AutoTokenizer.from_pretrained("ibm-fms/Bamba-9B-v1")
|
69 |
|
70 |
message = ["Mamba is a snake with following properties "]
|
71 |
inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False)
|