divykum commited on
Commit
e6ea002
·
verified ·
1 Parent(s): 0cbf233

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -21,7 +21,7 @@ We introduce Bamba-9B, a decoder-only language model based on the [Mamba-2](http
21
  The current release includes the following models:
22
  | **Stage** | **Bamba 9B** | **Quantized** | **Note** |
23
  |----------------------|----------------------------------------------------------------------|-------------------------------------------------------------------------|-------------------------------------------------------------------|
24
- | **Base Model** | [ibm-fms/Bamba-9B](https://huggingface.co/ibm-fms/Bamba-9B) | [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Stage 2 pretraining |
25
  | **Base Model** | [ibm-fms/Bamba-9B-2T](https://huggingface.co/ibm-fms/Bamba-9B-2T) | [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Stage 1 pretraining |
26
  | **Base Model** | [ibm-fms/Bamba-9B-1.8T](https://huggingface.co/ibm-fms/Bamba-9B-1.8T)| [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Intermediate checkpoints during Stage 1, more to come |
27
  | **SFT** | coming soon | coming soon | to be released in the next drop |
@@ -64,8 +64,8 @@ You can utilize our newly contributed HF integration to run inference on our Bam
64
  ```python
65
  from transformers import AutoModelForCausalLM, AutoTokenizer
66
 
67
- model = AutoModelForCausalLM.from_pretrained("ibm-fms/Bamba-9B")
68
- tokenizer = AutoTokenizer.from_pretrained("ibm-fms/Bamba-9B")
69
 
70
  message = ["Mamba is a snake with following properties "]
71
  inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False)
 
21
  The current release includes the following models:
22
  | **Stage** | **Bamba 9B** | **Quantized** | **Note** |
23
  |----------------------|----------------------------------------------------------------------|-------------------------------------------------------------------------|-------------------------------------------------------------------|
24
+ | **Base Model** | [ibm-fms/Bamba-9B-v1](https://huggingface.co/ibm-fms/Bamba-9B-v1) | [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Stage 2 pretraining |
25
  | **Base Model** | [ibm-fms/Bamba-9B-2T](https://huggingface.co/ibm-fms/Bamba-9B-2T) | [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Stage 1 pretraining |
26
  | **Base Model** | [ibm-fms/Bamba-9B-1.8T](https://huggingface.co/ibm-fms/Bamba-9B-1.8T)| [ibm-fms/Bamba-9B-fp8](https://huggingface.co/ibm-fms/Bamba-9B-fp8) | Intermediate checkpoints during Stage 1, more to come |
27
  | **SFT** | coming soon | coming soon | to be released in the next drop |
 
64
  ```python
65
  from transformers import AutoModelForCausalLM, AutoTokenizer
66
 
67
+ model = AutoModelForCausalLM.from_pretrained("ibm-fms/Bamba-9B-v1")
68
+ tokenizer = AutoTokenizer.from_pretrained("ibm-fms/Bamba-9B-v1")
69
 
70
  message = ["Mamba is a snake with following properties "]
71
  inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False)