Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,11 @@ library_name: transformers
|
|
10 |
## Introduction
|
11 |
AMD-Llama-135m is a language model trained on AMD MI250 GPUs. Based on LLaMA2 model architecture, this model can be smoothly loaded as LlamaForCausalLM with huggingface transformers. Furthermore, we use the same tokenizer as LLaMA2, enabling it to be a draft model of speculative decoding for LLaMA2 and CodeLlama.
|
12 |
|
|
|
|
|
|
|
|
|
|
|
13 |
```python
|
14 |
import sys
|
15 |
import os
|
|
|
10 |
## Introduction
|
11 |
AMD-Llama-135m is a language model trained on AMD MI250 GPUs. Based on LLaMA2 model architecture, this model can be smoothly loaded as LlamaForCausalLM with huggingface transformers. Furthermore, we use the same tokenizer as LLaMA2, enabling it to be a draft model of speculative decoding for LLaMA2 and CodeLlama.
|
12 |
|
13 |
+
|
14 |
+
## Quickstart
|
15 |
+
AMD-Llama-135m-code-GGUF can be loaded and used via Llama.cpp, here is a program with GUI.
|
16 |
+
|
17 |
+
|
18 |
```python
|
19 |
import sys
|
20 |
import os
|