LMLK commited on
Commit
71ff2e0
1 Parent(s): bd66412

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -10,6 +10,11 @@ library_name: transformers
10
  ## Introduction
11
  AMD-Llama-135m is a language model trained on AMD MI250 GPUs. Based on LLaMA2 model architecture, this model can be smoothly loaded as LlamaForCausalLM with huggingface transformers. Furthermore, we use the same tokenizer as LLaMA2, enabling it to be a draft model of speculative decoding for LLaMA2 and CodeLlama.
12
 
 
 
 
 
 
13
  ```python
14
  import sys
15
  import os
 
10
  ## Introduction
11
  AMD-Llama-135m is a language model trained on AMD MI250 GPUs. Based on LLaMA2 model architecture, this model can be smoothly loaded as LlamaForCausalLM with huggingface transformers. Furthermore, we use the same tokenizer as LLaMA2, enabling it to be a draft model of speculative decoding for LLaMA2 and CodeLlama.
12
 
13
+
14
+ ## Quickstart
15
+ AMD-Llama-135m-code-GGUF can be loaded and used via Llama.cpp, here is a program with GUI.
16
+
17
+
18
  ```python
19
  import sys
20
  import os