Update README.md
Browse files
README.md
CHANGED
@@ -1,14 +1,14 @@
|
|
1 |
---
|
2 |
license: llama2
|
3 |
inference: false
|
4 |
-
tags: [green, llmware-chat,
|
5 |
---
|
6 |
|
7 |
-
# llama-2-chat-ov
|
8 |
|
9 |
-
**llama-2-chat-ov** is an OpenVino int4 quantized version of Llama-2-Chat, providing a fast inference implementation, optimized for AI PCs using Intel GPU, CPU and NPU.
|
10 |
|
11 |
-
[**llama-2-chat**](https://huggingface.co/meta-llama/Llama-2-
|
12 |
|
13 |
|
14 |
### Model Description
|
@@ -16,8 +16,8 @@ tags: [green, llmware-chat, p7, ov, emerald]
|
|
16 |
- **Developed by:** meta-llama
|
17 |
- **Quantified by:** llmware
|
18 |
- **Model type:** llama2
|
19 |
-
- **Parameters:**
|
20 |
-
- **Model Parent:** meta-llama/Llama-2-
|
21 |
- **Language(s) (NLP):** English
|
22 |
- **License:** LLama2 Community License
|
23 |
- **Uses:** Chat and general purpose LLM
|
|
|
1 |
---
|
2 |
license: llama2
|
3 |
inference: false
|
4 |
+
tags: [green, llmware-chat, p13, ov, emerald]
|
5 |
---
|
6 |
|
7 |
+
# llama-2-13b-chat-ov
|
8 |
|
9 |
+
**llama-2-13b-chat-ov** is an OpenVino int4 quantized version of Llama-2-13B-Chat, providing a fast inference implementation, optimized for AI PCs using Intel GPU, CPU and NPU.
|
10 |
|
11 |
+
[**llama-2-13b-chat**](https://huggingface.co/meta-llama/Llama-2-13b-chat-hf) is the official 13b chat finetuned version of Llama2, and is one of the classic and best all-around chat models from 2023.
|
12 |
|
13 |
|
14 |
### Model Description
|
|
|
16 |
- **Developed by:** meta-llama
|
17 |
- **Quantified by:** llmware
|
18 |
- **Model type:** llama2
|
19 |
+
- **Parameters:** 13 billion
|
20 |
+
- **Model Parent:** meta-llama/Llama-2-13b-chat-hf
|
21 |
- **Language(s) (NLP):** English
|
22 |
- **License:** LLama2 Community License
|
23 |
- **Uses:** Chat and general purpose LLM
|