Quantizations of https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b

Inference Clients/UIs


From original readme

Prompt Format

The model follows the Alpaca prompt format:

### Instruction:
<prompt>

### Response:
<leave a newline blank for model to respond>

or

### Instruction:
<prompt>

### Input:
<additional context>

### Response:
<leave a newline blank for model to respond>
Downloads last month
331
GGUF
Model size
13B params
Architecture
llama

1-bit

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Examples
Inference API (serverless) has been turned off for this model.