File size: 1,081 Bytes
07c6da0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2cb7930
 
 
 
07c6da0
 
 
 
 
2cb7930
 
 
 
 
 
 
 
 
 
 
 
 
 
07c6da0
2cb7930
 
07c6da0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
---
license: apache-2.0
datasets:
- iamtarun/python_code_instructions_18k_alpaca
language:
- en
pipeline_tag: text-generation
tags:
- code
---

# rahuldshetty/tinyllama-python-gguf

Quantized GGUF model files for [tinyllama-python](https://huggingface.co/rahuldshetty/tinyllama-python).

- Base model: [unsloth/tinyllama-bnb-4bit](https://huggingface.co/unsloth/tinyllama-bnb-4bit)
- Dataset: [iamtarun/python_code_instructions_18k_alpaca](https://huggingface.co/datasets/iamtarun/python_code_instructions_18k_alpaca)
- Training Script: [unslothai: Alpaca + TinyLlama + RoPE Scaling full example.ipynb](https://colab.research.google.com/drive/1AZghoNBQaMDgWJpi4RbffGM1h6raLUj9?usp=sharing)

| Name | Quant method | Size |
| ---- | ---- | ---- |
| [tinyllama-python-unsloth.Q2_K.gguf](https://huggingface.co/rahuldshetty/tinyllama-python-gguf/resolve/main/tinyllama-python-unsloth.Q2_K.gguf) | fp16 | 432 MB  |


## Prompt Format

```
### Instruction:
{instruction}

### Response:
```

## Example

```
### Instruction:
Write a function to find cube of a number.

### Response:
```