cognitivecomputations/dolphin-2.9-llama3-8b AWQ

Curated and trained by Eric Hartford, Lucas Atkins, and Fernando Fernandes, and Cognitive Computations

Discord: https://discord.gg/8fbBeC7ZGx

My appreciation for the sponsors of Dolphin 2.9:

This model is based on Llama-3-8b, and is governed by META LLAMA 3 COMMUNITY LICENSE AGREEMENT

The base model has 8k context, and the full-weight fine-tuning was with 4k sequence length.

It took 2.5 days on 8x L40S provided by Crusoe Cloud

This model was trained FFT on all parameters, using ChatML prompt template format.

example:

<|im_start|>system
You are Dolphin, a helpful AI assistant.<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
Downloads last month
16
Safetensors
Model size
1.98B params
Tensor type
I32
·
FP16
·
Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for solidrust/dolphin-2.9-llama3-8b-AWQ

Quantized
(240)
this model

Datasets used to train solidrust/dolphin-2.9-llama3-8b-AWQ

Collection including solidrust/dolphin-2.9-llama3-8b-AWQ