Behold, one of the first fine-tunes of Mistral's 7B 0.2 Base model. SatoshiN is trained on 4 epochs 2e-4 learning rate (cosine) of a diverse custom data-set, combined with a polishing round of that same data-set at a 1e-4 linear learning rate. It's a nice assistant that isn't afraid to ask questions, and gather additional information before providing a response to user prompts.
SatoshiN | Base-Model
Wikitext Perplexity: 6.27 | 5.4
**Similar to SOTA, this model runs a bit hot, try using lower temperatures below .5 if experiencing any nonsense)
- Downloads last month
- 175
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.