Model Description

This model is based of llama-68m and was initialized from this model with tied embeddings and then further pretrained on the uncleaned c4 dataset. This model is obviously a beta based on the name and has not been trained on the full training set which is the first 1 million samples of the uncleaned c4 dataset as mentioned above.

Downloads last month
37
Safetensors
Model size
43.5M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for cpayne1303/llama-43m-beta

Base model

JackFram/llama-68m
Finetuned
(15)
this model

Dataset used to train cpayne1303/llama-43m-beta