Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
2
19
Burning ray
adarksky
Follow
0 followers
ยท
5 following
AI & ML interests
None yet
Recent Activity
reacted
to
merve
's
post
with ๐ฅ
21 days ago
small but mighty ๐ฅ you can fine-tune SmolVLM on an L4 with batch size of 4 and it will only take 16.4 GB VRAM ๐ซฐ๐ป also with gradient accumulation simulated batch size is 16 โจ I made a notebook that includes all the goodies: QLoRA, gradient accumulation, gradient checkpointing with explanations on how they work ๐ https://github.com/huggingface/smollm/blob/main/finetuning/Smol_VLM_FT.ipynb
liked
a model
30 days ago
Qwen/Qwen2.5-Coder-32B-Instruct
updated
a model
about 1 month ago
adarksky/pokemon-DDPM
View all activity
Organizations
spaces
2
Sort:ย Recently updated
Sleeping
๐ฌ
Summer24 Fine Tuning
Sleeping
๐
What Panda
models
4
Sort:ย Recently updated
adarksky/pokemon-DDPM
Unconditional Image Generation
โข
Updated
Nov 11
โข
31
adarksky/bart-base-rel-therapy
Text2Text Generation
โข
Updated
Nov 11
โข
23
adarksky/president-gpt2
Text Generation
โข
Updated
Jul 4
โข
18
adarksky/biden-gpt2
Text Generation
โข
Updated
Jul 3
โข
11
datasets
None public yet