Samuel Azran
SamuelAzran
·
AI & ML interests
None yet
Recent Activity
liked
a model
about 12 hours ago
Qwen/QVQ-72B-Preview
liked
a model
12 days ago
openai/whisper-large-v2
liked
a model
18 days ago
ruliad/deepthought-8b-llama-v0.01-alpha
Organizations
None yet
SamuelAzran's activity
New Gemma 2 27B?
2
#3 opened 6 months ago
by
SamuelAzran
Was it train after the latest Huggingface Transformers Gemma fix? if not any update plans?
#4 opened 10 months ago
by
SamuelAzran
Should not be called mixtral, the models made into the moe are yi based
9
#2 opened 12 months ago
by
teknium
How does the MoE work?
3
#5 opened 12 months ago
by
PacmanIncarnate
One or two models during inference?
3
#3 opened 12 months ago
by
Venkman42
You know Mixtral, Llama 2 70b, GPT3.5... Are All Much Better
1
#13 opened about 1 year ago
by
deleted
Awesome- Could you help with pointers on doing same for Other languages(Swedish)?
3
#2 opened about 1 year ago
by
Olofp
QLora or full fine-tuning?
1
#1 opened about 1 year ago
by
SamuelAzran
Was system message used during training?
1
#8 opened about 1 year ago
by
SamuelAzran
NEW! OpenLLMLeaderboard 2023 fall update
20
#356 opened about 1 year ago
by
clefourrier
Did you do full model fine tuning (all layers) or only adapters?
1
#2 opened about 1 year ago
by
SamuelAzran
Can you release a chat version soon ?
11
#8 opened about 1 year ago
by
dong0213
Great work, but why only 2048 context length?
1
#4 opened over 1 year ago
by
SamuelAzran
Would it work well with sequence length > 2048?
2
#1 opened over 1 year ago
by
SamuelAzran
Thank you very much!
10
#2 opened over 1 year ago
by
AiCreatornator
Error running the example code
21
#6 opened almost 2 years ago
by
will33am