Yosef Worku Alemneh
rasyosef
AI & ML interests
Pretraining, Supervised Fine Tuning, Direct Preference Optimization, Retrieval Augmented Generation (RAG), Function Calling
Recent Activity
updated
a model
12 days ago
rasyosef/bert-amharic-text-embedding-medium
updated
a model
12 days ago
rasyosef/roberta-amharic-text-embedding-base
updated
a model
12 days ago
rasyosef/roberta-amharic-text-embedding-medium
Organizations
None yet
rasyosef's activity
Adding Evaluation Results
#1 opened 5 months ago
by
leaderboard-pr-bot
![](https://cdn-avatars.huggingface.co/v1/production/uploads/655506df9dc61e22c5f9c732/IZGvup0FdVlioPPIPnzZv.jpeg)
Adding Evaluation Results
#3 opened 5 months ago
by
leaderboard-pr-bot
![](https://cdn-avatars.huggingface.co/v1/production/uploads/655506df9dc61e22c5f9c732/IZGvup0FdVlioPPIPnzZv.jpeg)
Phi-2-Instruct-APO: aligned with Anchored Preference Optimization
16
#3 opened 5 months ago
by
rasyosef
[Query-ISSUE] tokenizer.vocab_size is 128000, however len(tokenizer) is 128256, which prevents me from using those other tokens.
1
#34 opened 4 months ago
by
HV-Khurdula
![](https://cdn-avatars.huggingface.co/v1/production/uploads/noauth/4ML45IiW_K2p6S7kesM3Z.jpeg)
What are the start and stop tokens of this model?
1
#40 opened 3 months ago
by
aryaash
Is the BOS token id of 128000 hardcoded into the llama 3.2 tokenizer?
2
#17 opened 4 months ago
by
rasyosef
Mistral-NeMo-Minitron-8B-Chat
5
#5 opened 6 months ago
by
rasyosef
APO Trainer in TRL?
1
#2 opened 5 months ago
by
rasyosef
ChatML template does not work properly
10
#2 opened 6 months ago
by
WasamiKirua
![](https://cdn-avatars.huggingface.co/v1/production/uploads/651d625aff8745149ae562aa/2FNNEU-UlPVYMmhyfNuCp.jpeg)
Collaboration
1
#1 opened 6 months ago
by
deleted
Error when trying to run
1
#1 opened 6 months ago
by
ctranslate2-4you
What changed for people using this model in english?
3
#3 opened 6 months ago
by
migueltalka
Phi 2 Instruct: an instruction following Phi 2 SLM that has undergone SFT and DPO
#132 opened 6 months ago
by
rasyosef
Phi 1.5 Instruct: an instruction following Phi 1.5 model that has undergone SFT and DPO
#89 opened 7 months ago
by
rasyosef
Update README.md
1
#2 opened 7 months ago
by
seyyaw
![](https://cdn-avatars.huggingface.co/v1/production/uploads/1616110059668-6031719ed2c57896177ce767.jpeg)
Duplicate?
1
#2 opened 9 months ago
by
israel
Model card is about Mixtral-8x7B instead of Mixtral-8x22B
1
#3 opened 10 months ago
by
rasyosef
New tokens generated with FP16 inference are only exclamation marks "!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!"
13
#89 opened about 1 year ago
by
rasyosef