Kalle Hilsenbek

Bachstelze

AI & ML interests

Combining BERT with instructions for explainable AI: gitlab.com/Bachstelze/instructionbert

Recent Activity

updated a dataset 1 minute ago
Bachstelze/BabyLM-10M-2025-shuffled
updated a model 2 days ago
Bachstelze/smolSynformer
updated a model 2 days ago
Bachstelze/smolSynformerPeft
View all activity

Organizations

None yet

Bachstelze's activity

commented on Announcing AI Energy Score Ratings 2 months ago
view reply

Thanks for your effort in energy efficiency. You worked up my curiosity!
Why do smolLM-135m and smolLm-1.7B nearly have the same score besides a 10 times model size difference? Does the identical context size mostly cause it?
Could you please enable encoder-decoder models? They should be in theory more efficient because the input has to be encoded only once and can be reused in every decoding step.

upvoted an article 3 months ago
view article
Article

Is Attention Interpretable in Transformer-Based Large Language Models? Let’s Unpack the Hype

4
New activity in answerdotai/ModernBERT-base 3 months ago

ModernBART wen?

6
#38 opened 3 months ago by
Fizzarolli