README / README.md
cataluna84's picture
Initial Commit
1fae59e verified
|
raw
history blame
437 Bytes
metadata
title: README
emoji: πŸ“Š
colorFrom: purple
colorTo: gray
sdk: static
pinned: false

Multilingual language models are typically large, requiring significant computational resources.

Can we create multilingual models that maintain performance comparable to their larger models while reducing size, latency and inference speeds?

Potential Techniques:

  • Pruning
    • SparseGPT
    • ShortGPT
  • Knowledge Distillation
  • Quantization