README / README.md
cataluna84's picture
Update README.md
fc602df verified
|
raw
history blame
2.23 kB
metadata
title: README
emoji: πŸš€
colorFrom: purple
colorTo: gray
sdk: static
pinned: false

Multilingual language models have many deployment challenges. Deployment Challenges

Can we create multilingual models that maintain performance comparable to their larger models while reducing size, latency and inference speeds running in production with huge batch sizes? MemoryVariations through time

Techniques:

  • Pruning

  • Knowledge Distillation

    • Hidden State-Based Distillation ~ DistillKit | GitHub
    • Distil-Whisper: Robust Knowledge Distillation via Large-Scale Pseudo Labelling
    • On-Policy Distillation of Language Models: Learning from Self-Generated Mistakes
    • Minitron: Compact Language models via Pruning & Knowledge Distillation
    • DistiLLM: Towards Streamlined Distillation for Large Language Models
  • Quantization

    • Quantization Aware Training (QAT)
    • Post Training Quantization (PTQ)
      • KV Cache Quantization
      • Weight & Activation Quantization
  • Low-Rank Factorization

  • Fine-Tuning | GitHub

Techniques

Datasets:

Initial 7 datasets unified, having 6.62M rows which includes the following:

  • Bangla_Alpaca_Orca : Bangle
  • Urdu_Instruct_News_Article_Generation: Urdu
  • Urdu_Instruct_News_Headline_Generation: Urdu
  • Urdu_Instruct_News_Category_Classification: Urdu
  • cidar: Arabic
  • Six_Millions_Instruction_Dataset_For_Arabic_Llm_Ft: Arabic
  • instructv3: English

Get in touch with the team: