Model Card: Turkish Triplet Embedding Model (Nomic MoE)

Model Description

This is an embedding model trained on a Turkish triplet corpus, utilizing the dataset emrecan/all-nli-tr. The model is based on Nomic Mixture of Experts (MoE) and achieves strong performance compared to other existing Turkish embedding models.

Intended Use

  • Semantic similarity tasks
  • Text clustering
  • Information retrieval
  • Sentence and document-level embedding generation

Training Details

  • Architecture: Nomic Mixture of Experts (MoE)
  • Dataset: emrecan/all-nli-tr
  • Training Objective: Triplet loss for contrastive learning

Evaluation & Performance

Compared to other Turkish embedding models, this model demonstrates strong performance in capturing semantic relationships within the language. Further evaluations and benchmarks will be shared as they become available.

How to Use

You can use this model with Hugging Face's transformers or sentence-transformers library:

from sentence_transformers import SentenceTransformer

model = SentenceTransformer("your-huggingface-model-name")
embeddings = model.encode(["Merhaba dünya!", "Bugün hava çok güzel."])
print(embeddings)

License & Citation

Please refer to the repository for licensing details and citation instructions.

Downloads last month
21
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The HF Inference API does not support zero-shot-classification models for sentence-transformers library.

Model tree for teoyidu/nomic-moe-turkish-v1

Dataset used to train teoyidu/nomic-moe-turkish-v1