Model Card: Turkish Triplet Embedding Model (Nomic MoE)

Model Description

This is an embedding model trained on a Turkish triplet corpus, utilizing the dataset emrecan/all-nli-tr. The model is based on Nomic Mixture of Experts (MoE) and achieves strong performance compared to other existing Turkish embedding models.

Intended Use

  • Semantic similarity tasks
  • Text clustering
  • Information retrieval
  • Sentence and document-level embedding generation

Training Details

  • Architecture: Nomic Mixture of Experts (MoE)
  • Dataset: emrecan/all-nli-tr
  • Training Objective: Triplet loss for contrastive learning

Evaluation & Performance

Compared to other Turkish embedding models, this model demonstrates strong performance in capturing semantic relationships within the language. Further evaluations and benchmarks will be shared as they become available.

How to Use

You can use this model with Hugging Face's transformers or sentence-transformers library:

from sentence_transformers import SentenceTransformer

model = SentenceTransformer("your-huggingface-model-name")
embeddings = model.encode(["Merhaba dünya!", "Bugün hava çok güzel."])
print(embeddings)

License & Citation

Please refer to the repository for licensing details and citation instructions.

Downloads last month
13
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for teoyidu/nomic-moe-turkish-v1

Dataset used to train teoyidu/nomic-moe-turkish-v1