SPLADE sparse retrieval models based on BERT-Tiny (4M) and BERT-Mini (11M) distilled from a Cross-Encoder on the MSMARCO dataset
Yosef Worku Alemneh
rasyosef
AI & ML interests
Pretraining, Supervised Fine Tuning, Direct Preference Optimization, Retrieval Augmented Generation (RAG), Function Calling
Recent Activity
updated
a model
about 18 hours ago
rasyosef/splade-small
updated
a collection
about 19 hours ago
SPLADE-Tiny-MSMARCO
published
a model
about 19 hours ago
rasyosef/splade-small
Organizations
Llama 3.2 Amharic
Llama 3.2 decoder transformer models trained on Amharic text
-
rasyosef/Llama-3.2-400M-Amharic
Text Generation • 0.4B • Updated • 101 • 3 -
rasyosef/Llama-3.2-400M-Amharic-Instruct-Poems-Stories-Wikipedia
Text Generation • 0.4B • Updated • 36 • 1 -
rasyosef/Llama-3.2-400M-Amharic-Instruct
Text Generation • 0.4B • Updated • 2 -
rasyosef/Llama-3.2-180M-Amharic
Text Generation • 0.2B • Updated • 3 • 1
Amharic GPT2
GPT2 transformer decoder models pretrained on 290 million tokens of Amharic text
Phi 2 Chat Models
These models underwent supervised fine-tuning and direct preference optimization for instruction following on top of Microsoft's Phi 2 base LLM
Amharic Text Embedding Models
Text Embedding and ColBERT models based on Amharic RoBERTa and BERT for Amharic passage retrieval
-
Optimized Text Embedding Models and Benchmarks for Amharic Passage Retrieval
Paper • 2505.19356 • Published -
rasyosef/roberta-amharic-text-embedding-base
Sentence Similarity • 0.1B • Updated • 23 -
rasyosef/colbert-roberta-amharic-base
Sentence Similarity • 0.1B • Updated • 15 -
rasyosef/roberta-amharic-text-embedding-medium
Sentence Similarity • 0.0B • Updated • 11
Amharic BERT and RoBERTa
BERT and RoBERTa transformer encoder models pretrained on 290 million tokens of Amharic text
Phi 1.5 Chat Models
These models underwent supervised fine-tuning and direct preference optimization for instruction following on top of Microsoft's Phi 1.5 base LLM
Minitron Chat Models
Instruction-tuned (chat) versions of Nvidia's Minitron base models created through supervised fine-tuning (SFT)
SPLADE-Tiny-MSMARCO
SPLADE sparse retrieval models based on BERT-Tiny (4M) and BERT-Mini (11M) distilled from a Cross-Encoder on the MSMARCO dataset
Amharic Text Embedding Models
Text Embedding and ColBERT models based on Amharic RoBERTa and BERT for Amharic passage retrieval
-
Optimized Text Embedding Models and Benchmarks for Amharic Passage Retrieval
Paper • 2505.19356 • Published -
rasyosef/roberta-amharic-text-embedding-base
Sentence Similarity • 0.1B • Updated • 23 -
rasyosef/colbert-roberta-amharic-base
Sentence Similarity • 0.1B • Updated • 15 -
rasyosef/roberta-amharic-text-embedding-medium
Sentence Similarity • 0.0B • Updated • 11
Llama 3.2 Amharic
Llama 3.2 decoder transformer models trained on Amharic text
-
rasyosef/Llama-3.2-400M-Amharic
Text Generation • 0.4B • Updated • 101 • 3 -
rasyosef/Llama-3.2-400M-Amharic-Instruct-Poems-Stories-Wikipedia
Text Generation • 0.4B • Updated • 36 • 1 -
rasyosef/Llama-3.2-400M-Amharic-Instruct
Text Generation • 0.4B • Updated • 2 -
rasyosef/Llama-3.2-180M-Amharic
Text Generation • 0.2B • Updated • 3 • 1
Amharic BERT and RoBERTa
BERT and RoBERTa transformer encoder models pretrained on 290 million tokens of Amharic text
Amharic GPT2
GPT2 transformer decoder models pretrained on 290 million tokens of Amharic text
Phi 1.5 Chat Models
These models underwent supervised fine-tuning and direct preference optimization for instruction following on top of Microsoft's Phi 1.5 base LLM
Phi 2 Chat Models
These models underwent supervised fine-tuning and direct preference optimization for instruction following on top of Microsoft's Phi 2 base LLM
Minitron Chat Models
Instruction-tuned (chat) versions of Nvidia's Minitron base models created through supervised fine-tuning (SFT)