-
-
-
-
-
-
Inference Providers
Active filters:
dpo
SameedHussain/gemma-2-2b-it-Flight-Multi-Turn-V2-DPO
Text Generation
•
Updated
•
2
Siddartha10/outputs_dpo
Text Generation
•
0.1B
•
Updated
•
2
SongTonyLi/gemma-2b-it-SFT-D1_chosen-then-DPO-D2a-HuggingFaceH4-ultrafeedback_binarized-Xlarge
Text Generation
•
3B
•
Updated
•
5
CharlesLi/OpenELM-1_1B-DPO-full-llama-improve-openelm
Text Generation
•
1B
•
Updated
•
2
maxmyn/c4ai-takehome-model-dpo
Text Generation
•
0.1B
•
Updated
•
3
CharlesLi/OpenELM-1_1B-DPO-full-max-4-reward
Text Generation
•
1B
•
Updated
•
2
CharlesLi/OpenELM-1_1B-DPO-full-max-12-reward
Text Generation
•
1B
•
Updated
•
3
DUAL-GPO/phi-2-ipo-chatml-lora-i1
DUAL-GPO/phi-2-ipo-chatml-lora-10k-30k-i1
DUAL-GPO/phi-2-ipo-chatml-lora-20k-40k-i1
DUAL-GPO/phi-2-ipo-chatml-lora-30k-50k-i1
rasyosef/phi-2-apo
LBK95/Llama-2-7b-hf-DPO-LookAhead3_FullEval_TTree1.4_TLoop0.7_TEval0.2_Filter0.2_V2.0
coscotuff/SLFT_Trials_2
Text Generation
•
0.1B
•
Updated
•
5
preethu19/tiny-chatbot-dpo
Avinaash/a100_epoch1IPOBest
Text Generation
•
0.1B
•
Updated
•
2
ravithejads/test_model_sft
Text Generation
•
0.1B
•
Updated
Avinaash/a100_epoch2IPOBest
Text Generation
•
0.1B
•
Updated
•
3
Avinaash/a100_epoch1DPOCurated
Text Generation
•
0.1B
•
Updated
•
2
Avinaash/a100_epoch3DPOCurated
Text Generation
•
0.1B
•
Updated
•
2
Avinaash/a100_epoch3IPOBest
Text Generation
•
0.1B
•
Updated
•
1
Avinaash/a100_epoch2DPOCurated
Text Generation
•
0.1B
•
Updated
•
2
sarthakrw/dpo_model
Text Generation
•
0.1B
•
Updated
•
1
VivekChauhan06/SmolLM-FT-CoEdIT-DPO
Text Generation
•
0.1B
•
Updated
•
2
Avinaash/beta0.3_LR_2e-05_Epoch1_DPO_CuratedDataset
Text Generation
•
0.1B
•
Updated
•
2
Avinaash/beta0.7_LR_2e-05_Epoch1_DPO_CuratedDataset
Text Generation
•
0.1B
•
Updated
•
2
Avinaash/beta0.9_LR_2e-05_Epoch1_DPO_CuratedDataset
Text Generation
•
0.1B
•
Updated
•
2
Avinaash/beta1_LR_2e-05_Epoch1_DPO_CuratedDataset
Text Generation
•
0.1B
•
Updated
•
2
Avinaash/beta1.2_LR_2e-05_Epoch1_DPO_CuratedDataset
Text Generation
•
0.1B
•
Updated
•
2
Avinaash/beta1.4_LR_2e-05_Epoch1_DPO_CuratedDataset
Text Generation
•
0.1B
•
Updated
•
2