Inference Providers
Active filters: dpo
jpacifico/Chocolatine-2-4B-Instruct-DPO-v2.1
Text Generation
• 4B • Updated • 2.17k
• 8
F16/z-image-turbo-flow-dpo
Feature Extraction
• Updated • 162
sweepai/sweep-next-edit-v2-7B
Text Generation
• 8B • Updated • 105
• 14
mlabonne/NeuralDaredevil-8B-abliterated
Text Generation
• 8B • Updated • 16.3k
• • 267
Text Generation
• 8B • Updated • 693
• 2
bartowski/NeuralDaredevil-8B-abliterated-GGUF
Text Generation
• 8B • Updated • 1.23k
• 12
sapienzanlp/Minerva-7B-instruct-v1.0
Text Generation
• Updated • 4.62k
• 22
RISys-Lab/RedSage-Qwen3-8B-DPO
Text Generation
• 8B • Updated • 841
• 5
mradermacher/RedSage-Qwen3-8B-DPO-GGUF
Text Generation
• 8B • Updated • 324
• 2
HuiyuWang/dpo-qwen-cot-merged
Text Generation
• 4B • Updated • 50
• 1
mradermacher/G-Health-14B-instruct-i1-GGUF
15B • Updated • 15.8k
• 1
MerlinSafety/HybridIntelligence-0.5B
Text Generation
• 0.5B • Updated • 43
• 6
mradermacher/Qwen3.5-35B-A3B-heretic-v2-eq-v1-GGUF
35B • Updated • 2.2k
• 1
danielcherubini/Qwen3.5-DeltaCoder-9B-GGUF
Text Generation
• 9B • Updated • 6.88k
• 7
empero-ai/openNemo-9B-Claude-Opus-4.6-distill
Text Generation
• Updated • 3.17k
• 2
nivvis/Qwen3.5-35B-A3B-EQ-v5
Text Generation
• 35B • Updated • 1.15k
• 5
Fintacorp55/football-llm-q4
Text Generation
• 8B • Updated • 325
• 1
ToastyPigeon/Qwen3.5-27B-Marvin-DPO-V2
Text Generation
• 28B • Updated • 118
• 11
henrik3/sweep-next-edit-v2-7B-GGUF
Text Generation
• 8B • Updated • 614
• 1
Shubhamw11/gemma-3-270m-dpo-negative
Updated • 57
• 1
henrik3/sweep-next-edit-v2-7B-NVFP4
Text Generation
• 5B • Updated • 960
• 1
jasonhuang3/101-caldpo-dataset-dpo-llama3-2-3b-instruct-lora
Updated • 1
lyogavin/Anima33B-DPO-Belle-1k
Text Generation
• Updated • 1
lyogavin/Anima33B-DPO-Belle-1k-merged
Text Generation
• Updated • 5
• 12
daekeun-ml/Llama-2-ko-DPO-13B
Text Generation
• 13B • Updated • 8
• 19
lewtun/zephyr-7b-dpo-full
Text Generation
• 7B • Updated • 5
alignment-handbook/zephyr-7b-dpo-full
Text Generation
• 7B • Updated • 29
• 3
alignment-handbook/zephyr-7b-dpo-qlora
Updated • 14
• 9
Text Generation
• Updated • 7
• 7
argilla/notus-7b-v1-lora-adapter
Text Generation
• Updated • 3