Introducing Direct Preference Optimization (DPO) Support on OpenPipe

Introducing Direct Preference Optimization (DPO) Support on OpenPipe

Posts

One Right Answer or Many? A Useful Distinction for Evaluating and Fine-Tuning LLMs

Jan 14, 2025

Analyzing OpenAI’s Reinforcement Fine-Tuning: Less Data, Better Results

Dec 30, 2024

Using Reinforcement Learning and $4.80 of GPU Time to Find the Best HN Post Ever (RLHF Part 1)

Oct 28, 2024

A Founder’s Guide to AI Fine-Tuning

Oct 11, 2024

Introducing Direct Preference Optimization (DPO) Support on OpenPipe

Oct 1, 2024

Fine-tuning Best Practices Chapter 2: Models

Aug 28, 2024

Fine-tuning Best Practices Series Introduction and Chapter 1: Training Data

Aug 1, 2024

Announcing Llama 3.1 and GPT-4o Mini fine-tuning through OpenPipe!

Jul 24, 2024

OpenPipe Mixture of Agents: Outperform GPT-4 at 1/25th the Cost

Jun 20, 2024

The Ten Commandments of Fine-Tuning in Prod (a Mastering LLMs Conference Talk)

May 23, 2024

What we've learned in 3 days of Llama 3

Apr 21, 2024

Fine-Tuning in a Nutshell

Mar 28, 2024

We Raised $6.7M to Replace GPT-4 with Your Own Fine-Tuned Models

Mar 25, 2024

Mixtral Curious? Comparing Mistral 7B and Mixtral for fine-tuning

Feb 29, 2024

S-LoRA: Serving Thousands of Models From One GPU for Fun and Profit

Jan 17, 2024

Axis logo
Axis logo

Axis Improves Generation Quality and Lowers Costs With Fine Tuning

Jan 4, 2024

Product Updates December 2023

Jan 3, 2024

How we built “Mistral 7B Fine-Tune Optimized,” the best 7B model for fine-tuning

Dec 18, 2023

Announcing Automatic Evals for Fine-tuned Models

Dec 1, 2023

Is AI the Next Crypto? Insights from 2M HN comments

Nov 8, 2023

Llama 2 vs Mistral: Believe the Hype

Nov 5, 2023

White llama on orange background
White llama on orange background

Fine-tune your own Llama 2 to replace GPT-3.5/4

Sep 12, 2023

From Prompts to Models

Aug 28, 2023