Posts
Using Reinforcement Learning and $4.80 of GPU Time to Find the Best HN Post Ever (RLHF Part 1)
Oct 28, 2024
A Founder’s Guide to AI Fine-Tuning
Oct 11, 2024
Introducing Direct Preference Optimization (DPO) Support on OpenPipe
Oct 1, 2024
Fine-tuning Best Practices Chapter 2: Models
Aug 28, 2024
Fine-tuning Best Practices Series Introduction and Chapter 1: Training Data
Aug 1, 2024
Announcing Llama 3.1 and GPT-4o Mini fine-tuning through OpenPipe!
Jul 24, 2024
OpenPipe Mixture of Agents: Outperform GPT-4 at 1/25th the Cost
Jun 20, 2024
The Ten Commandments of Fine-Tuning in Prod (a Mastering LLMs Conference Talk)
May 23, 2024
What we've learned in 3 days of Llama 3
Apr 21, 2024
Fine-Tuning in a Nutshell
Mar 28, 2024
We Raised $6.7M to Replace GPT-4 with Your Own Fine-Tuned Models
Mar 25, 2024
Mixtral Curious? Comparing Mistral 7B and Mixtral for fine-tuning
Feb 29, 2024
S-LoRA: Serving Thousands of Models From One GPU for Fun and Profit
Jan 17, 2024
Axis Improves Generation Quality and Lowers Costs With Fine Tuning
Jan 4, 2024
Product Updates December 2023
Jan 3, 2024
How we built “Mistral 7B Fine-Tune Optimized,” the best 7B model for fine-tuning
Dec 18, 2023
Announcing Automatic Evals for Fine-tuned Models
Dec 1, 2023
Is AI the Next Crypto? Insights from 2M HN comments
Nov 8, 2023
Llama 2 vs Mistral: Believe the Hype
Nov 5, 2023
Fine-tune your own Llama 2 to replace GPT-3.5/4
Sep 12, 2023
From Prompts to Models
Aug 28, 2023