Skip to main content
OpenAI

GPT-4o and GPT-4 Fine-Tuning

Experimental access

GPT‑4o and GPT‑4 fine-tuning (text only) were initially developed as part of an experimental access program. As of August 20, 2024, gpt-4o-2024-08-06 fine-tuning has moved from experimental access to General Availability (GA).

We recommend developers opt to fine-tune gpt-4o-2024-08-06 over GPT‑4—it's 2x faster, more than 10x cheaper for inference and 3x cheaper to train, and has higher rate limits. For those still interested in fine-tuning GPT‑4, keep in mind that GPT‑4 fine-tuning may require more work to achieve meaningful improvements over the base model, compared to the substantial gains realized with GPT‑3.5 Turbo fine-tuning.

Pricing

gpt-4o-2024-05-13 and GPT‑4 fine-tuning are offered at the following:

Training
Input usage
Output usage
$45.00 / 1M tokens
$7.50 / 1M tokens
$22.50 / 1M tokens
Training
Input usage
Output usage
$90.00 / 1M tokens
$45.00 / 1M tokens
$90.00 / 1M tokens