GPT-3.5 Turbo fine-tuning and API updates
Summary
OpenAI has released fine-tuning capabilities for GPT-3.5 Turbo, allowing developers to customize models for specific use cases with improved performance, steerability, and output formatting. The update enables fine-tuned GPT-3.5 Turbo to match GPT-4 performance on certain tasks while reducing prompt sizes by up to 90%.
View Cached Full Text
Cached at: 04/20/26, 02:54 PM
Similar Articles
Customizing GPT-3 for your application
OpenAI has launched fine-tuning capabilities for GPT-3, allowing developers to customize the model on their own data via a single CLI command, resulting in improved accuracy, reduced costs, and lower latency for production use cases. Early customers like Keeper Tax, Viable, and Sana Labs report significant accuracy improvements after fine-tuning.
Fine-tuning now available for GPT-4o
OpenAI launches fine-tuning for GPT-4o and GPT-4o mini, allowing developers to customize models with their own datasets at lower costs. The feature includes free training tokens (1M/day for GPT-4o and 2M/day for GPT-4o mini through September 23) and is available to all paid-tier developers.
Introducing GPT-4.1 in the API
OpenAI launches GPT-4.1, GPT-4.1 mini, and GPT-4.1 nano models via API with major improvements in coding (54.6% on SWE-bench), instruction following, and 1M token context windows at lower costs. GPT-4.5 Preview will be deprecated on July 14, 2025.
GPT-4 API general availability and deprecation of older models in the Completions API
OpenAI announced GPT-4 API general availability and deprecated older completion models (GPT-3 base models and text-davinci-003), requiring developers to migrate to new models like gpt-3.5-turbo-instruct or newer by January 4, 2024. Fine-tuned models will need to be retrained on new base models with priority access offered for GPT-3.5 Turbo and GPT-4 fine-tuning.
Introducing GPT-5.1 for developers
OpenAI releases GPT-5.1, a new model in the GPT-5 series that dynamically adapts thinking time based on task complexity, offering 2-3x faster performance than GPT-5 while maintaining frontier intelligence. The release includes extended prompt caching (24-hour retention), new coding tools (apply_patch and shell), and a 'no reasoning' mode for latency-sensitive applications.