Tag
OpenAI introduces Prompt Caching, an automatic feature that reduces API costs by 50% and improves latency by reusing recently cached input tokens on GPT-4o, GPT-4o mini, o1-preview, and o1-mini models. The feature automatically applies to prompts longer than 1,024 tokens without requiring developer integration changes.
OpenAI launches fine-tuning for GPT-4o and GPT-4o mini, allowing developers to customize models with their own datasets at lower costs. The feature includes free training tokens (1M/day for GPT-4o and 2M/day for GPT-4o mini through September 23) and is available to all paid-tier developers.
OpenAI announces new Edit and Insert capabilities for GPT-3 and Codex, enabling mid-file code completion and text editing. The Insert feature is being piloted in GitHub Copilot and is now available in beta via the completions API.