New models and developer products announced at DevDay

OpenAI News
New models and developer products announced at DevDay

_Update: We previously stated that applications using the_`_gpt-3.5-turbo_`_name will automatically be upgraded to the new model version on December 11. We have edited the blogpost to remove this line since this will no longer be happening._

Today, we shared dozens of new additions and improvements, and reduced pricing across many parts of our platform. These include:

We’ll begin rolling out new features to OpenAI customers starting at 1pm PT today.

_Learn more about__OpenAI DevDay announcements for ChatGPT_⁠_._

## GPT-4 Turbo with 128K context

We released the first version of GPT‑4 in March and made GPT‑4 generally available to all developers in July. Today we’re launching a preview of the next generation of this model, GPT‑4 Turbo⁠(opens in a new window).

GPT‑4 Turbo is more capable and has knowledge of world events up to April 2023. It has a 128k context window so it can fit the equivalent of more than 300 pages of text in a single prompt. We also optimized its performance so we are able to offer GPT‑4 Turbo at a 3x cheaper⁠ price for input tokens and a 2x cheaper price for output tokens compared to GPT‑4.

GPT‑4 Turbo is available for all paying developers to try by passing `gpt-4-1106-preview` in the API and we plan to release the stable production-ready model in the coming weeks.

### Function calling updates

Function calling⁠(opens in a new window) lets you describe functions of your app or external APIs to models, and have the model intelligently choose to output a JSON object containing arguments to call those functions. We’re releasing several improvements today, including the ability to call multiple functions in a single message: users can send one message requesting multiple actions, such as “open the car window and turn off the A/C”,which would previously require multiple roundtrips with the model (learn more⁠(opens in a new window)). We are also improving function calling accuracy: GPT‑4 Turbo is more likely to return the right function parameters.

### Improved instruction following and JSON mode

GPT‑4 Turbo performs better than our previous models on tasks that require the careful following of instructions, such as generating specific formats (e.g., “always respond in XML”). It also supports our new JSON mode⁠(opens in a new window), which ensures the model will respond with valid JSON. The new API parameter `response_format` enables the model to constrain its output to generate a syntactically correct JSON object. JSON mode is useful for developers generating JSON in the Chat Completions API outside of function calling.

### Reproducible outputs and log probabilities

The new `seed` parameter enables reproducible outputs by making the model return consistent completions most of the time. This beta feature is useful for use cases such as replaying requests for debugging, writing more comprehensive unit tests, and generally having a higher degree of control over the model behavior. We at OpenAI have been using this feature internally for our own unit tests and have found it invaluable. We’re excited to see how developers will use it. Learn more⁠(opens in a new window).

We’re also launching a feature to return the log probabilities for the most likely output tokens generated by GPT‑4 Turbo and GPT‑3.5 Turbo in the next few weeks, which will be useful for building features such as autocomplete in a search experience.

### Updated GPT-3.5 Turbo

In addition to GPT‑4 Turbo, we are also releasing a new version of GPT‑3.5 Turbo that supports a 16K context window by default. The new 3.5 Turbo supports improved instruction following, JSON mode, and parallel function calling. For instance, our internal evals show a 38% improvement on format following tasks such as generating JSON, XML and YAML. Developers can access this new model by calling `gpt-3.5-turbo-1106` in the API. Older models will continue to be accessible by passing `gpt-3.5-turbo-0613` in the API until June 13, 2024. Learn more⁠(opens in a new window).

## Assistants API, Retrieval, and Code Interpreter

Today, we’re releasing the Assistants API⁠(opens in a new window), our first step towards helping developers build agent-like experiences within their own applications. An assistant is a purpose-built AI that has specific instructions, leverages extra knowledge, and can call models and tools to perform tasks. The new Assistants API provides new capabilities such as Code Interpreter and Retrieval as well as function calling to handle a lot of the heavy lifting that you previously had to do yourself and enable you to build high-quality AI apps.

This API is designed for flexibility; use cases range from a natural language-based data analysis app, a coding assistant, an AI-powered vacation planner, a voice-controlled DJ, a smart visual canvas—the list goes on. The Assistants API is built on the same capabilities that enable our new GPTs product⁠: custom instructions and tools such as Code interpreter, Retrieval, and function calling.

A key change introduced by this API is persistent and infinitely long threads, which allow developers to hand off thread state management to OpenAI and work around context window constraints. With the Assistants API, you simply add each new message to an existing `thread`.

Assistants also have access to call new tools as needed, including:

As with the rest of the platform, data and files passed to the OpenAI API are never used to train our models⁠ and developers can delete the data when they see fit.

You can try the Assistants API beta without writing any code by heading to the Assistants playground⁠(opens in a new window).

The Assistants API is in beta and available to all developers starting today. Please share what you build with us (@OpenAI⁠(opens in a new window)) along with your feedback which we will incorporate as we continue building over the coming weeks. Pricing for the Assistants APIs and its tools is available on our pricing page⁠.

## New modalities in the API

### GPT-4 Turbo with vision

GPT‑4 Turbo can accept images as inputs in the Chat Completions API, enabling use cases such as generating captions, analyzing real world images in detail, and reading documents with figures. For example, BeMyEyes uses this technology to help people who are blind or have low vision with daily tasks like identifying a product or navigating a store. Developers can access this feature by using `gpt-4-vision-preview` in the API. We plan to roll out vision support to the main GPT‑4 Turbo model as part of its stable release. Pricing⁠ depends on the input image size. For instance, passing an image with 1080×1080 pixels to GPT‑4 Turbo costs $0.00765. Check out our vision guide⁠(opens in a new window).

Developers can integrate DALL·E 3, which we recently launched⁠ to ChatGPT Plus and Enterprise users, directly into their apps and products through our Images API by specifying `dall-e-3` as the model. Companies like Snap, Coca-Cola, and Shutterstock have used DALL·E 3 to programmatically generate images and designs for their customers and campaigns. Similar to the previous version of DALL·E, the API incorporates built-in moderation to help developers protect their applications against misuse. We offer different format and quality options, with prices starting at $0.04 per image generated. Check out our guide to getting started⁠(opens in a new window) with DALL·E 3 in the API.

### Text-to-speech (TTS)

Developers can now generate human-quality speech from text via the text-to-speech API. Our new TTS model offers six preset voices to choose from and two model variants, `tts-1` and `tts-1-hd`. `tts` is optimized for real-time use cases and `tts-1-hd` is optimized for quality. Pricing starts at $0.015 per input 1,000 characters. Check out our TTS guide⁠(opens in a new window) to get started.

## Model customization

### GPT-4 fine tuning experimental access

We’re creating an experimental access program for GPT‑4 fine-tuning. Preliminary results indicate that GPT‑4 fine-tuning requires more work to achieve meaningful improvements over the base model compared to the substantial gains realized with GPT‑3.5 fine-tuning. As quality and safety for GPT‑4 fine-tuning improves, developers actively using GPT‑3.5 fine-tuning will be presented with an option to apply to the GPT‑4 program within their fine-tuning console⁠(opens in a new window).

For organizations that need even more customization than fine-tuning can provide (particularly applicable to domains with extremely large proprietary datasets—billions of tokens at minimum), we’re also launching a Custom Models program, giving selected organizations an opportunity to work with a dedicated group of OpenAI researchers to train custom GPT‑4 to their specific domain. This includes modifying every step of the model training process, from doing additional domain specific pre-training, to running a custom RL post-training process tailored for the specific domain. Organizations will have exclusive access to their custom models. In keeping with our existing enterprise privacy policies, custom models will not be served to or shared with other customers or used to train other models.Also, proprietary data provided to OpenAI to train custom models will not be reused in any other context. This will be a very limited (and expensive) program to start—interested orgs can apply here⁠.

## Lower prices and higher rate limits

We’re decreasing several prices⁠ across the platform to pass on savings to developers (all prices below are expressed per 1,000 tokens):

### Higher rate limits

To help you scale your applications, we’re doubling the tokens per minute limit for all our paying GPT‑4 customers. You can view your new rate limits in your rate limit page⁠(opens in a new window). We’ve also published our usage tiers⁠(opens in a new window) that determine automatic rate limits increases, so you know what to expect in how your usage limits will automatically scale. You can now request increases to usage limits from your account settings⁠(opens in a new window).

OpenAI is committed to protecting our customers with built-in copyright safeguards in our systems. Today, we’re going one step further and introducing Copyright Shield—we will now step in and defend our customers, and pay the costs incurred, if you face legal claims around copyright infringement. This applies to generally available features of ChatGPT Enterprise and our developer platform.

## Whisper v3 and Consistency Decoder

We are releasing Whisper large-v3,⁠(opens in a new window) the next version of our open source automatic speech recognition model (ASR) which features improved performance across languages. We also plan to support Whisper v3 in our API in the near future.

We are also open sourcing the Consistency Decoder⁠(opens in a new window), a drop in replacement for the Stable Diffusion VAE decoder. This decoder improves all images compatible with the by Stable Diffusion 1.0+ VAE, with significant improvements in text, faces and straight lines.

_Learn more about our__OpenAI DevDay announcements for ChatGPT_⁠_._

View all product articles

Global news partnerships: Le Monde and Prisa Media Company Mar 13, 2024

Review completed & Altman, Brockman to continue to lead OpenAI Company Mar 8, 2024

OpenAI announces new members to board of directors Company Mar 8, 2024

Our Research * Research Index * Research Overview * Research Residency * OpenAI for Science * Economic Research

Latest Advancements * GPT-5.3 Instant * GPT-5.3-Codex * GPT-5 * Codex

Safety * Safety Approach * Security & Privacy * Trust & Transparency

ChatGPT * Explore ChatGPT(opens in a new window) * Business * Enterprise * Education * Pricing(opens in a new window) * Download(opens in a new window)

Sora * Sora Overview * Features * Pricing * Sora log in(opens in a new window)

API Platform * Platform Overview * Pricing * API log in(opens in a new window) * Documentation(opens in a new window) * Developer Forum(opens in a new window)

For Business * Business Overview * Solutions * Contact Sales

Company * About Us * Our Charter * Foundation * Careers * Brand

Support * Help Center(opens in a new window)

More * News * Stories * Livestreams * Podcast * RSS

Terms & Policies * Terms of Use * Privacy Policy * Other Policies

(opens in a new window)(opens in a new window)(opens in a new window)(opens in a new window)(opens in a new window)(opens in a new window)(opens in a new window)

OpenAI © 2015–2026 Manage Cookies

English United States

Originally published on OpenAI News.