Models
Models Overview
Flagship Models
GPT-4o (New)
- Description: Our fastest and most affordable flagship model
- Capabilities: Text and image input, text output
- Context Length: 128k tokens
- Pricing:
- Input: $5 per million tokens
- Output: $15 per million tokens
GPTs API (Special Provided)
- Description: Special GPTs Model only provided by AgiUni
- Capabilities: You can have 1:1 experience like using any GPTs in ChatGPT GPTs store. The model has full access for text, image and file input, can output text or generate images, have full access to internet.
- Context Length: 32k tokens
- Pricing:
- Input: $10 per million tokens
- Output: $30 per million tokens
GPT-4-ALL (Special Provided)
- Description: Special GPT-4 Model only provided by AgiUni
- Capabilities: have full access for text, image and file input, can output text or generate images, have full access to internet.
- Context Length: 32k tokens
- Pricing:
- Input: $10 per million tokens
- Output: $30 per million tokens
GPT-3.5 Turbo
- Description: Our fast, inexpensive model for simple tasks
- Capabilities: Text input, text output
- Context Length: 16k tokens
- Pricing:
- Input: $0.50 per million tokens
- Output: $1.50 per million tokens
Models Overview
The OpenAI API is powered by a diverse set of models with different capabilities and price points. You can also make customizations to our models for your specific use case with fine-tuning.
Model | Description |
---|---|
GPT-4o | The fastest and most affordable flagship model |
GPT-4-ALL | The special provided flagship model which can do anything for you |
GPT-4 Turbo and GPT-4 | The previous set of high-intelligence models |
GPT-3.5 Turbo | A fast, inexpensive model for simple tasks |
DALL·E | A model that can generate and edit images given a natural language prompt |
TTS | A set of models that can convert text into natural sounding spoken audio |
Whisper | A model that can convert audio into text |
Embeddings | A set of models that can convert text into a numerical form |
Moderation | A fine-tuned model that can detect whether text may be sensitive or unsafe |
GPT base | A set of models without instruction following that can understand as well as generate natural language or code |
Deprecated | A full list of models that have been deprecated along with the suggested replacement |
We have also published open source models including Point-E, Whisper, Jukebox, and CLIP.
Continuous Model Upgrades
gpt-4o
, gpt-4-turbo
, gpt-4
, and gpt-3.5-turbo
point to their respective latest model version. You can verify this by looking at the response object after sending a request. The response will include the specific model version used (e.g., gpt-3.5-turbo-0613
).
We also offer pinned model versions that developers can continue using for at least three months after an updated model has been introduced. With the new cadence of model updates, we are also giving people the ability to contribute evals to help us improve the model for different use cases. If you are interested, check out the OpenAI Evals repository.
Learn more about model deprecation on our deprecation page.
GPT-4o
GPT-4o (“o” for “omni”) is our most advanced model. It is multimodal (accepting text, image, or even file inputs and outputting text and images), and it has the same high intelligence as GPT-4 Turbo but is much more efficient—it generates text 2x faster and is 50% cheaper. Additionally, GPT-4o has the best vision and performance across non-English languages of any of our models. GPT-4o is available in the OpenAI API to paying customers. Learn how to use GPT-4o in our text generation guide.
Model | Description | Context Window | Training Data |
---|---|---|---|
gpt-4o | New GPT-4o: Our most advanced, multimodal flagship model that’s cheaper and faster than GPT-4 Turbo. Currently points to gpt-4o-2024-05-13 . |
128,000 tokens | Up to Oct 2023 |
gpt-4o-2024-05-13 | gpt-4o currently points to this version. | 128,000 tokens | Up to Oct 2023 |
GPT-4 Turbo and GPT-4
GPT-4 is a large multimodal model (accepting text or image inputs and outputting text) that can solve difficult problems with greater accuracy than any of our previous models, thanks to its broader general knowledge and advanced reasoning capabilities. GPT-4 is available in the OpenAI API to paying customers. Like gpt-3.5-turbo
, GPT-4 is optimized for chat but works well for traditional completions tasks using the Chat Completions API. Learn how to use GPT-4 in our text generation guide.
Model | Description | Context Window | Training Data |
---|---|---|---|
gpt-4-turbo | The latest GPT-4 Turbo model with vision capabilities. Vision requests can now use JSON mode and function calling. Currently points to gpt-4-turbo-2024-04-09 . |
128,000 tokens | Up to Dec 2023 |
gpt-4-turbo-2024-04-09 | GPT-4 Turbo with Vision model. Vision requests can now use JSON mode and function calling. gpt-4-turbo currently points to this version. | 128,000 tokens | Up to Dec 2023 |
gpt-4-turbo-preview | GPT-4 Turbo preview model. Currently points to gpt-4-0125-preview . |
128,000 tokens | Up to Dec 2023 |
gpt-4-0125-preview | GPT-4 Turbo preview model intended to reduce cases of “laziness” where the model doesn’t complete a task. Returns a maximum of 4,096 output tokens. Learn more. | 128,000 tokens | Up to Dec 2023 |
gpt-4-1106-preview | GPT-4 Turbo preview model featuring improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Returns a maximum of 4,096 output tokens. This is a preview model. Learn more. | 128,000 tokens | Up to Apr 2023 |
gpt-4-vision-preview | Legacy GPT-4 model with the ability to understand images, in addition to all other GPT-4 Turbo capabilities. This is a preview model, we recommend developers to now use gpt-4o which includes vision capabilities. Currently points to gpt-4-1106-vision-preview . Will be deprecated on December 6, 2024. |
128,000 tokens | Up to Apr 2023 |
gpt-4-1106-vision-preview | Legacy GPT-4 model with the ability to understand images, in addition to all other GPT-4 Turbo capabilities. This is a preview model, we recommend developers to now use gpt-4o which includes vision capabilities. Returns a maximum of 4,096 output tokens. Learn more. Will be deprecated on December 6, 2024. |
128,000 tokens | Up to Apr 2023 |
gpt-4 | Currently points to gpt-4-0613 . See continuous model upgrades. |
8,192 tokens | Up to Sep 2021 |
gpt-4-0613 | Snapshot of gpt-4 from June 13th 2023 with improved function calling support. | 8,192 tokens | Up to Sep 2021 |
gpt-4-32k | Legacy Currently points to gpt-4-32k-0613 . See continuous model upgrades. This model was never rolled out widely in favor of GPT-4 Turbo. Will be deprecated on June 6, 2025. |
32,768 tokens | Up to Sep 2021 |
gpt-4-32k-0613 | Legacy Snapshot of gpt-4-32k from June 13th 2023 with improved function calling support. This model was never rolled out widely in favor of GPT-4 Turbo. Will be deprecated on June 6, 2025. | 32,768 tokens | Up to Sep 2021 |
For many basic tasks, the difference between GPT-4 and GPT-3.5 models is not significant. However, in more complex reasoning situations, GPT-4 is much more capable than any of our previous models.
GPT-4-all
GPT-4-all (gpt-4-all) is a special model we provided. It is multimodal (accepting text, image, or even file inputs and outputting text and images), and it has the same high intelligence as the model "gpt-4" you used in ChatGPT website.
Model | Description | Context Window | Training Data |
---|---|---|---|
gpt-4-all | Ours special multimodal flagship model that has the same effect as ChatGPT 4.0. | 328,000 tokens | Up to June 2024 |
Multilingual Capabilities
GPT-4 outperforms both previous large language models and as of 2023, most state-of-the-art systems (which often have benchmark-specific training or hand-engineering). On the MMLU benchmark, an English-language suite of multiple-choice questions covering 57 subjects, GPT-4 not only outperforms existing models by a considerable margin in English, but also demonstrates strong performance in other languages.