GPT-3.5 Turbo

GPT-3.5 Turbo

Low-cost model for simple chat, classification, and extraction.

Community Sentiment

Mostly Positive
Based on Reddit reviews

Community Verdict

Best for Fine-Tuning
Based on Reddit reviews

Input Modalities

Text

Output Modalities

Text

Price / 1M tokens

$0.5$1.5
InputOutput

Best For

Fine-tuning use cases
Long-term support for legacy applications
Budget-conscious chat applications
High-volume, low-complexity tasks

Avoid For

All new development (use GPT-4o-mini instead)
Complex reasoning tasks (use GPT-5 instead)
Vision/multimodal tasks (not supported)
Audio processing (not supported)
16,385 context window
4,096 max output tokens
Sep 1, 2021 knowledge cutoff
Reasoning not supported

Parameters

While OpenAI documents a unified parameter set for the Chat Completions and Responses APIs, each model supports only a limited subset of model-specific parameters and values. This table lists the supported model-specific GPT-3.5 Turbo API parameters and allowed values.
ParameterDescriptionPathSupported Values
ModelSelects the model version for the requestmodel
gpt-3.5-turbogpt-3.5-turbo-0125
Message rolesDefines the role of the message in the inputmessages[].role
developersystemuserassistant
Max output tokensMax output tokens the model may generatemax_completion_tokens16 .. 4,096
Output formatSpecifies the output format, including structured JSON schemasresponse_format
textjson_object
TemperatureControls how random or deterministic the output istemperature0 .. 2
Top PControls how diverse the output tokens aretop_p0 .. 1
Presence penaltyEncourages the model to introduce new topicspresence_penalty-2 .. 2
Frequency penaltyReduces repetition of the same words or phrasesfrequency_penalty-2 .. 2
Reasoning effortControls the depth of internal reasoning used by the modelreasoning_effortNot supported
Reasoning summaryControls whether the model produces a concise or detailed reasoning summaryreasoning_summaryNot supported
VerbosityControls how brief or detailed the generated response isverbosityNot supported

Pricing

GPT-3.5 Turbo API pricing is based on token usage for input and output. Prices are listed per 1M tokens, with lower rates for cached input. Tool-specific features may add per-call fees.
Text tokens
Per 1M tokens
Input$0.50
Cached inputN/A
Output$1.50
Example costs (GPT-3.5 Turbo)
TASK
APPROX COST
Simple chat (1k tokens)
~$0.0005
Classification batch (10k items)
~$0.05
Text extraction
~$0.01–$0.05

Modalities

What the model can accept and produce
TextInput and output
ImagesNot supported
AudioNot supported
VideoNot supported

Features

Platform-level capabilities
StreamingNot supported
Function callingNot supported
Structured outputsNot supported
Fine-tuningSupported
DistillationNot supported
Predicted outputsNot supported

Tools

Tools supported by this model when using the Responses API.
Web searchNot supported
File searchSupported
Image generationNot supported
Code interpreterSupported
MCPSupported
Computer useNot supported

Snapshots

GPT-3.5 Turbo model snapshots ensure stable behavior by locking a specific version. See all available snapshots and aliases below.
GPT-3.5 Turbo
gpt-3.5-turbo
↪ gpt-3.5-turbo-0125
gpt-3.5-turbo-0125
gpt-3.5-turbo-1106Deprecated
gpt-3.5-turbo-instructDeprecated

FAQs

Ship prompts that pass the tests
Don't wait until they break in production
© 2026 LangFast. All rights reserved. Privacy Policy. Terms of Service.