gpt-3.5-turbo-16k

Common Name: GPT-3.5 Turbo 16k

OpenAI
Released on Nov 6, 2023 12:00 AMTool Invocation

GPT-3.5 Turbo variant with extended 16K token context window for longer conversations and documents.

Specifications

Context16,385
Maximum Output4,096
Inputtext
Outputtext

Performance (7-day Average)

Uptime
TPS
RURT

Pricing

Input$3.30/MTokens
Output$4.40/MTokens
Batch Input$1.65/MTokens
Batch Output$2.20/MTokens

Usage Statistics

No usage data available for this model during the selected period
View your usage statistics for this model

Similar Models

$4.40/$17.60/M
ctx32Kmax4Kavailtps
InOut

This is our first general-availability realtime model, capable of responding to audio and text inputs in realtime over WebRTC, WebSocket, or SIP connections.

$4.40/$17.60/M
ctx32Kmax4Kavailtps
InOut

This is our first general-availability realtime model, capable of responding to audio and text inputs in realtime over WebRTC, WebSocket, or SIP connections.

$1.65/$2.20/M
ctx16Kmax4Kavailtps
InOut

GPT-3.5 model optimized for single-turn instruction following via completion API endpoint.

$1.65/$2.20/M
ctx16Kmax4Kavailtps
InOutCap

September 2023 snapshot of GPT-3.5 Turbo Instruct for legacy completion API use cases.

Documentation

No documentation available
This model (gpt-3.5-turbo-16k) uses a dedicated API. Please refer to the official documentation for usage examples.