What Does The OpenAI 16K Context Window Mean?

OpenAI launched a new model called gpt-3.5-turbo-16k, which means you can submit a document to OpenAI of 16,000 words in one go.

Cobus Greyling
4 min readJun 15, 2023

--

I’m currently the Chief Evangelist @ HumanFirst. I explore and write about all things at the intersection of AI and language; ranging from LLMs, Chatbots, Voicebots, Development Frameworks, Data-Centric latent spaces and more.

In the image below, is an extract of a 12,139 words, 14 pages document which I submitted to the new OpenAI model, with the result at the bottom:

I found it intriguing that the GPT-3.5-Turbo-16k model is a chat-only model and is not supported by the Completions Endpoint. The following is the error message I received:

InvalidRequestError: This is a chat model and not supported in the v1/completions endpoint. Did you mean to use v1/chat/completions?

Given this, I had to choose the chat endpoint and the ChatML notation for input.

As an example, below is a working Python application accessing the 16k model:

import os
import openai
openai.api_key = "xxxxxxxxxxxxxxxxxxxx"
completion = openai.ChatCompletion.create(
model="gpt-3.5-turbo-16k",
messages = [{"role": "system", "content" : "You are a chatbot which can search text and provide a summarised answer."},
{"role": "user", "content" : "How are you?"},
{"role": "assistant", "content" : "I am doing well"},
{"role": "user", "content" : "What is the distance between New York and Montreal?"}
]
)
print(completion)

When I attempted to submit a document which exceeded the maximum context window, I received an informative error message detailing what was possible from a context window perspective:

InvalidRequestError: This model’s maximum context length is 16385 tokens. However, your messages resulted in 18108 tokens. Please reduce the length of the messages.

16k context means the model can now support ~20 pages of text in a single request.

- OpenAI

The OpenAI model has alleviated one of the long-running ailments of LLMs, the limited context window size, which necessitates pre-processing of data.

The sheer context that this model can manage is both astounding and remarkable, with speeds and accuracy that make it an invaluable development.

With 256 words for a one sentence summary and 12 key points, the summary clearly details the significance of this new progress.

⭐️ Please follow me on LinkedIn for updates on Conversational AI ⭐️

I’m currently the Chief Evangelist @ HumanFirst. I explore and write about all things at the intersection of AI and language; ranging from LLMs, Chatbots, Voicebots, Development Frameworks, Data-Centric latent spaces and more.

LinkedIn

--

--

Cobus Greyling
Cobus Greyling

Written by Cobus Greyling

I’m passionate about exploring the intersection of AI & language. www.cobusgreyling.com

No responses yet