There is no limit on OUR side for the context length, however, ChatGPT has its own limitations;
The maximum context length that can be handled depends on the version of ChatGPT being integrated:
- ChatGPT-4 (32k version): Handles up to 32,768 tokens, allowing for larger inputs and outputs, such as detailed conversations or extensive document processing.
- ChatGPT-4 (8k version): Supports up to 8,192 tokens, suitable for most standard use cases.
- ChatGPT-3.5: Handles up to 4,096 tokens, making it ideal for simpler and shorter interactions.
Tokens include both the input text and the generated output. As a rough guide, one token equals about four characters of English text.
For optimal performance, ensure your input and expected output together stay within the token limit of your chosen model.