Recommended context: 65,536 tokens (can be increased)
Adequate Output Length: We recommend using an output length of 65,536 tokens for most queries, which is adequate for instruct models.