Unlike Anthropic, OpenAI models don't have a `max_tokens` setting for API calls, so I assume the max token output limit is automatically applied to API calls.
Otherwise the max token output limit stated on the models page would be meaningless.
OpenAI has a `max_tokens` setting. For the /chat/completions api, it defaults to the maximum for a desired model, but for the /completions api, it defaults to 16.
gpt-4o-2024-08-06 has 16,384 tokens output limit instead of 4,096 tokens.
https://platform.openai.com/docs/models/gpt-4o
We don't need the GPT-4o Long Output anymore.