Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Really important update that was not mentioned:

gpt-4o-2024-08-06 has 16,384 tokens output limit instead of 4,096 tokens.

https://platform.openai.com/docs/models/gpt-4o

We don't need the GPT-4o Long Output anymore.



But is this also the default or just the max? Is the default 4k or 16k?

Also, the question of the default value applies both at the server level and at the SDK level.


Unlike Anthropic, OpenAI models don't have a `max_tokens` setting for API calls, so I assume the max token output limit is automatically applied to API calls.

Otherwise the max token output limit stated on the models page would be meaningless.


OpenAI has a `max_tokens` setting. For the /chat/completions api, it defaults to the maximum for a desired model, but for the /completions api, it defaults to 16.

https://platform.openai.com/docs/api-reference/chat/create#c...


Ops. Not sure how I missed that.


Long Output is 64K though.




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: