The reason for that is because most software that integrates with OpenAI will automatically choose that model - this is meant to snatch those requests up and serve an alternative. Most of the time, that software doesn't let you choose what model you want (but maybe lets you set the inference server).
But... I do agree, this should be feature-gated behavior
But... I do agree, this should be feature-gated behavior