DeepSeek is an amazing product but has few issues:
1. Data is used for training
2. Context window is rather small and doesn't fit as well large codebase
I keep saying this over and over in all the content I create, the valu of coding with AI will come from working on big, complex, legacy codebases. Not from flashy demo where you create a to-do app.
For that you need solid models with big context and private inference.
Shouldn't we be comparing with other open source model? In particular since this is about llama3.3 then they have the exact context limit which is 128k [1]. Also
Their cost is $0.7 per 1M token.
DeepSeek is $0.14 / 1M tokens ( cache miss)