With similar capabilities from these great models, the only difference that matters, is price. It’s way cheaper to run Deepseek V3 or R1, even on pay as you go inference providers like Azure and Cloudflare. The point is getting the required results.
After tons of research, I realized that I too got swept up by the AI hype. When I narrowed down the use cases of what I want to achieve with AI, I found cheaper, more specialized models that can do what I need at lower costs and faster speeds (distilled models are naturally faster due to their size).
While it’s great to be able to use the latest and the greatest, if all your’re doing is just summarizing some text and return a JSON, you can find smaller models to run in the cloud that will achieve the same results. Just do some testing, fine-tuning and you’re good to go. Just use AI to refine your AI.