I'm not exactly sure what the point of this is. Deepseek already has instructions to use its API with many CLI's including Claude Code directly:
https://api-docs.deepseek.com/quick_start/agent_integrations...
>DeepSeek V4 Pro scores 96.4% on LiveCodeBench and costs $0.87/M output tokens
This is a heavily subsidized price and will only last until the end of the month: "The deepseek-v4-pro model is currently offered at a 75% discount, extended until 2026/05/31 15:59 UTC." [0]
The "supported backends" table is also deceiving -- while OpenRouter's server's may be in the US, the only way to get the $0.44/$0.87 pricing is to pass through to the DeepSeek API, which of course is China-based. [1]
I do think the model is quite good, I myself use it through Ollama Cloud for simple tasks. But I think some folks have bought in a little too much to the marketing hype around it.
[0] https://api-docs.deepseek.com/quick_start/pricing [1] https://openrouter.ai/deepseek/deepseek-v4-pro/providers
If you're looking for Claude Code alternatives, I would first suggest looking into pi.dev or opencode for your harness. And then for models, you can choose from OpenCode Go (IMO most cost effect at this moment), OpenRouter, or direct from DeepSeek. Better if you go the Kimi route IMO and just buy a subscription from kimi.com
Not sure you can replace Claude with DeepSeek V4 that easily and have same results.
From what I see while building my own agentic system in Elixir, the problem is in training for your specific harness/contracts. Claude/GPT-style models seem to be trained around very specific contracts used by the harness like tool call formats, planning structure, patching, reading files, recovering from errors, and knowing when to stop.
In practice, you either need a very strong general model that can infer and follow those contracts (expensive), or a weaker model that has been fine-tuned / trained specifically on your own agent contracts. Otherwise, the whole thing becomes flaky very quickly. And I suspect with Deepseek V4 you may get last options.
> Claude Code is the best autonomous coding agent.
If you look at the [email protected] leaderboard, you'll quickly see it's actually one of the weakest agentic harnesses. Anthropic's own models score lower with Claude Code than with virtually any other harness.
So it's quite the opposite. Claude Code is arguably the worst harness to run models with.
> DeepSeek V4 Pro scores 96.4% on LiveCodeBench and costs $0.87/M output tokens.
Yes and this is a temporary discount which increases to 3.48 USD on 2026/05/31 15:59 UTC.
It's surprisingly easy to hit $200 worth of tokens even at ~$1/M token though. No matter how many times I do the math the coding plans are the better value.
Just want to say that I faced this very problem the last week, I discovered OpenCode agent and it works great, with DeepSeek and other models. Try it out guys.
If you're okay with sonnet level performance, this sounds like a straight upgrade. But I find that sonnet messes up too much, that it ends up not being worth cost optimizing down to using it or another sonnet-level model. Glad to have this as an option though
Did... Did you just ask an AI to one-shot something that normally amounts to no more than setting two env variables?
I've been using DeepSeek v4 pro as an alternative to Claude models and for the first time I can see it as a real replacement. With the other Chinese models, I was missing something, but DeepSeek seems good enough for the kind of development I want to do.
I have a question. does anyone have a problem with switihng context between AI and your terminal
Next claude news (trump style): Recent versions of Claude code no longer allow talking to other models, or helping with any code that has the goal of moving away from anthropic models.
obviously vibe coded ( co authored ) + the prices dont even match
Is claude code the best coding harness? Anyone running evals on that?
Claude code can already use the DeepSeek API, so what are the advantages of this tool?
Interesting setup
do you have any benchmarks on: - token usage over time - failures/retry rates
would be great to see how it behaves in production
Is there a way to do this directly by using claudecode CLI (which I already have installed) and openrouter??
I just spent half my day getting CUDA and LLAMA to work with my 5070TI.
I was able to use it in agent mode with Roo, I stopped after having it write out a plan, but I'll continue when I have more time.
Deepseek feels less likely to do a straight up rug pull since you can self host with enough money, but I'm still more excited about local solutions.
Usually I just need grunt work done. I'm not solving difficult problems.
Using a bunch of CLIs to work with DeepSeek V4, I've found that Langcli is the best fit for DeepSeek V4. For programming tasks, the cache hit rate is above 95%.
Not only can it seamlessly and dynamically switch between DeepSeek V4 Flash, V4 Pro, and other mainstream models within the same context, but it is also 100% compatible with Claude Code.
I'm wondering why DeepSeek didn't create an AI coding agent like Kimi Code.
This has become a problem for me. I like trying new things. But I also know that in about a week, there's going to be a better/cheaper setup. And a week after that. And ideally I'd like to get some coding done when I'm not tinkering with the tools.
So I think I'll stay with CC for now.
I wonder if openrouter will replicate that 120x caching, I suppose they will?
You don't need Deep Claude. Claude Code is working with any model that exposes an endpoint for an Anthropic compatible API.
I am using Claude Code with GLM 5.1, MiniMax M2.7, Kimi K2.6 and Xiaomi MiMo V2.5 Pro.
Honestly with the likes of Opencode / pi / hermes I don't really find the "Claude Code agent loop" part particularly interesting.
The edge Anthropic has on others lies on its models performance. CLI tooling (and obviously pricing) is definitely not better than others.
why not opencode with deepseek?
Does it play Doom?
Nice, it's quite usefull to have a project like this which streamlines the setup necessary to use other "brains" in claude code "body". I personally will give this a try, but Ijust find the message on pricing a bit disingenuous, the deepseek price of "$0.87/M output tokens" is a discount, and this setup anyways needs a calude.ai subscription offering claude code, which now is 100$/month min.
the wrapper is basically env var glue. You’re still betting the whole loop on Anthropic's closed client.
Cost engineering [1] will be the next hot topic for AI.
[1] A fancier way of saying "reducing cost."
Can I... somehow run this locally? DeepSeek is opensource? Do I even need their API key?
(I have no experience with running anything locally, maybe it's a stupid question)
does it support aws bedrock provider
And if I don't care about cost, what about actual performance?
Is there some way to make claude/codex beep when it finishes a task.
Why wouldn't you use something open source like OpenCode, which already support DSv4 and has more features than CC?
I am now thinking how far can agentic AI can go how far we can achieve
layer on layer on layer to refactor bunch of lines xD
Oh shoot now the next CC upgrade will blow your subscription for doing this
anthropic messed up big time harness works with any muh commodity LLM, meanwhile VCs were duped on the myth of FOOM AGI, probably not a cooincidence Anthropic is enmeshed with the scifi fan fic forum known as lesswrong. The world wants useful tools. The bay area bubble in contrast thrives on Mythos.
[dead]
[flagged]
[flagged]
[dead]
[dead]
[flagged]