I have been using Deepseek v4 pro for personal projects and home infra related work for last couple of weeks. It's quality of work is not bad at all, it is fairly fast and given the fraction of the cost compared to Claude, I can keep going which makes it a very compelling option. Looking forward to trying out Kimi 2.6, thanks for the recommendation.
Compaction is basically seamless which is a major weak point of Claude. At effort=low, Claude is better than codex but still slower. If you don't mind trading the upfront quality of work with additional micromanaging but at a faster speed, it is fine. I also think because of that very reason, you absorb more of the code.
Yes. Will be interesting to see how this evolves. Depending on the task, wouldn't be surprised if, between the cost of an AI tool and the cost/effort of auditing it, you go full circle and don't actually get an efficiency gain
> We’re re-organizing R&D to create roughly 60 smaller, more empowered teams with end-to-end ownership, nearly doubling the number of independent teams.
This seems to be the new emerging theme. I wonder if the idea is to let these smaller teams fight it out and those who survive, do so at the expense of others? If so, instead of meritocracy, usually politically savvy ones win. Or maybe the motivation is deliberate fragmentation allowing for easily picking and choosing for further reduction? Or maybe I'm just too cynical for my own good.
Leaking of good things is just good "PR" - people even pay good money for that to "accidentally" happen. Wonder why nobody thinks about actually doing the good thing and not bother with the rest.
> Mostly I just nudge it along. "Did you think about X? What about Y? Let's test Z"
Exactly - you need to constantly have your sceptics glasses on and you need to be exacting in terms of the structure you want things to follow. Having and enforcing "taste" is important and you need to be willing to spend time on that phase because the quality of the payoff entirely depends on it.
I recently planned for a major refactor. The discussion with claude went on for almost two days. The actual implementation was done in 10 minutes. It probably has made some mistakes that I will have to check for during the review but given that the level of detail that plan document had, it is certainly 90-95% there. After pouring-in of that much opinion, it is a fairly good representation of what I would have written while still being faster than me doing everything by hand.
In my experience you need exactly what you said, and I would add that he probably would have spent half day to do the refactoring himself and it would be sure he did right.
I don't think you have to know the answer. If the person you replied to knew the answer, there wouldn't have been a big, lengthy discussion.
But yes, being an expert in the problem domain helps. Or at least knowing enough to know what the right questions are and what plausible answers look like.
I just had a similar situation where an hour or two of conversation turned into a five-minute robot coding task. The problem required a solution and the number of possible solutions is vast, but that list can be refined, and then once the course of action is set, sometimes the course itself isn't all that complicated.
reply