Its either of two things, either this person is going by market forces, saying whatever makes sense to please the market or the individual has not idea of what it takes to build software by self
I think it’s fairly normal at least in my career - rush to ship something, lots of “we’ll polish this later,” two years go by, get called into vp/cto/whoever’s office when the debt comes calling like “what the fuck why is this like this???” and I have to say “that ‘later’ we decided is now I guess”
The script I have fairly seen being played is where the one doing MVP gets rewarded and moves on with a promotion. The weight of completing and stabilizing the MVP falls on some one else who is not vocal enough in terms of influence. Ironically the flashy MVP does not includes monitoring, logging, security, edge-cases, CI-CD, DR, scaling which is why vibe coding is getting so popular and everyone seems to be under the impression that engineers are not needed anymore.
The only good justification of it can be that the companies can claim that the age verification was done as per Terms of Service, so in the future no parent or parent group can come after them for the content. Along with better targeted advertising by identifying the target audiences.
Logically parents are probably best suited to gate the content for their children how they see it fit.
These type of experiments are bound to have biases depending on who is doing it and who is funding it. The experiment is being funded for a particular reason itself to move the narrative in a desired direction. This is probably a good reason to have government funded research in these type of sensitive areas.
Thanks for jumping in the conversation. Logically it does makes sense to attribute the authors correctly, however in this context it might be helpful if you can provide any details about the users complaining that their PR's are being marked as co-authored even when they have not used the copilot? Is that intentional or a missed check in the implementation.
Also for layman readers like me who might not be actively involved, it might have been helpful to add the issue/referenced conversation why this change was made on the PR itself
The fact that non-AI changes are attributed to Copilot is a bug. The intent was to allow customers to add attribution of AI-generated code. As with any bug, it was not intetional.
Interestingly a product manager creates a PR with small but sort of policy change without any backstory/explanation, it gets reviewed by a single developer and merged without a single comment. The bar to make changes to a production software used by so many people has gown down considerably.
A GLOBAL do not track on the browsers works largely cause the target is all the websites being browsed and the tracking associated with it for advertising purposes. However telemetry is altogether a different thing, blocking it by default can be one idea, however using one standard variable to express the intent for all the tools is not practically viable
I think we are normalizing the PR process here, in reality its more convoluted and a good reflection of the team/organization itself. The relationship between the author and reviewer can have negative impact on the rationale and desired outcome of the PR itself.
To run the process smoothly, one can just hope that the team/tech lead is an ideal developer. Otherwise they are in a position where no one senior than them is available for the code review and any one junior would just rubber stamp their PR's.
My assumption is that a lot of these checks and changes lately are not well though out. They are knee jerk reaction to address something which was not anticipated in the original design. A lot of these changes to address scaling and abuse challenges probably fall into bucket of applying bandages on top of bandages. Maybe if Claude could build something to validate the baseline quality of the product to ensure these things are discovered early on.
Worse than that, these are all vibe coded changes. If you look at any public Anthropocene codebase, they are all vibe coded messes with no coherent vision. I was looking at the Claude Code GitHub Action and it is a mess of options that don’t exist together, unclear documentation, and usage story being terribly unclear.
People say that a mostly-vibed project will collapse under its own weight. I personally doubt it, but I will be amused if the first big one falls this way is Claude Code itself.
Unfortunately it will all probably sort of work, But best not to dwell too much on how the sausage is made, it is pretty unpleasant. There will be some interesting job titles in the future however.
I just read Vernor Vinge's "A deepness in the sky" And the way he modeled their compute systems felt depressingly believable, they have thousand of years of libraries floating around, sort of loosely tacked together. and specialist programmer-archaeologists are the ones who who dig deep and try to understand the system.
> Unfortunately it will all probably sort of work, But best not to dwell too much on how the sausage is made, it is pretty unpleasant.
Interestingly, most long-running codebases are like that, no?
It's just that producing (incl. reviewing/testing and all those, even AI-assisted) that amount of code in a significantly shorter period of time highlights this discrepancy much more to us.
I've seen ancient codebases that you need to be blessed by a priest to even touch but they keep chugging away and having new features added. I wouldn't hold my breath for a collapse, just a quagmire that we continually have to wade through to get anything done.
Isn't it also true that the deeper and thicker the quagmire, the more tokens one will have to use to wade through it?
This seems like a path to eventual LLM lock-in once the codebase gets messy enough. These things could end up being like 0% interest credit cards for technical debt. I guess it all depends on how the token usage scales over time. My guess is it will be steeper than linear.
What continues to perplex me is that these people claim that they will be able to contain AGI yet can't roll out a regex match? If AGI is possible then we're most certainly not containing anything.
Just give it a little time. AGI will be redefined to whatever is current and a new AI acronym will be coined for what everyone expected true AI to be in the first place.
Artificial Human Intelligence. Actually they'll probably drop the Artificial part. Human Scale Intelligence.
The meaning behind the acronym is so wrong that I already forgot what it stands for. This is aggravated by the fact that every single marketing page of this Arm brand refuses to mention what the acronym stands for.
Thanks to being at the forefront of AGI, Arm has had a spark of genius. The G in AGI stands for AI.
Of course the A is obviously Agentic and the I is Infrastructure.
Given what we know about their development practices, they almost certainly implemented this check by writing text along the lines of “Please ensure requests from Openclaw always go to extra usage” into a Claude prompt. Perhaps some junior engineer who didn’t understand the problem reviewed the generated code, or perhaps nobody at all reviewed it.
Of course they are not well thought out. The biggest limiting factor on software quality has always been PM and executive prioritization. If they decide that you should build garbage anti-user features, that's what you'll build for them.
Letting SWEs execute on that prioritization faster was never going to get us better software, it was just going us more enshittification faster.
AI improving productivity is great, except that the C-suite controlling where that productivity goes are people that are consistently ranking in the top of the 'Never should be trusted with a lot of power' list. All they want is to make more paperclips.
Isnt this illegal right away? A normal entity would have been punished for this otherwise this just opens up the door to make code changes to overcharge people and just claim it as mistake
reply