The Claude Code Leak (build.ms)

by mergesort 181 comments 200 points
Read article View on HN

181 comments

[−] thaumaturgy 44d ago
I wonder what happened to the person that wrote "Coding as Creative Expression" (https://build.ms/2022/5/21/coding-as-creative-expression/)?

I'm not (just) being glib. That earlier article displays some introspection and thoughtful consideration of an old debate. The writing style is clearly personal, human.

Today's post is not so much. It has LLM fingerprints on it. It's longer, there are more words. But it doesn't strike me as having the same thoughtful consideration in it. I would venture to guess that the author tried to come up with some new angles on the news of the Claude Code leak, because it's a hot topic, and jotted some notes, and then let an LLM flesh it out.

Writing styles of course change over time, but looking at these two posts side by side, the difference is stark.

[−] mergesort 44d ago
Hey there, author of the post here! I actually wrote this piece myself on my phone while I was out for a walk this morning. It was initially meant to be a quick note more than a full blog post —- whereas Coding As A Creative Expression took me a couple of days to write.

I made a commitment to write more this year and put my thoughts out quicker than I used to, so that’s likely the primary reason it’s not as deep of a piece of writing as the post you’re referencing. But I do want to note that this wasn’t written using AI, it just wasn’t intended to be as rich of a post.

The reason it came out longer is that I’ve honestly been thinking about these ideas for a while, and there is so much to say about this subject. I didn’t have any particular intention of hopping on a news cycle, but once I started writing the juices were flowing and I found myself coming up with five separate but interrelated thoughts around this story that I thought were worth sharing.

[−] tpoacher 44d ago
Reminds me of the classic Mark Twain quote: "Apologies, I didn't have time to write a short letter, so I wrote a long one."
[−] adzm 44d ago

> I didn't have time to write a short letter, so I wrote a long one.

First known use in English comes from a 1658 translation of Blaise Pascal in 1657

> Je n’ai fait celle-ci plus longue que parce que je n’ai pas eu le loisir de la faire plus courte.

translated to

> I had not made this longer then the rest, but that I had not the leisure to make it shorter then it is.

(note the archaic then)

This was a popular piece of wit at the time.

Mark Twain wrote something similar a hundred years later

> You'll have to excuse my lengthiness - the reason I dread writing letters is because I am so apt to get to slinging wisdom & forget to let up. Thus much precious time is lost.

But it's still quite different.

There is a great article about this one on quoteinvestigator! https://quoteinvestigator.com/2012/04/28/shorter-letter/

[−] jachee 44d ago
Hmm… this started as an admonition for using then and than interchangeably. I see folks get it wrong unintentionally a lot… then i re-read your parenthetical and pulled up etymonline and had my mind blown a bit.[0]

Seems everything old is new again.

[0] https://www.etymonline.com/word/than

[−] bostik 44d ago

>

wrote this piece myself on my phone while I was out for a walk

If you have a strategy for jotting down (or dictating) notes while walking about, I would be curious how you manage that. I spend plenty of time walking outside, and tend to get (at the time) ideas that I'd like to explore further, most of which have evaporated from my mind by the time I get back home. Or even before I can get my phone out to jot down the keywords to help me recall the details later.

Cannot even imagine how someone would manage both walking and writing at the same time.

[−] nyulmalac 44d ago
What I tried is to record my voice and then post-process the transcript. There are solutions which work without internet connection. I am non-native english and mix 3 languages, so transcript is shitty quality. Nevertheless the really good ideas stay with me and can be easily recalled by a few keywords. And you need to do the post-processing shortly after you reached home else it fades away…
[−] DANmode 44d ago

> Cannot even imagine how someone would manage both walking and writing at the same time.

Some are just born with it.

[−] mergesort 43d ago
And here I've been using Maybelline for nothing!
[−] staindk 43d ago
Have a look at the Pebble Index ring. Might be a good fit for you.
[−] mergesort 43d ago
Hey there! I'm not sure I have a universally applicable answer, but I can do my best to map out some things about my process and flow that hopefully help a bit and answer your question.

- I've had an iPhone for half my life (I'm 36 and got one when I was 19), so I've gotten pretty acclimated to typing on the go. I try switching to dictation every couple of months but the iPhone's dictation trips up over enough words that I find it more frustrating than typing as I walk.

- I don't do this but if you're worried about the thoughts disappearing I would absolutely recommend recording a voice note. As I'll touch on in a moment — do not let those thoughts disappear! Even the act of codifying them into something tangible allows you to process them more deeply.

- I live in NYC but I start most mornings by taking a walk along a relatively quiet street, so I rarely end up having to worry about bumping into someone. That is definitely not universally applicable advice. (:

- I look up as I'm typing and let autocorrect take the wheel. That works at least 95% of the time, so if I make the occasional typo it doesn't really matter, I'll just fix it in post.

- It helps to have an app with a great text editing experience. I've found that there are very few out there that are fluid, many have incredibly subtle hitches that make it hard to quickly jot down thoughts onto a canvas. I really love Craft (https://craft.do) and have been using it for years, so at this point it feels more like an extension of me than an app.

- This is surely unique to everyone but my writing tends to start from a few keystone thoughts. Once I have one written down, I let myself almost free associate, writing down whatever comes to mind from that initial thought to make sure I do not forget. I can always edit after the fact, and often the editing process leads to more interesting insights as well. But the main thing I want to avoid is losing those sparks, in the same way that you're mention your thoughts evaporating. Don't let those go, just get 'em on paper and sort through 'em afterwards.

- That's all a lot easier to do on my phone than if I approached the problem as "type an essay on my phone", so I'll almost always edit a post on my computer before publishing. Yesterday was more of an exception than the rule though because I was bouncing around between doctors all day, so I wrote all of this on my phone [not expecting it to blow up or get a ton of scrutiny].

Not sure if anything's missing but I'm happy to share anything that may be helpful! Clearly this post wasn't perfect, but I've been much happier since I started letting myself write out long-form thoughts on my phone and sharing them as blog post rather than firing them off as pithy tweets that decay into the ether once the algorithm says it's time for them to go.

[−] anfogoat 44d ago

>­ I actually wrote this piece myself on my phone while I was out for a walk this morning.

Apropos of nothing, this is astonishing me to no end. The ergonomics of 1) using a phone keyboard for anything but a word or two and 2) doing so while walking pretty much guarantee that I'd probably need a half a day to recover if I attempted the same.

[−] sirpilade 44d ago
As a more general comment, I hope we’ll stop trying to discriminate whether a certain text was or not written/polished/extended by AI and focus more on content and author’s responsibility for what they do or do not say with that text.
[−] dpark 43d ago
I agree. Not because I think that most AI content is worth reading, but because it can be criticized on more grounded merits. People wrote blogspam by hand for two decades before AI started generating it. It wasn’t high value when a human wrote it either.

On many (most?) posts, far more energy is spent arguing about whether a post is AI than discussing if there’s anything of value in the post.

[−] JonChesterfield 44d ago
I've stopped reading anything on blogs on the basis that it's now probably llm spew and life is too short for the signal to noise ratio that implies.

With the exception of things that places like HN seems to consider worth reading, which is why I'm looking through the comments to this and others to find recommendations.

[−] bengale 44d ago
Another swing and a miss from the AI police.
[−] stbev 44d ago
Have you noticed that comments like "this post seems written with AI" are now appearing on all posts, even those written without AI?

We're starting to become wary due to the abuse of AI and proliferation of sloppy content, but also because we often have trouble distinguishing authentic from sloppy content.

Another feature of this AI era that I hate.

[−] hxugufjfjf 44d ago
Agreed. Its so tedious that the top comment section on every HN post the last six months is "this seems be written by LLM" with a bunch of back and forth on whether it is or not.
[−] amarant 44d ago
Once I was tempted to say that those comments themselves were written by llms, but honestly I've yet to find a model quite so uncreative as that.
[−] watwut 44d ago
As a daily user of HN, this is not true. Maybe you are clicking on different headlines, but the ones I am clicking on dont have that as top comment all that often. They do not even have it as a comment somewhere all that often.
[−] grufkork 44d ago
I mean, a lot of it is. Green user, signed up 49mins ago, 5 comments, which erodes trust in real people as well. I’ve noticed I’ve just felt less engaged and more anxious about all kinds of online content. While most platforms were previously botted, had adverts, etc… You could always find niche corners where there were only people talking about things they genuinely cared about. Now you can fill out even those spaces automatically.
[−] bakugo 44d ago
Probably because most HN posts in the last six months have been written by LLMs. Not all, but that doesn't matter, the trust has been eroded to the point that clicking on an article on the front page of HN and not being immediately met with the sloppiest slop imaginable is now a standout event in my mind.
[−] roysting 44d ago
I’ve been accused of being AI. My first impression when it happened was that because I often deal in information that people don’t like hearing, because it challenges their frame of mind, i.e., what they were trained on, “this is AI” is just another convenient tool to either dismiss uncomfortable challenge, i.e., cognitive dissonance, and/or another means to keep the mental herd they are part of or control in line with dogma.

“This is AI” seems to just be an evolution of other thought terminating cliches where the negative conditioning associated with something is used in an abusive and manipulative way to evade challenge or the truth itself. It is a common tactic of abusive people, the “beyond the pale” moralizing.

[−] benterix 44d ago
Yeah, whenever I see "It's not... it's...", I catch myself instinctively dismissing the content as AI slop, but upon reflection I'm not so sure, it used to be a normal phrase.

But I do take extra care to avoid LLM-speak as much as I can.

[−] dpark 43d ago
“It's not delivery, it's DiGiorno!” — Probably AI, according to HN commenters
[−] grey-area 44d ago
It does read as if were written on a phone but it doesn’t read like LLM text to me.

What is interesting and has possibly bled over from heavy LLM use by the author is the style of simplistic bullet point titles for the argument with filler in between. It does read like they wrote the 5 bullet points then added the other text (by hand).

[−] raincole 44d ago
What changed is you, the reader. In 2026 we treat the smallest signs as evidence of LLM writing. Too long? LLM. Too short? LLM. Too grammatically correct? Must be LLM.
[−] jiusanzhou 44d ago
The copyright angle is the most underrated part of this story. Anthropic built their models on other people's code under the fair use argument, but the moment their own code leaks they reach for DMCA takedowns. You can't have it both ways. The clean room reimplementations are the natural consequence of the legal framework they themselves advocated for.
[−] himata4113 44d ago
I personally found it really amusing how they weaponized the legal system to DMCA all the claude code source code repositories. Code ingested into the model is not copyrightable, but produced code apparently is when by legal definition computer generated code can not be copyrighted and that's one of their primary arguments in legal cases.
[−] leduyquang753 44d ago

> Many software developers have argued that working like a pack of hyenas and shipping hundreds of commits a day without reading your code is an unsustainable way to build valuable software, but this leak suggests that maybe this isn’t true — bad code can build well-regarded products.

The product hasn't been around long enough to decide whether such an approach is "sustainable". It is currently in a hype state and needs more time for that hype to die down and the true value to show up, as well as to see whether it becomes the 9th circle of hell to keep in working order.

[−] kstenerud 44d ago

> It should serve as a warning to developers that the code doesn’t seem to matter, even in a product built for developers.

Code doesn't matter IN THE EARLY DAYS.

This is similar to what I've observed over 25 years in the industry. In a startup, the code doesn't really matter; the market fit does.

But as time goes on your codebase has to mature, or else you end up using more and more resources on maintenance rather than innovation.

[−] anematode 44d ago

> But then the clean room implementations started showing up. People had taken Anthropic’s source code and rewritten Claude Code from scratch in other languages like Python and Rust.

Seems like the phrase "clean room" is the new "nonplussed"... how does this make any sense?

[−] twelfthnight 44d ago
Seems equally valid to come out of this with the takeaway that code quality _does_ matter, because poor coding practices are what led to the leak.

Sure, the weights are where the real value lives, but if the quality is so lax they leak their whole codebase, maybe they are just lucky they didn’t leak customer data or the model weights? If that did happen, the entire business might evaporate overnight.

[−] slopinthebag 44d ago
Claude Code proves you don't need quality code — you just need hundreds of billions of dollars to produce a best-in-class LLM and then use your legal team to force the extreamly subsidised usage of it through your own agent harness. Or in other words, shitty software + massive moat = users.

Seriously, if Anthropic were like oAI and let you use their subscription plans with any agent harness, how many users would CC instantly start bleeding? They're #39 in terminal bench and they get beaten by a harness that provides a single tool: tmux. You can literally get better results by giving Opus 4.6 only a tmux session and having it do everything with bash commands.

It seems premature to make sweeping claims about code quality, especially since the main reason to desire a well architected codebase is for development over the long haul.

[−] r0x0r007 44d ago
I dont understand why so many people have the need to emphasize the code vs product battle. There is no battle. Coding/developing/software engineering is a skill, that just like any other skill has certain requirements and best practices that have to be followed in order to make a quality, maintainable and adaptable application that can stand the test of time from the software perspective. Product, features, marketing bla bla that is entrepreneurship part, and is not related to software, other than directing the software requirements,but not beacuse product people think about requirements, but instead just because they come naturally from the required features they envision.Just because programmers can write code doesnt mean they can ship good products.Just because a plumber can lay pipes doesnt mean he can run his own company or invent a new way of laying pipes. But I will tell you that a bad plumber who lays pipes and doesnt know how to connect them, bend them or shield them will surely have inferior product/service in the long term. And by the way, success of a company is measured over time, we will see where claude code will be in 10 years time when the hype dips a little, then we can say yeah the code was bad but everyone loved it and uses it still. I mean they leaked entire code online and this guy says yeah code was bad but who cares, what world are we living in? The fact that anything got leaked is a serious breach of best practices and security also at this point, something a company that used to work for DoD(W) shouldnt be doing, it can even be considered a national threat at this point. I know mistakes happen, I do them all the time, but then again the 'best' companies should be almost immune to mistakes cause stakes are high.But of course, move fast and break things is more important.Am I wrong?
[−] komali2 44d ago

> bad code can build well-regarded products.

Yes, exactly. Products.

It seems like me and all the engineers I've known always have this established dichotomy: engineers, who want to write good code and to think a lot about user needs, and project managers/ executives/sales people, who want to make the non-negative numbers on accounting documents larger.

The truth is that to write "good software," you do need to take care, review code, not single-shot vibe code and not let LLMs run rampant. The other truth is that good software is not necessary good product; the converse is also true: bad product doesn't necessarily mean bad software. However there's not really a correlation, as this article points out: terrible software can be great product! In fact if writing terrible software lets you shit out more features, more quickly, you'll probably come ahead in business world than someone carefully writing good software but releasing more slowly. That's because the priorities and incentives in business world are often in contradiction to priorities and incentives in human world.

I think this is hard to grasp for those of us who have been taught our whole lives that money is a good scorekeeper for quality and efficacy. In reality it's absolutely not. Money is Disney bucks recording who's doing Disney World in the most optimal way. Outside of Disney World, your optimal in-park behavior is often suboptimal for out-of-park needs. The problem is we've mistaken Disney World for all of reality, or, let Walt Disney enclose our globe within the boundaries of his park.

> The object which labor produces confronts it as something alien, as a power independent of the producer.

[−] MrDresden 44d ago
And this is why so much software today runs extremely hot.

It's creators clearly care not for the efficiency of how it is built, which translates directly into how it runs.

This blog post is effectively being apologetic about the fact that this is alright, since at least they got product market fit. Except Anthropic is never going to go back and clean up the mess once (if) they become profitable.

I doubt anyone will like how things will be in 5 years time if this trend of releasing badly engineered spaghetti continues.

[−] sakopov 44d ago
In my opinion, the “code is garbage” argument is a moot point. Anthropic is in the business of removing humans from the SDLC. As long as their models can understand and update the code they generate it can remain garbage. They’re not optimizing for human comprehension of the output. They don’t even want you looking at it. And eventually the models will get good enough that you won’t have to.
[−] tylerloveamber 40d ago
The five-layer permission system discussion is interesting from a governance angle. Most small teams deploying Claude Code have no idea those permission layers exist — they approved the tool based on the marketing page, not the actual trust model.

The practical question for any CEO: if your developer's machine is running an agent with filesystem access, do you know what it can touch? The leaked code shows the answer is more nuanced than "it only touches what you tell it to."

Wrote a non-technical breakdown of what this means for AI tool policy (specifically the autonomous permissions mode and memory system that were hidden behind feature flags): https://www.aipolicydesk.com/blog/claude-code-leak-what-ceo-...

[−] ptnpzwqd 44d ago
I feel the conclusions here are a bit thin.

Code quality tends to have an impact on more than just aesthetics - and Claude Code certainly feels like a buggy mess from an end user's perspective.

Of course people still use Claude Code, but that is certainly because of the underlying models first and foremost. Most products don't have such a moat and would not nearly see as much tolerance from end users. If the Max subscriptions could be used with other harnesses, I am sure Anthropic would have to compete harder on the quality of the harness (to be fair, most AI based tooling seems pretty alpha these days, but eventually things will stabilize).

Polish is not everything, clearly, but it is a factor, and I feel Claude Code is maybe the worst example to use here, as it doesn't at all generalize to most other products.

[−] grey-area 44d ago
Points from the article.

1. The code is garbage and this means the end of software.

Now try maintaining it.

2. Code doesn’t matter (the same point restated).

No, we shouldn’t accept garbage code that breaks e.g. login as an acceptable cost of business.

3. It’s about product market fit.

OK, but what happens after product market fit when your code is hot garbage that nobody understands?

4. Anthropic can’t defend the copyright of their leaked code.

This I agree with and they are hoist by their own petard. Would anyone want the garbage though?

5. This leak doesn’t matter

I agree with the author but for different reasons - the value is the models, which are incredibly expensive to train, not the badly written scaffold surrounding it.

We also should not mistake current market value for use value.

Unlike the author who seems to have fully signed up for the LLM hype train I don’t see this as meaning code is dead, it’s an illustration of where fully relying on generative AI will take you - to a garbage unmaintainable mess which must be a nightmare to work with for humans or LLMs.

[−] rustystump 44d ago
I think this misses the target.

First, the twitter quote is standard toxic clapback nonsense. Gambling makes billions and does not add any value. Even facebook can argue it adds more value than gambling so this one is a dud.

People use claud code because of claud the model and not claud the harness. Cursor or a hacked up agent loop using opus or whatever are about as good. The magic is in the model not the harness here. This isnt to say the hardness is doesnt do anything.

The other bit this misses is that yes the product matters more then the code, and if the product burns battery/ram/etc doing nothing because the ai has crappy code or maybe something leaks or has a security issue, then that impacts the product.

[−] krisgenre 44d ago
Reminds me of a question that I asked couple of years ago - https://news.ycombinator.com/item?id=37176689
[−] redhale 43d ago
I think the argument here ignores a critical fact: a huge factor in Claude Code's popularity is the Claude Max plans. These plans give you potentially thousands of dollars of tokens for a capped $200.

Speaking for myself, I long for the day I can dump the comparatively garbage experience of Claude Code for something more enjoyable and OSS like OpenCode. But the fact is that it is simply not economically viable to do so.

So the PMF is not really for Claude Code alone -- it is for Claude Code + Claude Max.

[−] ggrab 44d ago
I don't think it's vibe coded garbage. Sure, the 3000-line print.ts is terrible, but there's some good patterns in there that were definitely prompted in by some experienced engineers -- the feature flag setup, the ..I_VERIFIED_THIS_IS_NOT_PATH_OR_CODE funny type hints, the overall structure. Just the usual signs this started as a PoC but quickly evolved into something much bigger. The codebase is a really interesting read.
[−] greyman 44d ago
From a moral perspective, I would argue that this is still theft of IP, even if it's a "clean room reimplementation". The code carries valuable information about what works and what doesn't — knowledge that Anthropic had to discover through real work and iteration. It's the same as a Chinese factory duplicating a product: they skipped the entire R&D phase and saved time and money.
[−] moezd 44d ago
Code quality matters when it becomes the code you have to maintain after six months. I'm honestly surprised by how some features of Claude Code seem to be held together with gum and duct tape.

That being said, if you're just beginning and looking for your market fit, or pitching to investors with a flashy demo, it doesn't need to be an architectural miracle, in fact it will waste your time.

[−] kykat 43d ago
Saying that code quality doesn't matter is simply false, if people enjoyed Claude code so much, anthropic wouldn't have to block direct API access to users.

Their "product market fit" is the LLM model itself, not the harness. The harness is completely replaceable in my opinion, it's just that Claude is the cheapest way to access the models.

[−] uduni 44d ago
" The real value in the AI ecosystem isn’t the model or the harness — it’s the integration of both working seamlessly together. "

Wut? The value in the ecosystem is the model. Harnesses are simple. Great models work nearly identically in every harness

[−] Grisu_FTP 44d ago
Afaik you can run Claude Code locally but every single demo i see uses it exclusively with apis, so are the local models already good enough to be worth it or is the only reasonable use for claude code with cloud models?
[−] hbarka 44d ago
OP should expand on #1, why he thinks it’s garbage. Claude Code is the REPL harness Anthropic built, can read, write, edit, bash. Pi, Gemini, Codex do the same, but they are not hinted as garbage. Where’s the beef?
[−] piker 44d ago

> The success of Claude Code and Cursor at the higher end of the market shows that even the people pickiest about their software (developers) will use your software regardless of how good the code is.

Seems wrong. Devs will whine, moan and nitpick about even free software but they can understand failure modes, navigate around bugs and file issues on GitHub. The quality bar is 10-100x amongst non-techno-savvy folks and enterprise users that are paying for your software. They’re far more “picky”.

[−] motbus3 44d ago
I don't know how many years of software engineering told us that new projects hastily done needs to be redone hastily later.
[−] boomlinde 44d ago
How likely is it that the broken release workflow that produced the leak was Claude's own work?
[−] Finbarr 44d ago
Who cares that the code is garbage? As the models get bigger and more powerful it will be trivial to fully refactor the whole codebase. It’s coming sooner than you think.
[−] CrzyLngPwd 44d ago
99.9999% of consumers never give code a single thought.

Most corporations never give code a single thought.

In the race to market, quality always suffers, and with such high stakes, it should surprise no one that AI companies are vibe-coding their own slop.

[−] jeremie_strand 44d ago
[dead]
[−] panavm 44d ago
[flagged]
[−] hyperlambda 44d ago
[flagged]
[−] riteshkew1001 41d ago
[flagged]
[−] sunilkumarai 44d ago
[dead]
[−] shailesh-1403 44d ago
[dead]
[−] gtsop 44d ago
[dead]
[−] michaelashley29 44d ago
[dead]
[−] ValveFan6969 44d ago
[dead]
[−] s99850 44d ago
[flagged]
[−] pregseahorses 44d ago
They just said this was an April Fools joke.
[−] miki123211 44d ago

> So of course the first thing people did was point and laugh.

This just validates my theory that open-sourcing old code that people have sentimental attachments for, and that you won't ever make any money off of, again is actually a terrible idea.

Everything about this leak is a long list of arguments why you shouldn't ever open source anything.

We, the developer community, have really dropped the ball here.