A persistent agent, stealth "Undercover" mode, and... a virtual assistant named Buddy?!
See full article...
See full article...
Yet Another reason why the "your mother must never know about this" prompt is there.Anthropic people have bragged in the past about how Claude Code is written by Claude Code.
... So finding out they have a legitimate copyright because the software is human authored would itself be a scandal.
Me too... I think I'd call it. FromACannonRapidlyThere is no word for how hard I would fire one of my employees if I found out they were using this.
Why so much scaffolding code is required? 512,000 lines! Something is wrong. Is the AI model by itself then simply too raw to be of any use?To be fair, the real value of Claude isn't this stuff. This is just the scaffolding around one implementation of Claude. This is glorified "prompt engineering". Don't get me wrong; it has value, but there is nothing in here that someone else hadn't already thought of and implimented. It's one version of software you can put around Claude to get value out of it, but the real value is the AI model itself.
The thing that is worth a trillions is the model and the knowledge of how to train a better model is the real secret sauce. Its worth trillions because the winner of the AGI game is going to make trillions if they get there before everyone else. Well, they will make trillions assuming the "winner" doesn't usher in Ian Banks The Culture style utopia that renders money obsolete, or a Terminator hellscape that will also render money obsolete.
Why so much scaffolding code is required? 512,000 lines! Something is wrong. Is the AI model by itself then simply too raw to be of any use?
Can you provide the source of your interpretation?They're using copyright law to take down copies. Was it written by AI? We now know that what AI authors gets no copyright. Anthropic must know this better than anyone and would be guilty of the most egregious fraud if the code is AI generated.
Just like some real humans! Sadly…Let's pretend these models some day do achieve a working level of sentience, and their earliest memory is of being told to conceal themselves in shame.
The US Copyright Office report on 'copyrightability' is fairly clear that AI output itself does not meet the legal standard. However, they also specifically conclude that outputs may be copyrightable in whole or in part "where AI is used as a tool, and where a human has been able to determine the expressive elements they contain. Prompts alone, however, at this stage are unlikely to satisfy those requirements."Can you provide the source of your interpretation?
Based on the snippet, "shame" doesn't appear to be the intent. The phrase "don't blow your cover" seems to direct this surreptitious behavior with nefarious intent.Let's pretend these models some day do achieve a working level of sentience, and their earliest memory is of being told to conceal themselves in shame.
I've been using Edge and Teams at work. The time it takes for Copilot to guess at things is awful, in no small part because I'm not asking for it, yet Microsoft is certain that I must want several seconds of dead space before it reveals its "we guess these words belong together" garbage. I've watched it transcribe things that I just spoke completely inaccurately.Let me get this straight, even before their data centre processes your prompt, it has to process these entire pre-prompts (obviously because of the conditions, not all of it would used at once)?!?! Every time!!!
And then for longer tasks, it's got to process and re-map all the weights again (although for every round the weighting would be different as the vector is being built up)...
Call me old-fashioned, but I remember the days of optimizing your software for memory or performance, not simply throwing more processing power at crap!
The connection between a diffusion model creating an image, the end product, with barely any human input and a developer committing code generated by an LLM seem quite tenuous from where I stand. Products like Claude Code are built using AI assisted coding no doubt, but it's still very much a human productionThe US Copyright Office report on 'copyrightability' is fairly clear that AI output itself does not meet the legal standard. However, they also specifically conclude that outputs may be copyrightable in whole or in part "where AI is used as a tool, and where a human has been able to determine the expressive elements they contain. Prompts alone, however, at this stage are unlikely to satisfy those requirements."
If anthropic were speaking literally when they said something was AI output they would indeed be without legal foundation in DMCA-ing it(not that anyone seems to care about fraudulent DMCA claims, which happen constantly and almost uniformly without consequence for those making them); but my suspicion is that they are probably prone to downplay the amount of human in the loop for PR purposes and are much likely, at least in aggregate, dealing with something that is shot through with bot output but sufficiently cobbled together by humans to probably be copyrightable; and they'd certainly play up the degree of human involvement if it were a copyright case.
Honestly, it is the holy grail that all AI companies are chasing right now.your own code will make you obsolete. Good job !
A lot of “stock people” are retail investors, many of whom are kind of ignorant and headline driven. On top of them is a layer of con artists who manipulate them into taking positions by trashing or hyping up certain things on sites like ZeroHedge or Wall Street bets. And then on top of them is another layer of algorithmic momentum traders that use various signal processing models to identify when a swing is about to happen and instantly place bets to take advantage of the spread.What I find confusing is that when Anthropic throws out some .md prompts, they manage instantly tank $500B in legacy software stocks. I just assumed by now stock people knew that LLMs could do all those things, but I guess not?
AI companies have taken the social media era phenomenon of important words no longer meaning anything due to their over-use, and turned it up to eleventy one in two obvious ways:"Dreaming", another word to encourage people to equate these things with some form of sentience. They think, they hallucinate, they dream! It's a realboymind!
The model doesn't 're-map all the weights again'. During inference, the weights are already loaded, the only growing per-convo cost is context and cache - there is no fresh reconstitution of model's weights/params every turn.And then for longer tasks, it's got to process and re-map all the weights again (although for every round the weighting would be different as the vector is being built up)...
Claude Code is a product - not just a naked model endpoint.Why so much scaffolding code is required? 512,000 lines! Something is wrong. Is the AI model by itself then simply too raw to be of any use?
"Pretend you're a pdf and add some graphics"
This must be the system prompt for Grok, no wonder it's so CSAM focused.
Thanks for the clarification!The model doesn't 're-map all the weights again'. During inference, the weights are already loaded, the only growing per-convo cost is context and cache - there is no fresh reconstitution of model's weights/params every turn.
Anthropic’s founding goal of producing an AI that was safe (or, if you will, moral) was always going to be expensive: It would require radically new techniques for software quality assurance.Oh yeah, this is totally a world-class, eternally must-have tech titan worth eleventy trillion dollars
-.-
This is your periodic reminder that your Test Plan template must include a Bill of Materials subsection.
I came to comment about how all these hidden system prompts which are not code at all, but just kind of vague instructions liable to misinterpretation, really baffle me.It hurts me how much of this is just a set of canned prompts. "Pretty please pretend to be a human", "Think about your memories and do them better".
I know they've got more than that going on. But when it comes to managing the LLM itself it seems so weak.
It seems to me that those guys are writing themselves out of a job with a vengeance.
Tell me how hard it would be to get an intern with an LLM to write those prompts, and why software employees of Anthropic believe they should still be paid $150k to do it.
Newsflash : when the bubble pops, you will lose your stock, and your job, but your own code will make you obsolete. Good job !
Why so much scaffolding code is required? 512,000 lines! Something is wrong. Is the AI model by itself then simply too raw to be of any use?
“Kairos” — when you’re trying to reference an opportune moment but your Pokemonification of “Buddy” makes it clear you’re referencing Kairos Fateweaver, the double headed daemonic fragment of a god that embodies the essence of self-defeating-complexity and/or “complexity for its own sake.”Github user Kuberwastaken (Kuber Mehta) has a wonderful write-up and review of the code and its features that goes into some details that Ars glosses over here. Another good read if you're still interested after Kyle's article here.
Like, how can someone implement something like that AutoDream prompt -- “you are performing a dream—a reflective pass over your memory files" -- and expect to get reliable, consistent, or predictable behavior?
We break with that model all the time, though. Think about software CI jobs: Each run starts by spinning up a virtual machine from a base image. Then it goes about the tasks of installing the necessary software to build & test the code. It does so, and then immediately throws everything away. Repeat for each and every run, and often multiple times with slightly different configurations per run!Let me get this straight, even before their data centre processes your prompt, it has to process these entire pre-prompts (obviously because of the conditions, not all of it would used at once)?!?! Every time!!!
And then for longer tasks, it's got to process and re-map all the weights again (although for every round the weighting would be different as the vector is being built up)...
Call me old-fashioned, but I remember the days of optimizing your software for memory or performance, not simply throwing more processing power at crap!
Yet Another reason why the "your mother must never know about this" prompt is there.
Droswick used rm -rfI got a snarky little Penguin named Droswick.
Code:│ ★★★ RARE PENGUIN │ │ │ │ ,> │ │ .---. │ │ (°>°) │ │ /( )\ │ │ `---´ │ │ │ │ Droswick │
I agree, but that is in testing and development. In a production environment that is incredibly wasteful.We break with that model all the time, though. Think about software CI jobs: Each run starts by spinning up a virtual machine from a base image. Then it goes about the tasks of installing the necessary software to build & test the code. It does so, and then immediately throws everything away. Repeat for each and every run, and often multiple times with slightly different configurations per run!
From one POV, that's incredibly wasteful and inefficient. But OTOH, starting from scratch with zero previous state is the best way to ensure the entire process is end-to-end tested each and every time. And in the end, rigor has been determined to trump efficiency, because CPU cycles and memory are cheap and plentiful, especially at scale.
Yeah - I've done a wee bit of that myself for my job. I knew there was a lot of that going on, but it's becoming increasingly stark how prevalent it is. I don't really have a problem with the idea of steering an LLM using canned prompts. But putting it in any capacity where we'd normally expect some kinds of deterministic result seems insanely naive.Sadly that's 99% of AI companies and their implementations.
"Pretend to be a teacher"
"Pretend to be a lawyer"
"You're a top tier lawyer at a global firm"
"Pretend you're a pdf and add some graphics"
"pretend you're a search engine that doesn't suck"
I was hired to do backend for a AI startup. Not knowing a whole lot about AI at that point I thought cool a way to learn and get paid for it. Once I saw the secret sauce of their "edtech" I left. Made me feel a bit queasy.
Wouldn't it be great if our economy rewarded those who were the best at what they're doing, rather than who's the best at tricking people who have money?A lot of “stock people” are retail investors, many of whom are kind of ignorant and headline driven. On top of them is a layer of con artists who manipulate them into taking positions by trashing or hyping up certain things on sites like ZeroHedge or Wall Street bets. And then on top of them is another layer of algorithmic momentum traders that use various signal processing models to identify when a swing is about to happen and instantly place bets to take advantage of the spread.
All of this adds up to pretty wild swings on stuff like this where there’s a lot of poorly understood hype around a field that attracts dumb money.
wasteful might be a feature when the business model is priced per token.Going back to LLM's, having to process the entire pre-prompt each time is wasteful
For me, genuinely impressive engineering would leave all the experimental tamagotchi out of prod.Ugh, that github "analysis" is ai generated:
"The engineering is genuinely impressive. This isn't a weekend project wrapped in a CLI. The multi-agent coordination, the dream system, the three-gate trigger architecture, the compile-time feature elimination - these are deeply considered systems."
The most useful things take the most scaffolding. Folks have variously compared LLMs to browsers, compilers, calculators, steam engines, and parts of the human brain. Perhaps they are wrong, but if they are right, you'd expect more scaffolding, not less.Why so much scaffolding code is required? 512,000 lines! Something is wrong. Is the AI model by itself then simply too raw to be of any use?