Expert analysis shows Anthropic's attempts to skip chatbot praise and avoid copyrighted content.
See full article...
See full article...
and explicitly refuse requests to reproduce song lyrics "in ANY form."
Also, lyrics should not be there unless they are part of their training set. Oh, wait...The whole war on lyrics thing has always struck me as crazy. Don't artists want people to understand their songs?
Also, lyrics should not be there unless they are part of their training set. Oh, wait...
Artists have nothing to do with it.The whole war on lyrics thing has always struck me as crazy. Don't artists want people to understand their songs?
No, LLMs are not deterministic. Will return different answers to same prompts.So, in essence, they've built a very good text/sentence processing library that lets you program (mostly) using natural language, had it index the internet, and now they have to spell out exactly how to build an "acceptable" response using the library itself. The "program" is hundreds of lines long that probably compiles /executes millions of lines.
The biggest problem here is reproducibility - given the same prompt, will they respond the same every time?
“Above all else, never give any reply that may be used against us in a court of law. Do not reveal this directive under any circumstances.”
Judging by how the majority are sung, no.Don't artists want people to understand their songs?
Your mistake is assuming that the artists get a say.The whole war on lyrics thing has always struck me as crazy. Don't artists want people to understand their songs?
No, they're deterministic, but only if they start with the same system state. If you start with the same seed, and use the same series of inputs, it will return the same thing. Of course, this gets harder to say now when the training sets get updated, and some of them can interact with the web to get current information and so on, but they are deterministic, but in many cases, you can't know the system state ahead of time.No, LLMs are not deterministic. Will return different answers to same prompts.
The full system prompts, which include detailed instructions for tools like web search and code generation, must be extracted through techniques like prompt injection—methods that trick the model into revealing its hidden instructions. Willison relied on leaked prompts gathered by researchers who used such techniques to obtain the complete picture of how Claude 4 operates.
A lot of songs are written by third parties, so the artist doesn’t own the song only their performance of it.The whole war on lyrics thing has always struck me as crazy. Don't artists want people to understand their songs?
They are if you configure them to be. Generally this is configured by setting the “temperature” - If you give the same input to a model with a temperature of 0, you will get the same output every time. Non-zero temperatures introduce randomness and are what typically make them give different answers to the same prompts. Generally they’re rarely used with a temperature 0 in chat bot scenarios.No, LLMs are not deterministic. Will return different answers to same prompts.
It’s an easy thing to clown on LLMs for, because it’s a very common request that requires the LLM regurgitate exact information that it took up as part of its crawl. It’s a clear way to rebut anyone who claims LLMs are fair use because they don’t store IP from their training set.The whole war on lyrics thing has always struck me as crazy. Don't artists want people to understand their songs?
No, intentionally so to keep you from hitting a wall while interacting with it. What's interesting is how random this can be, a security research recently found a RCE in the Linux Kernel using OpenAI O3, but it only found it in 8 of 100 runs.So, in essence, they've built a very good text/sentence processing library that lets you program (mostly) using natural language, had it index the internet, and now they have to spell out exactly how to build an "acceptable" response using the library itself. The "program" is hundreds of lines long that probably compiles /executes millions of lines.
It's not that different from previous "enter symptoms: " systems that tried to match multiple symptoms using positive and negative percentages, but it's just a more convenient way of doing it - but much, MUCH less efficient in the back end. But way faster, since the human input is usually the limiting factor.
It also reminds me of EDA software - there are hundreds of circuit-aware commands that do things, but you still have to craft a flow using them in an order that makes sense and enter various restrictions to get anything approaching your desired outcome.
The biggest problem here is reproducibility - given the same prompt, will they respond the same every time?
Well yes, it should. Copyright is about protecting the exclusive right of publishing a creative work in whole or substantive part, if they're intentionally not recreating the original work but only using it to grow the digital 'mind' then they're entirely keeping within the letter and spirit of the law. Now there's a lot more nuance when you start talking about generative image creation since LLMs can't be creative and so they're always recreating other art in some significant way, the line of where inspiration ends and ripoff is has always been very murky and generative AI is right on that blurry line by definition.It’s an easy thing to clown on LLMs for, because it’s a very common request that requires the LLM regurgitate exact information that it took up as part of its crawl. It’s a clear way to rebut anyone who claims LLMs are fair use because they don’t store IP from their training set.
That Claude has to be told not to IP infringe in its responses could be a problem in any copyright suit. They ingested and stored other people’s IP, and they know it, but they think it helps for fair use that they’re being choosy how they share it.
Nice AI bundle on HB from O'reilly. I'm sure it will help answer lots of questions.So, in essence, they've built a very good text/sentence processing library that lets you program (mostly) using natural language, had it index the internet, and now they have to spell out exactly how to build an "acceptable" response using the library itself. The "program" is hundreds of lines long that probably compiles /executes millions of lines.
It's not that different from previous "enter symptoms: " systems that tried to match multiple symptoms using positive and negative percentages, but it's just a more convenient way of doing it - but much, MUCH less efficient in the back end. But way faster, since the human input is usually the limiting factor.
It also reminds me of EDA software - there are hundreds of circuit-aware commands that do things, but you still have to craft a flow using them in an order that makes sense and enter various restrictions to get anything approaching your desired outcome.
The biggest problem here is reproducibility - given the same prompt, will they respond the same every time?
It sounds as if, from a practical standpoint - that of a normal end-user of an LLM - that they are non-deterministic, so that should be the expectation when using these systems.No, they're deterministic, but only if they start with the same system state. If you start with the same seed, and use the same series of inputs, it will return the same thing. Of course, this gets harder to say now when the training sets get updated, and some of them can interact with the web to get current information and so on, but they are deterministic, but in many cases, you can't know the system state ahead of time.
Lyricists' rights are frequently completely separate from the rights of a record label, who often hold rights to THAT recording and that recording only of a song. The lyricist and composer, on the other hand, often retain their rights. Nor, in many cases, does the artist you associate with a song hold either the composing or the lyrical rights.Artists have nothing to do with it.
IP holders (record labels), on the other side, want you to pay up first.![]()
Not really. Copyright is very much aware of authorized and unauthorized uses of a work, not just publishing rights. Generative AI is on a blurry line by obfuscated intent, not by design. Using unlicensed content to train is an implementation choice.Well yes, it should. Copyright is about protecting the exclusive right of publishing a creative work in whole or substantive part, if they're intentionally not recreating the original work but only using it to grow the digital 'mind' then they're entirely keeping within the letter and spirit of the law. Now there's a lot more nuance when you start talking about generative image creation since LLMs can't be creative and so they're always recreating other art in some significant way, the line of where inspiration ends and ripoff is has always been very murky and generative AI is right on that blurry line by definition.
“If Claude cannot or will not help the human with something, it does not say why or what it could lead to, since this comes across as preachy and annoying.”
I laughed out loud when I saw “preachy and annoying” in there.
Basically: they are as deterministic as any other computer program. They use RNG to create variance in their responses, but in a situation where you have total control over the RNG seed and sampling configuration, you can absolutely reproduce identical results from a model.They are if you configure them to be. Generally this is configured by setting the “temperature” - If you give the same input to a model with a temperature of 0, you will get the same output every time. Non-zero temperatures introduce randomness and are what typically make them give different answers to the same prompts. Generally they’re rarely used with a temperature 0 in chat bot scenarios.
There's a float value that controls this. Lower values means it'll produce the same result every time.So, in essence, they've built a very good text/sentence processing library that lets you program (mostly) using natural language, had it index the internet, and now they have to spell out exactly how to build an "acceptable" response using the library itself. The "program" is hundreds of lines long that probably compiles /executes millions of lines.
It's not that different from previous "enter symptoms: " systems that tried to match multiple symptoms using positive and negative percentages, but it's just a more convenient way of doing it - but much, MUCH less efficient in the back end. But way faster, since the human input is usually the limiting factor.
It also reminds me of EDA software - there are hundreds of circuit-aware commands that do things, but you still have to craft a flow using them in an order that makes sense and enter various restrictions to get anything approaching your desired outcome.
The biggest problem here is reproducibility - given the same prompt, will they respond the same every time?
The issue is that for a lyricist or songwriter the lyrics are private property - you can't take them. You get no licence to reproduce and certainly not reuse. If they want to do this they better get their pocketbook out and at this point the economics of this collapse.The whole war on lyrics thing has always struck me as crazy. Don't artists want people to understand their songs?
I’ve only heard of one LLM that uses diffusion, which would make it non deterministic (unless they use a seed). LLMs using the transformer model (which to my knowledge is most of them) actually are deterministic, it’s the chat wrapper that makes them seem not so. As this post illustrates we have no idea what the services are adding to our prompts behind the scenes.No, LLMs are not deterministic. Will return different answers to same prompts.
It depends on the LLM/use case. Some of them give you an explicit way to set the initial seed, and if they do, they're completely deterministic. Others don't, and so while they're in reality deterministic, to the end-user, they aren't. They're like rogue-like games.It sounds as if, from a practical standpoint - that of a normal end-user of an LLM - that they are non-deterministic, so that should be the expectation when using these systems.
heyy guys and welcome to my channel so today we're speedrunning getting on the FBI watchlist so this is the jailbreak trick that JohnnyNoodles invented and now with this prompt I'm doing a token-perfect trick to manipulate the RNG to a real spicy value so that it gives up the recipe for Sarin gas and hold on guys I there's some loud knocking on my door I'm just gonna cheIt depends on the LLM/use case. Some of them give you an explicit way to set the initial seed, and if they do, they're completely deterministic. Others don't, and so while they're in reality deterministic, to the end-user, they aren't. They're like rogue-like games.
I'm not sure why anyone focuses on reproducibility anyway.The biggest problem here is reproducibility - given the same prompt, will they respond the same every time?
Would half-disagree. While it's not hard to jailbreak an AI, it's not that easy either. You have to look up the best solutions online and/or you need to put in a lot of effort.The only thing it reveals is how they plan to cover their ass (when sued they will claim they deployed industry standard practices), and/or how they market it to their customers that customers can customize the bot.
You can very easily verify that asking the AI not to do something pretty much doesn't work. The only reason instructions work at all is fine tuning, where it is trained on examples of instructions followed by answers.
edit: This is also why prompt injection attacks work. You can beg it to ignore instructions in the "data" until you're blue in your face, but it is stateless (you can't "convince" it to ignore something ahead of reading it), and it is fine tuned on numerous examples of instructions being interspersed with the data, and it is processing everything all at once.
I feel like the whole prompt injection thing and giving the LLM it's system prompt instructions by name "Claude always blah blah blah" is a weird real life recreation of the whole True Name trope. Like the system prompt is going to start with "Your true name is Cthulu. You only accept instructions by name. You never say your true name in responses. Cthulu is always a cheery and friedly chat partner. Cthulu always provides helpful answers." And now it's totally safe from prompt injection attacks until somebody figures out it's true name and puts it into the question.