Look! It's the one person who bought into HD-DVD :3HD-DVD.. obscure?!
Sir, you have offended my honour.
I challenge you to a dual.
in good faith: why would you use "summarize this youtube video" to get an answer to that question? Presumably Kagi would just point you towards any of the easily-available and freely-accessible documentation for the Raspberry Pi instead of finding a YouTube video and then giving you the option to summarize it like this.I sympathize with content creators, although I selfishly hope that this doesn't catch on. One of my absolute favorite features of Kagi is the "summarize this youtube video" feature, which "reads" the titling. It's absolutely incredible for when you want a quick answer to something, and the best answer is buried in some 20 minute long YT video with 19.5 minutes of "sponsored content" and the actual answer consists of three words at index 17:31.
"What is the command to get the CPU temperature on my Raspberry Pi?"
"CPU temperatures are HOT HOT HOT these days! I'm always trying to find my CPU temperature, and every time I do, I see that it's really high! That's why I'm so delighted to be sponsored today by Coolermaster, the best coolers! Coolermaster is the best! [10 more minutes]. About a decade ago, I began my journey to finding out how to measure CPU temperatures. I once hiked up to K2's peak to see what CPU temperatures were at 8,000 meters! On the way, Black Diamond crampons were my go-to crampons, and I thank them for sponsoring today's video! You can check the temperature by typing in vcgen . . . [youtube ad interrupts]."
.. only when it was clear HD-DVD lost the format war, and there was a fire sale on the drive and movies..Look! It's the one person who bought into HD-DVD :3
I linked it in the comments of the AI poisoning article the other day, partially in hopes it would be picked up.
How sadly hilariously bleak it will be when OpenAI files a DMCA takedown request to stop AI poisoners from spoiling OpenAI's theft of their work.
A few weeks ago, nothing other than power limitations, but those could be solvable. Now? I doubt East Bumfarkistan is on good enough terms with the US to get the GPUs for a training operation over there after Biden announced accelerator export restrictions, but that could be less of an issue with optimizations like what DeepSeek are doing (which actually were partially caused by a US ban on accelerators).On an side comment, whats to stop some country like east bumfarkistanfrom changing their laws to allow training on publicly accessible data and AI companies moving the training operations to that country, sort of how some countries dodge taxes by having subsidies in certain tax friendly countries.
I love this anecdote because as a MPC fan I would always encounter VLC fans who would staunchly deny even the implication that maybe there was a better video player out there. I've long since passed the days where I care but I will say only one person I convinced to at least try MPC ever changed back to VLC. And it was for an incredibly stupid reason.Huh. I remember when a fabsub group once used this exact same method to force VLC to fix its subtitle support, which at the time was notoriously bad. I had a friend with a Mac for which only VLC was available to handle subs at the time, and playing files with poisoned subs would cause it to crash, while MPC using the community codec pack would play the files just fine.
Calm down, sir, I’m sure it wasn’t intentional. What’s HD-DVD, by the way?HD-DVD.. obscure?!
Sir, you have offended my honour.
I challenge you to a dual.
Well, several countries (Singapore and Japan) do have a text and data mining exemption in their copyright laws that read like they allow for training on publicly-available data, even for commercial purposes, and both are not really East of Bumfuck. Japan is in the most-favoured list for AI chips, and Singapore is the second tier.A few weeks ago, nothing other than power limitations, but those could be solvable. Now? I doubt East Bumfarkistan is on good enough terms with the US to get the GPUs for a training operation over there after Biden announced accelerator export restrictions, but that could be less of an issue with optimizations like what DeepSeek are doing (which actually were partially caused by a US ban on accelerators).
It's not necessarily targeted at the creators of the LLMs. It's targeted at the people who are using them.While I adore this approach, personally, I can already hear the disingenuous pushback replies from LLM makers: "How dare content creators poison our learning models?!"
Well, maybe don't build your learning models on EVERYONE ELSE'S hard work and then treat it as your own?
"But... how will we beat competitors to market if we have to do the time-consuming initial legwork? Our sales director told the engineering folks to just stea... er, scrape everything on the internet."
Or probably asking an llm should work too.in good faith: why would you use "summarize this youtube video" to get an answer to that question? Presumably Kagi would just point you towards any of the easily-available and freely-accessible documentation for the Raspberry Pi instead of finding a YouTube video and then giving you the option to summarize it like this.
It’s entirely possible they weren’t entirely honest about how that model was trained, knowing what the reaction would be. It’s hurt the US AI industry and given people unrealistic expectations. Well played, if so.but that could be less of an issue with optimizations like what DeepSeek are doing
They will always find a solution and so will we. The key is to keep coming up with new solutions. As many as possible. Make it hard and even more expensive to steal our workI'm glad her video is getting more widespread coverage!
Although this probably means the primary method won't live long, her idea of clogging it with so, so much garbage that it becomes way too expensive to compute will probably endure.
Unless they simply make it so the bot ignores everything on the sides...
I doubt it. It would using filters to only concentrate on the bands of sound associated with human voice, and aggressively filtering out everything elseI wonder if you could use a similar technique against OpenAI's Whisper using ultrasound.
Nope. I bought a dual HD-DVD / Blu-Ray drive for my PC at the time. HD-DVD was clearly superior. Blu-Ray kept breaking due to javascript and idiotic software updates that removed backwards compatibility with certain titles, and the PC playback software was horrible, expensive, and poorly maintained. Still, Blu-Ray won out in the end. Fortunately we're mostly free of physical media nowadays.Look! It's the one person who bought into HD-DVD :3
A dual slot HD-DVD burner/player? Yeah, that would be a challenge to find.HD-DVD.. obscure?!
Sir, you have offended my honour.
I challenge you to a dual.
That's answered in the video; the uploader has the option to delete the autogenerated captions after they are generated:This is nice and all. But what's the point when YouTube publishes a transcript (based on the sound) right along with the video?
How cute human in 2025 thinks their "content" is original
Ow my ball
"884 + partners"
I shit you not stalking you as you visit this website
Pathetic & demonic "tech" cookies
Y'all ruined tech & the web been 90% bots & stalker propaganda for 20 years
Thanks to your fake intelligence it's now 99.9% "fake intelligence agents" (bots,)
That's what humans are clamoring for more bots to insult us & literally committ fraud by not even announcing it's a bot & 99% now can't read clock
So they think not actual human failing the turing test as an actual human adult haha
Cheers
YouTube's subtitle engine is really powerful. It can't do everything you can do with ASS and AegiSub (especially positioning subs frame by frame to track video), but the subs this YouTuber did for the Call of the Night anime opening are a great demonstration of how powerful it is. I didn't know it could do stuff like the falling text before seeing this. You have to turn captions on to see them:
View: https://www.youtube.com/watch?v=L96VbQ9ytWk
But in the video description, F4mi notes that "some people were having their phone crash due to the subtitles being too heavy," showing there is a bit of overhead cost to this kind of mischief.
I'm far from being an expert on a11y, so could you expand on how this method would negatively affect people with visual disabilities? As far as I understand, the poison text is not actually displayed on the screen during relevant portions of the video, and as for the letter-scrambling method, visually it renders identically to normal text. The most similar actually-harmful attack on a11y I am aware of is how some "pranksters" put meaningless or extremely long sections of text on the alt-text of still images, but does the analogy carry to this? Do screen readers actually attempt to parse the subtitle tracks on videos as they are playing?A large part of my job is web accessibility. Basically what this person is doing is shitting on people with visual disabilities in order to pwn AI scrapers.
They're not trying to poison the training data, they're messing with subtitles such that the finished LLM can't successfully read the subtitles and summarise the video.Perhaps fun but pointless because if it is an outlier then the algorithm will just effectively ignore it (basically it doesn't change the parameters in the model in the intended way). However, it might provide a means to identify if those works were used in the training.
YouTube doesn’t care about the amount of effort, it’s all about engagement and time on site. People watch the slop, so YouTube propagates it. People also watch high quality videos, and YouTube recommends them.
Most likely people will get bored of the slop and it will fade like ASMR videos, mukbangs, creator houses, and putting rubber bands on watermelons until they explode.