I've got a local Stable Diffusion client running. It's been interesting to check out. I've actually used it twice so far to help make images for Ars stories. It's a tool like anything else in my mind.
The whole "you're about to be replaced by an AI" narrative definitely feels far out at this point though. We'll see how solving these issues goes, I'm of the opinion it's going to be harder than some think, but as it is you really should try it for yourself to form an opinion.
2 cents is definitely cheap, but part of generating these images in a useful fashion is doing a ton of them. Sometimes you get lucky fast, but mostly it's a game of numbers. So those costs come with a multiplier attached to them in the real world.
I've got a local Stable Diffusion client running. It's been interesting to check out. I've actually used it twice so far to help make images for Ars stories. It's a tool like anything else in my mind.
The whole "you're about to be replaced by an AI" narrative definitely feels far out at this point though. We'll see how solving these issues goes, I'm of the opinion it's going to be harder than some think, but as it is you really should try it for yourself to form an opinion.
2 cents is definitely cheap, but part of generating these images in a useful fashion is doing a ton of them. Sometimes you get lucky fast, but mostly it's a game of numbers. So those costs come with a multiplier attached to them in the real world.
Do you think the API is for ingesting training material?So much copyrighted material to ingest, now easier via api. This will be interesting to watch.
I've got a local Stable Diffusion client running. It's been interesting to check out. I've actually used it twice so far to help make images for Ars stories. It's a tool like anything else in my mind.
The whole "you're about to be replaced by an AI" narrative definitely feels far out at this point though. We'll see how solving these issues goes, I'm of the opinion it's going to be harder than some think, but as it is you really should try it for yourself to form an opinion.
2 cents is definitely cheap, but part of generating these images in a useful fashion is doing a ton of them. Sometimes you get lucky fast, but mostly it's a game of numbers. So those costs come with a multiplier attached to them in the real world.
I've got a local Stable Diffusion client running. It's been interesting to check out. I've actually used it twice so far to help make images for Ars stories. It's a tool like anything else in my mind.
Check the promoted comment at the end of this story for an example:I've got a local Stable Diffusion client running. It's been interesting to check out. I've actually used it twice so far to help make images for Ars stories. It's a tool like anything else in my mind.
I'm sure I'm not the only one curious to know what that process looked like. What was the initial prompt and products? How did it help the creative process?
Didn’t OpenAI bill themselves as, “we’ll do this scary AI stuff in the open so it can be used for good, we gotta beat the bad actors to it and do it right and ethical!”How funny it is that "OpenAI" is closed source and gated while Stable Diffusion is wide open. Apparently they're also no longer a non-profit.
I'm not convinced the technology isn't going to continue to rapidly improve. Issues like modelling a specific style or subject are already being addressed with techniques like DreamBooth and eDiffi.I've got a local Stable Diffusion client running. It's been interesting to check out. I've actually used it twice so far to help make images for Ars stories. It's a tool like anything else in my mind.
The whole "you're about to be replaced by an AI" narrative definitely feels far out at this point though. We'll see how solving these issues goes, I'm of the opinion it's going to be harder than some think, but as it is you really should try it for yourself to form an opinion.
2 cents is definitely cheap, but part of generating these images in a useful fashion is doing a ton of them. Sometimes you get lucky fast, but mostly it's a game of numbers. So those costs come with a multiplier attached to them in the real world.
Well, as I said, I encourage people to use it for themselves, and not go by cherry picked images in news stories. It's easy to see the 'good stuff' and not realize how much time and effort it can often take to get there.I'm not convinced the technology isn't going to continue to rapidly improve. Issues like modelling a specific style or subject are already being addressed with techniques like DreamBooth and eDiffi.
There are things it's pretty good at. And things it freaking sucks at (hands lol).
It doesn't have to be good, it just has to be not bad enough.Well, as I said, I encourage people to use it for themselves, and not go by cherry picked images in news stories. It's easy to see the 'good stuff' and not realize how much time and effort it can often take to get there.
Also, if you're happy with "oh that looks neat" it's often not hard to get that. If you have something specific in mind trying to actually achieve what you're envisioning can be very difficult.
There are things it's pretty good at. And things it freaking sucks at (hands lol). And those latter things aren't necessarily easy to solve.
I'm not convinced yet myself.
Given that I've seen word around that they originally wanted to charge around several dollars for a single generated image, this seems like a massive pivot. I'd guess that Stable Diffusion's sudden popularity and integration with services that presumably may have been considering Dall-E, have forced their hand.
Only on the Moon would a horse be able to stand with its front legs crossed like that. And, since the bridle doesn't go over the top of her nose (can't see any plumbing back there, so I'm assuming it's a her), ain't much holding the bit in her mouth.
You were right to be suspicious. Turns out that's exactly what it meant (along with perhaps some tax shenanigans from the non-profit status, can't say for sure).Didn’t OpenAI bill themselves as, “we’ll do this scary AI stuff in the open so it can be used for good, we gotta beat the bad actors to it and do it right and ethical!”How funny it is that "OpenAI" is closed source and gated while Stable Diffusion is wide open. Apparently they're also no longer a non-profit.
I was suspicious that it didn’t just mean “fake being open until you’re able to rake it cash from your service that now has open source vibes that encourage trust”
Only on the Moon would a horse be able to stand with its front legs crossed like that. And, since the bridle doesn't go over the top of her nose (can't see any plumbing back there, so I'm assuming it's a her), ain't much holding the bit in her mouth.
Yeah, there's a lot of uncanny going on in those images. In particular, the shadows on the legs look somehow off to me, although I can't quite put my finger on it. It's like the light for those shadows is above the viewer's right shoulder, where all the other shadows are generated by a light at the far left.
What really bugged me, though, was that the corgis are blurry in all the pictures but the sunglasses are very sharp on two of the four pictures. What's up with that? Why are the corgis all blurry?
Can we get a Stable Diffusion beginners guide from Ars, please? Step by step, how to install on a local computer and how to use. I don’t understand what’s involved and I can’t find anything that is easy to follow. Most what I found is chaotic at best.
Can we get a Stable Diffusion beginners guide from Ars, please? Step by step, how to install on a local computer and how to use. I don’t understand what’s involved and I can’t find anything that is easy to follow. Most what I found is chaotic at best.
Diffusion Bee is a GUI-based Stable Diffusion installation for M1 Macs. It's a ~ 8 Gb one-click install.
https://diffusionbee.com
There are things it's pretty good at. And things it freaking sucks at (hands lol).
As an avid midjourney user, I am haunted by all of failed hand images it has produced. Truly terrifying.
This will just fuel more weaponized meme chaos. 3 Years ago, I never would've imagined I'd come to hate memes as much as I do today. And it all started with misappropriated terribly drawn french cartoon frogs.
I feel like the main contribution here is trying to sketch something up to get the real requirements. "No, it should be a black stallion. And it should be more front-facing. And the horse should be galloping. And the astronaut should be leaning forward like he's a jockey in a race. Dark horse in the space race, get it?"A few observations regarding the astronaut on horse photo: astronaut's hand looks really weird (as Aurich mentioned that is typical for AI). Horse's legs and tail are really weird. Tail looks more like a stocky tentacle. The horse's head also has an interesting structure on its face above the eyes like some sort of unicorn ancestor or a dinosaur with cranial bone plates.
Overall, though -- the program certainly produced a photo which matches the prompt. Interesting tech.
edit: also what is that creepy porthole looking thing on the horse's side? At first I thought it was part of a saddle but now it kinda looks like the horse has a hole in it. Creepy![]()
That sounds like a really useful bit of design software. Time to write my investor pitch...I feel like the main contribution here is trying to sketch something up to get the real requirements. "No, it should be a black stallion. And it should be more front-facing. And the horse should be galloping. And the astronaut should be leaning forward like he's a jockey in a race. Dark horse in the space race, get it?"A few observations regarding the astronaut on horse photo: astronaut's hand looks really weird (as Aurich mentioned that is typical for AI). Horse's legs and tail are really weird. Tail looks more like a stocky tentacle. The horse's head also has an interesting structure on its face above the eyes like some sort of unicorn ancestor or a dinosaur with cranial bone plates.
Overall, though -- the program certainly produced a photo which matches the prompt. Interesting tech.
edit: also what is that creepy porthole looking thing on the horse's side? At first I thought it was part of a saddle but now it kinda looks like the horse has a hole in it. Creepy![]()
Because I can't tell you how much time I've spent with users that is like "you asked for an astronaut on a horse, I delivered an astronaut on a horse" and they start going "yes, but not like that what I meant was..." That's with computer code though, I can't do art worth shit but I imagine creators deal with the same thing.
And yes you can manage this with contracts and change orders but it's still better to manage expectations up front as much as possible. Or they can click through a thousands of generated images with a hundred variations of the same prompt hoping to win the lottery.
So how is machine learning any different than an artist analyzing and incorporating ideas from a great a copyrighted work?So much copyrighted material to ingest, now easier via api. This will be interesting to watch.
There are things it's pretty good at. And things it freaking sucks at (hands lol).
As an avid midjourney user, I am haunted by all of failed hand images it has produced. Truly terrifying.
Hands and proportional legs or feet tend to be the bane of all artists, it's not a surprise that AI would screw them up more frequently even if the training set had a million images from artists who mastered hands.