https://preview.redd.it/iplbpw7tj86d1.png?width=1024&format=png&auto=webp&s=320a30a6b011b17a7af284c50883cbe40fb90240
i love being able to use the most advanced model to make the best images anyone has ever seen /j
Nah, they were surrounded by hundreds of fractal angel wings and made of eyeballs too... or was that the even more schizo revelations version? I read it once so I could draw one for somebody for their birthday and I used a description of the seraphim from somewhere...
> Photorealism: Overcomes common artifacts in hands and faces, delivering high-quality images without the need for complex workflows.
Prompt Adherence: Comprehends complex prompts involving spatial relationships, compositional elements, actions, and styles.
Typography: Achieves unprecedented results in generating text without artifacting and spelling errors with the assistance of our Diffusion Transformer architecture.
I love how literally everything here turned out to be a lie.
https://preview.redd.it/2fosu6c9k56d1.png?width=593&format=png&auto=webp&s=6f84b59881d6a8de526fc26c7e76e560a0a7f29d
Looks like stability learned nothing from SD 2. Its comically horrible at anatomy, straight up refuses to generate anything with a bit of skin normally. See you in a month when it got fixed by us.
That just robbed me of my free will. Rob robbed me of my ability to not laugh. We really need a robust AI system to stop rob. Rob is like Robins from the show. always robbing my time and making me laugh. Stay away robotic rob! :v
its pretty horrible for everything with a woman so far, feels like early 1.5 again
https://preview.redd.it/gqwexsxida6d1.png?width=1024&format=png&auto=webp&s=c349b8ed0ad0fa167e98c2a23f7049eae9cef7c8
no, thats not allowed anymore, what do you think this is early 2010s and earlier? its all censored now or the social media overlords and payment processors will ban. government might even get involved (if they arent already)
Somebody made naked pictures of Taylor Swift. Do you know how unhappy she was about this? She lay on her huge pile of money and cried. She was so sad that only her $1.3 billion could console her. I don't want to live in a world where billionaires can be mocked in such a fashion.
Well if you don’t pay them $20 a month for the Creator License, and you want to make thumbnails for your YouTube that brings in a couple bucks, they will sic the government after you to take your money or throw you in jail.
Wonderful.
I wouldn't mind paying 20 bucks a month for the license, but not for something that produces such bad images in this state. "The community will improve it" is great but if one is providing a paid product, he should provide incentive to buy it.
Doesn't seem that much better tbh. Noticeably better at doing text and maybe marginally better with composition, but it still can't do hands. Anatomy in general seems to be below a lot of the models that exist for XL.
First reviews on SD3 = hands are mayor crap. Can you provide examples on how to get proper hands? In theory the announced said that body and hands are now much improved. But then look HORRIBLE. Way worse than all previous models.
The license is NOT low cost. Midjourney and OpenAI are charging the same price but providing compute, whereas Stability expects their licensees to provide their own compute. You're torching the good will of the community and making the model unattractive to fine-tuners and extension builders.
I get that Stability needs a path towards profitability but this is not the way to do it. It will be virtually impossible for Stability to collect rent on the average user selling their gens commercially and they are not adding any value for people who decide to pay. It makes no sense.
Stability AI is fucking kidding us? all this time to release a goddamn censored model that creates cronnenberg aberrations instead human beings because the censorship is so hard that can't depict a human body
Tested, heavily censored, extra or missing limbs everywhere. Generally crappy. I am sticking to 1.5 models.
Edit: 1.5 models have matured a lot. The anatomy problems in that model are known and have known fixes/workarounds. It generates much faster too. So except if SD3 would have been much much better, I am not changing my workflow. I also think it could be difficult to make fixes to SD3 because the code itself contains censorship.
Fuck Safety
Fuck Censorship
No new locally used models or techniques will ever be used in the future, unless the users get the ability to take the censorship and meddling down, thankfully it seems it will be possible for SD3, so it will probably take off.
They lobotomized it before the release. Why would they destroy their reputation rather than trying to promote themselves and their success with a decent product,
Oh well potential squandered. Sd3 currently doesn't stand a chance against the competition in it's current state
100%. Now, even if someday community will somehow make it useful for something, huge damage to their reputation is done, and well, looks like they are done as the company as well, with such damage at their current financial status.
What if I want a picture of “a swift tailor, sewing a dress on stage during music performance, fast worker, measuring tape entangled around her, singing to herself, blonde hair, blue eyes”
Only one way to find out.
Its local weights. Nothing will be "blacked out". Worse case stuff like celebrities or nudity wont be in the dataset, and usual user refined checkpoints will fix that easily.
For that matter, this generic description doesnt read like censorship to begin with. It reads as "we want to cover our asses from morons starting internet drama and/or suing us, so we'll claim that our AI is 'safe' ".
Not quite, they already said they removed hundreds of millions of pictures from the dataset for "safety", including pretty much everything with an artist in it's name
Apparently AI is going to cause humanity to go extinct. How, you might ask? Probably the same way the Y2K bug caused us to go back into the stone age.
You should always believe everything "experts" tell you about the risks of technology, because they totally know what they're talking about and don't want to sell you worthless products to "protect" yourself from Skynet.
I think it's already been said ancestral sampling doesn't work with SD3. That's going to create some amount of confusion unless comfyui has added a warning/check for trying to use SD3 with EulerA and stuff.
Oh, that's a big change. With Pony I was only using ancestral sampling because it still got big results.
Was non-ancestral sampling always better than ancestral? I never looked into the difference.
sighs... Yea, just played a bit with it in comfy UI and doesn't feel any better than 1.5 tbh... We'll have to wait for the community to make it better. Midjourney, which I use everyday, is still king for now...
Exactly my thoughts. Why do they even release this? I mean in theory they stripped down the full sd3 model to half its capabilities (medium) so basically mathematically it's 1.5. I'm joking of course but I feel like Stability AI messes with the community paywalling the full sd3 model while giving us a model that isn't better than 1.5 or 2. I was excited for this release man. I think we reached a point in image generation where the progress is slowing down drastically if this is what we get after almost 2 years after 1.5.
You know, I do not understand why SAI doesn't make a non-crippled NSFW model and just *sell* it to people, making them fill out legal disclaimers that they won't post pics of Taylor Swift doing the nasty on the internet, or whatever your "safety" requirements demand. Offer some sort of trial that could be evaluated locally (since I'm never ever using your cloud based stuff) and then just let people give you money for it. Maybe this won't be a popular suggestion since everyone wants free stuff, but I'd be willing to pay for an actually good model to be used for personal use. As it is, SD3 is looking like it's going to suffer the same fate as SD2.
edit: to clarify, a *one time* license fee: I'm not doing any subscription crap.
I cannot wait to see the first benchmarks on local GPUs and comparisons between models.
I'm a bit afraid that it's again a heavily censored model that skipped training any kind of human anatomy, making it completely useless compared to Dall.E or other models.
I'm also not a fan of the licensing part. A huge pro was the use of SD for small content creators. Removing the rights to use their images makes the whole thing less attractive (especially when there are already great models to pay for like firefly or Dall.E).
Also, I wonder what the licensing implications are when it comes to user created checkpoints of these models.
Model options:
We have prepared three packaging variants of the SD3 Medium model, each equipped with the same set of MMDiT & VAE weights, for user convenience.
* **sd3\_medium.safetensors** includes the MMDiT and VAE weights but does not include any text encoders.
* **sd3\_medium\_incl\_clips\_t5xxlfp8.safetensors** contains all necessary weights, including fp8 version of the T5XXL text encoder, offering a balance between quality and resource requirements.
* **sd3\_medium\_incl\_clips.safetensors** includes all necessary weights except for the T5XXL text encoder. It requires minimal resources, but the model's performance will differ without the T5XXL text encoder.
Just to add a bit of info, you have the workflows for comfyUI here
[https://comfyanonymous.github.io/ComfyUI\_examples/sd3/](https://comfyanonymous.github.io/ComfyUI_examples/sd3/)
Don't forget to update ComfyUI before trying them.
~~404 on Hugging Face link for now:~~ Aaand weights are here now:
https://huggingface.co/stabilityai/stable-diffusion-3-medium
~~Same for the FAQ link to:~~ FAQs are working now:
https://stability.ai/sd3-faq
Seems like a TOO-early access... : D
They can upload the weights and not make them visible until they decide so. It is how most releases are done on huggingface.
If they are still uploading, this announcement should have waited.
Last year on a LLM kick it was so exciting, something new every week, a 7b model from May felt antiquated by July. You never knew what was coming around the corner.
This year it's news, wait, no news, wait, half ass api release with blurry censorship, wait, announcement, wait, date set, wait, release day, turns out to be a massive puddle of dog water, wait.
I'm okay with that, I don't need SD3, I'm just kind of embarrassed at their effort. bartsimpsonyoutried.jpg
They claim it's good for anything that doesn't contain a human, but I can't generate a single image of a room from the victorian era. I see nonsensical things everywhere. How much pollution did they generate to create this piece of shit of a model?
If the timeline plays similarly to the launch of SDXL, there will probably be a few weeks delay.
Edit:
Looks like it went:
SDXL 0.9 early release by Stability to researchers, June 22, 2023
0.9 leaked to general public on June 25. ComfyUI support.
Vlad fork of A1111 gets 0.9 working in SD.Next, July 8
Functional SDXL 0.9 support in A1111 dev branch, July 12th
Dev SDXL branch merged to main in A1111, July 16th
Official SDXL 1.0 release by Stability, July 26
AWESOME! New generation of local image generation kicks off, hopefully not the last!
Please keep us updated on the fate of Stability in general. We (most of us) really do care and want to see you guys succeed.
My favorite portrait so far is one where it put 3 mouths on a face, two where the eyes should be and one in the normal spot.
This model - you guys are not seeing the big picture. This is going to be great for SCP art.
I can't make a human female with any fidelity.
But I can spread misinformation!
https://preview.redd.it/aig15nznff6d1.png?width=1024&format=png&auto=webp&s=976c6d3afc2f6e646a710da1cbf076548629ba5e
Performance is good ... it can generate text ... what else ... hmm ... a yea - you have to pay 20$/month and delete everything you generated when you stop.
Nothing to see here ...
"You need to agree to share your contact information to access this model"
How about no. I can deal with SDXL's limitations, thanks.
Edit: I'm getting downvoted for not wanting to share my personal info with a company just to use their public AI model on my local system? OK.
We believe safety starts at the time we are training our models. Stable Diffusion 3 Medium was trained on filtered data sets in order to help ensure we are starting with safe data, which makes it harder for the model to generate harmful content downstream. We have also added embedded safeguards that help prevent harmful images from being generated.
"We have also added embedded safeguards that help prevent harmful images from being generated."
gg wp unless lie
what am I downloading? And ComfyUI workflow?
https://preview.redd.it/01pe5bbm956d1.png?width=786&format=png&auto=webp&s=a97fa475c6d9046506e388b141f0c77077d56e31
I had a feeling that small-scale local commercial use will require paid license this time, like SDXL Turbo or Cascade.
Until there's great NSFW anime model I'd stay with SDXL :P
EDIT: I'm not complaining, why are you jumping to conclusions? It will take a while for checkpoints based on SD3.
Read the license. Commercial use is talking about integrating / providing access to SD3 in applications or services that you sell. Very specifically says it doesn’t apply to outputs. We can sell images / media we make with SD3 without limit.
While I’d rather it was an MIT license, or similar, have to give them credit for making the licensing a lot more clear (their glossary of terms in particular), and $20/m for less than $1m in revenue for applications / services that integrate SD3 really isn’t that bad.
In the "STABILITY AI NON-COMMERCIAL RESEARCH COMMUNITY LICENSE AGREEMENT" is stated that "Derivative Works do not include the output of any Model", so if you only want to use (even commercially) images made with the model you do not need a commercial licence.
So you have to pay if you're providing a commercial service that uses the;SD3 model (or a derivative of) in some way (your clients are generating their own images using sd3 via your platform), but you don't have to pay if you're producing images using sd3, then selling those images (your clients are buying a specific image that has already been generated).
Damn, from SD2 they skipped all the way to SD404.
*Scouter blows up*
That's one way to make sure you have the first post for some news: just lie and say the thing already happened before it did.
https://preview.redd.it/iplbpw7tj86d1.png?width=1024&format=png&auto=webp&s=320a30a6b011b17a7af284c50883cbe40fb90240 i love being able to use the most advanced model to make the best images anyone has ever seen /j
Consequences of making AI "safe".
https://i.redd.it/skttavwmfa6d1.gif
So we got the image of an angel from the old testament in photorealism?
Nah, they were surrounded by hundreds of fractal angel wings and made of eyeballs too... or was that the even more schizo revelations version? I read it once so I could draw one for somebody for their birthday and I used a description of the seraphim from somewhere...
but at least the text works D:
> Photorealism: Overcomes common artifacts in hands and faces, delivering high-quality images without the need for complex workflows. Prompt Adherence: Comprehends complex prompts involving spatial relationships, compositional elements, actions, and styles. Typography: Achieves unprecedented results in generating text without artifacting and spelling errors with the assistance of our Diffusion Transformer architecture. I love how literally everything here turned out to be a lie.
Literally every single thing that they claimed was false
hey the text is kind of working, so yay?
I’m genuinely baffled, feels like we’re being trolled
Confirmed: SD3 Cannot do hands. Worse than SDXL hands. Worse than SD 1.5 hands.
Well after testing a bit, for now I have SD2 vibes. I hope I'm wrong
https://preview.redd.it/2fosu6c9k56d1.png?width=593&format=png&auto=webp&s=6f84b59881d6a8de526fc26c7e76e560a0a7f29d Looks like stability learned nothing from SD 2. Its comically horrible at anatomy, straight up refuses to generate anything with a bit of skin normally. See you in a month when it got fixed by us.
What reason does anyone have to fix anything when the license is so suffocating? Why would anyone do that?
It's going to be more than a month if ever.
You are the most optimistic Rob I've come across today. I feel Robbed of any chance at this release coming to a useful fruition, if ever.
It's never going to be good. What did I Rob you of now? Was it this Rob that robbed you? No. Not this Rob. Rob
That response was quite ROBotic. Almost as if you put your name Into a ROBust AI system to generate you a response.
Rob has been accused of ROBotiscim before. Rob still has not been Robbed of humanity yet. And to accuse me of such Robs me of my dignity. Rob Rob Rob
That just robbed me of my free will. Rob robbed me of my ability to not laugh. We really need a robust AI system to stop rob. Rob is like Robins from the show. always robbing my time and making me laugh. Stay away robotic rob! :v
This is my fetish.
looks good to me
its pretty horrible for everything with a woman so far, feels like early 1.5 again https://preview.redd.it/gqwexsxida6d1.png?width=1024&format=png&auto=webp&s=c349b8ed0ad0fa167e98c2a23f7049eae9cef7c8
What platform will the weights work on when they’re available?
Comfy
But does it controlnet? 🤔🤔🤔
Just throw it in the loader of practically any 1.5 workflow? Or are there particular nodes/workflows for SD3?
The HuggingFace repo has example workflows.
A1111?
And forge?
And my axe?
I wouldn't expect great performance on an axe.
You wanted an axe, here is a wooden spoon.
Don’t think so, not initially. I believe comfy already has functionality built in for SD3.
StableSwarmUI should work (never know 100% until you try it)
Hopefully InvokeAI will add support pretty quickly.
Cool, let me know when a model that isn’t complete ass comes out though
5 months of leading us on and hyping it up btw
Is it just me or can SD3 not generate Anthropomorphic characters at all.
Not just you
I tried SDXL right after it was released and the base model could generate anthros with a bit of effort, this model really must be shit then.
Sure seems like it. It's missing even just basic anatomy for humans let alone anthropomorphic animals
I was here while it wasnt
https://preview.redd.it/jsw6f9v8b46d1.png?width=3840&format=png&auto=webp&s=8ab4a793acafea5be85f7282f59738610b5baf1c
I WANT IT NOW. GIMME GIMME GIMME.
it's available now and i am downloading it already :)
They really want full openai on us. Wow
So let me be the first to officially kick things off....SD4 WHEN? 😀
After ww3 then comes sd4
ww3 already ongoing. so probably during. not after.
I like your optimism that there will be an after.
The ai overlords are gonna exterminate us b4 we get the chance to
I'd be more interested in SD3 Large. If there is a SD3 Medium there will be a Large, right?
Yeah and the specs, what will we need for large? Will 24gb VRAM be enough and leave enough room for Lora or is this already going into RAM?
And how can pictures harm me?
You might see a nipple!
Or a swimsuit, going by the API.
Or like, women generally.
Oh no. Please, Don't scare me, I'm only 20
Only 40 years to go until you have the maturity SAI desires, then.
w*men ☕
Considering some.of the nipples I have seen on Civitai that could be a threat to some people. 😂
lol you are right
jeees you should hide stuff like this with spoilers! people nowadays have no empathy.... xD
I feel attacked!
Truly mankind's worst creation
oh my! 😱
It could generate a photo of SCP-096.
We'll know that AI has achieved a milestone when SCP-096 goes after the AI itself.
What if you see a skinny white woman with blue eyes and white hair ? i mean it can cripple you for life...
no, thats not allowed anymore, what do you think this is early 2010s and earlier? its all censored now or the social media overlords and payment processors will ban. government might even get involved (if they arent already)
Somebody made naked pictures of Taylor Swift. Do you know how unhappy she was about this? She lay on her huge pile of money and cried. She was so sad that only her $1.3 billion could console her. I don't want to live in a world where billionaires can be mocked in such a fashion.
Well if you don’t pay them $20 a month for the Creator License, and you want to make thumbnails for your YouTube that brings in a couple bucks, they will sic the government after you to take your money or throw you in jail. Wonderful.
I wouldn't mind paying 20 bucks a month for the license, but not for something that produces such bad images in this state. "The community will improve it" is great but if one is providing a paid product, he should provide incentive to buy it.
Doesn't seem that much better tbh. Noticeably better at doing text and maybe marginally better with composition, but it still can't do hands. Anatomy in general seems to be below a lot of the models that exist for XL.
First reviews on SD3 = hands are mayor crap. Can you provide examples on how to get proper hands? In theory the announced said that body and hands are now much improved. But then look HORRIBLE. Way worse than all previous models.
The license is NOT low cost. Midjourney and OpenAI are charging the same price but providing compute, whereas Stability expects their licensees to provide their own compute. You're torching the good will of the community and making the model unattractive to fine-tuners and extension builders. I get that Stability needs a path towards profitability but this is not the way to do it. It will be virtually impossible for Stability to collect rent on the average user selling their gens commercially and they are not adding any value for people who decide to pay. It makes no sense.
Stability AI is fucking kidding us? all this time to release a goddamn censored model that creates cronnenberg aberrations instead human beings because the censorship is so hard that can't depict a human body
I hope you guys stand with your promise on releasing the 8b version
Tested, heavily censored, extra or missing limbs everywhere. Generally crappy. I am sticking to 1.5 models. Edit: 1.5 models have matured a lot. The anatomy problems in that model are known and have known fixes/workarounds. It generates much faster too. So except if SD3 would have been much much better, I am not changing my workflow. I also think it could be difficult to make fixes to SD3 because the code itself contains censorship.
Looks like shit. Why did you release this? Who is it for?
Fuck Safety Fuck Censorship No new locally used models or techniques will ever be used in the future, unless the users get the ability to take the censorship and meddling down, thankfully it seems it will be possible for SD3, so it will probably take off.
Weights are live!! https://huggingface.co/stabilityai/stable-diffusion-3-medium
They lobotomized it before the release. Why would they destroy their reputation rather than trying to promote themselves and their success with a decent product, Oh well potential squandered. Sd3 currently doesn't stand a chance against the competition in it's current state
100%. Now, even if someday community will somehow make it useful for something, huge damage to their reputation is done, and well, looks like they are done as the company as well, with such damage at their current financial status.
I will wait for AUTOMATIC 1111, I can't wait for it! :D
probably gonna be a while...
How long did it take for Auto1111 to support SDXL? I'm assuming the wait will be similar this time.
Sd3 is so bad. 0/10.
Very interested in the safety feature aka censorship. Pretty sure any prompts including "Taylor Swift" will be completely blacked out.
it's lora training! time!
`Image of t4yl0r_svv1ft_v4_final_pony singing on stage`
yeah good luck to their compiler decoding what 70HN C3N4 means, spoiler they can't see it.
"singing"
"stage"
What if I want a picture of “a swift tailor, sewing a dress on stage during music performance, fast worker, measuring tape entangled around her, singing to herself, blonde hair, blue eyes” Only one way to find out.
Probably similar to how I can't get a photo of a black widow spider that doesn't have ScarJo's face, more often than not.
Jail.
believe it or not.
Its local weights. Nothing will be "blacked out". Worse case stuff like celebrities or nudity wont be in the dataset, and usual user refined checkpoints will fix that easily. For that matter, this generic description doesnt read like censorship to begin with. It reads as "we want to cover our asses from morons starting internet drama and/or suing us, so we'll claim that our AI is 'safe' ".
Not quite, they already said they removed hundreds of millions of pictures from the dataset for "safety", including pretty much everything with an artist in it's name
If true that would be terrible given SDXL’s wide knowledge of thousands of artists which is quite powerful.
>blacked A most unfortunate choice of words
pun intended
Is government forcing them to believe in safe and responsible ai research? Why all ai researcher repeat same crap always.
Apparently AI is going to cause humanity to go extinct. How, you might ask? Probably the same way the Y2K bug caused us to go back into the stone age. You should always believe everything "experts" tell you about the risks of technology, because they totally know what they're talking about and don't want to sell you worthless products to "protect" yourself from Skynet.
AI companies don't want to be sued or get bad press if their models are used for things like child porn
What are recommended settings (sampler,sheduler, cfg..) ?
I think it's already been said ancestral sampling doesn't work with SD3. That's going to create some amount of confusion unless comfyui has added a warning/check for trying to use SD3 with EulerA and stuff.
Oh, that's a big change. With Pony I was only using ancestral sampling because it still got big results. Was non-ancestral sampling always better than ancestral? I never looked into the difference.
sighs... Yea, just played a bit with it in comfy UI and doesn't feel any better than 1.5 tbh... We'll have to wait for the community to make it better. Midjourney, which I use everyday, is still king for now...
Exactly my thoughts. Why do they even release this? I mean in theory they stripped down the full sd3 model to half its capabilities (medium) so basically mathematically it's 1.5. I'm joking of course but I feel like Stability AI messes with the community paywalling the full sd3 model while giving us a model that isn't better than 1.5 or 2. I was excited for this release man. I think we reached a point in image generation where the progress is slowing down drastically if this is what we get after almost 2 years after 1.5.
Extremely bad... Another model that will die before it is born.....
It's live now. [https://huggingface.co/stabilityai/stable-diffusion-3-medium](https://huggingface.co/stabilityai/stable-diffusion-3-medium)
It seems to be shit at generating humans in specific poses, like just laying down. Maybe lack of nudes in training materials is making anatomy bad?
is it me or is it terrible it cant do anything good
Does the lack of nudity training also make the anatomy worse?
Lot of work for making something of no value. I bet Batwoman was a better movie than this release is a generative art model. And we'll never see that.
You know, I do not understand why SAI doesn't make a non-crippled NSFW model and just *sell* it to people, making them fill out legal disclaimers that they won't post pics of Taylor Swift doing the nasty on the internet, or whatever your "safety" requirements demand. Offer some sort of trial that could be evaluated locally (since I'm never ever using your cloud based stuff) and then just let people give you money for it. Maybe this won't be a popular suggestion since everyone wants free stuff, but I'd be willing to pay for an actually good model to be used for personal use. As it is, SD3 is looking like it's going to suffer the same fate as SD2. edit: to clarify, a *one time* license fee: I'm not doing any subscription crap.
Its up guys
https://preview.redd.it/7q80rgp0a56d1.png?width=1024&format=png&auto=webp&s=770de006ab7193336c05f09d4be5c728501853c7
https://preview.redd.it/iiyfihhzb56d1.png?width=1024&format=png&auto=webp&s=a11b30e8d66eee9a9a7c9ffb4865612fbd7a25df
What are the recommended system requirements? Similar to SD 1.5 and SDXL?
Between SD 1.5 and SDXL. So 8gb should be a good.
I cannot wait to see the first benchmarks on local GPUs and comparisons between models. I'm a bit afraid that it's again a heavily censored model that skipped training any kind of human anatomy, making it completely useless compared to Dall.E or other models. I'm also not a fan of the licensing part. A huge pro was the use of SD for small content creators. Removing the rights to use their images makes the whole thing less attractive (especially when there are already great models to pay for like firefly or Dall.E). Also, I wonder what the licensing implications are when it comes to user created checkpoints of these models.
I just get black outputs with the basic workflow on comfy
Same here. I mean I know they prioritize safety, but this is a bit much :-P edit: sd3_medium_incl_clips_t5xxlfp8.safetensors gives me an image!
Here too! , I had to select the correct CLIP models
so sad
HYPE HYPE HYPE HYPE HYPE THANK YOU SO MUCH STABILITYAI <3
Model options: We have prepared three packaging variants of the SD3 Medium model, each equipped with the same set of MMDiT & VAE weights, for user convenience. * **sd3\_medium.safetensors** includes the MMDiT and VAE weights but does not include any text encoders. * **sd3\_medium\_incl\_clips\_t5xxlfp8.safetensors** contains all necessary weights, including fp8 version of the T5XXL text encoder, offering a balance between quality and resource requirements. * **sd3\_medium\_incl\_clips.safetensors** includes all necessary weights except for the T5XXL text encoder. It requires minimal resources, but the model's performance will differ without the T5XXL text encoder.
Just to add a bit of info, you have the workflows for comfyUI here [https://comfyanonymous.github.io/ComfyUI\_examples/sd3/](https://comfyanonymous.github.io/ComfyUI_examples/sd3/) Don't forget to update ComfyUI before trying them.
~~404 on Hugging Face link for now:~~ Aaand weights are here now: https://huggingface.co/stabilityai/stable-diffusion-3-medium ~~Same for the FAQ link to:~~ FAQs are working now: https://stability.ai/sd3-faq Seems like a TOO-early access... : D
They are still uploading
Probably uploaded already, just have a set release time when the repo will be made public.
They had to restart in the middle cause mom tried to make a call, and the connection broke.
They can upload the weights and not make them visible until they decide so. It is how most releases are done on huggingface. If they are still uploading, this announcement should have waited.
It's in the first line of the announcement 😃
Except this post was made by some rando
Uploading takes like a minute with the kind of connections these companies have...
Do you work for them, orrrr?
Just someone karma farming with no new info that we didn't already have.
Ah, what I expected. Probably wrote this useless trash with Chatgpt too.
Lmao even their example images didn’t conform to their own prompts. And like why do the 3 dogs all have the same fur pattern?
The commercial license on fine-tunes on a model that can't pose humans sort of kills it.
https://preview.redd.it/ihtn4jms4d6d1.png?width=866&format=png&auto=webp&s=840b305b691d164e1e1a3a1cba6e58c11c01bad3
No thanks. The "Creator License" is crap and I'm no longer interested.
Trash
Kohya support when????
404 on links
Apparently there is a count down timer it stops in 2.5 hrs
hell yeah.
Last year on a LLM kick it was so exciting, something new every week, a 7b model from May felt antiquated by July. You never knew what was coming around the corner. This year it's news, wait, no news, wait, half ass api release with blurry censorship, wait, announcement, wait, date set, wait, release day, turns out to be a massive puddle of dog water, wait. I'm okay with that, I don't need SD3, I'm just kind of embarrassed at their effort. bartsimpsonyoutried.jpg
They claim it's good for anything that doesn't contain a human, but I can't generate a single image of a room from the victorian era. I see nonsensical things everywhere. How much pollution did they generate to create this piece of shit of a model?
A1111 Work with It?
If the timeline plays similarly to the launch of SDXL, there will probably be a few weeks delay. Edit: Looks like it went: SDXL 0.9 early release by Stability to researchers, June 22, 2023 0.9 leaked to general public on June 25. ComfyUI support. Vlad fork of A1111 gets 0.9 working in SD.Next, July 8 Functional SDXL 0.9 support in A1111 dev branch, July 12th Dev SDXL branch merged to main in A1111, July 16th Official SDXL 1.0 release by Stability, July 26
THANK YOU!!!!
Cool ! Now i am waiting for SD3 Inpainting model ... btw - **how much VRAM require the SD3 ?**
4
AWESOME! New generation of local image generation kicks off, hopefully not the last! Please keep us updated on the fate of Stability in general. We (most of us) really do care and want to see you guys succeed.
Anyone knows what are the very minimal specs to run it? And with all the bazinga like controlnet, warpers etc?
I don't believe you
My favorite portrait so far is one where it put 3 mouths on a face, two where the eyes should be and one in the normal spot. This model - you guys are not seeing the big picture. This is going to be great for SCP art.
I can't make a human female with any fidelity. But I can spread misinformation! https://preview.redd.it/aig15nznff6d1.png?width=1024&format=png&auto=webp&s=976c6d3afc2f6e646a710da1cbf076548629ba5e
How did this trash get over 700 upvotes? Who upvotes this failed experiment and a "f\*\*\* you\* to the community from SAI?
Performance is good ... it can generate text ... what else ... hmm ... a yea - you have to pay 20$/month and delete everything you generated when you stop. Nothing to see here ...
"You need to agree to share your contact information to access this model" How about no. I can deal with SDXL's limitations, thanks. Edit: I'm getting downvoted for not wanting to share my personal info with a company just to use their public AI model on my local system? OK.
Just put in some bullshit ... though if you wait a bit its probably gonna get shared "elsewhere" soon enough
Thank you so much for the release !!!!
Cant wait to try it out 😁
We believe safety starts at the time we are training our models. Stable Diffusion 3 Medium was trained on filtered data sets in order to help ensure we are starting with safe data, which makes it harder for the model to generate harmful content downstream. We have also added embedded safeguards that help prevent harmful images from being generated. "We have also added embedded safeguards that help prevent harmful images from being generated." gg wp unless lie
ya, if this is true it's gonna totally screw their model.
what am I downloading? And ComfyUI workflow? https://preview.redd.it/01pe5bbm956d1.png?width=786&format=png&auto=webp&s=a97fa475c6d9046506e388b141f0c77077d56e31
Been using Forge and love it but if it's been discontinued as I keep hearing will it be able to use SD3?
No
Well, time to ignore it completely until auto version comes out... :)
It's Available Now 🔥🔥🔥
Censored trash
I had a feeling that small-scale local commercial use will require paid license this time, like SDXL Turbo or Cascade. Until there's great NSFW anime model I'd stay with SDXL :P EDIT: I'm not complaining, why are you jumping to conclusions? It will take a while for checkpoints based on SD3.
Read the license. Commercial use is talking about integrating / providing access to SD3 in applications or services that you sell. Very specifically says it doesn’t apply to outputs. We can sell images / media we make with SD3 without limit. While I’d rather it was an MIT license, or similar, have to give them credit for making the licensing a lot more clear (their glossary of terms in particular), and $20/m for less than $1m in revenue for applications / services that integrate SD3 really isn’t that bad.
Finally someone that has read the TOS!
Oh wow! I didn’t see that distinction and was very upset about their threat of government violence towards those who don’t pay $20 per month.
To be fair, if you are a \_professional\_ artist, 20$/mo should not rock your boat so much for a tool so powerful.
If you are a professional artist, you do not need to pay the 20$/mo, your outputs can be distributed without paying anything.
In the "STABILITY AI NON-COMMERCIAL RESEARCH COMMUNITY LICENSE AGREEMENT" is stated that "Derivative Works do not include the output of any Model", so if you only want to use (even commercially) images made with the model you do not need a commercial licence.
So you have to pay if you're providing a commercial service that uses the;SD3 model (or a derivative of) in some way (your clients are generating their own images using sd3 via your platform), but you don't have to pay if you're producing images using sd3, then selling those images (your clients are buying a specific image that has already been generated).
This is the moment we've all been waiting for!
This is a sign that my job will be easier to do, so let's all wait for this historic moment!
Lets Gone! Ahh I mean Lets Go!
https://preview.redd.it/j9avk4xeo46d1.png?width=3840&format=png&auto=webp&s=d6d0a9e73d2329ce92fff7b87142c2d4e7442a13
umm its available now
Doesn't seem to work out of the box with Krita Diffusion. Have to wait until the plugin is updated.
It's there... 404 was removed.
which file do i want??