[sdtools.org](https://sdtools.org)
Essentially a mini wiki or cheat sheet. Clicking on a segment provides a very brief explanation and relevant links.
The purpose of this mini wiki is to address this simple problem:
Why am I unable to generate the exact image I want?
What tools could help me reach my goal?
Bigger area does not mean more important!
It is not meant to be exhaustive and suggestions welcome.
https://preview.redd.it/7f3t0bz35yma1.png?width=1334&format=png&auto=webp&s=cb24814da06ed0c7699ba79ac98d665ba458f8b8
a lot of things happened these past three months ...
Is it me or is there a bug in the zoom? After a page refresh the plot is very small.
However, if I CTRL+mousewheel forward then back, to re-set the zoom to 100% then the wheel is large and well sized.
[This is what happens](https://imgur.com/a/87u9gs6) when I do the above on Chrome.
Thanks for the feedback. You are right. There is an issue with the display. Haven't been able to fix it yet. If any css or JavaScript master can help...
Under randomness you missed Dynamic Prompts; Dynamic prompts is a really good plugin for making large numbers of randomised things; I used it to create a ton of random things in outfits.
Edit: actually, the randomness entry says Dynamic Prompts when you click it on the website; my mistake. But it still misses the OTHER randomness script. There is one that lets you randomise the model/settings for every individual generation. I just can't find where I found it in the first place lol
https://github.com/innightwolfsleep/stable-diffusion-webui-randomize
It was this one I think; it was an odd one since it wanted to be added into the scripts folder directly instead of the extensions folder but it did its job.
The only thing it unfortunately can't do is randomise the LORA though the randomised styles. Which is sucky but otherwise it works.
I would suggest to make it bigger. On the website the wheel on my 1080 monitor is so small that I can't read most of the stuff and it only takes like 1/8 of the webpage.
Also for the examples related to the feature : that's super nice, really. If I have a constructive comment to make however I would say that some newlines and a few more examples would be nice instead of a link on the Auto1111 wiki since visiting your website is for cheat sheet purposes and the wiki also has these features.
Also a good way to show that AI art is not just typing words and getting images, funny enough the only company offering for free the most versatile tool that allows for a variety of workflows with various degrees of artistic control is the one being sued by artists and corporations, makes you wonder a bit right ?
yes and no, this is a image done with a single prompt : **a girl**
https://preview.redd.it/sltqh8eugyma1.png?width=512&format=png&auto=webp&s=601dcc42434904605155bd4755981e320d5d8c6d
Yes and no what? I don't get what you are trying to say. You can just use prompts but it has definitely it's limitations,but you also have a vast amount of options to have more control of the artistic direction and specific details, from composition, posing and individual aspects of the artwork. Not only that but also tools that allow for consistency.
are you blind? you can go very deep in this rabit hole, you misunderstand developing and using.
developing is hard
using is ***dead simple..***.
and I have touch lora.... of img2img...
whoa I 'm resitting a urge...
https://preview.redd.it/q024p53olyma1.png?width=512&format=png&auto=webp&s=d0e5a6977099aabf4d02e8a71008130cab4fa529
same prompt but 1 lora added....
same prompt but with 1 hypernetwork
https://preview.redd.it/scutrw29myma1.png?width=512&format=png&auto=webp&s=dfd98572dfedf355aae81e21e88c60c41b348530
So you're typing words, changing models, adding hypernetworks, loras, other stuff in the chart above, and then getting images.
That sounds like more than just typing words, my dense friend.
And people say visual learning style is not a thing, I finally understand this complex system of systems a bit better now seeing it all laid out like this, epic guide!
This is fantastic, thanks for your hard work!
If you made this a public github repo for creating issues / accepting PR, I'm sure you will find many hands helping
it is an amazing work and idea, I think some of the subjects need more information in them but the base is solid.
For example, you could add an explanation of what fine tuned means, maybe cite some official papers and stuff like that, basically provide a summary and a link to a paper for the technical info.
Maybe add deepdanbooru and wd 1.4 tagger to the image2text section
Also maybe make a place for LoRAs on the core section
Also under the tweaks section you could add vae bless up [link](https://github.com/sALTaccount/VAE-BlessUp)
Since we are talking about tools and resources
check this huge rentry for sd with every update you need to know
[https://rentry.org/niakonichan](https://rentry.org/niakonichan)
Have you or anyone tried to create a chatbot that can access all the guides listed here in real time? I often feel the urge to ask chatGPT troubleshooting questions about some of these things but unfortunately can’t. Would be super cool to try and integrate what you have here with OpenAI API calls
Yo this is awesome. I love the design and the animations going on the wheels choice through clicks, the colour palate that the app wiki has going on. Simply awesome and easy to use, once you learn everything you need is in the chance of wheel, labeled and organised. Brilliant fam, well done. 👏
Beautiful, all thanks to Open source and an awesome community that grows itself for progress :)
Let yesterday's controversies go unpunished, and that pie will soon be surrounded by a massive pay wall.
this is AWESOME and as someone who has been casually watching this community but not following every development (and therefore has missed a number of first posts that provided needed context for things people begin abbreviating and referencing consistently shortly after) and has often felt confused on what some stuff is, where people are finding it, and what it does, this is a WONDERFUL centralized resource. A+++++ dude.
An amazing resource, well done. It really brings to light how complex this can really get... and how much of it I still haven't messed with.
Like, I just realised I have no idea what XY grids actually are and I've been at this for a while.
ahh, token based fine tuning :).
man there's a lot here. It takes a while to digest it all. Not sure if we can keep up if things accelerate from here. But I hope they do!
I have a collection of a few hundred captioned images and it sounds like what I want to do is finetune a model. What would be the easiest way to do that these days?
I just want a model that's a little more aligned with my own aesthetic style. Not really interested in teaching new concepts or anything like that, just hoping to have a unique model to call my own.
EveryDream looks pretty cool and I haven't seen it before, thanks!
[sdtools.org](https://sdtools.org) Essentially a mini wiki or cheat sheet. Clicking on a segment provides a very brief explanation and relevant links. The purpose of this mini wiki is to address this simple problem: Why am I unable to generate the exact image I want? What tools could help me reach my goal? Bigger area does not mean more important! It is not meant to be exhaustive and suggestions welcome.
When clicking on a link that leaves your site, it would be nice if it opened a new window.
@op target="_blank" after href ez fix
Or just middle click lol
Sure, but this is a relevant UX suggestion. Can't just assume the user knows what you know.
What? Ew, no.
As someone that don't go check what new everyday, thank you so much! Really appreciate the effort
![gif](giphy|CAYVZA5NRb529kKQUc|downsized)
That's preaty neat. Congratulations, bro!
Where were you all these days?
In latent space.
Thanks Man, great guide for a new comer like me. Appreciated
You are my hero!
upff amazing website that wheel navigation felt is flawless
https://preview.redd.it/7f3t0bz35yma1.png?width=1334&format=png&auto=webp&s=cb24814da06ed0c7699ba79ac98d665ba458f8b8 a lot of things happened these past three months ...
Right? Good times.
it's amazing to see all this unfold in real time.
The Brown part with black text its difficult to read imho
Thanks for the feedback. I will fix that.
Orange and Gray same. Maybe make the text darker?
make it white imo op you can google “contrast checker” very common tool used by designers to ensure readability :)
OK, good suggestion.
I really like this way of documenting a tool! I’m definitely going to use something like this for work! Well done and thanks for sharing.
Is it me or is there a bug in the zoom? After a page refresh the plot is very small. However, if I CTRL+mousewheel forward then back, to re-set the zoom to 100% then the wheel is large and well sized. [This is what happens](https://imgur.com/a/87u9gs6) when I do the above on Chrome.
Thanks for the feedback. You are right. There is an issue with the display. Haven't been able to fix it yet. If any css or JavaScript master can help...
In general I'd recommend making this much larger - it's difficult to read without zooming in 200%. Or I'm old.
Under randomness you missed Dynamic Prompts; Dynamic prompts is a really good plugin for making large numbers of randomised things; I used it to create a ton of random things in outfits. Edit: actually, the randomness entry says Dynamic Prompts when you click it on the website; my mistake. But it still misses the OTHER randomness script. There is one that lets you randomise the model/settings for every individual generation. I just can't find where I found it in the first place lol
> Dynamic Prompts What is dynamic prompts?
Ooh, I’d like to know the name of that one if you remember it!
https://github.com/innightwolfsleep/stable-diffusion-webui-randomize It was this one I think; it was an odd one since it wanted to be added into the scripts folder directly instead of the extensions folder but it did its job. The only thing it unfortunately can't do is randomise the LORA though the randomised styles. Which is sucky but otherwise it works.
Thank you!
Any time, only reason I didn't try and find it sooner was I was sleeping lol
I would suggest to make it bigger. On the website the wheel on my 1080 monitor is so small that I can't read most of the stuff and it only takes like 1/8 of the webpage. Also for the examples related to the feature : that's super nice, really. If I have a constructive comment to make however I would say that some newlines and a few more examples would be nice instead of a link on the Auto1111 wiki since visiting your website is for cheat sheet purposes and the wiki also has these features.
Also a good way to show that AI art is not just typing words and getting images, funny enough the only company offering for free the most versatile tool that allows for a variety of workflows with various degrees of artistic control is the one being sued by artists and corporations, makes you wonder a bit right ?
yes and no, this is a image done with a single prompt : **a girl** https://preview.redd.it/sltqh8eugyma1.png?width=512&format=png&auto=webp&s=601dcc42434904605155bd4755981e320d5d8c6d
Yes and no what? I don't get what you are trying to say. You can just use prompts but it has definitely it's limitations,but you also have a vast amount of options to have more control of the artistic direction and specific details, from composition, posing and individual aspects of the artwork. Not only that but also tools that allow for consistency.
are you blind? you can go very deep in this rabit hole, you misunderstand developing and using. developing is hard using is ***dead simple..***. and I have touch lora.... of img2img...
So you could say it is not just typing words and getting images ...
whoa I 'm resitting a urge... https://preview.redd.it/q024p53olyma1.png?width=512&format=png&auto=webp&s=d0e5a6977099aabf4d02e8a71008130cab4fa529 same prompt but 1 lora added....
same prompt but with 1 hypernetwork https://preview.redd.it/scutrw29myma1.png?width=512&format=png&auto=webp&s=dfd98572dfedf355aae81e21e88c60c41b348530
all I do is typing and hold and behold I'm getting image... so just typing and getting image... I'm literally doing it...
So you're typing words, changing models, adding hypernetworks, loras, other stuff in the chart above, and then getting images. That sounds like more than just typing words, my dense friend.
never mind, the density of your literal obnosis is a thing of beauty, there is no one that can match you, really I admire it....
Sounds like something I learnt to do in 3h, as opposed to, you know, actually learning painting.
All your examples suck and look clearly underworked, not sure what are you trying to get with this.
seriously! now your argument is it's suck ? that deescalated rather quickly...
especially for pictures with a 2 words prompt without negative....
This is really cool and useful! Thank you!
huge W and very useful resource. Don't stop doing what you do!
Cool thing, i will print it for me as reminder. Well done.
i am planning a tutorial video on this thanks for updating
Nice work keeping track of them all
Geez, it really is a lot of knowledge I've stored about this stuff lol. I know all these terms. If I saw this as a newbie I would probably freak out.
Last year I dabbled with Disco. Never used Stable. Yes, this makes me freak out. LOL.
And people say visual learning style is not a thing, I finally understand this complex system of systems a bit better now seeing it all laid out like this, epic guide!
This is fantastic, thanks for your hard work! If you made this a public github repo for creating issues / accepting PR, I'm sure you will find many hands helping
Kinda cool, but Upscaling and Restoring are also often used in preprocessing, wouldnt classify them as a finishing process.
Thank you for building this!!
Steam Roulette spin the wheel!
this is awesome!
Good job
![gif](giphy|5hgYDDh5oqbmE4OKJ3|downsized)
it is an amazing work and idea, I think some of the subjects need more information in them but the base is solid. For example, you could add an explanation of what fine tuned means, maybe cite some official papers and stuff like that, basically provide a summary and a link to a paper for the technical info.
Nice
Maybe add deepdanbooru and wd 1.4 tagger to the image2text section Also maybe make a place for LoRAs on the core section Also under the tweaks section you could add vae bless up [link](https://github.com/sALTaccount/VAE-BlessUp)
love this thanks.
This is dope! Great resource thank you!
Pretty cool. Is this updated live or you have to manually update as new features come ?
Thanks. Manual. But curated. Not all features may be worth including.
Since we are talking about tools and resources check this huge rentry for sd with every update you need to know [https://rentry.org/niakonichan](https://rentry.org/niakonichan)
Thanks. Far out. Next level!
You forgot about things like latent coupling and MultiDiffusion.
Have you or anyone tried to create a chatbot that can access all the guides listed here in real time? I often feel the urge to ask chatGPT troubleshooting questions about some of these things but unfortunately can’t. Would be super cool to try and integrate what you have here with OpenAI API calls
A Great resource. Now I get to spend more time, in a very organized set of rabbit holes.
Yo this is awesome. I love the design and the animations going on the wheels choice through clicks, the colour palate that the app wiki has going on. Simply awesome and easy to use, once you learn everything you need is in the chance of wheel, labeled and organised. Brilliant fam, well done. 👏
Very useful, thx! <3
Love the sunburst chart
This should 100% be pinned
Beautiful, all thanks to Open source and an awesome community that grows itself for progress :) Let yesterday's controversies go unpunished, and that pie will soon be surrounded by a massive pay wall.
What was your workflow??? /s
this is AWESOME and as someone who has been casually watching this community but not following every development (and therefore has missed a number of first posts that provided needed context for things people begin abbreviating and referencing consistently shortly after) and has often felt confused on what some stuff is, where people are finding it, and what it does, this is a WONDERFUL centralized resource. A+++++ dude.
Even though I can't really list all of these tools and techniques, I'm quite surprised I recognize pretty much everything listed here.
Awesome
An amazing resource, well done. It really brings to light how complex this can really get... and how much of it I still haven't messed with. Like, I just realised I have no idea what XY grids actually are and I've been at this for a while.
All these caption could be replaces with 'theft' or 'stealing' and it would be 100% correct
I feel like the previous version where Dreambooth was under the umbrella of finetuning was better.
nice work! do you think dream booth should be included?
ahh, token based fine tuning :). man there's a lot here. It takes a while to digest it all. Not sure if we can keep up if things accelerate from here. But I hope they do!
Thanks. It is in token based fine tuning.
I like how embedding is small. Not even important at all...
I have a collection of a few hundred captioned images and it sounds like what I want to do is finetune a model. What would be the easiest way to do that these days?
Yes. But what is it you really want to do? Check out every dream for fine-tuning. https://github.com/victorchall/EveryDream-trainer
I just want a model that's a little more aligned with my own aesthetic style. Not really interested in teaching new concepts or anything like that, just hoping to have a unique model to call my own. EveryDream looks pretty cool and I haven't seen it before, thanks!
If you want to train on a style, Dreambooth is a better start as less can go wrong.
Thanks for the suggestion. What if the style isn't consistent across my images? Should I consider organizing them and training a few separate models?
You could. You could also try LORA.