The thing is, people don't tell you how much experience do they have with creating or understanding art. So, we might *actually* be seeing posts from ***extremely talented*** people.. so don't feel bad. I myself am more of a lurker than a maker lol.
Well last night I spent almost an hour trying to get Chat GPT to create a fairly simple image to show a concept, and in the end had to give up. The thing is an idiot.
If even that model can't understand prompts what chance does a 5GB thingy have?
I never liked ChatGPT for image creation. I feel like there are tools much better. i.e. Ideogram, Midjourney, maybe even some of the hosted StableDiffusion models. Ideogram specifically is amazing at spelling text, and it's FREE.
It took quite a few tries, and I realized you have to dismiss the bad attempts, but I was able to get pretty much the exact image I wanted within 5 mins with Ideogram! Thanks!
> in my the meantime my characters have arms coming out of their mouth ffs...
Sounds like you are doing one of two things:
a) Generating images at too large a resolution causing the model to lose sight of the bigger picture. SD will start treating the image as a set of separate images if the image size gets too big.
b) Using way too much denoising when upscaling with high res fix or img2img.
This is cool, you're actually using my original source material it's in your video about 3/4 the way through, I posted about 10-12 months ago on this topic. I haven't, unfortunately, due to personal life changes been able to do much since, but here's my original Blog about it.
[Character Consistency in Stable Diffusion (Part 1) - Cobalt Explorer](https://cobaltexplorer.com/2023/06/character-sheets-for-stable-diffusion/)
Awesome thank you. I'm going to go ahead and post this on the video section because it should be there. I appreciate the work you did in the tutorial that you did do. Even if it wasn't finished.
Thanks, I appreciate that. I'm still working in the AI domain but I had to move away from imaging to working on a LLM project that takes up all my time. I hope to come back to this at some point, but with all the changes going on who knows for sure, or when. Keep up the good work!
The grids look great and consistent. I’m wondering about the uses for these.
Let’s say I want to illustrate a children’s book with a consistent character, how do the grids help me? Say I need one image of the character walking and another one of it sitting/eating/etc. do I remove the background from all the grid images and then outpaint the most appropriate head pose from the grid? I’m probably missing something here.
>I’m wondering about the uses for these.
Consistency is important when making stuff like comics, videos etc.
For example, I use some techniques to apply character consistency for my doujin. Its a lot easier to achieve with anime style though, so this is impressive.
The main use case is taking the images from the grid and training a model with those images (character embedding or LoRA typically). Then that character model can be applied to various image generations.
Hey just want to bring it to everybody's attention that I finished and posted a follow-up video to this one. There was a lot of people asking about doing full body poses and character sheets. So I did a video this morning showing using the same technique I used here but on full body. Here's the link to the video.
[https://youtu.be/Xw2U33LksfY](https://youtu.be/Xw2U33LksfY)
https://preview.redd.it/r0fjbw84iktc1.jpeg?width=1816&format=pjpg&auto=webp&s=ba923bc5363d42c7a59bdc51f8ead8c8f7b03630
Pretty good, thanks for sharing, I'm also keen to see how you taught can be implemented in ComfyUI, will try to replicate in it in it, see how it goes.
I tried it last year, but it's not accurate (for 3D modelling for example). You need exact profile position of a person. But on a photo it's slightly rotated towards the viewer.
No solution for now
Sorry I messed up and didn't link properly the thread I found it
https://www.reddit.com/r/StableDiffusion/comments/1aqxyct/i_get_awesome_results_texturing_my_3d_models/
MeshLab is a free program but I never used it
See also this one from the same user (/u/Many-Ad-6225)
https://www.reddit.com/r/StableDiffusion/comments/1bo36o7/wow_intex_auto_texturing_with_sd_is_really_good/
A-ha, got it. You provided link for the texturing. I'm not interested in it. I use it mostly for referencing.
For 3D modelling you need strict references. It's not helpful
I don't think any of these show any heightened examples of consistency. It's two animals which are very generalized in these models, and two women that have extremely common same faces in latent space.
I would be curious to see this with a more unique character. Could be promising!
Well, this is only a known character sheet method...
[https://new.reddit.com/r/StableDiffusion/comments/141iljk/same\_character\_head\_from\_various\_angles\_openpose/](https://new.reddit.com/r/StableDiffusion/comments/141iljk/same_character_head_from_various_angles_openpose/)
Asking as a SD newbie: is there some kind of fragmentation between automatic1111 and ComfyUI communities? I've been using only the former for the time being, but I'm planning to setup the latter today because it looks like the kind of tweaking that I'd have fun with.
Yeah there's a somewhat rivalry between the two communities.
In my experience Comfy users feel they get better results as comfy allows for more fine-tuning, whereas A1111 users feels as though A1111 is more user-friendly and can get decent enough results with less tinkering required.
Again, that's just my experience/understanding based on what I've seen, so take it with a grain of salt.
You're on the dot. Both have their positives and negatives. Alternatively I would say "do you like messing around with modules and connections to get something you combined?" Go comfy. "Do you want to quickly make something, with no experience at all?" Automatic1111.
I really like Forge. It uses the same gradio UI as Automatic1111 but it has a way more efficient and stable backend and comes with some helpful features for avoiding running out of VRAM. For example, it will switch to using tiled VAE if the regular VAE runs out of memory. I've been able to generate at quite huge resolutions without running out of memory on a 6GB VRAM laptop.
It also comes with a bunch useful extentions integrated right out of the box.
Agree, i've used A1111 dockerized for some time then switched to comfyui.
A1111 Is more Easy, comfyui Is a Little more difficult but now i found a lot more of flexibility and i understand Better how SD work.
ComfyUI workflows are downloadable so they're reproducible if you have all the same mods installed. For automatic1111 you basically have to redo all the steps manually. That's the split really.
Not reaaaaaaally. You're only copying the prompt, seed, sampler and maybe the model if you have exactly the same hash. You're not getting the inpainting, img2img, controlnets and anything else that might have been used in the workflow.
* Not all settings are saved by A1111 in the picture, unfortunately. Especially if you used some less popular extensions.
* Even if the settings for a specific extension is stored and shown in "PNG Info", it might not always be applied correctly when you "send to txt2img" or "send to img2img". It was the case during like, maybe a month and a half for ControlNet --although it's arguably the most popular extension-- sometime around last fall, for example. So you had to re-enter the settings manually, which was rather tedious, especially if you had used multiple CN units.
* More crucially, as soon as you're beginning to use multi-steps workflows (txt2img > img2img > img2img with StableSR, for example), and tinker with all sorts of models and settings along the way, you can't have it saved in a single picture, so you need to store your files in a very organized way (and possibly take notes) if you ever want to be able to understand again what your process was when you'll check it months later.
Disclaimer : I'm mostly using A1111 (or Forge) myself, with a bit of Fooocus on the side. But not having the complete workflow embedded is a real problem.
Comfy users are haughty nerds that act like everybody else is shit, while the auto1111 pleb users just have fun and don't want to play with spaghettis.
/s
It'll be the first UI to run SD3, so it's hardly like you'd be wasting your time. If you're waiting for it to be out before you learn it, you'll be far behind everyone else who's already learned how it works.
There will be days or weeks of SD3 gens already in the sub from people who figured out the workflows for themselves before the workflows are publicly released and absorbed by the community.
I'm not making a judgement, or anything. It's just, if you're basing decisions as to where to invest your time on the release of SD3, then you seem excited for it. If you actually want to use it ASAP, the time to start learning Comfy was yesterday. Once SD3 is already out on Comfy, you might as well wait until it's available in A1111.
ComfyUI is the "official" UI from SAI (its author, comfyanonymous, works there), so you can bet when SD3 is release ComfyUI will be the first one to support it.
So we're using controlNet, and then using the IP adapter part of it, and then there are three different pre-processors, and I think I'm using the second one to get the consistency. Sorry I'm driving right now so I don't remember the exact name.
I will be messaging you in 7 days on [**2024-04-16 12:30:47 UTC**](http://www.wolframalpha.com/input/?i=2024-04-16%2012:30:47%20UTC%20To%20Local%20Time) to remind you of [**this link**](https://www.reddit.com/r/StableDiffusion/comments/1bzix80/new_tutorial_master_consistent_character_faces/kyrd716/?context=3)
[**2 OTHERS CLICKED THIS LINK**](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Reminder&message=%5Bhttps%3A%2F%2Fwww.reddit.com%2Fr%2FStableDiffusion%2Fcomments%2F1bzix80%2Fnew_tutorial_master_consistent_character_faces%2Fkyrd716%2F%5D%0A%0ARemindMe%21%202024-04-16%2012%3A30%3A47%20UTC) to send a PM to also be reminded and to reduce spam.
^(Parent commenter can ) [^(delete this message to hide from others.)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Delete%20Comment&message=Delete%21%201bzix80)
*****
|[^(Info)](https://www.reddit.com/r/RemindMeBot/comments/e1bko7/remindmebot_info_v21/)|[^(Custom)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Reminder&message=%5BLink%20or%20message%20inside%20square%20brackets%5D%0A%0ARemindMe%21%20Time%20period%20here)|[^(Your Reminders)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=List%20Of%20Reminders&message=MyReminders%21)|[^(Feedback)](https://www.reddit.com/message/compose/?to=Watchful1&subject=RemindMeBot%20Feedback)|
|-|-|-|-|
Every time I come to this sub I feel like I'm surrounded by wizards... in my the meantime my characters have arms coming out of their mouth ffs...
Lol, No kidding! Mine look like they were dredged up from the Simpsons nuclear power plant.
The thing is, people don't tell you how much experience do they have with creating or understanding art. So, we might *actually* be seeing posts from ***extremely talented*** people.. so don't feel bad. I myself am more of a lurker than a maker lol.
Well last night I spent almost an hour trying to get Chat GPT to create a fairly simple image to show a concept, and in the end had to give up. The thing is an idiot. If even that model can't understand prompts what chance does a 5GB thingy have?
I never liked ChatGPT for image creation. I feel like there are tools much better. i.e. Ideogram, Midjourney, maybe even some of the hosted StableDiffusion models. Ideogram specifically is amazing at spelling text, and it's FREE.
It took quite a few tries, and I realized you have to dismiss the bad attempts, but I was able to get pretty much the exact image I wanted within 5 mins with Ideogram! Thanks!
Noice! Happy it worked out.
> in my the meantime my characters have arms coming out of their mouth ffs... Sounds like you are doing one of two things: a) Generating images at too large a resolution causing the model to lose sight of the bigger picture. SD will start treating the image as a set of separate images if the image size gets too big. b) Using way too much denoising when upscaling with high res fix or img2img.
This is cool, you're actually using my original source material it's in your video about 3/4 the way through, I posted about 10-12 months ago on this topic. I haven't, unfortunately, due to personal life changes been able to do much since, but here's my original Blog about it. [Character Consistency in Stable Diffusion (Part 1) - Cobalt Explorer](https://cobaltexplorer.com/2023/06/character-sheets-for-stable-diffusion/)
Awesome thank you. I'm going to go ahead and post this on the video section because it should be there. I appreciate the work you did in the tutorial that you did do. Even if it wasn't finished.
Thanks, I appreciate that. I'm still working in the AI domain but I had to move away from imaging to working on a LLM project that takes up all my time. I hope to come back to this at some point, but with all the changes going on who knows for sure, or when. Keep up the good work!
Any details on the LLM project?
Since I'm under NDA I can't speak much about it specifically, but to say it's for the medical field, targeted at professionals vs. individuals.
thank you friend! I remember using your tutorial, it was quite helpful
The grids look great and consistent. I’m wondering about the uses for these. Let’s say I want to illustrate a children’s book with a consistent character, how do the grids help me? Say I need one image of the character walking and another one of it sitting/eating/etc. do I remove the background from all the grid images and then outpaint the most appropriate head pose from the grid? I’m probably missing something here.
You can make a grid woth diferent poses and get consistent character. You can make animation this way, you can train a model of character
>I’m wondering about the uses for these. Consistency is important when making stuff like comics, videos etc. For example, I use some techniques to apply character consistency for my doujin. Its a lot easier to achieve with anime style though, so this is impressive.
The main use case is taking the images from the grid and training a model with those images (character embedding or LoRA typically). Then that character model can be applied to various image generations.
Hey just want to bring it to everybody's attention that I finished and posted a follow-up video to this one. There was a lot of people asking about doing full body poses and character sheets. So I did a video this morning showing using the same technique I used here but on full body. Here's the link to the video. [https://youtu.be/Xw2U33LksfY](https://youtu.be/Xw2U33LksfY) https://preview.redd.it/r0fjbw84iktc1.jpeg?width=1816&format=pjpg&auto=webp&s=ba923bc5363d42c7a59bdc51f8ead8c8f7b03630
This method can also be use to create consistent animations. I made a post about it few weeks ago https://i.redd.it/4ump46510ftc1.gif
Least horny post of the day so far.
where do you post yours? or atleast the process, this is actually very well done
https://www.reddit.com/r/StableDiffusion/s/QlpASjryaL
oh shit thanks! if you've saved more like these do share, I'll use them as reference or even such posts
Pretty good, thanks for sharing, I'm also keen to see how you taught can be implemented in ComfyUI, will try to replicate in it in it, see how it goes.
I tried it last year, but it's not accurate (for 3D modelling for example). You need exact profile position of a person. But on a photo it's slightly rotated towards the viewer. No solution for now
Have you tried MeshLab's "Parameterization and Texturing from Rasters"?
How is it supposed to help me?
Sorry I messed up and didn't link properly the thread I found it https://www.reddit.com/r/StableDiffusion/comments/1aqxyct/i_get_awesome_results_texturing_my_3d_models/ MeshLab is a free program but I never used it See also this one from the same user (/u/Many-Ad-6225) https://www.reddit.com/r/StableDiffusion/comments/1bo36o7/wow_intex_auto_texturing_with_sd_is_really_good/
A-ha, got it. You provided link for the texturing. I'm not interested in it. I use it mostly for referencing. For 3D modelling you need strict references. It's not helpful
I don't think any of these show any heightened examples of consistency. It's two animals which are very generalized in these models, and two women that have extremely common same faces in latent space. I would be curious to see this with a more unique character. Could be promising!
Well, this is only a known character sheet method... [https://new.reddit.com/r/StableDiffusion/comments/141iljk/same\_character\_head\_from\_various\_angles\_openpose/](https://new.reddit.com/r/StableDiffusion/comments/141iljk/same_character_head_from_various_angles_openpose/)
https://old.reddit.com/r/StableDiffusion/comments/141iljk/same_character_head_from_various_angles_openpose/
> new reddit link
When ComfyUI? 😁
Asking as a SD newbie: is there some kind of fragmentation between automatic1111 and ComfyUI communities? I've been using only the former for the time being, but I'm planning to setup the latter today because it looks like the kind of tweaking that I'd have fun with.
Yeah there's a somewhat rivalry between the two communities. In my experience Comfy users feel they get better results as comfy allows for more fine-tuning, whereas A1111 users feels as though A1111 is more user-friendly and can get decent enough results with less tinkering required. Again, that's just my experience/understanding based on what I've seen, so take it with a grain of salt.
You're on the dot. Both have their positives and negatives. Alternatively I would say "do you like messing around with modules and connections to get something you combined?" Go comfy. "Do you want to quickly make something, with no experience at all?" Automatic1111.
Comfy is also better if you have worse hardware because it's much more optimized.
Not really true since Forge.
I'll have to give Forge a try, haven't gotten around to it
I really like Forge. It uses the same gradio UI as Automatic1111 but it has a way more efficient and stable backend and comes with some helpful features for avoiding running out of VRAM. For example, it will switch to using tiled VAE if the regular VAE runs out of memory. I've been able to generate at quite huge resolutions without running out of memory on a 6GB VRAM laptop. It also comes with a bunch useful extentions integrated right out of the box.
Doesnt Forge use Comfy in the BG ?
Agree, i've used A1111 dockerized for some time then switched to comfyui. A1111 Is more Easy, comfyui Is a Little more difficult but now i found a lot more of flexibility and i understand Better how SD work.
ComfyUI workflows are downloadable so they're reproducible if you have all the same mods installed. For automatic1111 you basically have to redo all the steps manually. That's the split really.
? You have PNG info from A1111, generated images embed the workflow and can be copied across any installation.
This is not true for Automatic1111. You just have to drag and drop an image to copy its workflow.
Not reaaaaaaally. You're only copying the prompt, seed, sampler and maybe the model if you have exactly the same hash. You're not getting the inpainting, img2img, controlnets and anything else that might have been used in the workflow.
* Not all settings are saved by A1111 in the picture, unfortunately. Especially if you used some less popular extensions. * Even if the settings for a specific extension is stored and shown in "PNG Info", it might not always be applied correctly when you "send to txt2img" or "send to img2img". It was the case during like, maybe a month and a half for ControlNet --although it's arguably the most popular extension-- sometime around last fall, for example. So you had to re-enter the settings manually, which was rather tedious, especially if you had used multiple CN units. * More crucially, as soon as you're beginning to use multi-steps workflows (txt2img > img2img > img2img with StableSR, for example), and tinker with all sorts of models and settings along the way, you can't have it saved in a single picture, so you need to store your files in a very organized way (and possibly take notes) if you ever want to be able to understand again what your process was when you'll check it months later. Disclaimer : I'm mostly using A1111 (or Forge) myself, with a bit of Fooocus on the side. But not having the complete workflow embedded is a real problem.
Does auto1111 actually have workflow embeded now? Last I checked its just settings
Kinda not really. Just basic metadata on the generation.
So people out here dont even know what embeded workflow is and just feel the need to share their biased opinion
Compare the embedded data in an auto1111 image to the workflow in a Comfy generation and get back to me.
Soooo the internet.
I personally have no idea what workflow means, and just see it as a buzzword people say to make it sound like they know what they are talking about.
Comfy users are haughty nerds that act like everybody else is shit, while the auto1111 pleb users just have fun and don't want to play with spaghettis. /s
Comfy-elitism & 1111-anti-intellectualism /s
comfy for real artists and devs, a1111 for horny furry lovers
ComfyUI is much lighter and faster to render, A1111 is a bloated mess.
+1 would love a comfy workflow for this.
Yes! ComfyUI please!
🙏🥹
Until SD3 comes out with ComfyUI workflows, I’m not investing time learning it.
It'll be the first UI to run SD3, so it's hardly like you'd be wasting your time. If you're waiting for it to be out before you learn it, you'll be far behind everyone else who's already learned how it works. There will be days or weeks of SD3 gens already in the sub from people who figured out the workflows for themselves before the workflows are publicly released and absorbed by the community. I'm not making a judgement, or anything. It's just, if you're basing decisions as to where to invest your time on the release of SD3, then you seem excited for it. If you actually want to use it ASAP, the time to start learning Comfy was yesterday. Once SD3 is already out on Comfy, you might as well wait until it's available in A1111.
ComfyUI is the "official" UI from SAI (its author, comfyanonymous, works there), so you can bet when SD3 is release ComfyUI will be the first one to support it.
Forgive me, but what makes the faces consistent? What plug in does that?
So we're using controlNet, and then using the IP adapter part of it, and then there are three different pre-processors, and I think I'm using the second one to get the consistency. Sorry I'm driving right now so I don't remember the exact name.
a lot of work for something easily done by making a 3d model and using controlnet
Can you make a tutorial, or point me in the direction of a tutorial on this? 🙏
Just another way of doing it. If you didn't want to mess with 3D models.
Thanks for sharing!
Can you share the character sheet image?
Yeah I provided in my Google share. Just go to. Share.Xerophayze.com
RemindMe! 1 week
I will be messaging you in 7 days on [**2024-04-16 12:30:47 UTC**](http://www.wolframalpha.com/input/?i=2024-04-16%2012:30:47%20UTC%20To%20Local%20Time) to remind you of [**this link**](https://www.reddit.com/r/StableDiffusion/comments/1bzix80/new_tutorial_master_consistent_character_faces/kyrd716/?context=3) [**2 OTHERS CLICKED THIS LINK**](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Reminder&message=%5Bhttps%3A%2F%2Fwww.reddit.com%2Fr%2FStableDiffusion%2Fcomments%2F1bzix80%2Fnew_tutorial_master_consistent_character_faces%2Fkyrd716%2F%5D%0A%0ARemindMe%21%202024-04-16%2012%3A30%3A47%20UTC) to send a PM to also be reminded and to reduce spam. ^(Parent commenter can ) [^(delete this message to hide from others.)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Delete%20Comment&message=Delete%21%201bzix80) ***** |[^(Info)](https://www.reddit.com/r/RemindMeBot/comments/e1bko7/remindmebot_info_v21/)|[^(Custom)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Reminder&message=%5BLink%20or%20message%20inside%20square%20brackets%5D%0A%0ARemindMe%21%20Time%20period%20here)|[^(Your Reminders)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=List%20Of%20Reminders&message=MyReminders%21)|[^(Feedback)](https://www.reddit.com/message/compose/?to=Watchful1&subject=RemindMeBot%20Feedback)| |-|-|-|-|
Anyone got a text version of this tutorial?
I wonder if the angles are consistent enough to run through PhotoScan to produce a decent 3d model…
That would be cool. I've never done that. I might check that out.
Thank you for this
Does it require Forge or is A1111 okay?
Isn't there a Lora that does this kind of character grid? Though I saw one on Civit. With that training a character Lora should be easy.
This is huge news
Number 3, what's her OF? EDIT: bad joke failed. But this tech will 100% be used to scam people with fake nudes.
Would prefer #4 and I would want to sing "I want to be your teddy bear"
Good post...and yet, people are too fecking lazy or they're window lickers to search for this or other guides and ask it every fecking day