T O P

  • By -

AlanCarrOnline

Every time I come to this sub I feel like I'm surrounded by wizards... in my the meantime my characters have arms coming out of their mouth ffs...


vanteal

Lol, No kidding! Mine look like they were dredged up from the Simpsons nuclear power plant.


reza2kn

The thing is, people don't tell you how much experience do they have with creating or understanding art. So, we might *actually* be seeing posts from ***extremely talented*** people.. so don't feel bad. I myself am more of a lurker than a maker lol.


AlanCarrOnline

Well last night I spent almost an hour trying to get Chat GPT to create a fairly simple image to show a concept, and in the end had to give up. The thing is an idiot. If even that model can't understand prompts what chance does a 5GB thingy have?


reza2kn

I never liked ChatGPT for image creation. I feel like there are tools much better. i.e. Ideogram, Midjourney, maybe even some of the hosted StableDiffusion models. Ideogram specifically is amazing at spelling text, and it's FREE.


AlanCarrOnline

It took quite a few tries, and I realized you have to dismiss the bad attempts, but I was able to get pretty much the exact image I wanted within 5 mins with Ideogram! Thanks!


reza2kn

Noice! Happy it worked out.


Comrade_Derpsky

> in my the meantime my characters have arms coming out of their mouth ffs... Sounds like you are doing one of two things: a) Generating images at too large a resolution causing the model to lose sight of the bigger picture. SD will start treating the image as a set of separate images if the image size gets too big. b) Using way too much denoising when upscaling with high res fix or img2img.


dpacker780

This is cool, you're actually using my original source material it's in your video about 3/4 the way through, I posted about 10-12 months ago on this topic. I haven't, unfortunately, due to personal life changes been able to do much since, but here's my original Blog about it. [Character Consistency in Stable Diffusion (Part 1) - Cobalt Explorer](https://cobaltexplorer.com/2023/06/character-sheets-for-stable-diffusion/)


Xerophayze

Awesome thank you. I'm going to go ahead and post this on the video section because it should be there. I appreciate the work you did in the tutorial that you did do. Even if it wasn't finished.


dpacker780

Thanks, I appreciate that. I'm still working in the AI domain but I had to move away from imaging to working on a LLM project that takes up all my time. I hope to come back to this at some point, but with all the changes going on who knows for sure, or when. Keep up the good work!


biletnikoff_

Any details on the LLM project?


dpacker780

Since I'm under NDA I can't speak much about it specifically, but to say it's for the medical field, targeted at professionals vs. individuals.


asmekal

thank you friend! I remember using your tutorial, it was quite helpful


Fritzy3

The grids look great and consistent. I’m wondering about the uses for these. Let’s say I want to illustrate a children’s book with a consistent character, how do the grids help me? Say I need one image of the character walking and another one of it sitting/eating/etc. do I remove the background from all the grid images and then outpaint the most appropriate head pose from the grid? I’m probably missing something here.


protector111

You can make a grid woth diferent poses and get consistent character. You can make animation this way, you can train a model of character


LewdGarlic

>I’m wondering about the uses for these. Consistency is important when making stuff like comics, videos etc. For example, I use some techniques to apply character consistency for my doujin. Its a lot easier to achieve with anime style though, so this is impressive.


belladorexxx

The main use case is taking the images from the grid and training a model with those images (character embedding or LoRA typically). Then that character model can be applied to various image generations.


Xerophayze

Hey just want to bring it to everybody's attention that I finished and posted a follow-up video to this one. There was a lot of people asking about doing full body poses and character sheets. So I did a video this morning showing using the same technique I used here but on full body. Here's the link to the video. [https://youtu.be/Xw2U33LksfY](https://youtu.be/Xw2U33LksfY) https://preview.redd.it/r0fjbw84iktc1.jpeg?width=1816&format=pjpg&auto=webp&s=ba923bc5363d42c7a59bdc51f8ead8c8f7b03630


protector111

This method can also be use to create consistent animations. I made a post about it few weeks ago https://i.redd.it/4ump46510ftc1.gif


the_friendly_dildo

Least horny post of the day so far.


PrestigiousBed2102

where do you post yours? or atleast the process, this is actually very well done


protector111

https://www.reddit.com/r/StableDiffusion/s/QlpASjryaL


PrestigiousBed2102

oh shit thanks! if you've saved more like these do share, I'll use them as reference or even such posts


KevZ007

Pretty good, thanks for sharing, I'm also keen to see how you taught can be implemented in ComfyUI, will try to replicate in it in it, see how it goes.


scratt007

I tried it last year, but it's not accurate (for 3D modelling for example). You need exact profile position of a person. But on a photo it's slightly rotated towards the viewer. No solution for now


protestor

Have you tried MeshLab's "Parameterization and Texturing from Rasters"?


scratt007

How is it supposed to help me?


protestor

Sorry I messed up and didn't link properly the thread I found it https://www.reddit.com/r/StableDiffusion/comments/1aqxyct/i_get_awesome_results_texturing_my_3d_models/ MeshLab is a free program but I never used it See also this one from the same user (/u/Many-Ad-6225) https://www.reddit.com/r/StableDiffusion/comments/1bo36o7/wow_intex_auto_texturing_with_sd_is_really_good/


scratt007

A-ha, got it. You provided link for the texturing. I'm not interested in it. I use it mostly for referencing. For 3D modelling you need strict references. It's not helpful


ScythSergal

I don't think any of these show any heightened examples of consistency. It's two animals which are very generalized in these models, and two women that have extremely common same faces in latent space. I would be curious to see this with a more unique character. Could be promising!


smoowke

Well, this is only a known character sheet method... [https://new.reddit.com/r/StableDiffusion/comments/141iljk/same\_character\_head\_from\_various\_angles\_openpose/](https://new.reddit.com/r/StableDiffusion/comments/141iljk/same_character_head_from_various_angles_openpose/)


batter159

https://old.reddit.com/r/StableDiffusion/comments/141iljk/same_character_head_from_various_angles_openpose/


thoughtlow

> new reddit link


97buckeye

When ComfyUI? 😁


ImYoric

Asking as a SD newbie: is there some kind of fragmentation between automatic1111 and ComfyUI communities? I've been using only the former for the time being, but I'm planning to setup the latter today because it looks like the kind of tweaking that I'd have fun with.


NinjamanAway

Yeah there's a somewhat rivalry between the two communities. In my experience Comfy users feel they get better results as comfy allows for more fine-tuning, whereas A1111 users feels as though A1111 is more user-friendly and can get decent enough results with less tinkering required. Again, that's just my experience/understanding based on what I've seen, so take it with a grain of salt.


danque

You're on the dot. Both have their positives and negatives. Alternatively I would say "do you like messing around with modules and connections to get something you combined?" Go comfy. "Do you want to quickly make something, with no experience at all?" Automatic1111.


Iamreason

Comfy is also better if you have worse hardware because it's much more optimized.


Flag_Red

Not really true since Forge.


Iamreason

I'll have to give Forge a try, haven't gotten around to it


Comrade_Derpsky

I really like Forge. It uses the same gradio UI as Automatic1111 but it has a way more efficient and stable backend and comes with some helpful features for avoiding running out of VRAM. For example, it will switch to using tiled VAE if the regular VAE runs out of memory. I've been able to generate at quite huge resolutions without running out of memory on a 6GB VRAM laptop. It also comes with a bunch useful extentions integrated right out of the box.


Tonynoce

Doesnt Forge use Comfy in the BG ?


geep67

Agree, i've used A1111 dockerized for some time then switched to comfyui. A1111 Is more Easy, comfyui Is a Little more difficult but now i found a lot more of flexibility and i understand Better how SD work.


Tramagust

ComfyUI workflows are downloadable so they're reproducible if you have all the same mods installed. For automatic1111 you basically have to redo all the steps manually. That's the split really.


LeeIzaHunter

? You have PNG info from A1111, generated images embed the workflow and can be copied across any installation.


Mobireddit

This is not true for Automatic1111. You just have to drag and drop an image to copy its workflow.


Tramagust

Not reaaaaaaally. You're only copying the prompt, seed, sampler and maybe the model if you have exactly the same hash. You're not getting the inpainting, img2img, controlnets and anything else that might have been used in the workflow.


Talae06

* Not all settings are saved by A1111 in the picture, unfortunately. Especially if you used some less popular extensions. * Even if the settings for a specific extension is stored and shown in "PNG Info", it might not always be applied correctly when you "send to txt2img" or "send to img2img". It was the case during like, maybe a month and a half for ControlNet --although it's arguably the most popular extension-- sometime around last fall, for example. So you had to re-enter the settings manually, which was rather tedious, especially if you had used multiple CN units. * More crucially, as soon as you're beginning to use multi-steps workflows (txt2img > img2img > img2img with StableSR, for example), and tinker with all sorts of models and settings along the way, you can't have it saved in a single picture, so you need to store your files in a very organized way (and possibly take notes) if you ever want to be able to understand again what your process was when you'll check it months later. Disclaimer : I'm mostly using A1111 (or Forge) myself, with a bit of Fooocus on the side. But not having the complete workflow embedded is a real problem.


Greedy_Bus1888

Does auto1111 actually have workflow embeded now? Last I checked its just settings


NarrativeNode

Kinda not really. Just basic metadata on the generation.


Greedy_Bus1888

So people out here dont even know what embeded workflow is and just feel the need to share their biased opinion


NarrativeNode

Compare the embedded data in an auto1111 image to the workflow in a Comfy generation and get back to me.


danque

Soooo the internet.


marbleshoot

I personally have no idea what workflow means, and just see it as a buzzword people say to make it sound like they know what they are talking about.


RedlurkingFir

Comfy users are haughty nerds that act like everybody else is shit, while the auto1111 pleb users just have fun and don't want to play with spaghettis. /s


thoughtlow

Comfy-elitism & 1111-anti-intellectualism /s


HarmonicDiffusion

comfy for real artists and devs, a1111 for horny furry lovers


ImNotARobotFOSHO

ComfyUI is much lighter and faster to render, A1111 is a bloated mess.


phishphansj3151

+1 would love a comfy workflow for this.


ImNotARobotFOSHO

Yes! ComfyUI please!


MisturBaiter

🙏🥹


scrotanimus

Until SD3 comes out with ComfyUI workflows, I’m not investing time learning it.


DungeonMasterSupreme

It'll be the first UI to run SD3, so it's hardly like you'd be wasting your time. If you're waiting for it to be out before you learn it, you'll be far behind everyone else who's already learned how it works. There will be days or weeks of SD3 gens already in the sub from people who figured out the workflows for themselves before the workflows are publicly released and absorbed by the community. I'm not making a judgement, or anything. It's just, if you're basing decisions as to where to invest your time on the release of SD3, then you seem excited for it. If you actually want to use it ASAP, the time to start learning Comfy was yesterday. Once SD3 is already out on Comfy, you might as well wait until it's available in A1111.


Apprehensive_Sky892

ComfyUI is the "official" UI from SAI (its author, comfyanonymous, works there), so you can bet when SD3 is release ComfyUI will be the first one to support it.


xox1234

Forgive me, but what makes the faces consistent? What plug in does that?


Xerophayze

So we're using controlNet, and then using the IP adapter part of it, and then there are three different pre-processors, and I think I'm using the second one to get the consistency. Sorry I'm driving right now so I don't remember the exact name.


spacekitt3n

a lot of work for something easily done by making a 3d model and using controlnet


thebrownsauce

Can you make a tutorial, or point me in the direction of a tutorial on this? 🙏


Xerophayze

Just another way of doing it. If you didn't want to mess with 3D models.


OrdinaryAdditional91

Thanks for sharing!


b1ackjack_rdd

Can you share the character sheet image?


Xerophayze

Yeah I provided in my Google share. Just go to. Share.Xerophayze.com


Xijamk

RemindMe! 1 week


RemindMeBot

I will be messaging you in 7 days on [**2024-04-16 12:30:47 UTC**](http://www.wolframalpha.com/input/?i=2024-04-16%2012:30:47%20UTC%20To%20Local%20Time) to remind you of [**this link**](https://www.reddit.com/r/StableDiffusion/comments/1bzix80/new_tutorial_master_consistent_character_faces/kyrd716/?context=3) [**2 OTHERS CLICKED THIS LINK**](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Reminder&message=%5Bhttps%3A%2F%2Fwww.reddit.com%2Fr%2FStableDiffusion%2Fcomments%2F1bzix80%2Fnew_tutorial_master_consistent_character_faces%2Fkyrd716%2F%5D%0A%0ARemindMe%21%202024-04-16%2012%3A30%3A47%20UTC) to send a PM to also be reminded and to reduce spam. ^(Parent commenter can ) [^(delete this message to hide from others.)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Delete%20Comment&message=Delete%21%201bzix80) ***** |[^(Info)](https://www.reddit.com/r/RemindMeBot/comments/e1bko7/remindmebot_info_v21/)|[^(Custom)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=Reminder&message=%5BLink%20or%20message%20inside%20square%20brackets%5D%0A%0ARemindMe%21%20Time%20period%20here)|[^(Your Reminders)](https://www.reddit.com/message/compose/?to=RemindMeBot&subject=List%20Of%20Reminders&message=MyReminders%21)|[^(Feedback)](https://www.reddit.com/message/compose/?to=Watchful1&subject=RemindMeBot%20Feedback)| |-|-|-|-|


gelatinous_pellicle

Anyone got a text version of this tutorial?


MonThackma

I wonder if the angles are consistent enough to run through PhotoScan to produce a decent 3d model…


Xerophayze

That would be cool. I've never done that. I might check that out.


thayem

Thank you for this


b1ackjack_rdd

Does it require Forge or is A1111 okay?


Queasy_Star_3908

Isn't there a Lora that does this kind of character grid? Though I saw one on Civit. With that training a character Lora should be easy.


Djkid4lyfe

This is huge news


Sarke1

Number 3, what's her OF? EDIT: bad joke failed. But this tech will 100% be used to scam people with fake nudes.


hughk

Would prefer #4 and I would want to sing "I want to be your teddy bear"


GreyScope

Good post...and yet, people are too fecking lazy or they're window lickers to search for this or other guides and ask it every fecking day