Unreal Engine 5 Metahuman and Stable Diffusion
CoffeeVectors wanted to see what kind of workflow he could build bridging Epic’s MetaHuman with AI image synthesis platforms like Stable Diffusion. His experiments were really interesting and instructive to other beginners.
He started with a simple face generated by stable diffusion:
He then fed that image into Metahuman as a new starting point and cycled through a few generations. With each cycle, you can change the prompt and steer things in slightly different directions. It becomes less about a single initial prompt and more about understanding/modifying the larger system of settings interacting with each other. The results were actually quite good:
I thought he had a good observation on creating prompts here:
Don’t get tunnel vision on the prompts. There’s more to a car than the engine. Prompts are important but they’re not everything. With platforms like Dall-E 2 where underlying variables aren’t exposed, the prompts do play a dominant role.
But with Stable Diffusion and Midjourney, there are more controls available to you that affect the output. If you’re not getting what you want from prompts alone in Stable Diffusion, for instance, it could be because you need to shop around the sampler methods and CFG Scale values. Even the starting resolution affects the images you get because it changes the initial noise pattern
Definitely worth the read if you’re interested.