A Blender addon for using Stable Diffusion to render texture bakes for objects.
It will retain your existing UV coords and bake the result to them.
Supports both Automatic1111 and ComfyUI backends
Limitations: The intermediate object has its uv's projected from the view. So even though the result will be mapped 'around' the object, it won't render what isn't seen. Currently doesn't take subdivisions into consideration for the baking.
Multiple objects since v1.5.0
Example video: https://youtu.be/_VmOCM1rY7g More complex example: https://youtu.be/TsxcImL1cCk
Why use this method?
- Control. This is probably even quicker than sketching in a 2D program
- Post processing. Use your textured objects as part of a scene, animate them, light them, put them in a stew.
Is this cheating?
Maybe. And if so, it's a good thing. Some inspiration:
Ian Hubert on World Building in Blender: https://youtu.be/whPWKecazgM
Steps:
-
Download the 'sd_render' folder and install it as an addon.
-
Add a material to your object
-
Unwrap the UV map.
-
Set up a viewer node and render depth. Save result (Note: You can use other control nets, see repo for example with 'scribble'. In that case, you still need to add a viewer node, but just let it show the normal view.)
-
Adjust the camera to maximize rendered area and for optimal projection
-
Select object and hit "Render" in the "Stable Diffusion Render" panel in the "Render" tab
Switching backend: This happens in the addon preferences
After switching, you need to restart for the configuration properties to update
If you have no use for the rendered image in Blender and just want to create a scene for stable diffusion, you can uncheck the button that returns the image. That way you won't have unnecessary image instances in Blender.