animatediff |
Alt AnimateDiff: Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning |
Text-to-Video Generation |
link |
alt_diffusion |
Alt Diffusion |
Text-to-Image Generation |
link |
alt_diffusion |
Alt Diffusion |
Image-to-Image Text-Guided Generation |
link |
audio_diffusion |
Audio Diffusion |
Unconditional Audio Generation |
link |
audioldm |
AudioLDM: Text-to-Audio Generation with Latent Diffusion Models |
Text-to-Audio Generation |
link |
audioldm2 |
AudioLDM 2: Learning Holistic Audio Generation with Self-supervised Pretraining |
Text-to-Audio Generation |
link |
auto_pipeline |
Stable Diffusion |
Text-to-Image Generation |
link |
controlnet |
ControlNet with Stable Diffusion |
Image-to-Image Text-Guided Generation |
link |
controlnet |
ControlNet with Stable Diffusion |
Text-Guided Image-to-Image Generation |
link |
controlnet |
ControlNet with Stable Diffusion |
Text-Guided Image Inpainting |
link |
controlnet |
ControlNet with Stable Diffusion |
Text-to-Image Generation(Stable Diffusion) |
link |
controlnet |
ControlNet with Stable Diffusion |
Text-to-Image Generation(SDXL) |
link |
consistency_models |
Consistency Models |
Text-Guided Image-to-Image Generation |
link |
dance_diffusion |
Dance Diffusion |
Unconditional Audio Generation |
link |
ddpm |
Denoising Diffusion Probabilistic Models |
Unconditional Image Generation |
link |
ddim |
Denoising Diffusion Implicit Models |
Unconditional Image Generation |
link |
deepfloyd_if |
Deep-Floyd |
Text-Guided Image-to-Image Generation |
link |
deepfloyd_if |
Deep-Floyd |
Text-Guided Image Inpainting |
link |
deepfloyd_if |
Deep-Floyd |
Text-to-Image Generation |
link |
DiffusionPipeline |
Mixture of Diffusers |
Text-to-Image Generation |
link |
DiffusionPipeline |
SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis |
Text-to-Image Generation with Refiner |
link |
DiffusionPipeline |
Shap-E: Generating Conditional 3D Implicit Functions |
Text-to-3D Generation(img2img) |
link |
DiffusionPipeline |
Shap-E: Generating Conditional 3D Implicit Functions |
Text-to-3D Generation |
link |
dit |
Scalable Diffusion Models with Transformers |
Class Conditional Image Generation(DIT) |
link |
dit |
Scalable Diffusion Models with Transformers |
Class Conditional Image Generation(DIT_3b) |
link |
dit |
Scalable Diffusion Models with Transformers |
Class Conditional Image Generation(DIT_7b) |
link |
kandinsky |
Kandinsky-2 |
Text-Guided Image-to-Image Generation |
link |
kandinsky |
Kandinsky-2 |
Text-Guided Image Inpainting |
link |
kandinsky |
Kandinsky-2 |
Text-to-Image Generation |
link |
kandinsky2_2 |
Kandinsky-2 |
Text-Guided Image-to-Image Generation |
link |
kandinsky2_2 |
Kandinsky-2 |
Text-Guided Image Inpainting |
link |
kandinsky2_2 |
Kandinsky-2 |
Text-to-Image Generation |
link |
latent_diffusion |
High-Resolution Image Synthesis with Latent Diffusion Models |
Text-to-Image Generation |
link |
latent_diffusion |
High-Resolution Image Synthesis with Latent Diffusion Models |
Text-to-Image Generation(dit_3b) |
link |
latent_diffusion |
High-Resolution Image Synthesis with Latent Diffusion Models |
Text-to-Image Generation(uvit_small) |
link |
latent_diffusion |
High-Resolution Image Synthesis with Latent Diffusion Models |
Super Superresolution |
link |
latent_diffusion_uncond |
High-Resolution Image Synthesis with Latent Diffusion Models |
Unconditional Image Generation |
link |
lvdm |
Latent Video Diffusion Models for High-Fidelity Long Video Generation |
Text-to-Video Generation |
link |
paint_by_example |
Paint by Example: Exemplar-based Image Editing with Diffusion Models |
Image-Guided Image Inpainting |
link |
pndm |
Pseudo Numerical Methods for Diffusion Models on Manifolds |
Unconditional Image Generation |
link |
repaint |
Repaint |
Image Inpainting |
link |
score_sde_ve |
Score-Based Generative Modeling through Stochastic Differential Equations |
Unconditional Image Generation |
link |
semantic_stable_diffusion |
Semantic Guidance |
Text-Guided Generation |
link |
spectrogram_diffusion |
Multi-instrument Music Synthesis with Spectrogram Diffusion |
Unconditional audio Generation |
link |
stable_diffusion |
Stable Diffusion |
Text-to-Image Generation |
link |
stable_diffusion |
Stable Diffusion |
Text-Guided Image-to-Image Generation |
link |
stable_diffusion |
Stable Diffusion |
Text-Guided Image Inpainting |
link |
stable_diffusion |
Stable Diffusion |
Clip_Guided Image Mixing |
link |
stable_diffusion |
Stable Diffusion |
Image Variation |
link |
stable_diffusion_2 |
Stable Diffusion 2 |
Text-to-Image Generation |
link |
stable_diffusion_2 |
Stable Diffusion 2 |
Text-Guided Image-to-Image Generation |
link |
stable_diffusion_2 |
Stable Diffusion 2 |
Text-Guided Image Inpainting |
link |
stable_diffusion_2 |
Stable Diffusion 2 |
Text-Guided Image Upscaling |
link |
stable_diffusion_3 |
Stable Diffusion 3 |
Text-Guided Image-to-Image Generation |
link |
stable_diffusion_3 |
Stable Diffusion 3 |
Text-to-Image Generation |
link |
stable_diffusion_xl |
SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis |
Text-Guided Image-to-Image Generation |
link |
stable_diffusion_xl |
SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis |
Text-Guided Image Inpainting |
link |
stable_diffusion_xl |
SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis |
Text-to-Image Generation |
link |
stable_diffusion_safe |
Safe Stable Diffusion |
Text-to-Image Generation |
link |
stochastic_karras_ve |
Elucidating the Design Space of Diffusion-Based Generative Models |
Unconditional Image Generation |
link |
Stable_Video_Diffusion |
Stable Video Diffusion |
Image-to-Video Generation |
link |
t2i_adapter |
T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models |
Text-to-Image Generation |
link |
t2i_adapter |
T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models |
Text-to-Image Generation(Stable Diffusion) |
link |
text_to_video_synthesis |
zeroscope_v2_XL |
Text-to-Video Generation |
link |
text_to_video_synthesis |
ModelScope Text-to-Video Technical Report |
Text-to-Video Generation |
link |
text_to_video_synthesis |
Stable Diffusion |
Text-to-Video Generation |
link |
unclip |
UnCLIP |
Text-to-Image Generation |
link |
undiffuser |
One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale |
Image-to-Text Generation |
link |
undiffuser |
One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale |
Text-to-Image Generation |
link |
undiffuser |
One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale |
Image Variation |
link |
undiffuser |
One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale |
Text Variation |
link |
undiffuser |
One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale |
Image Text Joint Generation |
link |
Image-to-Video |
unknown |
Image-to-Video Generation |
link |
versatile_diffusion |
Versatile Diffusion |
Text-to-Image Generation |
link |
versatile_diffusion |
Versatile Diffusion |
Image Variation |
link |
versatile_diffusion |
Versatile Diffusion |
Dual Text and Image Guided Generation |
link |
vq_diffusion |
VQ Diffusion |
Text-to-Image Generation |
link |
wuerstchen |
Wuerstchen: An Efficient Architecture for Large-Scale Text-to-Image Diffusion Modelsn |
Text-to-Image Generation |
link |