Skip to content

Latest commit

 

History

History
85 lines (82 loc) · 33.9 KB

File metadata and controls

85 lines (82 loc) · 33.9 KB

Pipelines汇总

下表总结了目前所有支持的Pipelines,以及相应的论文、任务、推理脚本。

Pipeline 源链接 任务 推理脚本
animatediff Alt AnimateDiff: Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning Text-to-Video Generation link
alt_diffusion Alt Diffusion Text-to-Image Generation link
alt_diffusion Alt Diffusion Image-to-Image Text-Guided Generation link
audio_diffusion Audio Diffusion Unconditional Audio Generation link
audioldm AudioLDM: Text-to-Audio Generation with Latent Diffusion Models Text-to-Audio Generation link
audioldm2 AudioLDM 2: Learning Holistic Audio Generation with Self-supervised Pretraining Text-to-Audio Generation link
auto_pipeline Stable Diffusion Text-to-Image Generation link
controlnet ControlNet with Stable Diffusion Image-to-Image Text-Guided Generation link
controlnet ControlNet with Stable Diffusion Text-Guided Image-to-Image Generation link
controlnet ControlNet with Stable Diffusion Text-Guided Image Inpainting link
controlnet ControlNet with Stable Diffusion Text-to-Image Generation(Stable Diffusion) link
controlnet ControlNet with Stable Diffusion Text-to-Image Generation(SDXL) link
consistency_models Consistency Models Text-Guided Image-to-Image Generation link
dance_diffusion Dance Diffusion Unconditional Audio Generation link
ddpm Denoising Diffusion Probabilistic Models Unconditional Image Generation link
ddim Denoising Diffusion Implicit Models Unconditional Image Generation link
deepfloyd_if Deep-Floyd Text-Guided Image-to-Image Generation link
deepfloyd_if Deep-Floyd Text-Guided Image Inpainting link
deepfloyd_if Deep-Floyd Text-to-Image Generation link
DiffusionPipeline Mixture of Diffusers Text-to-Image Generation link
DiffusionPipeline SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis Text-to-Image Generation with Refiner link
DiffusionPipeline Shap-E: Generating Conditional 3D Implicit Functions Text-to-3D Generation(img2img) link
DiffusionPipeline Shap-E: Generating Conditional 3D Implicit Functions Text-to-3D Generation link
dit Scalable Diffusion Models with Transformers Class Conditional Image Generation(DIT) link
dit Scalable Diffusion Models with Transformers Class Conditional Image Generation(DIT_3b) link
dit Scalable Diffusion Models with Transformers Class Conditional Image Generation(DIT_7b) link
kandinsky Kandinsky-2 Text-Guided Image-to-Image Generation link
kandinsky Kandinsky-2 Text-Guided Image Inpainting link
kandinsky Kandinsky-2 Text-to-Image Generation link
kandinsky2_2 Kandinsky-2 Text-Guided Image-to-Image Generation link
kandinsky2_2 Kandinsky-2 Text-Guided Image Inpainting link
kandinsky2_2 Kandinsky-2 Text-to-Image Generation link
latent_diffusion High-Resolution Image Synthesis with Latent Diffusion Models Text-to-Image Generation link
latent_diffusion High-Resolution Image Synthesis with Latent Diffusion Models Text-to-Image Generation(dit_3b) link
latent_diffusion High-Resolution Image Synthesis with Latent Diffusion Models Text-to-Image Generation(uvit_small) link
latent_diffusion High-Resolution Image Synthesis with Latent Diffusion Models Super Superresolution link
latent_diffusion_uncond High-Resolution Image Synthesis with Latent Diffusion Models Unconditional Image Generation link
lvdm Latent Video Diffusion Models for High-Fidelity Long Video Generation Text-to-Video Generation link
paint_by_example Paint by Example: Exemplar-based Image Editing with Diffusion Models Image-Guided Image Inpainting link
pndm Pseudo Numerical Methods for Diffusion Models on Manifolds Unconditional Image Generation link
repaint Repaint Image Inpainting link
score_sde_ve Score-Based Generative Modeling through Stochastic Differential Equations Unconditional Image Generation link
semantic_stable_diffusion Semantic Guidance Text-Guided Generation link
spectrogram_diffusion Multi-instrument Music Synthesis with Spectrogram Diffusion Unconditional audio Generation link
stable_diffusion Stable Diffusion Text-to-Image Generation link
stable_diffusion Stable Diffusion Text-Guided Image-to-Image Generation link
stable_diffusion Stable Diffusion Text-Guided Image Inpainting link
stable_diffusion Stable Diffusion Clip_Guided Image Mixing link
stable_diffusion Stable Diffusion Image Variation link
stable_diffusion_2 Stable Diffusion 2 Text-to-Image Generation link
stable_diffusion_2 Stable Diffusion 2 Text-Guided Image-to-Image Generation link
stable_diffusion_2 Stable Diffusion 2 Text-Guided Image Inpainting link
stable_diffusion_2 Stable Diffusion 2 Text-Guided Image Upscaling link
stable_diffusion_3 Stable Diffusion 3 Text-Guided Image-to-Image Generation link
stable_diffusion_3 Stable Diffusion 3 Text-to-Image Generation link
stable_diffusion_xl SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis Text-Guided Image-to-Image Generation link
stable_diffusion_xl SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis Text-Guided Image Inpainting link
stable_diffusion_xl SDXL: Improving Latent Diffusion Models for High-Resolution Image Synthesis Text-to-Image Generation link
stable_diffusion_safe Safe Stable Diffusion Text-to-Image Generation link
stochastic_karras_ve Elucidating the Design Space of Diffusion-Based Generative Models Unconditional Image Generation link
Stable_Video_Diffusion Stable Video Diffusion Image-to-Video Generation link
t2i_adapter T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models Text-to-Image Generation link
t2i_adapter T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models Text-to-Image Generation(Stable Diffusion) link
text_to_video_synthesis zeroscope_v2_XL Text-to-Video Generation link
text_to_video_synthesis ModelScope Text-to-Video Technical Report Text-to-Video Generation link
text_to_video_synthesis Stable Diffusion Text-to-Video Generation link
unclip UnCLIP Text-to-Image Generation link
undiffuser One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale Image-to-Text Generation link
undiffuser One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale Text-to-Image Generation link
undiffuser One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale Image Variation link
undiffuser One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale Text Variation link
undiffuser One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale Image Text Joint Generation link
Image-to-Video unknown Image-to-Video Generation link
versatile_diffusion Versatile Diffusion Text-to-Image Generation link
versatile_diffusion Versatile Diffusion Image Variation link
versatile_diffusion Versatile Diffusion Dual Text and Image Guided Generation link
vq_diffusion VQ Diffusion Text-to-Image Generation link
wuerstchen Wuerstchen: An Efficient Architecture for Large-Scale Text-to-Image Diffusion Modelsn Text-to-Image Generation link