SegMoE: Segmind Mixture of Diffusion Experts
About this article
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Back to Articles SegMoE: Segmind Mixture of Diffusion Experts Published February 3, 2024 Update on GitHub Upvote 7 +1 Yatharth Gupta Warlord-K Follow guest Vishnu V Jaddipal Icar Follow guest Harish Prabhala harishp Follow guest SegMoE is an exciting framework for creating Mixture-of-Experts Diffusion models from scratch! SegMoE is comprehensively integrated within the Hugging Face ecosystem and comes supported with diffusers 🔥! Among the features and integrations being released today: Models on the Hub, with their model cards and licenses (Apache 2.0) Github Repository to create your own MoE-style models. Table of Contents What is SegMoE About the name Inference Samples Using 🤗 Diffusers Using a Local Model Comparison Creating your Own SegMoE Disclaimers and ongoing work Additional Resources Conclusion What is SegMoE? SegMoE models follow the same architecture as Stable Diffusion. Like Mixtral 8x7b, a SegMoE model comes with multiple models in one. The way this works is by replacing some Feed-Forward layers with a sparse MoE layer. A MoE layer contains a router network to select which experts process which tokens most efficiently. You can use the segmoe package to create your own MoE models! The process takes just a few minutes. For further information, please visit the Github Repository. We take inspiration from the popular library mergekit to design segmoe. We thank the contributors of mergekit for such a useful library. For more details on MoEs, see the Hugging Face 🤗 ...