Working
🔨 researching next-generation modeling paradigms; building scalable foundation model systems
Pinned Loading
-
XueFuzhao/OpenMoE
XueFuzhao/OpenMoE PublicA family of open-sourced Mixture-of-Experts (MoE) Large Language Models
-
NVIDIA/Megatron-LM
NVIDIA/Megatron-LM PublicOngoing research training transformer models at scale
-
deepseek-ai/DeepSeek-MoE
deepseek-ai/DeepSeek-MoE PublicDeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
-
EvolvingLMMs-Lab/lmms-engine
EvolvingLMMs-Lab/lmms-engine PublicA simple, unified multimodal models training engine. Lean, flexible, and built for hacking at scale.
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.