Mixture of Experts
Omar Sanseviero @osanseviero Grok weights are out. Download them quickly at https://huggingface.co/xai-org/grok-1 huggingface-cli download xai-org/grok-1 –repo-type model –include ckpt/tensor* –local-dir checkpoints/ckpt-0 –local-dir-use-symlinks False Learn about mixture of experts at https://hf.co/blog/moe Replying to @osanseviero It seems there is a conflict between saying “Grok-1 open-weights model” and “Due to the large size of the model (314B parameters),
Read More »