Scattered Mixture-of-Experts Implementation - Presents ScatterMoE, an implementation of Sparse Mixture-of-Experts on GPU - Enables a higher throughput and lower memory footprint repo: https://t.co/kJZjY6Q4O7 abs: https://t.co/xxIx1a2vu2 https://t.co/IqWxmNln6W
— Aran Komatsuzaki (@arankomatsuzaki) Mar 14, 2024
from Twitter https://twitter.com/arankomatsuzaki
March 14, 2024 at 01:48AM
via IFTTT
No comments:
Post a Comment