Show simple item record

dc.contributor.authorLiao, Xudong
dc.contributor.authorSun, Yijun
dc.contributor.authorTian, Han
dc.contributor.authorWan, Xinchen
dc.contributor.authorJin, Yilun
dc.contributor.authorWang, Zilong
dc.contributor.authorRen, Zhenghang
dc.contributor.authorHuang, Xinyang
dc.contributor.authorLi, Wenxue
dc.contributor.authorTse, Kin Fai
dc.contributor.authorZhong, Zhizhen
dc.contributor.authorLiu, Guyue
dc.contributor.authorZhang, Ying
dc.contributor.authorYe, Xiaofeng
dc.contributor.authorZhang, Yiming
dc.contributor.authorChen, Kai
dc.date.accessioned2025-09-10T19:26:51Z
dc.date.available2025-09-10T19:26:51Z
dc.date.issued2025-08-27
dc.identifier.isbn979-8-4007-1524-2
dc.identifier.urihttps://hdl.handle.net/1721.1/162639
dc.descriptionSIGCOMM ’25, Coimbra, Portugalen_US
dc.description.abstractMixture-of-Expert (MoE) models outperform conventional models by selectively activating different subnets, named experts, on a per-token basis. This gated computation generates dynamic communications that cannot be determined beforehand, challenging the existing GPU interconnects that remain static during distributed training. In this paper, we advocate for a first-of-its-kind system, called MixNet, that unlocks topology reconfiguration during distributed MoE training. Towards this vision, we first perform a production measurement study and show that the MoE dynamic communication pattern has strong locality, alleviating the need for global reconfiguration. Based on this, we design and implement a regionally reconfigurable high-bandwidth domain that augments existing electrical interconnects using optical circuit switching (OCS), achieving scalability while maintaining rapid adaptability. We build a fully functional MixNet prototype with commodity hardware and a customized collective communication runtime. Our prototype trains state-of-the-art MoE models with in-training topology reconfiguration across 32 A100 GPUs. Large-scale packet-level simulations show that MixNet achieves performance comparable to a non-blocking fat-tree fabric while boosting the networking cost efficiency (e.g., performance per dollar) of four representative MoE models by 1.2×–1.5× and 1.9×–2.3× at 100 Gbps and 400 Gbps link bandwidths, respectively.en_US
dc.publisherACM|ACM SIGCOMM 2025 Conferenceen_US
dc.relation.isversionofhttps://doi.org/10.1145/3718958.3750465en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titleMixNet: A Runtime Reconfigurable Optical-Electrical Fabric for Distributed Mixture-of-Experts Trainingen_US
dc.typeArticleen_US
dc.identifier.citationXudong Liao, Yijun Sun, Han Tian, Xinchen Wan, Yilun Jin, Zilong Wang, Zhenghang Ren, Xinyang Huang, Wenxue Li, Kin Fai Tse, Zhizhen Zhong, Guyue Liu, Ying Zhang, Xiaofeng Ye, Yiming Zhang, and Kai Chen. 2025. MixNet: A Runtime Reconfigurable Optical-Electrical Fabric for Distributed Mixture-of-Experts Training. In Proceedings of the ACM SIGCOMM 2025 Conference (SIGCOMM '25). Association for Computing Machinery, New York, NY, USA, 554–574.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratoryen_US
dc.identifier.mitlicensePUBLISHER_POLICY
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2025-09-01T07:53:41Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2025-09-01T07:53:42Z
mit.licensePUBLISHER_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record