1Qwen1.5-MoE-A2.7B: A Small MoE Model with only 2.7B Activated Parameters yet Matching the Performance of State-of-the-Art 7B models(www.marktechpost.com)posted 6 months ago byAkisamb@programming.devMin machine_learning@programming.devView comments0 commentssavehidereportSort:HotTopControversialNewOldWritePreviewBiSubmitNo comments yet!