Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

More context at: "Moonshot's Kimi K2 uses a 1T-parameter MoE architecture with 32B active parameters and outperforms models like GPT-4.1 and DeepSeek-V3 on key benchmarks" <https://www.techmeme.com/250712/p11#a250712p11>


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: