> [...] Frontier models readily generate detailed image descriptions and… more
5 links
> While Mixture-of-Experts (MoE) scales capacity via conditional computation,… more
> "A generator is a parser of randomness." This perspective on generators for… more
> We introduce Moshi, a speech-text foundation model and full-duplex spoken… more
> LLMs are useful because they generalize so well. But can you have too much of… more