Comments on: Dell, Lenovo Also Waiting For Their AI Server Waves https://www.nextplatform.com/2023/12/11/dell-lenovo-also-waiting-for-their-ai-server-waves/ In-depth coverage of high-end computing at large enterprises, supercomputing centers, hyperscale data centers, and public clouds. Wed, 03 Jan 2024 15:19:31 +0000 hourly 1 https://wordpress.org/?v=6.7.1 By: Hubert https://www.nextplatform.com/2023/12/11/dell-lenovo-also-waiting-for-their-ai-server-waves/#comment-217509 Wed, 13 Dec 2023 04:58:34 +0000 https://www.nextplatform.com/?p=143383#comment-217509 In reply to Slim Jim.

I agree! In between French News reporting that Maroilles Lesire (AOP Gros de 750 grammes) has won “meilleur fromage du monde 2024” in Lyon, and similarly for “la bière Vandale brune” (named world’s best beer), both from Picardie, we did get a tiny news item here on Mistral AI’s new Mixtral high quality Sparse Mixture-of-Experts (MoE) LLM ( https://mistral.ai/news/mixtral-of-experts/ ).

Apparently, the MoE approach used by Mistral makes it possible to run the equivalent of Llama-2 and GPT-3.5 locally (eg. WinoGrande BBQ), on last-generation Macs for example, with openly available weights (reminds me a bit of Google’s SparseCores splitting of a large ANN into problem-specific sub-components). Advances of this type have the potential to shift part of the AI/ML computations back towards CPUs, including those for “composition” of the underlying experts, in gastronomic fashion (eh-eh-eh!).

]]>
By: Slim Jim https://www.nextplatform.com/2023/12/11/dell-lenovo-also-waiting-for-their-ai-server-waves/#comment-217472 Tue, 12 Dec 2023 09:33:01 +0000 https://www.nextplatform.com/?p=143383#comment-217472 The CoWos-HBM packaging supply chain could probably have exercised more foresight in preparing itself for this tsunami wave of demand, but at least we see the major OEMs navigating in similar boats which is fair. The 9 month gestation period could also be a good thing in some way, seeing how AI/ML is still in a state of flux where models seem to not always be fully baked, and somewhat brittle and unreliable for now.

Yann LeCun (at Meta) recently suggested, for example, that improvements may come from replacing (or combining?) auto-regressive token prediction with planning (classical AI methods of tree- and graph-based searches, with heuristics and backtracking) as shown by Noam Brown’s (now at OpenAI) success at multiplayer Texas hold’em poker, and Diplomacy (Pluribus, and Cicero). Google’s DeepMind Gemini seems to also be moving in this direction if I understand well. This suggests (to me) that CPU oomph may become important again, beyond shoveling data into GPUs, and this could affect the architecture of future AI servers (I think).

]]>