grem
'I love you too!' My family's creepy, unsettling week with an AI toy
'Let's talk about something fun!' Grem the AI chatbot toy. 'Let's talk about something fun!' Grem the AI chatbot toy. 'I love you too!' My family's creepy, unsettling week with an AI toy The cuddly chatbot Grem is designed to'learn' your child's personality, while every conversation they have is recorded, then transcribed by a third party. It wasn't long before I wanted this experiment to be over ... 'I'm going to throw that thing into a river!" my wife says as she comes down the stairs looking frazzled after putting our four-year-old daughter to bed. To be clear, "that thing" is not our daughter, Emma*. It's Grem, an AI-powered stuffed alien toy that the musician Claire Boucher, better known as Grimes, helped develop with toy company Curio. Designed for kids aged three and over and built with OpenAI's technology, the toy is supposed to "learn" your child's personality and have fun, educational conversations with them. It's advertised as a healthier alternative to screen time and is ...
- North America > United States (0.29)
- Europe > United Kingdom (0.15)
- Europe > Ukraine (0.05)
- (5 more...)
- Government > Regional Government (0.70)
- Leisure & Entertainment > Sports (0.69)
Armada: Memory-Efficient Distributed Training of Large-Scale Graph Neural Networks
Waleffe, Roger, Sarda, Devesh, Mohoney, Jason, Vlatakis-Gkaragkounis, Emmanouil-Vasileios, Rekatsinas, Theodoros, Venkataraman, Shivaram
We study distributed training of Graph Neural Networks (GNNs) on billion-scale graphs that are partitioned across machines. Efficient training in this setting relies on min-edge-cut partitioning algorithms, which minimize cross-machine communication due to GNN neighborhood sampling. Yet, min-edge-cut partitioning over large graphs remains a challenge: State-of-the-art (SoTA) offline methods (e.g., METIS) are effective, but they require orders of magnitude more memory and runtime than GNN training itself, while computationally efficient algorithms (e.g., streaming greedy approaches) suffer from increased edge cuts. Thus, in this work we introduce Armada, a new end-to-end system for distributed GNN training whose key contribution is GREM, a novel min-edge-cut partitioning algorithm that can efficiently scale to large graphs. GREM builds on streaming greedy approaches with one key addition: prior vertex assignments are continuously refined during streaming, rather than frozen after an initial greedy selection. Our theoretical analysis and experimental results show that this refinement is critical to minimizing edge cuts and enables GREM to reach partition quality comparable to METIS but with 8-65x less memory and 8-46x faster. Given a partitioned graph, Armada leverages a new disaggregated architecture for distributed GNN training to further improve efficiency; we find that on common cloud machines, even with zero communication, GNN neighborhood sampling and feature loading bottleneck training. Disaggregation allows Armada to independently allocate resources for these operations and ensure that expensive GPUs remain saturated with computation. We evaluate Armada against SoTA systems for distributed GNN training and find that the disaggregated architecture leads to runtime improvements up to 4.5x and cost reductions up to 3.1x.
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- Europe > Portugal > Lisbon > Lisbon (0.04)