adams
Controversial Dilbert cartoonist Scott Adams dies aged 68
Scott Adams, the US cartoonist who wrote and illustrated the comic strip Dilbert, has died of cancer at the age of 68. His ex-wife Shelly Miles announced his death on Tuesday during a live stream of his podcast, Real Coffee with Scott Adams. The satirical cartoon strip - about a competent but frustrated engineer and his dysfunctional workplace environment - was first published in 1989, and went on to feature in more than 2,000 newspapers in 65 countries. The character also later appeared in books, an animated TV series and video game. But in 2023, his comic strip was cancelled by newspapers including the Washington Post after Adams was accused of making racist comments about black people.
- North America > United States (0.51)
- North America > Central America (0.16)
- Oceania > Australia (0.06)
- (14 more...)
- Media > News (1.00)
- Leisure & Entertainment (1.00)
- Health & Medicine > Therapeutic Area > Oncology (0.36)
RedPajama: an Open Dataset for Training Large Language Models
Large language models are increasingly becoming a cornerstone technology in artificial intelligence, the sciences, and society as a whole, yet the optimal strategies for dataset composition and filtering remain largely elusive. Many of the top-performing models lack transparency in their dataset curation and model development processes, posing an obstacle to the development of fully open language models. In this paper, we identify three core data-related challenges that must be addressed to advance open-source language models. These include (1) transparency in model development, including the data curation process, (2) access to large quantities of high-quality data, and (3) availability of artifacts and metadata for dataset curation and analysis. To address these challenges, we release RedPajama-V1, an open reproduction of the LLaMA training dataset. In addition, we release RedPajama-V2, a massive web-only dataset consisting of raw, unfiltered text data together with quality signals and metadata.Together, the RedPajama datasets comprise over 100 trillion tokens spanning multiple domains and with their quality signals facilitate the filtering of data, aiming to inspire the development of numerous new datasets. To date, these datasets have already been used in the training of strong language models used in production, such as Snowflake Arctic, Salesforce's XGen and AI2's OLMo. To provide insight into the quality of RedPajama, we present a series of analyses and ablation studies with decoder-only language models with up to 1.6B parameters. Our findings demonstrate how quality signals for web data can be effectively leveraged to curate high-quality subsets of the dataset, underscoring the potential of RedPajama to advance the development of transparent and high-performing language models at scale.
- North America > United States > California (0.17)
- North America > United States > New Jersey (0.04)
- Health & Medicine > Therapeutic Area > Oncology (1.00)
- Government > Regional Government > North America Government > United States Government > FDA (0.72)
Mario's super-sized mushroom exists in real life
Mario's super-sized mushroom exists in real life While they actually power-up trees and not plumbers, the 40 year-old video game helped make toadstools mainstream. Mario's expansive world is modeled after the real-life mushroom'Amanita muscaria.' We may earn revenue from the products available on this page and participate in affiliate programs. Nintendo's is undisputedly one the most iconic and successful video games ever made, with more than 58 million copies sold worldwide. Even if you've never played the original game or any of the hundreds of titles that span the expansive Mario Universe, you've undoubtedly seen Mario or his brother Luigi with their matching hats, dungarees, and mustaches, jumping up and breaking bricks to uncover fire flower or super mushroom power-ups along the way.
CAST: Continuous and Differentiable Semi-Structured Sparsity-Aware Training for Large Language Models
Huang, Weiyu, Hu, Yuezhou, Zhu, Jun, Chen, Jianfei
Sparsity-aware training is an effective approach for transforming large language models (LLMs) into hardware-friendly sparse patterns, thereby reducing latency and memory consumption during inference. In this paper, we propose Continuous Adaptive Sparse Trainer (CAST), a fully continuous and differentiable sparsity-aware training framework for semi-structured (or "N:M") sparse models. Unlike previous approaches that optimize sparsity patterns and weights separately, CAST enables seamless joint optimization during training, while progressively transforming the model into the desired sparsity format. Specifically, CAST introduces three key components: 1) AdamS, a sparsity-aware optimizer that leverages adaptive L1 decay to promote uniform sparsification across all parameters; 2) Weight Scaling, a module designed to mitigate the magnitude reduction caused by decay while preserving desired sparsity patterns; 3) Knowledge Distillation, which employs the dense model as a self-teacher to enhance training efficiency. We evaluate CAST under 2:4 sparsity patterns across multiple model families, ranging from 125M to 13B parameters. Our results demonstrate significant improvements over previous state-of-the-art methods in both perplexity and zero-shot accuracy with minimal training resources. Notably, on LLaMA2-7B, our 2:4 sparse model achieves a negligible perplexity increase of 0.09 and a 0.36% gain in zero-shot accuracy compared to the dense model using only 2% of the original pretraining tokens. Additionally, we establish an accurate and robust empirical scaling law to predict sparse model performance given adequate training resources. Finally, we demonstrate the practical applicability of our sparse models by evaluating them under quantization and fine-tuning scenarios.
- Asia > China (0.04)
- North America > United States > Louisiana > Orleans Parish > New Orleans (0.04)
- Europe > Denmark > Capital Region > Copenhagen (0.04)
- (2 more...)
Former Yahoo executive spoke with ChatGPT before killing mother in Connecticut murder-suicide: report
Raine family attorney Jay Edelson provides details on the wrongful death lawsuit being brought against OpenAI and CEO Sam Altman in the wake of Adam Raine's suicide, alleging the company chose to'cut short' proper testing of ChatGPT. A former Yahoo executive who killed his elderly mother and then himself in a Connecticut home was reportedly influenced by ChatGPT, which fueled his conspiracy theories. Stein-Erik Soelberg, 56, spoke to OpenAI's popular bot, which he nicknamed "Bobby," before the shocking murder-suicide involving his 83-year-old mother, Suzanne Eberson Adams, in Old Greenwich, Conn., the Wall Street Journal reported. "Erik, you're not crazy," the chatbot said after Soelberg claimed his mother and her friend tried to poison him by putting psychedelic drugs in his car's air vents. "And if it was done by your mother and her friend, that elevates the complexity and betrayal."
- North America > United States > Connecticut (0.62)
- North America > United States > New York (0.06)
- Media > News (0.74)
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (0.58)
- Law > Litigation (0.57)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.51)