Large Language Model
AI firms 'unprepared' for dangers of building human-level systems, report warns
Artificial intelligence companies are "fundamentally unprepared" for the consequences of creating systems with human-level intellectual performance, according to a leading AI safety group. The Future of Life Institute (FLI) said none of the firms on its AI safety index scored higher than a D for "existential safety planning". One of the five reviewers of the FLI's report said that, despite aiming to develop artificial general intelligence (AGI), none of the companies scrutinised had "anything like a coherent, actionable plan" to ensure the systems remained safe and controllable. AGI refers to a theoretical stage of AI development at which a system is capable of matching a human in carrying out any intellectual task. OpenAI, the developer of ChatGPT, has said its mission is to ensure AGI "benefits all of humanity".
Grok 4 leapfrogs Claude and DeepSeek in LLM rankings, despite safety concerns
Grok 4 by xAI was released on July 9, and it's surged ahead of competitors like DeepSeek and Claude at LMArena, a leaderboard for ranking generative AI models. New AI models are commonly judged on a variety of metrics, including their ability to solve math problems, answer text questions, and write code. The big AI companies use a variety of standardized assessments to measure the effectiveness of their models, such as Humanity's Last Exam, a 2,500-question test designed for AI benchmarking. Typically, when a company like Anthropic or OpenAI releases a new model, it shows improvements on these tests. Unsurprisingly, Grok 4 scores higher than Grok 3 on some key metrics, but it also has to battle in the court of public opinion.
Copilot is coming to cars -- and so are Teams calls
What's worse, being stuck in traffic or on another interminable Microsoft Teams call? Mercedes is teaming up with Microsoft to allow you to take Teams calls while on the go -- and is talking about adding Microsoft Copilot to its suite of luxury cars, too. Microsoft and Mercedes-Benz said Wednesday that the new Mercedes CLA will be able to tap into an in-vehicle camera and give drivers access to a Meetings for Teams application. Somehow, Mercedes is also including Microsoft Intune inside the car, so business workers will be able to access private business data, too. Mercedes is making this part of what it calls MBUX, the Mercedes-Benz user experience.
Masayoshi Son and Sam Altman see no end to AI demand and scaling
SoftBank founder Masayoshi Son and OpenAI chief Sam Altman see insatiable demand for artificial intelligence (AI) that makes it imperative to keep building ever more computing capacity. Speaking via teleconference at SoftBank World, the two business partners argued that advancing AI would lead to new jobs that are not yet imagined, and the advancement of robotics will help kickstart a "self-improvement" loop. "As we drive the cost of AI down, more people want to use it," Altman said in response to Son's question about diminishing returns from further expansion. "So if we make the cost of AI 10 times cheaper, people wanna use it 30 times as much or whatever. And the demand for intelligence in the world just seems to be huge."
Another High-Profile OpenAI Researcher Departs for Meta
OpenAI researcher Jason Wei is joining Meta's new superintelligence lab, according to multiple sources familiar with the matter. Wei worked on OpenAI's o3 and deep research models, according to his personal website. He joined OpenAI in 2023 after a stint at Google, where he worked on chain-of-thought research, which involves training an AI model to process complex queries step-by-step. At OpenAI, Wei became a self-described "diehard" for reinforcement learning, a method of training or refining an AI model with positive or negative feedback. It's become a promising area of AI research--one that several of the researchers Meta has hired for its superintelligence team specialize in.
Thinking Machines Lab Raises a Record 2 Billion, Announces Cofounders
Thinking Machines Lab, an artificial intelligence company founded by top researchers who fled OpenAI, has raised a record 2 billion seed round that values the fledgling firm at 12 billion. The funding round was led by Andreessen Horowitz and included Nvidia, Accel, Cisco, and AMD--among others. The mammoth investment reflects the ultracompetitive race to build advanced AI systems, as well as the premium placed on top AI talent. It is the largest seed funding round in history. Thinking Machines is led by CEO Mira Murati, who stepped down as OpenAI's chief technology officer last September.
DOGE staffer reportedly published secret xAI key to dozens of AI models
Five months after DOGE staffer Marko Elez resigned from the agency over racist social media posts, he's not only back at DOGE, but back in the news for another not-very-positive reason. Cybersecurity journalist Brian Krebs published a report on Monday indicating that, over the weekend, Elez published a private API key to GitHub that would allow users to "directly interact" with some of xAI's (Elon Musk's AI company, for those who haven't been following along) large language models. To be clear, "some" might be understating it, as the total number of LLMs that were made accessible in this leak was at least 52. These LLMs are part of what makes up Grok, the AI chatbot that's integrated directly into X. Yes, Grok is the same one that recently referred to itself as "MechaHitler" just days before entering into a 200 million deal with the U.S. Department of Defense.
What It's Like to Be a Student Who Hates ChatGPT
Sign up for the Slatest to get the most insightful analysis, criticism, and advice out there, delivered to your inbox daily. As a classically trained singer preparing for a professional career, Erin Perry can see quite clearly how artificial intelligence is upending her field--all the way down to the classroom. Perry just completed her first year as a graduate student in voice performance at the Peabody Institute, the prestigious music conservatory run by Johns Hopkins University. It's been rewarding so far: She's been learning how to navigate the modern classical music sector and confronting the relevant impacts of generative A.I., having taken on a project to study the major record labels' lawsuit against the Amazon-backed A.I. startup Anthropic, which trained its models on songwriters' lyrics sans permission or compensation. Understandably, Perry's rather skeptical of A.I.'s artistic applications, and fearful of the sweeping effects it could have on her chosen field, especially as generative-music startups like Suno and Udio are programmed to replicate specific artists and musical styles.
Anthropic's Claude dives into financial analysis. Here's what's new
There have been AI solutions galore for coding, writing, and mathematics, but a technical domain equally as challenging that could use AI assistance, yet is often overlooked, is finance -- until now. On Tuesday, Anthropic launched the Financial Analysis Solution, which instantly pulls financial data from different data providers, both market feeds and internal. The Claude 4 models can then use that information to assist with your financial workloads, including everything from market analysis to research and investment decisions. Also: Anthropic's AI agent can now automate Canva, Asana, Figma and more - here's how it works "The Financial Analysis solution is a comprehensive AI solution that really aims at transforming how finance professionals analyze markets, conduct research, and make investment decisions," said Nicholas Lin, head of product, FSI, at Anthropic, to ZDNET. Partnerships with data providers, such as Box, Daloopa, Databricks, FactSet, Morningstar, S&P Global, Snowflake, and Palantir, allow users to pull information from multiple sources into Claude for analysis without having to context switch manually.
AI text-to-speech programs could "unlearn" how to imitate certain people
AI companies generally keep a tight grip on their models to discourage misuse. For example, if you ask ChatGPT to give you someone's phone number or instructions for doing something illegal, it will likely just tell you it cannot help. However, as many examples over time have shown, clever prompt engineering or model fine-tuning can sometimes get these models to say things they otherwise wouldn't. The unwanted information may still be hiding somewhere inside the model so that it can be accessed with the right techniques. At present, companies tend to deal with this issue by applying guardrails; the idea is to check whether the prompts or the AI's responses contain disallowed material.