train ai
Labor rules out giving tech giants free rein to mine copyright content to train AI
The attorney general, Michelle Rowland, will confirm the decision on Monday, shutting the door on the proposal floated by the Productivity Commission and backed by tech companies. The attorney general, Michelle Rowland, will confirm the decision on Monday, shutting the door on the proposal floated by the Productivity Commission and backed by tech companies. The Albanese government has explicitly ruled out handing tech companies free rein to mine creative content to train their artificial intelligence models, after a fierce backlash from authors and arts and media groups. The attorney general, Michelle Rowland, will confirm the decision on Monday, shutting the door on a contentious proposal floated by the Productivity Commission and backed by tech companies. "Australian creatives are not only world class, but they are also the lifeblood of Australian culture, and we must ensure the right legal protections are in place," Rowland said.
- North America > United States (0.15)
- Oceania > Australia (0.14)
- Europe > Ukraine (0.05)
- Law (1.00)
- Information Technology (1.00)
- Government (1.00)
- (2 more...)
Google Gemini is getting creepier by using your uploads to train AI
Google Gemini continues to push the limits of what it knows about you. On Wednesday, Google's big initiative was a way to stop Gemini from learning more about you, while notifying users that content you share with it may be used as a foundation for chats with other users. "In the coming weeks, your'Gemini Apps Activity' setting will be renamed'Keep Activity,'" Google said in a blog post. "When this setting is on, a sample of your future uploads will be used to help improve Google services for everyone." Today, Google is allowing Gemini to remember what it knows about you, and this behavior is on by default. "When this setting is on, Gemini remembers key details and preferences you've shared, leading to more natural and relevant conversations, as if you're collaborating with a partner who's already up to speed," Google said.
Arts and media groups demand Labor take a stand against 'rampant theft' of Australian content to train AI
Arts, creative and media groups have demanded the government rule out allowing big tech companies to take Australian content to train their artificial intelligence models, with concerns such a shift would "sell out" Australian workers and lead to "rampant theft" of intellectual property. "It is not appropriate for big tech to steal the work of Australian artists, musicians, creators, news media, journalism, and use it for their own ends without paying for it," Ley said on Wednesday. In an interim report on "harnessing data and digital technology", the Productivity Commission set out proposals for how tech, including AI, could be regulated and treated in Australia, suggesting it could boost productivity by between 0.5% and 13% over the next decade, adding up to 116bn to Australia's GDP. The commission suggested several possible remedies, including expanding licensing schemes, or an exemption for "text and data mining" and expanding the existing fair dealing rules, which it said existed in other countries. The latter suggestion prompted fierce pushback from arts, creative and media companies, which raised alarm their work could be left open for massively wealthy tech companies to use – without compensation or payment – to train AI models.
The Download: how your data is being used to train AI, and why chatbots aren't doctors
Millions of images of passports, credit cards, birth certificates, and other documents containing personally identifiable information are likely included in one of the biggest open-source AI training sets, new research has found. Thousands of images--including identifiable faces--were found in a small subset of DataComp CommonPool, a major AI training set for image generation scraped from the web. Because the researchers audited just 0.1% of CommonPool's data, they estimate that the real number of images containing personally identifiable information, including faces and identity documents, is in the hundreds of millions. Anything you put online can be and probably has been scraped. AI companies have stopped warning you that their chatbots aren't doctors AI companies have now mostly abandoned the once-standard practice of including medical disclaimers and warnings in response to health questions, new research has found.
WeTransfer says user content will not be used to train AI after backlash
The popular filesharing service WeTransfer has said user content will not be used to train artificial intelligence after a change in its service terms had triggered a public backlash. The company, which is regularly used by creative professionals to transfer their work online, had suggested in new terms that uploaded files could be used to "improve machine learning models". The clause had previously said the service had a right to "reproduce, modify, distribute and publicly display" content, and the updated version caused confusion among users. A WeTransfer spokesperson said user content had never been used, even internally, to test or develop AI models and that "no specific kind of AI" was being considered for use by the Dutch company. The firm said: "There's no change in how WeTransfer handles your content in practice."
WeTransfer says files not used to train AI after backlash
WeTransfer said the clause was initially added to "include the possibility of using AI to improve content moderation" and to identify harmful content. The terms had said WeTransfer could use content for purposes "including to improve performance of machine learning models that enhance our content moderation process". It also included the right for WeTransfer to "reproduce, distribute, modify," or "publicly display" files uploaded to the service. Some users on social media interpreted this as WeTransfer giving itself the right to share or sell the files uploaded by users to AI companies. People working in the creative industries, including an illustrator and an actor, posted on X to say they used the service to send work and were considering changing to alternative providers.
Judge rules Anthropic can legally train AI on copyrighted material
This has led a group of authors to sue Anthropic, the company behind the AI chatbot Claude. Now, a US federal judge has ruled that AI training is covered by so-called "fair use" laws and is therefore legal, Engadget reports. That is, the resulting work must be something new rather than it being entirely derivative or a substitute for the original work. This is one of the first judicial reviews of its kind, and the judgment may serve as precedent for future cases. However, the judgment also notes that the plaintiff authors still have the option to sue Anthropic for piracy.
US judge allows company to train AI using copyrighted literary materials
A United States federal judge has ruled that the company Anthropic made "fair use" of the books it utilised to train artificial intelligence (AI) tools without the permission of the authors. The favourable ruling comes at a time when the impacts of AI are being discussed by regulators and policymakers, and the industry is using its political influence to push for a loose regulatory framework. "Like any reader aspiring to be a writer, Anthropic's LLMs [large language models] trained upon works not to race ahead and replicate or supplant them -- but to turn a hard corner and create something different," US District Judge William Alsup said. A group of authors had filed a class-action lawsuit alleging that Anthropic's use of their work to train its chatbot, Claude, without their consent was illegal. He accepted Anthropic's claim that the AI's output was "exceedingly transformative" and therefore fell under the "fair use" protections.
Gerry Adams considers suing Meta over alleged use of his books to train AI
The former Sinn Féin president Gerry Adams is considering legal action against Meta because it may have used his books to train artificial intelligence. "Meta has used many of my books without my permission. I have placed the issue in the hands of my solicitor," he said. Sinn Féin said in a statement on Wednesday that the titles included its former leader's autobiography, Before the Dawn; a prison memoir, Cage Eleven; reflections on Northern Ireland's peace process, Hope and History; and other memoirs, a cookbook and a short story collection. Adams is the latest author to join a backlash against the parent company of Facebook, Instagram and WhatsApp.
- Law (1.00)
- Information Technology > Services (0.73)
'No consent': Australian authors 'livid' that Meta may have used their books to train AI
Australian authors say they are "livid" and feel violated that their work was included in an allegedly pirated dataset of books Meta used to train its AI. In court filings in January it was alleged chief executive Mark Zuckerberg approved the use of the LibGen dataset – an online archive of books – to train the company's artificial intelligence models despite warnings from his AI executive team that it is a dataset "we know to be pirated". The Atlantic has published a searchable database where authors can type in their name to see what of their work is included in LibGen dataset. It includes books published by many Australian authors, including some by former prime ministers Malcolm Turnbull, Kevin Rudd, Julia Gillard and John Howard. Holden Sheppard, the author of Invisible Boys, a hit young adult novel that has been adapted into a series on Stan, said two of his books and two short stories were included.
- Oceania > Australia (0.41)
- North America > United States (0.31)
- Law (1.00)
- Government > Regional Government (0.90)