Collaborating Authors


Hebbian Governance


TL;DR -- Each set of rules, even similar ones, behave very differently. Researchers are finding the rules that make a system learn and innovate -- in artificial neural networks, the algorithms used in A.I. Those are far superior at finding results, compared to the crude and corrupted'algorithm' of our nations' governance. Our governmental protocol currently forms a giant, dumb A.I. and that's why we can't solve problems that a single human brain finds obvious. We need specific, researched rules for collaboration and governance, such that our Leviathan behaves more like a neural network, a functioning brain. Research gives us a clear picture of which options seem good, so far, and where we should look for improvements.

MPs call on the government to take urgent action on Artificial Intelligence accountability


A new report by the All-Party Parliamentary Group (APPG) on the Future of Work has called on the government to bring forward robust proposals for artificial intelligence (AI) regulation. The APPG inquiry found that AI is transforming work and working lives across the country in ways that have plainly outpaced, or avoid, the existing regimes for regulation. Their recommendations are aimed at ensuring that the AI ecosystem is genuinely human-centred, principles-driven and accountable to shape a future of better work. They are centred around a proposal for an Accountability for Algorithms Act ('the AAA'). The AAA offers an overarching, principles-driven framework for governing and regulating AI in response to the fast-changing developments in workplace technology.

$10 million to build defence's AI capability and support critical Tech for Australia


The Morrison Government is investing $10 million in innovative artificial intelligence (AI) technologies that will strengthen Defence's military capability and support highly skilled jobs in Australia's defence industry. The investment supports the Government's new Blueprint for Critical Technologies and Action Plan, released by the Prime Minister yesterday. It also contributes to the development of a sovereign critical technology capability in AI, one of the Government's nine listed critical technologies of national interest. Minister for Defence Industry and Science and Technology Melissa Price today announced 10 new Defence Innovation Hub contracts funded under the Government's two-year, $32 million COVID economic stimulus package. The package was established to support jobs growth in the defence industry while navigating the challenges posed by the pandemic.

EU Parliament, countries want more innovation, less burden in AI Act


An internal report on Artificial Intelligence recently approved by a special committee of the European Parliament embodies a push from EU lawmakers and member states to make regulation on artificial intelligence less burdensome and more innovation-friendly. Christian Democrat MEP Axel Voss has been leading the charge against "overburdening" companies with excessive regulation, arguing that the EU regulatory environment should leave more room for innovation. That was the underlying motive of an own-initiative report on Artificial Intelligence in a Digital Age, recently approved in the AIDA committee, a parliamentary body set up in 2020, under Voss' leadership. "We need a better regulatory framework that learns also from the mistakes of the GDPR," Voss said while presenting the report. Instead of overburdening companies, the AI Act should give clear guidance and should leave space for innovation, he added.

Stanford HAI 2021 fall conference: four radical proposals for a better society


This year's Stanford HAI virtual fall conference took place on 9-10 November. It comprised a discussion of four policy proposals that respond to the issues and opportunities created by artificial intelligence. The premise is that each policy proposal poses a challenge to the status quo. These proposals were presented to panels of experts who debated the merits and issues surrounding each policy. The event was recorded and you can watch both days' sessions on YouTube.

EU artificial intelligence regulation risks undermining social safety net


The European Union's (EU) proposed plan to regulate the use of artificial intelligence (AI) threatens to undermine the bloc's social safety net, and is ill-equipped to protect people from surveillance and discrimination, according to a report by Human Rights Watch. Social security support across Europe is increasingly administered by AI-powered algorithms, which are being used by governments to allocate life-saving benefits, provide job support and control access to a variety of social services, said Human Rights Watch in its 28-page report, How the EU's flawed artificial intelligence regulation endangers the social safety net. Drawing on case studies from Ireland, France, the Netherlands, Austria, Poland and the UK, the non-governmental organisation (NGO) found that Europe's trend towards automation is discriminating against people in need of social security support, compromising their privacy, and making it harder for them to obtain government assistance. It added that while the EU's Artificial Intelligence Act (AIA) proposal, which was published in April 2021, does broadly acknowledge the risks associated with AI, "it does not meaningfully protect people's rights to social security and an adequate standard of living". "In particular, its narrow safeguards neglect how existing inequities and failures to adequately protect rights – such as the digital divide, social security cuts, and discrimination in the labour market – shape the design of automated systems, and become embedded by them."

(PDF) Time Series Segmentation Based on Stationarity Analysis to Improve New Samples Prediction


A wide range of applications based on sequential data, named time series, have become increasingly popular in recent years, mainly those based on the Internet of Things (IoT). Several different machine learning algorithms exploit the patterns extracted from sequential data to support multiple tasks. However, this data can suffer from unreliable readings that can lead to low accuracy models due to the low-quality training sets available. Detecting the change point between high representative segments is an important ally to find and thread biased subsequences. By constructing a framework based on the Augmented Dickey-Fuller (ADF) test for data stationarity, two proposals to automatically segment subsequences in a time series were developed.

Google pursues Pentagon cloud contract in spite of past employee concerns


Three years after its involvement with the military's controversial Project Maven program led to employee strife within its walls, Google reportedly hopes to once again work with the Pentagon. According to The New York Times, the company is "aggressively" pursuing the Defense Department's Joint Warfighting Cloud Capability cloud contract. A Google spokesperson confirmed to Engadget it was pursuing a bid. Announced at the start of July, the program is a replacement for the military's cancelled $10 billion Joint Enterprise Defense Infrastructure initiative. With JEDI, the Pentagon had planned to modernize its IT infrastructure with help from Microsoft.

AI rules: what the European Parliament wants


Parliament is working on the Commission proposal, presented on 21 April 2021, for turning Europe into the global hub for trustworthy AI. Ahead of the Commission's proposal on AI, the Parliament set up a special committee to analyse the impact of artificial intelligence on the EU economy. "Europe needs to develop AI that is trustworthy, eliminates biases and discrimination, and serves the common good, while ensuring business and industry thrive and generate economic prosperity," said the new committee chair Dragoș Tudorache. On 20 October 2020, Parliament adopted three reports outlining how the EU can best regulate AI while boosting innovation, ethical standards and trust in technology. One of the reports focuses on how to ensure safety, transparency and accountability, prevent bias and discrimination, foster social and environmental responsibility, and ensure respect for fundamental rights.

EAB talk probes balance between border biometrics benefits and rights risks


Achieving a balance between the benefits of biometrics in border processes and the risk that they may cause harm to fundamental rights will require gathering more data about the problems biometrics collection is meant to address, as well as detailed policy considerations, attendees heard in the European Association for Biometrics' (EAB's) virtual lunch talk this week. Bianca-Ioana Marcu of Vrije Universiteit Brussel (VUB) gave a presentation on'Biometrics, Facial Recognition and the Fundamental Rights of Migrants,' focussing on the EU migration management context, and considering non-technical impacts. Marcu noted a tangle of databases that could be involved in EU migration processes, but focussed on the EURODAC database. Biometrics have been adopted in immigration systems to apply efficiency, trust and reliability to the large numbers of people moving between countries, Marcu points out. Pressures on EU external borders, both from migration volumes and terrorism concerns, have resulted in a move towards "the establishment of a genuine security union," she says, facilitated by EU-wide information systems.