Goto

Collaborating Authors

 automation


Women in tech and finance at higher risk from AI job losses, report says

The Guardian

The Corporation of London is calling on employers to re-skill female workers not currently in technical roles. The Corporation of London is calling on employers to re-skill female workers not currently in technical roles. 'Mid-career' females also being sidelined by rigid hiring processes, says City of London Corporation Women working in tech and financial services are at greater risk of losing their jobs to increased use of AI and automation than their male peers, according to a report that found experienced females were also being sidelined as a result of "rigid hiring processes". "Mid-career" women - with at least five years' experience - are being overlooked for digital roles in the tech and financial and professional services sectors, where they are traditionally underrepresented, according to the report by the City of London Corporation. The governing body that runs the capital's Square Mile found female applicants were discriminated against by rigid, and sometimes automated, screening of their CVs, which did not take into account career gaps related to caring for children or relatives, or only narrowly considered their professional experience.


One of Our Favorite Smart Plugs for Apple Users Is 15 Off

WIRED

The Meross smart plug mini boasts excellent compatibility and slim construction. On the hunt for new smart plugs to upgrade your home automation? One of our favorite picks, the Meross MSS110 Smart Plug Mini, is currently marked down on Amazon. The two pack is discounted from $34 to $27, and the four pack is down to $34 from its usual price of $52 . These plugs help add smart functionality to otherwise dumb devices around your home, like lamps or fans, so they can be included in your routines.


Rethinking AI's future in an augmented workplace

MIT Technology Review

By focusing on the economic opportunities and economic data, fears about AI investment can turn into smart business decisions. There are many paths AI evolution could take. On one end of the spectrum, AI is dismissed as a marginal fad, another bubble fueled by notoriety and misallocated capital. On the other end, it's cast as a dystopian force, destined to eliminate jobs on a large scale and destabilize economies. Markets oscillate between skepticism and the fear of missing out, while the technology itself evolves quickly and investment dollars flow at a rate not seen in decades. All the while, many of today's financial and economic thought leaders hold to the consensus that the financial landscape will stay the same as it has been for the last several years.


Philips Hue 'SpatialAware' feature harmonizes all the lights in a room

Engadget

The company also introduced Apple Home support and natural language automations at CES. Philips Hue has introduced a new software feature called SpatialAware at CES 2026 designed to ensure that all the lights in a space are in harmony with each other. Available exclusively for the Hue Bridge Pro, it takes into account each light point in a room and tailors the colors to ensure a natural representation. In a sunset scene, for example, the lights on one side of the room emit warm yellow tones to mimic the setting sun, while the ceiling lights on the opposite side reflect darker shades, the company wrote on its blog. The new feature is set to launch in the spring of 2026. To use the feature, you scan a room with your smartphone camera and use augmented reality to determine the positions of individual lights.


WONDERBREAD: A Benchmark for Evaluating Multimodal Foundation Models on Business Process Management Tasks

Neural Information Processing Systems

Existing ML benchmarks lack the depth and diversity of annotations needed for evaluating models on business process management (BPM) tasks. BPM is the practice of documenting, measuring, improving, and automating enterprise workflows. However, research has focused almost exclusively on one task -- full end-to-end automation using agents based on multimodal foundation models (FMs) like GPT-4. This focus on automation ignores the reality of how most BPM tools are applied today -- simply documenting the relevant workflow takes 60% of the time of the typical process optimization project. To address this gap we present WONDERBREAD, the first benchmark for evaluating multimodal FMs on BPM tasks beyond automation. Our contributions are: (1) a dataset containing 2928 documented workflow demonstrations; (2) 6 novel BPM tasks sourced from real-world applications ranging from workflow documentation to knowledge transfer to process improvement; and (3) an automated evaluation harness. Our benchmark shows that while state-of-the-art FMs can automatically generate documentation (e.g.


Auditing for Human Expertise

Neural Information Processing Systems

High-stakes prediction tasks (e.g., patient diagnosis) are often handled by trained human experts. A common source of concern about automation in these settings is that experts may exercise intuition that is difficult to model and/or have access to information (e.g., conversations with a patient) that is simply unavailable to a would-be algorithm. This raises a natural question whether human experts add value which could not be captured by an algorithmic predictor.We develop a statistical framework under which we can pose this question as a natural hypothesis test. Indeed, as our framework highlights, detecting human expertise is more subtle than simply comparing the accuracy of expert predictions to those made by a particular learning algorithm. Instead, we propose a simple procedure which tests whether expert predictions are statistically independent from the outcomes of interest after conditioning on the available inputs ('features'). A rejection of our test thus suggests that human experts may add value to any algorithm trained on the available data, and has direct implications for whether human-AI'complementarity' is achievable in a given prediction task.We highlight the utility of our procedure using admissions data collected from the emergency department of a large academic hospital system, where we show that physicians' admit/discharge decisions for patients with acute gastrointestinal bleeding (AGIB) appear to be incorporating information that is not available to a standard algorithmic screening tool. This is despite the fact that the screening tool is arguably more accurate than physicians' discretionary decisions, highlighting that - even absent normative concerns about accountability or interpretability - accuracy is insufficient to justify algorithmic automation.



A Zero-Positive Learning Approach for Diagnosing Software Performance Regressions

Neural Information Processing Systems

The field of machine programming (MP), the automation of the development of software, is making notable research advances. This is, in part, due to the emergence of a wide range of novel techniques in machine learning. In this paper, we apply MP to the automation of software performance regression testing. A performance regression is a software performance degradation caused by a code change.


Wyze Rule: Federated Rule Dataset for Rule Recommendation Benchmarking

Neural Information Processing Systems

In the rapidly evolving landscape of smart home automation, the potential of IoT devices is vast. In this realm, rules are the main tool utilized for this automation, which are predefined conditions or triggers that establish connections between devices, enabling seamless automation of specific processes. However, one significant challenge researchers face is the lack of comprehensive datasets to explore and advance the field of smart home rule recommendations. These datasets are essential for developing and evaluating intelligent algorithms that can effectively recommend rules for automating processes while preserving the privacy of the users, as it involves personal information about users' daily lives. To bridge this gap, we present the Wyze Rule Dataset, a large-scale dataset designed specifically for smart home rule recommendation research. Wyze Rule encompasses over 1 million rules gathered from a diverse user base of 300,000 individuals from Wyze Labs, offering an extensive and varied collection of real-world data. With a focus on federated learning, our dataset is tailored to address the unique challenges of a cross-device federated learning setting in the recommendation domain, featuring a large-scale number of clients with widely heterogeneous data. To establish a benchmark for comparison and evaluation, we have meticulously implemented multiple baselines in both centralized and federated settings. Researchers can leverage these baselines to gauge the performance and effectiveness of their rule recommendation systems, driving advancements in the domain.


Joint Activity Design Heuristics for Enhancing Human-Machine Collaboration

Jalaeian, Mohammadreza, Morey, Dane A., Rayo, Michael F.

arXiv.org Artificial Intelligence

-- Joint activity describes when more than one agent (human or machine) contributes to the completion of a task or activity. Designing for joint activity focuses on explicitly supporting the interdependencies between agents necessary for effective coordination amon g agents engaged in the joint activity. This builds and expands upon designing for usability to further address how technologies can be designed to act as effective team players. Effective joint activity requires supporting, at minimum, five primary macroc ognitive functions within teams: Event Detection, Sensemaking, Adaptability, Perspective - Shifting, and Coordination. Supporting these functions is equally as important as making technologies usable. We synthesized fourteen heuristics from relevant literatu re including display design, human factors, cognitive systems engineering, cognitive psychology, and computer science to aid the design, development, and evaluation of technologies that support joint human - machine activity . Recent advances in Artificial Intelligence (AI) and Machine Learning (ML) technologies have accelerated human - machine interactions progress ing from simple tool - based engagements to complex cognitive collaborations [1] . Machines are being designed to perform an increasing set of functions and are being expected to engage more deeply in the collaborative joint activit ies related to these functions. This shift in machine capabilities and expectations demands a corresponding re - evaluation and broadening of design and evaluation principles to support joint human - machine activity in ways that lie outside the boundaries of trad itional usability methods and models [2] . Traditional usability heuristics, such as those proposed by [3], provide a strong foundation focusing primarily on surface - level interactions such as enhancing the ease of use, efficiency, and satisfaction in human - machine interaction . These heuristics are primarily oriented towards actions and responses but offer limited support for the essential macrocognitive functions associated with effective teamwork including event detection, sensemaking, adaptability, perspective shifting, and co ordination, all of which are vital in the close collaboration of humans and machine s with joint activities [2], [4], [5], [6] . These heuristics are primarily oriented towards actions and responses but offer limited support for the essential macrocognitive functions associated with effective teamwork including event detection, sensemaking, adaptability, perspective shifting, and co ordination . A ll of these macrocognitive functions are vital in the close collaboration of humans and machines with joint activities in high - stakes and dynamic environments with little room for error [2], [5] . This reliance on macrocognitive functions is evident in domains where the ability to process complex information and adapt to changing conditions is crucial.