Goto

Collaborating Authors

 shield





Starstruck

MIT Technology Review

Aomawa Shields '97 was equally enticed by the prospect of studying stars and the dream of becoming one herself. Today, she draws from her exploration of acting and astronomy to search for life on other planets. Few people, if any, contemplate stars--celestial or cinematic--the way Aomawa Shields does. An astronomer and astrobiologist, Shields explores the potential habitability of planets beyond our solar system. But she is also a classically trained actor--and that's helped shape her professional trajectory in unexpected ways. Today, Shields is an associate professor in the Department of Physics and Astronomy at the University of California, Irvine, where she oversees a research team that uses computer models to explore conditions on exoplanets, or planets that revolve around stars other than the sun.


Made in space? Start-up brings factory in orbit one step closer to reality

BBC News

It sounds like science fiction - a factory, located hundreds of kilometres above the Earth, churning out high-quality materials. But a Cardiff-based company is a step closer to making this a reality. Space Forge have sent a microwave-sized factory into orbit, and have demonstrated that its furnace can be switched on and reach temperatures of around 1,000C. They plan to manufacture material for semiconductors, which can be used back on Earth in electronics in communications infrastructure, computing and transport. Conditions in space are ideal for making semiconductors, which have the atoms they're made of arranged in a highly ordered 3D structure.


Bubble wrap-like material could help insulate glass windows

Popular Science

Only five millimeters of this experimental material called MOCHI can shield your hand from a flame. Breakthroughs, discoveries, and DIY tips sent every weekday. A well-placed window can brighten a room with natural light and offer scenic views of the outside world. Buildings consume around 40 percent of society's energy production, and much of that energy is wasted due to poor insulation in the winter and too much heat retention during the summer. Even the most eco-friendly windows inevitably add to this energy drain.


ContactRL: Safe Reinforcement Learning based Motion Planning for Contact based Human Robot Collaboration

Mulkana, Sundas Rafat, Yu, Ronyu, Guha, Tanaya, Li, Emma

arXiv.org Artificial Intelligence

Abstract-- In collaborative human-robot tasks, safety requires not only avoiding collisions but also ensuring safe, intentional physical contact. We present ContactRL, a reinforcement learning (RL) based framework that directly incorporates contact safety into the reward function through force feedback. This enables a robot to learn adaptive motion profiles that minimize human-robot contact forces while maintaining task efficiency. In simulation, ContactRL achieves a low safety violation rate of 0.2% with a high task success rate of 87.7%, outperforming state-of-the-art constrained RL baselines. In order to guarantee deployment safety, we augment the learned policy with a kinetic energy based Control Barrier Function (eCBF) shield. Real-world experiments on an UR3e robotic platform performing small object handovers from a human hand across 360 trials confirm safe contact, with measured normal forces consistently below 10N. These results demonstrate that ContactRL enables safe and efficient physical collaboration, thereby advancing the deployment of collaborative robots in contact-rich tasks.


Shielded Controller Units for RL with Operational Constraints Applied to Remote Microgrids

Nekoei, Hadi, Massé, Alexandre Blondin, Hassani, Rachid, Chandar, Sarath, Mai, Vincent

arXiv.org Artificial Intelligence

Reinforcement learning (RL) is a powerful framework for optimizing decision-making in complex systems under uncertainty, an essential challenge in real-world settings, particularly in the context of the energy transition. A representative example is remote microgrids that supply power to communities disconnected from the main grid. Enabling the energy transition in such systems requires coordinated control of renewable sources like wind turbines, alongside fuel generators and batteries, to meet demand while minimizing fuel consumption and battery degradation under exogenous and intermittent load and wind conditions. These systems must often conform to extensive regulations and complex operational constraints. To ensure that RL agents respect these constraints, it is crucial to provide interpretable guarantees. In this paper, we introduce Shielded Controller Units (SCUs), a systematic and interpretable approach that leverages prior knowledge of system dynamics to ensure constraint satisfaction. Our shield synthesis methodology, designed for real-world deployment, decomposes the environment into a hierarchical structure where each SCU explicitly manages a subset of constraints. We demonstrate the effectiveness of SCUs on a remote microgrid optimization task with strict operational requirements. The RL agent, equipped with SCUs, achieves a 24% reduction in fuel consumption without increasing battery degradation, outperforming other baselines while satisfying all constraints. We hope SCUs contribute to the safe application of RL to the many decision-making challenges linked to the energy transition.


SHIELD: Secure Hypernetworks for Incremental Expansion Learning Defense

Krukowski, Patryk, Gorczyca, Łukasz, Helm, Piotr, Książek, Kamil, Spurek, Przemysław

arXiv.org Artificial Intelligence

Continual learning under adversarial conditions remains an open problem, as existing methods often compromise either robustness, scalability, or both. We propose a novel framework that integrates Interval Bound Propagation (IBP) with a hypernetwork-based architecture to enable certifiably robust continual learning across sequential tasks. Our method, SHIELD, generates task-specific model parameters via a shared hypernetwork conditioned solely on compact task embeddings, eliminating the need for replay buffers or full model copies and enabling efficient over time. To further enhance robustness, we introduce Interval MixUp, a novel training strategy that blends virtual examples represented as $\ell_{\infty}$ balls centered around MixUp points. Leveraging interval arithmetic, this technique guarantees certified robustness while mitigating the wrapping effect, resulting in smoother decision boundaries. We evaluate SHIELD under strong white-box adversarial attacks, including PGD and AutoAttack, across multiple benchmarks. It consistently outperforms existing robust continual learning methods, achieving state-of-the-art average accuracy while maintaining both scalability and certification. These results represent a significant step toward practical and theoretically grounded continual learning in adversarial settings.


Adaptive GR(1) Specification Repair for Liveness-Preserving Shielding in Reinforcement Learning

Georgescu, Tiberiu-Andrei, Goodall, Alexander W., Alrajeh, Dalal, Belardinelli, Francesco, Uchitel, Sebastian

arXiv.org Artificial Intelligence

Shielding is widely used to enforce safety in reinforcement learning (RL), ensuring that an agent's actions remain compliant with formal specifications. Classical shielding approaches, however, are often static, in the sense that they assume fixed logical specifications and hand-crafted abstractions. While these static shields provide safety under nominal assumptions, they fail to adapt when environment assumptions are violated. In this paper, we develop the first adaptive shielding framework - to the best of our knowledge - based on Generalized Reactivity of rank 1 (GR(1)) specifications, a tractable and expressive fragment of Linear Temporal Logic (LTL) that captures both safety and liveness properties. Our method detects environment assumption violations at runtime and employs Inductive Logic Programming (ILP) to automatically repair GR(1) specifications online, in a systematic and interpretable way. This ensures that the shield evolves gracefully, ensuring liveness is achievable and weakening goals only when necessary. We consider two case studies: Minepump and Atari Seaquest; showing that (i) static symbolic controllers are often severely suboptimal when optimizing for auxiliary rewards, and (ii) RL agents equipped with our adaptive shield maintain near-optimal reward and perfect logical compliance compared with static shields.