Goto

Collaborating Authors

 laboratory


Mystery as Communion bread and wine 'miraculously' appear to turn into human tissue and blood

Daily Mail - Science & tech

Trump says he's'not afraid' of Vietnam-style ground combat in Iran Furious US troops erupt at CNN's $20m steak and lobster claims as grim photos expose reality Hollywood's top insider makes VERY catty observation about Kaitlan Collins Pam Bondi is formally subpoenaed by Congress as Trump's Epstein nightmare grows What the Jane Plan did to my body: The unfashionable retro diet's fans say it's life-changing, easy, better than fat jabs - and shifts weight fast. My husband tried a'cure' for his ALS... days later he went blind and couldn't move. The children screamed on video call as he died. Outrage after Pete Hegseth aide ousted for'leaks' lands new top secret intelligence job Everything JFK Jr told friends about his love affair with'sexual dynamo' Madonna... her unprintable pillow talk... and his perverse incest request that she couldn't go through with SARAH VINE: How telling that Meghan's joined the ranks of those peddling wellness and fake lifestyles to the gullible My chilling conversations with the Unabomber and America's worst serial killers when I ran a Supermax prison, revealed in The Crime Desk newsletter Oscars afterparty snitches reveal cringing details of how stars stopped talking to him... a brutal message from Kylie's gloating ex... and her'humiliating' admission to friends Joe Burrow cements his place as the NFL's most eligible bachelor as he is spotted cozying up to Tate McRae and Alix Earle at glitzy Oscars afterparty Dark secret past of husband killer Kouri Richins' Iraq war veteran lover revealed... and their toe-curling sex texts that helped convict her Mystery as Communion bread and wine'miraculously' appear to turn into human tissue and blood READ MORE: Scientists stunned as 500-year-old'miracle' image of Virgin Mary reveals impossible microscopic reflection Catholics believe that during Communion, bread and wine become the body and blood of Jesus Christ, though they continue to appear unchanged to the human eye. But there have been a handful of rare and debated cases in which the sacred elements appeared to take on a far more literal, physical form.


All major AI models risk encouraging dangerous science experiments

New Scientist

Researchers risk fire, explosion or poisoning by allowing AI to design experiments, warn scientists. The use of AI models in scientific laboratories risks enabling dangerous experiments that could cause fires or explosions, researchers have warned. Such models offer a convincing illusion of understanding but are susceptible to missing basic and vital safety precautions. In tests of 19 cutting-edge AI models, every single one made potentially deadly mistakes. Serious accidents in university labs are rare but certainly not unheard of.


MIT engineers design an aerial microrobot that can fly as fast as a bumblebee

Robohub

In the future, tiny flying robots could be deployed to aid in the search for survivors trapped beneath the rubble after a devastating earthquake. So far, aerial microrobots have only been able to fly slowly along smooth trajectories, far from the swift, agile flight of real insects -- until now. MIT researchers have demonstrated aerial microrobots that can fly with speed and agility that is comparable to their biological counterparts. A collaborative team designed a new AI-based controller for the robotic bug that enabled it to follow gymnastic flight paths, such as executing continuous body flips. With a two-part control scheme that combines high performance with computational efficiency, the robot's speed and acceleration increased by about 450 percent and 250 percent, respectively, compared to the researchers' best previous demonstrations.


Tracr: Compiled Transformers as a Laboratory for Interpretability

Neural Information Processing Systems

We show how to compile human-readable programs into standard decoder-only transformer models. Our compiler, Tracr, generates models with known structure. This structure can be used to design experiments. For example, we use it to study superposition in transformers that execute multi-step algorithms. Additionally, the known structure of Tracr-compiled models can serve as for evaluating interpretability methods. Commonly, because the programs learned by transformers are unknown it is unclear whether an interpretation succeeded. We demonstrate our approach by implementing and examining programs including computing token frequencies, sorting, and parenthesis checking.


Closed-Loop Robotic Manipulation of Transparent Substrates for Self-Driving Laboratories using Deep Learning Micro-Error Correction

Fontenot, Kelsey, Gorti, Anjali, Goel, Iva, Buonassisi, Tonio, Siemenn, Alexander E.

arXiv.org Artificial Intelligence

Self-driving laboratories (SDLs) have accelerated the throughput and automation capabilities for discovering and improving chemistries and materials. Although these SDLs have automated many of the steps required to conduct chemical and materials experiments, a commonly overlooked step in the automation pipeline is the handling and reloading of substrates used to transfer or deposit materials onto for downstream characterization. Here, we develop a closed-loop method of Automated Substrate Handling and Exchange (ASHE) using robotics, dual-actuated dispensers, and deep learning-driven computer vision to detect and correct errors in the manipulation of fragile and transparent substrates for SDLs. Using ASHE, we demonstrate a 98.5% first-time placement accuracy across 130 independent trials of reloading transparent glass substrates into an SDL, where only two substrate misplacements occurred and were successfully detected as errors and automatically corrected. Through the development of more accurate and reliable methods for handling various types of substrates, we move toward an improvement in the automation capabilities of self-driving laboratories, furthering the acceleration of novel chemical and materials discoveries.


LabOS: The AI-XR Co-Scientist That Sees and Works With Humans

Cong, Le, Smerkous, David, Wang, Xiaotong, Yin, Di, Zhang, Zaixi, Jin, Ruofan, Wang, Yinkai, Gerasimiuk, Michal, Dinesh, Ravi K., Smerkous, Alex, Shi, Lihan, Zheng, Joy, Lam, Ian, Wu, Xuekun, Liu, Shilong, Li, Peishan, Zhu, Yi, Zhao, Ning, Parakh, Meenal, Serrao, Simran, Mohammad, Imran A., Chen, Chao-Yeh, Xie, Xiufeng, Chen, Tiffany, Weinstein, David, Barbone, Greg, Caglar, Belgin, Sunwoo, John B., Li, Fuxin, Deng, Jia, Wu, Joseph C., Wu, Sanfeng, Wang, Mengdi

arXiv.org Artificial Intelligence

Modern science advances fastest when thought meets action. LabOS represents the first AI co-scientist that unites computational reasoning with physical experimentation through multimodal perception, self-evolving agents, and Extended-Reality(XR)-enabled human-AI collaboration. By connecting multi-model AI agents, smart glasses, and robots, LabOS allows AI to see what scientists see, understand experimental context, and assist in real-time execution. Across applications -- from cancer immunotherapy target discovery to stem-cell engineering and material science -- LabOS shows that AI can move beyond computational design to participation, turning the laboratory into an intelligent, collaborative environment where human and machine discovery evolve together.


Data-Centric Visual Development for Self-Driving Labs

Liu, Anbang, Hu, Guanzhong, Wang, Jiayi, Guo, Ping, Liu, Han

arXiv.org Artificial Intelligence

Self-driving laboratories offer a promising path toward reducing the labor-intensive, time-consuming, and often irreproducible workflows in the biological sciences. Yet their stringent precision requirements demand highly robust models whose training relies on large amounts of annotated data. However, this kind of data is difficult to obtain in routine practice, especially negative samples. In this work, we focus on pipetting, the most critical and precision sensitive action in SDLs. To overcome the scarcity of training data, we build a hybrid pipeline that fuses real and virtual data generation. The real track adopts a human-in-the-loop scheme that couples automated acquisition with selective human verification to maximize accuracy with minimal effort. The virtual track augments the real data using reference-conditioned, prompt-guided image generation, which is further screened and validated for reliability. Together, these two tracks yield a class-balanced dataset that enables robust bubble detection training. On a held-out real test set, a model trained entirely on automatically acquired real images reaches 99.6% accuracy, and mixing real and generated data during training sustains 99.4% accuracy while reducing collection and review load. Our approach offers a scalable and cost-effective strategy for supplying visual feedback data to SDL workflows and provides a practical solution to data scarcity in rare event detection and broader vision tasks.


AI-Driven Robotics for Optics

Uddin, Shiekh Zia, Vaidya, Sachin, Choudhary, Shrish, Chen, Zhuo, Salib, Raafat K., Huang, Luke, Englund, Dirk R., Soljačić, Marin

arXiv.org Artificial Intelligence

Optics is foundational to research in many areas of science and engineering, including nanophotonics, quantum information, materials science, biomedical imaging, and metrology. However, the design, assembly, and alignment of optical experiments remain predominantly manual, limiting throughput and reproducibility. Automating such experiments is challenging due to the strict, non-negotiable precision requirements and the diversity of optical configurations found in typical laboratories. Here, we introduce a platform that integrates generative artificial intelligence, computer vision, and robotics to automate free-space optical experiments. The platform translates user-defined goals into valid optical configurations, assembles them using a robotic arm, and performs micrometer-scale fine alignment using a robot-deployable tool. It then executes a range of automated measurements, including beam characterization, polarization mapping, and spectroscopy, with consistency surpassing that of human operators. This work demonstrates the first flexible, AI-driven automation platform for optics, offering a path towards remote operation, cloud labs, and high-throughput discovery in the optical sciences.


PREVENT: Proactive Risk Evaluation and Vigilant Execution of Tasks for Mobile Robotic Chemists using Multi-Modal Behavior Trees

Veeramani, Satheeshkumar, Zhou, Zhengxue, Munguia-Galeano, Francisco, Fakhruldeen, Hatem, Roddelkopf, Thomas, Al-Okby, Mohammed Faeik Ruzaij, Thurow, Kerstin, Cooper, Andrew Ian

arXiv.org Artificial Intelligence

Mobile robotic chemists are a fast growing trend in the field of chemistry and materials research. However, so far these mobile robots lack workflow awareness skills. This poses the risk that even a small anomaly, such as an improperly capped sample vial could disrupt the entire workflow. This wastes time, and resources, and could pose risks to human researchers, such as exposure to toxic materials. Existing perception mechanisms can be used to predict anomalies but they often generate excessive false positives. This may halt workflow execution unnecessarily, requiring researchers to intervene and to resume the workflow when no problem actually exists, negating the benefits of autonomous operation. To address this problem, we propose PREVENT a system comprising navigation and manipulation skills based on a multimodal Behavior Tree (BT) approach that can be integrated into existing software architectures with minimal modifications. Our approach involves a hierarchical perception mechanism that exploits AI techniques and sensory feedback through Dexterous Vision and Navigational Vision cameras and an IoT gas sensor module for execution-related decision-making. Experimental evaluations show that the proposed approach is comparatively efficient and completely avoids both false negatives and false positives when tested in simulated risk scenarios within our robotic chemistry workflow. The results also show that the proposed multi-modal perception skills achieved deployment accuracies that were higher than the average of the corresponding uni-modal skills, both for navigation and for manipulation.


Kinematic Analysis and Integration of Vision Algorithms for a Mobile Manipulator Employed Inside a Self-Driving Laboratory

Sulaiman, Shifa, Jensen, Tobias Busk, Bengtson, Stefan Hein, Bøgh, Simon

arXiv.org Artificial Intelligence

Recent advances in robotics and autonomous systems have broadened the use of robots in laboratory settings, including automated synthesis, scalable reaction workflows, and collaborative tasks in self-driving laboratories (SDLs). This paper presents a comprehensive development of a mobile manipulator designed to assist human operators in such autonomous lab environments. Kinematic modeling of the manipulator is carried out based on the Denavit Hartenberg (DH) convention and inverse kinematics solution is determined to enable precise and adaptive manipulation capabilities. A key focus of this research is enhancing the manipulator ability to reliably grasp textured objects as a critical component of autonomous handling tasks. Advanced vision-based algorithms are implemented to perform real-time object detection and pose estimation, guiding the manipulator in dynamic grasping and following tasks. In this work, we integrate a vision method that combines feature-based detection with homography-driven pose estimation, leveraging depth information to represent an object pose as a $2$D planar projection within $3$D space. This adaptive capability enables the system to accommodate variations in object orientation and supports robust autonomous manipulation across diverse environments. By enabling autonomous experimentation and human-robot collaboration, this work contributes to the scalability and reproducibility of next-generation chemical laboratories