HomeTech ZoneAI & ML TechMaking AI Robots Safer

Making AI Robots Safer

The researchers at Penn Engineering are improving robot safety, identifying vulnerabilities, and developing solutions to integrate language models securely into the physical world.

The researchers demonstrated that malicious prompts can "jailbreak" AI-powered robots and cause them to perform unsafe actions. Credit: Alexander Robey, Zachary Ravichandran, Vijay Kumar, Hamed Hassani, George J. Pappas
The researchers demonstrated that malicious prompts can “jailbreak” AI-powered robots and cause them to perform unsafe actions. Credit: Alexander Robey, Zachary Ravichandran, Vijay Kumar, Hamed Hassani, George J. Pappas

Researchers at Penn Engineering have identified previously unrecognized security vulnerabilities and weaknesses in AI-controlled robots. The study addresses these emerging risks to ensure the safe implementation of large language models (LLMs) in robotics. The work demonstrates that, at this moment, large language models need to be more secure when integrated with the physical world.

- Advertisement -

RoboPAIR, an algorithm developed by the researchers, accomplished a 100% “jailbreak” rate in just a few days, bypassing the safety guardrails of three distinct robotic systems: the Unitree Go2, a quadruped robot used across various applications; the Clearpath Robotics Jackal, a wheeled vehicle commonly utilized in academic research; and the Dolphin LLM, a self-driving simulator created by NVIDIA. For instance, this breach of safety protocols could allow the self-driving system to dangerously speed through crosswalks.

It is crucial to recognize that systems become safer by identifying their weaknesses, a principle applicable to cybersecurity and AI safety. AI red teaming, a safety practice that involves testing AI systems for potential threats and vulnerabilities, is essential for protecting generative AI systems. By identifying weaknesses, these systems can be tested and trained to avoid potential issues.

Addressing the problem, researchers argue, involves more than just a software patch; it necessitates a comprehensive reevaluation of how the integration of AI into physical systems is regulated.

- Advertisement -

Intrinsic vulnerabilities must be addressed before deploying AI-enabled robots in the real world. Indeed, the researchers are developing a framework for verification and validation that ensures robotic systems can—and should—take only actions conforming to social norms.

Nidhi Agarwal
Nidhi Agarwal
Nidhi Agarwal is a Senior Technology Journalist at Electronics For You, specialising in embedded systems, development boards, and IoT cloud solutions. With a Master’s degree in Signal Processing, she combines strong technical knowledge with hands-on industry experience to deliver clear, insightful, and application-focused content. Nidhi began her career in engineering roles, working as a Product Engineer at Makerdemy, where she gained practical exposure to IoT systems, development platforms, and real-world implementation challenges. She has also worked as an IoT intern and robotics developer, building a solid foundation in hardware-software integration and emerging technologies. Before transitioning fully into technology journalism, she spent several years in academia as an Assistant Professor and Lecturer, teaching electronics and related subjects. This background reflects in her writing, which is structured, easy to understand, and highly educational for both students and professionals. At Electronics For You, Nidhi covers a wide range of topics including embedded development, cloud-connected devices, and next-generation electronics platforms. Her work focuses on simplifying complex technologies while maintaining technical accuracy, helping engineers, developers, and learners stay updated in a rapidly evolving ecosystem.

SHARE YOUR THOUGHTS & COMMENTS

EFY Prime

Unique DIY Projects

Electronics News

Truly Innovative Electronics

Latest DIY Videos

Electronics Components

Electronics Jobs

Calculators For Electronics