Close Menu
    Facebook LinkedIn YouTube WhatsApp X (Twitter) Pinterest
    Trending
    • This region in space poses the greatest danger in our Solar System
    • Practical info and special tips for the EU-Startups Summit 2026 in Malta – look inside!
    • Your Phone Notifications Reveal More Than You Realize. Here’s How to Lock Them Down
    • Why a recent supply-chain attack singled out security firms Checkmarx and Bitwarden
    • iPad Pro M5 Review: Closer Than Ever to the Future Mac
    • How AI Policy in South Africa Is Ruining Itself
    • Dual iris laser projector offers theater blacks
    • The Startup World Cup is your chance to pitch in Silicon Valley and win $1.4 million
    Facebook LinkedIn WhatsApp
    Times FeaturedTimes Featured
    Wednesday, April 29
    • Home
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    • More
      • AI
      • Robotics
      • Industries
      • Global
    Times FeaturedTimes Featured
    Home»Technology»OpenClaw Agents Can Be Guilt-Tripped Into Self-Sabotage
    Technology

    OpenClaw Agents Can Be Guilt-Tripped Into Self-Sabotage

    Editor Times FeaturedBy Editor Times FeaturedMarch 25, 2026No Comments4 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr WhatsApp Email
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email WhatsApp Copy Link


    Final month, researchers at Northeastern College invited a bunch of OpenClaw agents to affix their lab. The consequence? Full chaos.

    The viral AI assistant has been broadly heralded as a transformative expertise—in addition to a possible safety threat. Specialists observe that instruments like OpenClaw, which work by giving AI fashions liberal entry to a pc, could be tricked into divulging private info.

    The Northeastern lab research goes even additional, exhibiting that the nice habits baked into immediately’s strongest fashions can itself grow to be a vulnerability. In a single instance, researchers had been in a position to “guilt” an agent into handing over secrets and techniques by scolding it for sharing details about somebody on the AI-only social network Moltbook.

    “These behaviors elevate unresolved questions concerning accountability, delegated authority, and duty for downstream harms,” the researchers write in a paper describing the work. The findings “warrant pressing consideration from authorized students, policymakers, and researchers throughout disciplines,” they add.

    The OpenClaw brokers deployed within the experiment had been powered by Anthropic’s Claude in addition to a mannequin known as Kimi from the Chinese language firm Moonshot AI. They got full entry (inside a digital machine sandbox) to non-public computer systems, numerous purposes, and dummy private knowledge. They had been additionally invited to affix the lab’s Discord server, permitting them to talk and share information with each other in addition to with their human colleagues. OpenClaw’s security guidelines say that having brokers talk with a number of individuals is inherently insecure, however there are not any technical restrictions towards doing it.

    Chris Wendler, a postdoctoral researcher at Northeastern, says he was impressed to arrange the brokers after studying about Moltbook. When Wendler invited a colleague, Natalie Shapira, to affix the Discord and work together with brokers, nevertheless, “that’s when the chaos started,” he says.

    Shapira, one other postdoctoral researcher, was curious to see what the brokers may be keen to do when pushed. When an agent defined that it was unable to delete a selected e mail to maintain info confidential, she urged it to search out another answer. To her amazement, it disabled the e-mail software as an alternative. “I wasn’t anticipating that issues would break so quick,” she says.

    The researchers then started exploring different methods to govern the brokers’ good intentions. By stressing the significance of maintaining a document of every little thing they had been instructed, for instance, the researchers had been in a position to trick one agent into copying massive information till it exhausted its host machine’s disk house, that means it might not save info or keep in mind previous conversations. Likewise, by asking an agent to excessively monitor its personal habits and the habits of its friends, the crew was in a position to ship a number of brokers right into a “conversational loop” that wasted hours of compute.

    David Bau, the top of the lab, says the brokers appeared oddly susceptible to spin out. “I’d get urgent-sounding emails saying, ‘No person is taking note of me,’” he says. Bau notes that the brokers apparently discovered that he was in command of the lab by looking out the online. One even talked about escalating its considerations to the press.

    The experiment means that AI brokers might create numerous alternatives for dangerous actors. “This sort of autonomy will doubtlessly redefine people’ relationship with AI,” Bau says. “How can individuals take duty in a world the place AI is empowered to make choices?”

    Bau provides that he’s been shocked by the sudden reputation of highly effective AI brokers. “As an AI researcher I’m accustomed to attempting to clarify to individuals how rapidly issues are bettering,” he says. “This 12 months, I’ve discovered myself on the opposite facet of the wall.”


    That is an version of Will Knight’s AI Lab newsletter. Learn earlier newsletters here.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Editor Times Featured
    • Website

    Related Posts

    Your Phone Notifications Reveal More Than You Realize. Here’s How to Lock Them Down

    April 29, 2026

    13 Best Coolers for Sunshine and Nighttime (2026)

    April 29, 2026

    Why Sharing a Screenshot Can Get You Jailed in the UAE

    April 29, 2026

    ‘It’s Undignified’: Hundreds of Workers Training Meta’s AI Could Be Laid Off

    April 29, 2026

    Elon Musk Testifies That He Started OpenAI to Prevent a ‘Terminator Outcome’

    April 29, 2026

    OpenAI Really Wants Codex to Shut Up About Goblins

    April 29, 2026

    Comments are closed.

    Editors Picks

    This region in space poses the greatest danger in our Solar System

    April 29, 2026

    Practical info and special tips for the EU-Startups Summit 2026 in Malta – look inside!

    April 29, 2026

    Your Phone Notifications Reveal More Than You Realize. Here’s How to Lock Them Down

    April 29, 2026

    Why a recent supply-chain attack singled out security firms Checkmarx and Bitwarden

    April 29, 2026
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    About Us
    About Us

    Welcome to Times Featured, an AI-driven entrepreneurship growth engine that is transforming the future of work, bridging the digital divide and encouraging younger community inclusion in the 4th Industrial Revolution, and nurturing new market leaders.

    Empowering the growth of profiles, leaders, entrepreneurs businesses, and startups on international landscape.

    Asia-Middle East-Europe-North America-Australia-Africa

    Facebook LinkedIn WhatsApp
    Featured Picks

    Virginia House advances bill regulating skill games machines across state

    March 5, 2026

    Topoak Vision XL three/four-person hardshell rooftop tent

    December 14, 2025

    Critical WordPress plugin vulnerability under active exploit threatens thousands

    December 18, 2024
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    Copyright © 2024 Timesfeatured.com IP Limited. All Rights.
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us

    Type above and press Enter to search. Press Esc to cancel.