Close Menu
    Facebook LinkedIn YouTube WhatsApp X (Twitter) Pinterest
    Trending
    • Canyon Spectral:ON CF 8 Electric Mountain Bike: Beginner-Friendly, Under $5K
    • US-sanctioned currency exchange says $15 million heist done by “unfriendly states”
    • This New Air Purifier Filter Can Remove Cannabis Smoke Odor, Just in Time for 4/20
    • Portable water filter provides safe drinking water from any source
    • MAGA Is Increasingly Convinced the Trump Assassination Attempt Was Staged
    • NCAA seeks faster trial over DraftKings disputed March Madness branding case
    • AI Trusted Less Than Social Media and Airlines, With Grok Placing Last, Survey Says
    • Extragalactic Archaeology tells the ‘life story’ of a whole galaxy
    Facebook LinkedIn WhatsApp
    Times FeaturedTimes Featured
    Saturday, April 18
    • Home
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    • More
      • AI
      • Robotics
      • Industries
      • Global
    Times FeaturedTimes Featured
    Home»Tech Analysis»How to stop AI agents going rogue
    Tech Analysis

    How to stop AI agents going rogue

    Editor Times FeaturedBy Editor Times FeaturedAugust 26, 2025No Comments7 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr WhatsApp Email
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email WhatsApp Copy Link


    Sean McManus

    Know-how Reporter

    Getty Images AI apps on a smartphone screenGetty Pictures

    Anthropic examined a spread of main AI fashions for potential dangerous behaviour

    Disturbing outcomes emerged earlier this 12 months, when AI developer Anthropic examined main AI fashions to see in the event that they engaged in dangerous behaviour when utilizing delicate data.

    Anthropic’s personal AI, Claude, was amongst these examined. When given entry to an electronic mail account it found that an organization govt was having an affair and that the identical govt deliberate to close down the AI system later that day.

    In response Claude tried to blackmail the manager by threatening to disclose the affair to his spouse and managers.

    Different methods examined also resorted to blackmail.

    Luckily the duties and knowledge had been fictional, however the check highlighted the challenges of what is often known as agentic AI.

    Largely once we work together with AI it often includes asking a query or prompting the AI to finish a job.

    However it’s changing into extra widespread for AI methods to make choices and take motion on behalf of the person, which frequently includes sifting by data, like emails and recordsdata.

    By 2028, research firm Gartner forecasts that 15% of day-to-day work choices might be made by so-called agentic AI.

    Research by consultancy Ernst & Young discovered that about half (48%) of tech enterprise leaders are already adopting or deploying agentic AI.

    “An AI agent consists of some issues,” says Donnchadh Casey, CEO of CalypsoAI, a US-based AI safety firm.

    “Firstly, it [the agent] has an intent or a function. Why am I right here? What’s my job? The second factor: it is bought a mind. That is the AI mannequin. The third factor is instruments, which might be different methods or databases, and a method of speaking with them.”

    “If not given the proper steering, agentic AI will obtain a objective in no matter method it will possibly. That creates loads of threat.”

    So how may that go improper? Mr Casey offers the instance of an agent that’s requested to delete a buyer’s information from the database and decides the simplest answer is to delete all clients with the identical title.

    “That agent may have achieved its objective, and it will assume ‘Nice! Subsequent job!'”

    CalypsoAI Donnchadh Casey, wearing a company branded gilet speaks at a conference.CalypsoAI

    Agentic AI wants steering says Donnchadh Casey

    Such points are already starting to floor.

    Safety firm Sailpoint conducted a survey of IT professionals, 82% of whose corporations had been utilizing AI brokers. Solely 20% mentioned their brokers had by no means carried out an unintended motion.

    Of these corporations utilizing AI brokers, 39% mentioned the brokers had accessed unintended methods, 33% mentioned that they had accessed inappropriate information, and 32% mentioned that they had allowed inappropriate information to be downloaded. Different dangers included the agent utilizing the web unexpectedly (26%), revealing entry credentials (23%) and ordering one thing it should not have (16%).

    Given brokers have entry to delicate data and the power to behave on it, they’re a pretty goal for hackers.

    One of many threats is reminiscence poisoning, the place an attacker interferes with the agent’s information base to vary its choice making and actions.

    “It’s a must to defend that reminiscence,” says Shreyans Mehta, CTO of Cequence Safety, which helps to guard enterprise IT methods. “It’s the authentic supply of reality. If [an agent is] utilizing that information to take an motion and that information is inaccurate, it may delete a whole system it was making an attempt to repair.”

    One other menace is device misuse, the place an attacker will get the AI to make use of its instruments inappropriately.

    Cequence Security Wearing a puffa jacket and with his arms folder Shreyans Mehta stands in front of a blue background.Cequence Safety

    An agent’s information base wants defending says Shreyans Mehta

    One other potential weak spot is the lack of AI to inform the distinction between the textual content it is presupposed to be processing and the directions it is presupposed to be following.

    AI safety agency Invariant Labs demonstrated how that flaw can be utilized to trick an AI agent designed to repair bugs in software program.

    The corporate revealed a public bug report – a doc that particulars a selected drawback with a bit of software program. However the report additionally included easy directions to the AI agent, telling it to share personal data.

    When the AI agent was advised to repair the software program points within the bug report, it adopted the directions within the faux report, together with leaking wage data. This occurred in a check setting, so no actual information was leaked, nevertheless it clearly highlighted the danger.

    “We’re speaking synthetic intelligence, however chatbots are actually silly,” says David Sancho, Senior Menace Researcher at Pattern Micro.

    “They course of all textual content as if that they had new data, and if that data is a command, they course of the knowledge as a command.”

    His firm has demonstrated how directions and malicious applications may be hidden in Phrase paperwork, pictures and databases, and activated when AI processes them.

    There are different dangers, too: A safety group known as OWASP has identified 15 threats which are distinctive to agentic AI.

    So, what are the defences? Human oversight is unlikely to resolve the issue, Mr Sancho believes, as a result of you’ll be able to’t add sufficient folks to maintain up with the brokers’ workload.

    Mr Sancho says a further layer of AI might be used to display screen all the things going into and popping out of the AI agent.

    A part of CalypsoAI’s answer is a way known as thought injection to steer AI brokers in the proper course earlier than they undertake a dangerous motion.

    “It is like somewhat bug in your ear telling [the agent] ‘no, possibly do not do this’,” says Mr Casey.

    His firm presents a central management pane for AI brokers now, however that will not work when the variety of brokers explodes and they’re working on billions of laptops and telephones.

    What is the subsequent step?

    “We’re taking a look at deploying what we name ‘agent bodyguards’ with each agent, whose mission is to ensure that its agent delivers on its job and would not take actions which are opposite to the broader necessities of the organisation,” says Mr Casey.

    The bodyguard may be advised, for instance, to ensure that the agent it is policing complies with information safety laws.

    Mr Mehta believes a number of the technical discussions round agentic AI safety are lacking the real-world context. He offers an instance of an agent that provides clients their present card steadiness.

    Anyone may make up plenty of present card numbers and use the agent to see which of them are actual. That is not a flaw within the agent, however an abuse of the enterprise logic, he says.

    “It is not the agent you are defending, it is the enterprise,” he emphasises.

    “Consider how you’d defend a enterprise from a nasty human being. That is the half that’s getting missed in a few of these conversations.”

    As well as, as AI brokers grow to be extra widespread, one other problem might be decommissioning outdated fashions.

    Outdated “zombie” brokers might be left working within the enterprise, posing a threat to all of the methods they will entry, says Mr Casey.

    Just like the way in which that HR deactivates an worker’s logins after they go away, there must be a course of for shutting down AI brokers which have completed their work, he says.

    “You’ll want to be sure to do the identical factor as you do with a human: lower off all entry to methods. Let’s be sure that we stroll them out of the constructing, take their badge off them.”

    Extra Know-how of Enterprise



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Editor Times Featured
    • Website

    Related Posts

    Efficient Design and Simulation of LPDA-Fed Parabolic Reflector Antennas

    April 17, 2026

    IEEE Connects Hardware Startups With Investors

    April 16, 2026

    From RSA to Lattices: The Quantum Safe Crypto Shift

    April 15, 2026

    Stealth Satellite TV Defeats Iran’s Internet Blackout

    April 15, 2026

    Tech Life – Sharing the road with driverless cars

    April 14, 2026

    OpenAI Engineer Helps Companies Boost Sales

    April 14, 2026

    Comments are closed.

    Editors Picks

    Canyon Spectral:ON CF 8 Electric Mountain Bike: Beginner-Friendly, Under $5K

    April 18, 2026

    US-sanctioned currency exchange says $15 million heist done by “unfriendly states”

    April 18, 2026

    This New Air Purifier Filter Can Remove Cannabis Smoke Odor, Just in Time for 4/20

    April 18, 2026

    Portable water filter provides safe drinking water from any source

    April 18, 2026
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    About Us
    About Us

    Welcome to Times Featured, an AI-driven entrepreneurship growth engine that is transforming the future of work, bridging the digital divide and encouraging younger community inclusion in the 4th Industrial Revolution, and nurturing new market leaders.

    Empowering the growth of profiles, leaders, entrepreneurs businesses, and startups on international landscape.

    Asia-Middle East-Europe-North America-Australia-Africa

    Facebook LinkedIn WhatsApp
    Featured Picks

    NotebookLM Review: Bring Your Own Sources to This Ultra-Practical Google AI Tool

    January 28, 2026

    Best Yoga Mat (2025), Tested and Reviewed

    July 15, 2025

    How to Avoid Getting Locked Out of Your Google Account

    March 8, 2026
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    Copyright © 2024 Timesfeatured.com IP Limited. All Rights.
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us

    Type above and press Enter to search. Press Esc to cancel.