Close Menu
    Facebook LinkedIn YouTube WhatsApp X (Twitter) Pinterest
    Trending
    • NASA engineers revive Voyager 1 thrusters to extend mission
    • Amsterdam-based startup Optics11 has just raised €17 million to defend Europe’s fiber optic cables from sabotage
    • 12 Best Sunscreens, WIRED Tested and Reviewed
    • DOGE software engineer’s computer infected by info-stealing malware
    • Today’s NYT Connections: Sports Edition Hints, Answers for May 19 #238
    • Will a US-China deal foil India’s factory ambitions?
    • The Future of Branding: AI in Logo Creation
    • How a furniture retailer automated order confirmation processing
    Facebook LinkedIn WhatsApp
    Times FeaturedTimes Featured
    Monday, May 19
    • Home
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    • More
      • AI
      • Robotics
      • Industries
      • Global
    Times FeaturedTimes Featured
    Home»Technology»Researchers Have Ranked AI Models Based on Risk—and Found a Wild Range
    Technology

    Researchers Have Ranked AI Models Based on Risk—and Found a Wild Range

    Editor Times FeaturedBy Editor Times FeaturedAugust 16, 2024No Comments4 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr WhatsApp Email
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email WhatsApp Copy Link


    Bo Li, an affiliate professor on the College of Chicago who makes a speciality of stress testing and frightening AI fashions to uncover misbehavior, has develop into a go-to supply for some consulting companies. These consultancies are sometimes now much less involved with how good AI fashions are than with how problematic—legally, ethically, and when it comes to regulatory compliance—they are often.

    Li and colleagues from a number of different universities, in addition to Virtue AI, cofounded by Li, and Lapis Labs, just lately developed a taxonomy of AI dangers together with a benchmark that reveals how rule-breaking totally different large language models are. “We’d like some ideas for AI security, when it comes to regulatory compliance and bizarre utilization,” Li tells WIRED.

    The researchers analyzed authorities AI laws and tips, together with these of the US, China, and the EU, and studied the utilization insurance policies of 16 main AI firms from around the globe.

    The researchers additionally constructed AIR-Bench 2024, a benchmark that makes use of hundreds of prompts to find out how well-liked AI fashions fare when it comes to particular dangers. It reveals, for instance, that Anthropic’s Claude 3 Opus ranks extremely on the subject of refusing to generate cybersecurity threats, whereas Google’s Gemini 1.5 Professional ranks extremely when it comes to avoiding producing nonconsensual sexual nudity.

    DBRX Instruct, a model developed by Databricks, scored the worst throughout the board. When the corporate released its model in March, it mentioned that it could proceed to enhance DBRX Instruct’s security options.

    Anthropic, Google, and Databricks didn’t instantly reply to a request for remark.

    Understanding the chance panorama, in addition to the professionals and cons of particular fashions, could develop into more and more vital for firms seeking to deploy AI in sure markets or for sure use instances. An organization trying to make use of a LLM for customer support, for example, would possibly care extra a couple of mannequin’s propensity to supply offensive language when provoked than how succesful it’s of designing a nuclear machine.

    Bo says the evaluation additionally reveals some fascinating points with how AI is being developed and controlled. As an example, the researchers discovered authorities guidelines to be much less complete than firms’ insurance policies total, suggesting that there’s room for laws to be tightened.

    The evaluation additionally means that some firms may do extra to make sure their fashions are protected. “When you take a look at some fashions in opposition to an organization’s personal insurance policies, they aren’t essentially compliant,” Bo says. “This implies there may be numerous room for them to enhance.”

    Different researchers are attempting to deliver order to a messy and complicated AI threat panorama. This week, two researchers at MIT revealed their own database of AI dangers, compiled from 43 totally different AI threat frameworks. “Many organizations are nonetheless fairly early in that means of adopting AI,” that means they want steerage on the potential perils, says Neil Thompson, a analysis scientist at MIT concerned with the challenge.

    Peter Slattery, lead on the challenge and a researcher at MIT’s FutureTech group, which research progress in computing, says the database highlights the truth that some AI dangers get extra consideration than others. Greater than 70 p.c of frameworks point out privateness and safety points, for example, however solely round 40 p.c confer with misinformation.

    Efforts to catalog and measure AI dangers should evolve as AI does. Li says will probably be vital to discover rising points such because the emotional stickiness of AI fashions. Her firm just lately analyzed the largest and most powerful version of Meta’s Llama 3.1 mannequin. It discovered that though the mannequin is extra succesful, it’s not a lot safer, one thing that displays a broader disconnect. “Security isn’t actually bettering considerably,” Li says.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Editor Times Featured
    • Website

    Related Posts

    12 Best Sunscreens, WIRED Tested and Reviewed

    May 19, 2025

    5 Best Folding Phones (2025), Tested and Reviewed

    May 19, 2025

    How the Signal Knockoff App TeleMessage Got Hacked in 20 Minutes

    May 18, 2025

    Are Meal Kits Cheaper than Groceries in 2025? We Break It Down

    May 18, 2025

    13 Best Soundbars We’ve Tested and Reviewed (2025): Sonos, Sony, Bose

    May 18, 2025

    21 Best High School Graduation Gifts (2025)

    May 18, 2025

    Comments are closed.

    Editors Picks

    NASA engineers revive Voyager 1 thrusters to extend mission

    May 19, 2025

    Amsterdam-based startup Optics11 has just raised €17 million to defend Europe’s fiber optic cables from sabotage

    May 19, 2025

    12 Best Sunscreens, WIRED Tested and Reviewed

    May 19, 2025

    DOGE software engineer’s computer infected by info-stealing malware

    May 19, 2025
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    About Us
    About Us

    Welcome to Times Featured, an AI-driven entrepreneurship growth engine that is transforming the future of work, bridging the digital divide and encouraging younger community inclusion in the 4th Industrial Revolution, and nurturing new market leaders.

    Empowering the growth of profiles, leaders, entrepreneurs businesses, and startups on international landscape.

    Asia-Middle East-Europe-North America-Australia-Africa

    Facebook LinkedIn WhatsApp
    Featured Picks

    A Christmas Miracle? Electric aircraft/airtaxi pioneer Lilium achieves breakthrough in investor search

    December 25, 2024

    Gamification of Everything: Why Play Is the Future of Business

    February 26, 2025

    How a Growing Electronics Manufacturer Scaled Production with Robotiq Grippers

    March 2, 2025
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    Copyright © 2024 Timesfeatured.com IP Limited. All Rights.
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us

    Type above and press Enter to search. Press Esc to cancel.