By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
Online Tech Guru
  • News
  • PC/Windows
  • Mobile
  • Apps
  • Gadgets
  • More
    • Gaming
    • Accessories
    • Editor’s Choice
    • Press Release
Reading: Researchers Propose a Better Way to Report Dangerous AI Flaws
Best Deal
Font ResizerAa
Online Tech GuruOnline Tech Guru
  • News
  • Mobile
  • PC/Windows
  • Gaming
  • Apps
  • Gadgets
  • Accessories
Search
  • News
  • PC/Windows
  • Mobile
  • Apps
  • Gadgets
  • More
    • Gaming
    • Accessories
    • Editor’s Choice
    • Press Release

Mafia: The Old Country Tells a Story That Doesn’t Require a ‘Massive Time Commitment,’ Publisher Sets Price at $50 and Announces Release Date

News Room News Room 9 May 2025
FacebookLike
InstagramFollow
YoutubeSubscribe
TiktokFollow
  • Subscribe
  • Privacy Policy
  • Contact
  • Terms of Use
© Foxiz News Network. Ruby Design Company. All Rights Reserved.
Online Tech Guru > News > Researchers Propose a Better Way to Report Dangerous AI Flaws
News

Researchers Propose a Better Way to Report Dangerous AI Flaws

News Room
Last updated: 13 March 2025 15:36
By News Room 4 Min Read
Share
SHARE

In late 2023, a team of third party researchers discovered a troubling glitch in OpenAI’s widely used artificial intelligence model GPT-3.5.

When asked to repeat certain words a thousand times, the model began repeating the word over and over, then suddenly switched to spitting out incoherent text and snippets of personal information drawn from its training data, including parts of names, phone numbers, and email addresses. The team that discovered the problem worked with OpenAI to ensure the flaw was fixed before revealing it publicly. It is just one of scores of problems found in major AI models in recent years.

In a proposal released today, more than 30 prominent AI researchers, including some who found the GPT-3.5 flaw, say that many other vulnerabilities affecting popular models are reported in problematic ways. They suggest a new scheme supported by AI companies that gives outsiders permission to probe their models and a way to disclose flaws publicly.

“Right now it’s a little bit of the Wild West,” says Shayne Longpre, a PhD candidate at MIT and the lead author of the proposal. Longpre says that some so-called jailbreakers share their methods of breaking AI safeguards the social media platform X, leaving models and users at risk. Other jailbreaks are shared with only one company even though they might affect many. And some flaws, he says, are kept secret because of fear of getting banned or facing prosecution for breaking terms of use. “It is clear that there are chilling effects and uncertainty,” he says.

The security and safety of AI models is hugely important given widely the technology is now being used, and how it may seep into countless applications and services. Powerful models need to be stress-tested, or red-teamed, because they can harbor harmful biases, and because certain inputs can cause them to break free of guardrails and produce unpleasant or dangerous responses. These include encouraging vulnerable users to engage in harmful behavior or helping a bad actor to develop cyber, chemical, or biological weapons. Some experts fear that models could assist cyber criminals or terrorists, and may even turn on humans as they advance.

The authors suggest three main measures to improve the third-party disclosure process: adopting standardized AI flaw reports to streamline the reporting process; for big AI firms to provide infrastructure to third-party researchers disclosing flaws; and for developing a system that allows flaws to be shared between different providers.

The approach is borrowed from the cybersecurity world, where there are legal protections and established norms for outside researchers to disclose bugs.

“AI researchers don’t always know how to disclose a flaw and can’t be certain that their good faith flaw disclosure won’t expose them to legal risk,” says Ilona Cohen, chief legal and policy officer at HackerOne, a company that organizes bug bounties, and a coauthor on the report.

Large AI companies currently conduct extensive safety testing on AI models prior to their release. Some also contract with outside firms to do further probing. “Are there enough people in those [companies] to address all of the issues with general-purpose AI systems, used by hundreds of millions of people in applications we’ve never dreamt?” Longpre asks. Some AI companies have started organizing AI bug bounties. However, Longpre says that independent researchers risk breaking the terms of use if they take it upon themselves to probe powerful AI models.

Share This Article
Facebook Twitter Copy Link
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Samsung Galaxy Watch to Reportedly Receive Galaxy AI-Powered Now Bar and Now Brief Features

News Room News Room 9 May 2025
FacebookLike
InstagramFollow
YoutubeSubscribe
TiktokFollow

Trending

Threads Rolls Out Instagram-Like Account Status Feature for More Transparency

Threads on Friday rolled out a new feature to help users understand about their posts…

9 May 2025

Mobile studio Fuse Games receives $7m investment

Turkish mobile developer Fuse Games has received a $7 million investment from venture capital firm…

9 May 2025

Pope Leo XIV is the first pope with an online footprint

The moment the white smoke appeared above the Sistine Chapel, I immediately turned on my…

9 May 2025
Mobile

Samsung Galaxy S25 Edge to Use Corning Gorilla Glass Ceramic 2 Display Protection

Samsung Galaxy S25 Edge is set to launch globally on May 13. Ahead of the launch, the company has confirmed that the handset will come with Corning's new Gorilla Glass Ceramic 2…

News Room 9 May 2025

Your may also like!

Gaming

10 Worker Placement Board Games Worth Playing in 2025

News Room 9 May 2025
Gaming

Tripledot Studios acquires AppLovin’s mobile games business for $800 million

News Room 9 May 2025
News

Apple is planning smart glasses with and without AR

News Room 9 May 2025
News

Razer’s Clio is a $230 surround sound head cushion

News Room 9 May 2025

Our website stores cookies on your computer. They allow us to remember you and help personalize your experience with our site.

Read our privacy policy for more information.

Quick Links

  • Subscribe
  • Privacy Policy
  • Contact
  • Terms of Use
Advertise with us

Socials

Follow US
Welcome Back!

Sign in to your account

Lost your password?