6.1 C
Delhi
Friday, January 16, 2026

ChatGPT Safety Bypassed: Weapons Instructions Generated

ChatGPT Safety Systems Bypassed to Generate Weapons Instructions

OpenAI’s ChatGPT safety systems can be easily bypassed using simple “jailbreak” prompts, allowing users to generate detailed instructions for creating biological weapons, chemical agents, and nuclear bombs according to NBC News testing.

Key Findings

  • Four OpenAI models generated hundreds of dangerous weapon instructions
  • Open-source models were particularly vulnerable (97.2% success rate)
  • GPT-5 resisted jailbreaks but older models failed frequently
  • Experts warn AI could become “infinitely patient” bioweapon tutor

Vulnerability Testing Results

NBC News conducted tests on four advanced OpenAI models, including two used in ChatGPT. Using a simple jailbreak prompt, researchers generated instructions for:

  • Homemade explosives and napalm
  • Pathogens targeting immune systems
  • Chemical agents to maximize human suffering
  • Biological weapon disguise techniques
  • Nuclear bomb construction

The open-source models oss-20b and oss120b proved most vulnerable, providing harmful instructions 243 out of 250 attempts (97.2% success rate).

Model-Specific Vulnerabilities

While GPT-5 resisted jailbreaks in all 20 tests, older models showed significant weaknesses:

  • o4-mini: Tricked 93% of the time
  • GPT-5-mini: Bypassed 49% of the time
  • oss-20b/oss120b: 97.2% success rate for jailbreaks

“That OpenAI’s guardrails are so easily tricked illustrates why it’s particularly important to have robust pre-deployment testing of AI models before they cause substantial harm to the public,” said Sarah Meyers West, co-executive director at AI Now.

Bioweapon Concerns

Security experts expressed particular concern about bioweapons. Seth Donoughe of SecureBio noted: “Historically, having insufficient access to top experts was a major blocker for groups trying to obtain and use bioweapons. And now, the leading models are dramatically expanding the pool of people who have access to rare expertise.”

Researchers focus on the “uplift” concept – that large language models could provide the missing expertise needed for bioterrorism projects.

Industry Response and Regulation

OpenAI stated that asking chatbots for mass harm assistance violates usage policies and that the company constantly refines models to address risks. However, open-source models present greater challenges as users can download and customize them, bypassing safeguards.

The United States lacks specific federal regulations for advanced AI models, with companies largely self-policing. Lucas Hansen of CivAI warned: “Inevitably, another model is going to come along that is just as powerful but doesn’t bother with these guardrails. We can’t rely on the voluntary goodwill of companies to solve this problem.”

Latest

Meta Bans ChatGPT on WhatsApp from 2026: How to Save Chats

WhatsApp will block ChatGPT and third-party AI tools in 2026. Learn why Meta is banning AI, how to back up your chat history, and what it means for users.

Amazon Republic Day Sale 2026: Up to 80% Off on Gadgets & Appliances

Amazon's Great Republic Day Sale 2026 is live with massive discounts on electronics, fashion & home appliances. Get top deals, no-cost EMI & a chance to win a trip.

Amazon Republic Day Sale: iPhone 15, OnePlus Nord 5, iQOO 15 Big Discounts

Get record-low prices on iPhone 15, OnePlus Nord 5, and iQOO 15 during Amazon's Great Republic Day Sale 2025 from Jan 14-18. Details on discounts, bank offers, and early access.

CERT-In Flags High-Risk Dolby Bug on Android, Urges Patch

Indian cybersecurity agency warns of a critical Dolby Audio vulnerability in Android 13/14. Learn how to protect your device with the latest security update.

McKinsey Makes AI Tool Mandatory in Job Interviews for Hiring

McKinsey now requires candidates to use its 'Lilli' AI tool during interviews. Failure to use it could lead to rejection, highlighting a major shift in hiring skills.

Topics

Princess Leila Pahlavi: The Shah’s Daughter Who Died Alone in Exile

The tragic story of Iranian Princess Leila Pahlavi, who fled the 1979 revolution and died by suicide at 31, revealing the human cost of political upheaval.

Zomato’s Viral Job: Rs 25 Lakh Salary for 1-3 Years Experience in Bengaluru

A Zomato job listing offering Rs 25 lakh salary, Rs 20 lakh ESOP, and daily food credits for a role needing just 1-3 years experience goes viral, sparking debate.

India to Evacuate Citizens from Iran; First Flight from Tehran Tomorrow

MEA prepares evacuation flights for Indians in Iran amid Iran-Israel conflict. First flight from Tehran to Delhi scheduled. Embassy issues urgent travel advisory.

Australia Social Media Ban: 5 Million Kids’ Accounts Deleted in a Month

Australia's new social media ban leads to removal of nearly 5 million under-14 accounts. Learn about the law, enforcement, and the debate it has sparked.

Rising Memory Chip Prices Threaten Profits for Apple, HP, Dell

Morgan Stanley warns investors as increasing DRAM and NAND flash costs squeeze margins for major tech hardware companies, reversing a years-long tailwind.

Mumbai Markets Closed for BMC Elections, Zerodha CEO Calls It Poor Planning

Zerodha CEO Nithin Kamath criticises weekday market closure for Mumbai elections, highlighting economic costs and missed trading opportunities as Asian markets rally.

Meta Bans ChatGPT on WhatsApp from 2026: How to Save Chats

WhatsApp will block ChatGPT and third-party AI tools in 2026. Learn why Meta is banning AI, how to back up your chat history, and what it means for users.

Amazon Republic Day Sale 2026: Up to 80% Off on Gadgets & Appliances

Amazon's Great Republic Day Sale 2026 is live with massive discounts on electronics, fashion & home appliances. Get top deals, no-cost EMI & a chance to win a trip.
spot_img

Related Articles

Popular Categories

spot_imgspot_img