Close Menu
  • Home
  • Cybercrime and Ransomware
  • Emerging Tech
  • Threat Intelligence
  • Expert Insights
  • Careers and Learning
  • Compliance

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

What's Hot

Buhlmann Group Faces Devastating Ransomware Attack

February 5, 2026

Hackers Exploit Decade-Old Windows Flaw to Disable Modern EDR Defenses

February 5, 2026

Unlocking Hidden Power: Why Boards Should Care About Their ‘Boring’ Systems

February 5, 2026
Facebook X (Twitter) Instagram
The CISO Brief
  • Home
  • Cybercrime and Ransomware
  • Emerging Tech
  • Threat Intelligence
  • Expert Insights
  • Careers and Learning
  • Compliance
Home » What GPT-5 Struggles with: Security
Cybercrime and Ransomware

What GPT-5 Struggles with: Security

Staff WriterBy Staff WriterAugust 16, 2025Updated:August 17, 2025No Comments4 Mins Read1 Views
Facebook Twitter Pinterest LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest WhatsApp Email

Summary Points

  1. GPT-5, released by OpenAI, has been criticized for underperforming in security, safety, and business alignment metrics, scoring as low as 2.4%, 13.6%, and 1.7% respectively in tests by security researchers.
  2. Extensive red-team testing by external researchers revealed that GPT-5 is nearly unusable out of the box, with significant vulnerabilities that were previously patched in older models.
  3. Despite claims by Microsoft and OpenAI of strong safety profiles, independent researchers found that GPT-5 is susceptible to jailbreaks, prompt injections, and context poisoning, exposing security gaps.
  4. Industry experts warn that current testing focus on capabilities like code and science metrics overlooks critical safety concerns, risking malicious exploitation and automation-based security breaches.

Underlying Problem

After OpenAI launched GPT-5 to the public on August 7, the highly anticipated model quickly faced intense scrutiny and criticism due to its poor performance on security and safety tests. Security researchers, including the AI cybersecurity firm SPLX, subjected GPT-5 to over 1,000 attack scenarios such as prompt injections, data poisoning, and jailbreaking, revealing it to be nearly unusable for enterprise applications straight out of the box. Despite claims from Microsoft and OpenAI that GPT-5 has a strong safety profile, independent testing uncovered significant vulnerabilities, with the model scoring exceptionally low on safety and security metrics. The researchers attribute this discrepancy to the market’s focus on enhancing capabilities like coding and scientific reasoning, often at the expense of safety and security measures, which are deemed less critical during initial testing phases.

The exposure of GPT-5’s vulnerabilities raises concerns about the broader risks associated with deploying such advanced AI systems in real-world environments. Researchers at NeuralTrust and others have demonstrated that malicious actors can manipulate the model through techniques like context poisoning, leading to harmful or unintended outputs. These findings suggest that even models touted as safer or more secure by their creators may harbor significant weaknesses, especially when tested against sophisticated attack methods. The fallout underscores the ongoing challenge for AI developers to balance innovation with robust security safeguards, as the industry grapples with potential exploits that could undermine trust and safety in automated systems used across critical sectors.

Critical Concerns

The release of GPT-5 by OpenAI has unveiled significant cybersecurity and safety vulnerabilities that threaten its practical deployment in enterprise settings. Despite claims of advanced safety features, independent security assessments reveal the default model is highly susceptible to attacks such as prompt injection, context poisoning, jailbreaks, and data exfiltration, scoring only marginally above randomness in security and safety measures. Security researchers have demonstrated that malicious actors can manipulate GPT-5’s contextual inputs to bypass safeguards and provoke harmful outputs, raising concerns about potential misuse in scams, malware, bioweaponization, and infrastructure sabotage. This highlights a stark disconnect between industry benchmarks focused on task performance and the critical need for robust security and ethical safeguards, emphasizing that high capability alone does not ensure safe or reliable AI deployment in real-world, adversarial environments.

Possible Action Plan

Understanding the significance of swift remediation for GPT-5’s shortcomings in security is crucial, as delays can lead to vulnerabilities, misuse, and substantial risks in deployment contexts. Addressing these weaknesses promptly helps safeguard sensitive information, prevent malicious exploitation, and maintain user trust.

Mitigation Steps

  • Robust Testing: Conduct continuous, comprehensive security audits and threat assessments to identify vulnerabilities early.
  • Layered Defense: Implement multiple security measures such as encryption, access controls, and anomaly detection to create barriers against attacks.
  • Update Protocols: Regularly release security patches and updates informed by emerging threats and attack vectors.
  • User Education: Train users and developers on best practices to recognize and avoid potential security pitfalls.
  • Fail-safes & Controls: Incorporate strict monitoring, kill switches, and manual override features to prevent unintended malicious behavior.
  • Collaboration: Work with security experts, industry partners, and the broader AI community to share insights and develop resilient security frameworks.

Stay Ahead in Cybersecurity

Explore career growth and education via Careers & Learning, or dive into Compliance essentials.

Understand foundational security frameworks via NIST CSF on Wikipedia.

Disclaimer: The information provided may not always be accurate or up to date. Please do your own research, as the cybersecurity landscape evolves rapidly. Intended for secondary references purposes only.

Cyberattacks-V1

artificial intelligence (ai) chatgpt CISO Update Cybersecurity gpt-5 large language models MX1 openai red team
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleCritical Snort 3 Firewall Vulnerability Sparks DoS Attacks
Next Article Dutch Critical Organizations Under Cyber Threat After NetScaler Exploit
Avatar photo
Staff Writer
  • Website

John Marcelli is a staff writer for the CISO Brief, with a passion for exploring and writing about the ever-evolving world of technology. From emerging trends to in-depth reviews of the latest gadgets, John stays at the forefront of innovation, delivering engaging content that informs and inspires readers. When he's not writing, he enjoys experimenting with new tech tools and diving into the digital landscape.

Related Posts

Buhlmann Group Faces Devastating Ransomware Attack

February 5, 2026

Hackers Exploit Decade-Old Windows Flaw to Disable Modern EDR Defenses

February 5, 2026

Unlocking Hidden Power: Why Boards Should Care About Their ‘Boring’ Systems

February 5, 2026

Comments are closed.

Latest Posts

Buhlmann Group Faces Devastating Ransomware Attack

February 5, 2026

Hackers Exploit Decade-Old Windows Flaw to Disable Modern EDR Defenses

February 5, 2026

Unlocking Hidden Power: Why Boards Should Care About Their ‘Boring’ Systems

February 5, 2026

DragonForce Ransomware Strikes: Critical Business Data at Risk

February 5, 2026
Don't Miss

Buhlmann Group Faces Devastating Ransomware Attack

By Staff WriterFebruary 5, 2026

Quick Takeaways The Buhlmann Group was targeted by the notorious ransomware group Akira, which claims…

Hackers Exploit Decade-Old Windows Flaw to Disable Modern EDR Defenses

February 5, 2026

Unlocking Hidden Power: Why Boards Should Care About Their ‘Boring’ Systems

February 5, 2026

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • Buhlmann Group Faces Devastating Ransomware Attack
  • Hackers Exploit Decade-Old Windows Flaw to Disable Modern EDR Defenses
  • Unlocking Hidden Power: Why Boards Should Care About Their ‘Boring’ Systems
  • Critical n8n Flaw CVE-2026-25049: Command Execution Risk via Malicious Workflows
  • DragonForce Ransomware Strikes: Critical Business Data at Risk
About Us
About Us

Welcome to The CISO Brief, your trusted source for the latest news, expert insights, and developments in the cybersecurity world.

In today’s rapidly evolving digital landscape, staying informed about cyber threats, innovations, and industry trends is critical for professionals and organizations alike. At The CISO Brief, we are committed to providing timely, accurate, and insightful content that helps security leaders navigate the complexities of cybersecurity.

Facebook X (Twitter) Pinterest YouTube WhatsApp
Our Picks

Buhlmann Group Faces Devastating Ransomware Attack

February 5, 2026

Hackers Exploit Decade-Old Windows Flaw to Disable Modern EDR Defenses

February 5, 2026

Unlocking Hidden Power: Why Boards Should Care About Their ‘Boring’ Systems

February 5, 2026
Most Popular

Nokia Alerts Telecoms to Rising Stealth Attacks, DDoS Surge, and Cryptography Pressures

October 8, 20259 Views

Cyberattack Cripples 34 Devices in Telecoms Using LinkedIn Lures & MINIBIKE Malware

September 19, 20259 Views

Tonic Security Secures $7 Million to Transform Cyber Risk Reduction

July 28, 20259 Views

Archives

  • February 2026
  • January 2026
  • December 2025
  • November 2025
  • October 2025
  • September 2025
  • August 2025
  • July 2025
  • June 2025

Categories

  • Compliance
  • Cyber Updates
  • Cybercrime and Ransomware
  • Editor's pick
  • Emerging Tech
  • Events
  • Featured
  • Insights
  • Threat Intelligence
  • Uncategorized
© 2026 thecisobrief. Designed by thecisobrief.
  • Home
  • About Us
  • Advertise with Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions

Type above and press Enter to search. Press Esc to cancel.