Close Menu
GeekBlog

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Meta will sell you refurbished Ray-Ban smart glasses for $76 off – how to find them

    August 30, 2025

    Garmin Fenix 8 Pro rumors swirl, and new leaks point to 4 new subscription tiers – mere months after the Connect+ debacle

    August 30, 2025

    SSA Whistleblower’s Resignation Email Mysteriously Disappeared From Inboxes

    August 29, 2025
    Facebook X (Twitter) Instagram Threads
    GeekBlog
    • Home
    • Mobile
    • Reviews
    • Tech News
    • Deals & Offers
    • Gadgets
      • How-To Guides
    • Laptops & PCs
      • AI & Software
    • Blog
    Facebook X (Twitter) Instagram
    GeekBlog
    Home»Tech News»Anthropic will nuke your attempt to use AI to build a nuke
    Tech News

    Anthropic will nuke your attempt to use AI to build a nuke

    Michael ComaousBy Michael ComaousAugust 23, 2025No Comments3 Mins Read0 Views
    Share Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
    A bomb and crosshair on a keyboard.
    Share
    Facebook Twitter LinkedIn Pinterest Email Copy Link

    • Anthropic has developed an AI-powered tool that detects and blocks attempts to ask AI chatbots for nuclear weapons design
    • The company worked with the U.S. Department of Energy to ensure the AI could identify such attempts
    • Anthropic claims it spots dangerous nuclear-related prompts with 96% accuracy and has already proven effective on Claude

    If you’re the type of person who asks Claude how to make a sandwich, you’re fine. If you’re the type of person who asks the AI chatbot how to build a nuclear bomb, you’ll not only fail to get any blueprints, you might also face some pointed questions of your own. That’s thanks to Anthropic’s newly deployed detector of problematic nuclear prompts.

    Like other systems for spotting queries Claude shouldn’t respond to, the new classifier scans user conversations, in this case flagging any that veer into “how to build a nuclear weapon” territory. Anthropic built the classification feature in a partnership with the U.S. Department of Energy’s National Nuclear Security Administration (NNSA), giving it all the information it needs to determine whether someone is just asking about how such bombs work or if they’re looking for blueprints. It’s performed with 96% accuracy in tests.

    Though it might seem over-the-top, Anthropic sees the issue as more than merely hypothetical. The chance that powerful AI models may have access to sensitive technical documents and could pass along a guide to building something like a nuclear bomb worries federal security agencies. Even if Claude and other AI chatbots block the most obvious attempts, innocent-seeming questions could in fact be veiled attempts at crowdsourcing weapons design. The new AI chatbot generations might help even if it’s not what their developers intend.


    You may like

    The classifier works by drawing a distinction between benign nuclear content, asking about nuclear propulsion, for instance, and the kind of content that could be turned to malicious use. Human moderators might struggle to keep up with any gray areas at the scale AI chatbots operate, but with proper training, Anthropic and the NNSA believe the AI could police itself. Anthropic claims its classifier is already catching real-world misuse attempts in conversations with Claude.

    Nuclear AI safety

    Nuclear weapons in particular represent a uniquely tricky problem, according to Anthropic and its partners at the DoE. The same foundational knowledge that powers legitimate reactor science can, if slightly twisted, provide the blueprint for annihilation. The arrangement between Anthropic and the NNSA could catch deliberate and accidental disclosures, and set up a standard to prevent AI from being used to help make other weapons, too. Anthropic plans to share its approach with the Frontier Model Forum AI safety consortium.

    The narrowly tailored filter is aimed at making sure users can still learn about nuclear science and related topics. You still get to ask about how nuclear medicine works, or whether thorium is a safer fuel than uranium.

    What the classifier attempts to circumvent are attempts to turn your home into a bomb lab with a few clever prompts. Normally, it would be questionable if an AI company could thread that needle, but the expertise of the NNSA should make the classifier different from a generic content moderation system. It understands the difference between “explain fission” and “give me a step-by-step plan for uranium enrichment using garage supplies.”

    Sign up for breaking news, reviews, opinion, top tech deals, and more.

    This doesn’t mean Claude was previously helping users design bombs. But it could help forestall any attempt to do so. Stick to asking about the way radiation can cure diseases or ask for creative sandwich ideas, not bomb blueprints.

    You might also like

    Anthropic attempt build nuke
    Follow on Google News Follow on Flipboard
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Telegram Email Copy Link
    Previous ArticleBluesky Goes Dark in Mississippi Over Age Verification Law
    Next Article No, iPadOS 26 isn’t a laptop killer, but these 4 things make it a huge leap forward
    Michael Comaous
    • Website

    Related Posts

    5 Mins Read

    Meta will sell you refurbished Ray-Ban smart glasses for $76 off – how to find them

    3 Mins Read

    Garmin Fenix 8 Pro rumors swirl, and new leaks point to 4 new subscription tiers – mere months after the Connect+ debacle

    4 Mins Read

    SSA Whistleblower’s Resignation Email Mysteriously Disappeared From Inboxes

    1 Min Read

    The fight against labeling long-term streaming rentals as “purchases” you “buy”

    2 Mins Read

    Libby is adding an AI book recommendation feature

    3 Mins Read

    TikTok now lets users send voice notes and images in DMs

    Top Posts

    8BitDo Pro 3 review: better specs, more customization, minor faults

    August 8, 202512 Views

    WIRED Roundup: ChatGPT Goes Full Demon Mode

    August 2, 202512 Views

    Framework Desktop Review: A Delightful Surprise

    August 7, 202511 Views
    Stay In Touch
    • Facebook
    • YouTube
    • TikTok
    • WhatsApp
    • Twitter
    • Instagram
    Latest Reviews

    Subscribe to Updates

    Get the latest tech news from FooBar about tech, design and biz.

    Most Popular

    8BitDo Pro 3 review: better specs, more customization, minor faults

    August 8, 202512 Views

    WIRED Roundup: ChatGPT Goes Full Demon Mode

    August 2, 202512 Views

    Framework Desktop Review: A Delightful Surprise

    August 7, 202511 Views
    Our Picks

    Meta will sell you refurbished Ray-Ban smart glasses for $76 off – how to find them

    August 30, 2025

    Garmin Fenix 8 Pro rumors swirl, and new leaks point to 4 new subscription tiers – mere months after the Connect+ debacle

    August 30, 2025

    SSA Whistleblower’s Resignation Email Mysteriously Disappeared From Inboxes

    August 29, 2025

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    Facebook X (Twitter) Instagram Pinterest Threads
    • About Us
    • Contact us
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    © 2025 geekblog. Designed by Pro.

    Type above and press Enter to search. Press Esc to cancel.