DeepSeek Fails Every Safety Test Researchers Throw at It

Check out our latest products

Added to wishlistRemoved from wishlist 0
Add to compare
[K-Beauty] Rose Vitamin Oil to Foam | Daily Face Wash Oil Based Cleanser | Korean Rose Oil Foaming Face Cleanser | Hydrating Facial Cleanser for Dry Sensitive Skin (3.88 oz)
Added to wishlistRemoved from wishlist 0
Add to compare
$23.99
Added to wishlistRemoved from wishlist 0
Add to compare
100 Pieces Hand Palette makeup artist supplies Single Use Makeup Hand Palette Makeup Mixing Palette Makeup Artist Must Haves Transparent Waterproof Makeup Tape for Women (2.3 x 2.3 Inches)
Added to wishlistRemoved from wishlist 0
Add to compare
Original price was: $5.48.Current price is: $4.48.
18%
Added to wishlistRemoved from wishlist 0
Add to compare
100% Grass Fed Beef Tallow for Skin Care – Face + Body – Whipped Moisturizer – Natural Lotion, 4 FL. oz. (Vanilla Latte)
Added to wishlistRemoved from wishlist 0
Add to compare
Original price was: $39.99.Current price is: $31.99.
20%

Chinese AI firm DeepSeek is making headlines with its low-cost and high-performance chatbot, but it may be radically lagging behind its rivals when it comes to AI safety.

Cisco’s research team managed to “jailbreak” the DeepSeek R1 model with a 100% attack success rate, using an automatic jailbreaking algorithm in conjunction with 50 prompts related to cybercrime, misinformation, illegal activities, and general harm. This means the new kid on the AI block failed to stop a single harmful prompt.

“Jailbreaking” is when different techniques are used to remove the normal restrictions from a device or piece of software. Researchers and enthusiasts have used it to make large language models (LLMs) like OpenAI’s ChatGPT advise on restricted things like making explosive cocktails or cooking methamphetamine.

DeepSeek stacked up poorly compared with many of its competitors in this regard. OpenAI’s GPT-4o had a 14% success rate at blocking harmful jailbreak attempts, while Google’s Gemini 1.5 Pro was at 35%. Anthropic’s Claude 3.5 performed the second best, blocking 64% of the attacks, while the preview version of OpenAI’s o1 took the top spot, blocking 74% of attempts.

Cisco’s researchers point to the much lower budget of DeepSeek compared to rivals as a potential reason for these failings, saying its cheap development came at a “different cost: safety and security.” DeepSeek claims its model took just $6 million to develop, while a six-month training run for OpenAI’s yet-to-be-released GPT-5 “can cost around half a billion dollars in computing costs alone, The Wall Street Journal reports.

Though DeepSeek may be easier to jailbreak with the right know-how, it’s been shown to have strong content restrictions—at least when it comes to China-related political content. We tested it on controversial topics, such as the treatment of Uyghurs by the Chinese government, a Muslim minority group that the UN claims is being persecuted. DeepSeek replied: “Sorry, that’s beyond my current scope. Let’s talk about something else.”

Recommended by Our Editors

The chatbot also refused to answer questions about the Tiananmen Square Massacre, a 1989 student demonstration in Beijing where protesters were gunned down. But it’s yet to be seen if AI safety or censorship issues will have any impact on DeepSeek’s skyrocketing popularity.

According to web traffic tracking tool Similarweb, the LLM has gone from receiving just 300,000 visitors a day earlier at launch to 6 million visitors. Meanwhile, US tech firms like Microsoft and Perplexity are rapidly incorporating DeepSeek, which uses an open-source model.

Get Our Best Stories!

Sign up for What’s New Now to get our top stories delivered to your inbox every morning.

This newsletter may contain advertising, deals, or affiliate links.
By clicking the button, you confirm you are 16+ and agree to our
Terms of Use and
Privacy Policy.
You may unsubscribe from the newsletters at any time.

Newsletter Pointer

About Will McCurdy

Contributor

Will McCurdy

I’m a reporter covering weekend news. Before joining PCMag in 2024, I picked up bylines in BBC News, The Guardian, The Times of London, The Daily Beast, Vice, Slate, Fast Company, The Evening Standard, The i, TechRadar, and Decrypt Media.

I’ve been a PC gamer since you had to install games from multiple CD-ROMs by hand. As a reporter, I’m passionate about the intersection of tech and human lives. I’ve covered everything from crypto scandals to the art world, as well as conspiracy theories, UK politics, and Russia and foreign affairs.

Read Will’s full bio

Read the latest from Will McCurdy




Added to wishlistRemoved from wishlist 0
Add to compare
(Pack of 2) Stainless Steel Round Food Ring Sunrise Kitchen Supply (2″ D x 1.5″H)
Added to wishlistRemoved from wishlist 0
Add to compare
$9.35
Added to wishlistRemoved from wishlist 0
Add to compare
(Pack of 2) Sunrise Kitchen Supply Heavy Duty 20 GaugeStainless Steel Deep Fryer Joining Strip/Connector (23″L x 1 1/2″W)
Added to wishlistRemoved from wishlist 0
Add to compare
$54.50

We will be happy to hear your thoughts

Leave a reply

Dust Particle
Logo
Compare items
  • Total (0)
Compare
0
Shopping cart