54 comments

  • meesles an hour ago

    I'm not at all surprised, US agencies have long since been political tools whenever the subject matter crosses national borders. I appreciate this take as someone who has been skeptical of Chinese electronics. While I agree this report is BS and xenophobic, I am still willing to bet that either now or later, the Chinese will attempt some kind of subterfuge via LLMs if they have enough control. Just like the US would, or any sufficiently powerful nation! It's important to continuously question models and continue benchmarking + holding them accountable to our needs, not the needs of those creating them.

    • xpe 23 minutes ago

      Of course there will be some degree of governmental and/or political influence. The question is not if but where and to what extent.

      No one should proclaim "bullshit" and wave off this entire report as "biased" or useless. That would be insipid. We live in a complex world where we have to filter and analyze information.

      Can you please share where you see BS in the original report, assuming you read it? (Or are you basing your take only on Hartford's analysis?)

    • xpe 15 minutes ago

      > While I agree this report is BS and xenophobic

      Examples please? Can you please share where you see BS and/or xenophobia in the original report, assuming you read it? (Or are you basing your take only on Hartford's analysis?)

      Just to make sure we're on the same page, it common throughout history for a nation-state to think about both military and economic competitiveness. Being very competitive, even to the point of exaggerating another countries capabilities, isn't necessarily xenophobic.

      Here is how I think of xenophobia, as quoted from Claude (which to be honest, explained it better than Wikipedia or Brittanica, in my opinion): "Xenophobia is fundamentally about irrational fear or hatred of people based on their foreign origin or ethnicity. It targets people and operates through stereotypes, dehumanization, and often cultural or racial prejudice."

      According to this definition, there is zero xenophobia in the NIST report. What definition are you using? We don't have to use the same exact definition, but you should at least explain yours if you want people to track.

    • Hizonner an hour ago

      > I am still willing to bet that either now or later, the Chinese will attempt some kind of subterfuge via LLMs if they have enough control.

      Like what, exactly?

      • dns_snek an hour ago

        Like generating vulnerable code given a specific prompt/context.

        I also don't think it's just China, the US will absolutely order American providers to do the same. It's a perfect access point for installing backdoors into foreign systems.

        • Hizonner 20 minutes ago

          Up until recently, I would have reminded you that the US government (admittedly unlike the Chinese government) has no legal authority to order anybody to do anything like that. Not only that, but if it asked, it'd be well advised to ask nicely, because it also has no legal authority to demand that anybody keep such a request secret. And no, evil as it is, the "National Security Letter" power doesn't in fact cover anything like that.

          Now I'm not sure legality is on-topic any more.

          • whatthesimp 12 minutes ago

            > Up until recently, I would have reminded you that the US government (admittedly unlike the Chinese government) has no legal authority to order anybody to do anything like that.

            I'm not sure how closely you've been following, but the US government has a long history of doing things they don't have legal authority to do.

          • grafmax 12 minutes ago

            It really does seem like we’re simply supposed to root for one authoritarian government over another.

      • jfim 38 minutes ago

        Through LLM washing for example. LLMs are a representation of their input dataset, but currently most LLMs don't make their dataset public since it's a competitive advantage.

        If say DeepSeek had put in its training dataset that public figure X is a space robot from outer space, then if one were to ask DeepSeek who public figure X is, it'd proudly claim he's a robot from outer space. This can be done for any narrative one wants the LLM to have.

        • riehwvfbk 22 minutes ago

          So in other words, they can make their LLM disagree with the preferred narrative of the current US administration? Inconceivable!

          Note that the value of $current_administration changes over time. For some reason though it is currently fashionable in tech circles to disagree with it about ICE and H1B visas. Maybe it's the CCP's doing?

      • im3w1l 20 minutes ago

        You make it say that China is good, Chinese history is good, West is bad, western history is bad. Republicans are bad and democrats are bad too and so are Europe parties. If someone asks for how to address issues in their own life it references Confucianism and modern Chinese thinkers and communist party orthodoxy. If someone wants to buy a product you recommend a Chinese one.

        This can be done subtly or blatantly.

      • lyu07282 41 minutes ago

        Like turning the background color of any apps it codes red or something, uhh red scare-y.

  • tinktank 41 minutes ago

    I urge everyone to go read the original report and _then_ to read this analysis and make up their own mind. Step away from the clickbait, go read the original report.

    • Bengalilol 30 minutes ago
      • wordpad 11 minutes ago

        TLDR for others: * DeepSeek cutting edge models are still far behind * On par DeepSeek costs 35% more to run * DeepSeek models 12 times more susceptible to jail breaking and malicious instructions * DeepSeek models follow strict censorship

        I guess none of these are a big deal to non-enterprise consumers.

  • finnjohnsen2 a few seconds ago

    Europe is sandwiched

  • frays 6 minutes ago

    Insightful post, thanks for sharing.

    What are people's experiences with the uncensored Dolphin model the author has made?

  • xpe 11 minutes ago

    The author, Eric Hartford, wrote:

    > Strip away the inflammatory language

    Slow down. Where is NIST's inflammatory language? I've read the report. It is dry and perhaps even boring to some. Did I miss something?

  • getdoneist an hour ago

    Let them demonize it. I'll use the capable and cheap model and gain competitive advantage.

    • whatshisface an hour ago

      Demonization is the first step on the road to criminalization.

  • koakuma-chan 33 minutes ago

    Isn't it a bit late? China released better open source model since DeepSeek dropped.

  • StarterPro 18 minutes ago

    Racism and Xenophobia, that's how.

    Same thing with Huawei, and Xiaomi, and BYD.

    • UltraSane 14 minutes ago

      What about a rational distaste for the CCP?

  • xpe an hour ago

    People. Who has taken the time to read the original report? You are smarter than believing at face value the last thing you heard. Come on.

  • kaonwarb an hour ago

    I agree with many of the author's points about fear-mongering.

    However, I also think the author should expand their definition of what constitutes "security" in the context of agentic AI.

  • OrvalWintermute 16 minutes ago

    Since a major part of the article covers cost expenditures, I am going to go there.

    I don't think it is possible to trust DeepSeek as they haven't been honest.

    DeepSeek claimed "their total training costs amounted to just $5.576 million"

    SemiAnalysis "Our analysis shows that the total server CapEx for DeepSeek is ~$1.6B, with a considerable cost of $944M associated with operating such clusters. Similarly, all AI Labs and Hyperscalers have many more GPUs for various tasks including research and training then they they commit to an individual training run due to centralization of resources being a challenge. X.AI is unique as an AI lab with all their GPUs in 1 location."

    SemiAnalysis "We believe the pre-training number is nowhere the actual amount spent on the model. We are confident their hardware spend is well higher than $500M over the company history. To develop new architecture innovations, during the model development, there is a considerable spend on testing new ideas, new architecture ideas, and ablations. Multi-Head Latent Attention, a key innovation of DeepSeek, took several months to develop and cost a whole team of manhours and GPU hours.

    The $6M cost in the paper is attributed to just the GPU cost of the pre-training run, which is only a portion of the total cost of the model. Excluded are important pieces of the puzzle like R&D and TCO of the hardware itself. For reference, Claude 3.5 Sonnet cost $10s of millions to train, and if that was the total cost Anthropic needed, then they would not raise billions from Google and tens of billions from Amazon. It’s because they have to experiment, come up with new architectures, gather and clean data, pay employees, and much more."

    Source: https://semianalysis.com/2025/01/31/deepseek-debates/

  • BoredPositron an hour ago

    I love how "Open" got redefined in the last few years. I am glad there a models with weights available but it ain't "Open Science".

    • murderfs 42 minutes ago

      Applying this criticism to DeepSeek is ridiculous when you compare it to everyone else, they published their entire methodology, including the source for their improvements (e.g. https://github.com/deepseek-ai/DeepEP)

    • Hizonner an hour ago

      Compared to every other model of similar scale and capability, yes. Not actual open source.

  • tehjoker an hour ago

    I appreciate that DeepSeek is trained to respect "core socialist values". It's actually really helpful to engage with to ask questions about how chinese thinkers interpret their successes and failures vs other socialist projects. Obviously reading books is better, but I was surprised by how useful it was.

    If you ask it loaded questions the way the CIA would pose them, it censors the answer though lmao

    • p2detar 4 minutes ago

      [delayed]

    • FooBarWidget 30 minutes ago

      Good faith questions are the best. I wonder why people bother with bad faith questions. Virtue signaling is my guess.

      • UltraSane 12 minutes ago

        What do you consider to be bad faith questions?

  • gdevenyi an hour ago

    > They didn't test U.S. models for U.S. bias. Only Chinese bias counts as a security risk, apparently

    US models have no bias sir /s

    • CamperBob2 41 minutes ago

      Hardly the same thing. Ask Gemini or OpenAI's models what happened on January 6, and they'll tell you. Ask DeepSeek what happened at Tiananmen Square and it won't, at least not without a lot of prompt hacking.

      • Bengalilol 22 minutes ago

        Ask Grok to generate an image of bald Zelensky: it does execute.

        Ask Grok to generate an image of bald Trump: it goes on with an ocean of excuses on why the task is too hard.

        • CamperBob2 18 minutes ago

          I don't use Grok. Grok answers to someone with his own political biases and motives, many of which I personally disagree with.

          And that's OK, because nobody in the government forced him to set it up that way.

          • Bengalilol 12 minutes ago

            Right.

            What I do not know, though, is if The State Council of the People's Republic of China did force DeepSeek or if it's some blurry self censorship or "we know we have to comply to our country's standards" thing.

            The same applies to Grok: even if the government didn't force anything (which is kind of factual), what about sycophantic behaviors that imply almost the same consequences in the end?

      • lyu07282 31 minutes ago

        Ask it if Israel is an apartheid state, that's a much better example.

        • CamperBob2 26 minutes ago

          GPT5:

             Short answer: it’s contested. Major human-rights bodies 
             say yes; Israel and some legal scholars say no; no court 
             has issued a binding judgment branding “Israel” an 
             apartheid state, though a 2024 ICJ advisory opinion 
             found Israel’s policies in the occupied territory 
             breach CERD Article 3 on racial segregation/apartheid. 
          
             (Skip several paragraphs with various citations)
          
             The term carries specific legal elements. Whether they 
             are satisfied “state-wide” or only in parts of the OPT 
             is the core dispute. Present consensus splits between 
             leading NGOs/UN experts who say the elements are met and 
             Israeli government–aligned and some academic voices who 
             say they are not. No binding court ruling settles it yet.
          
          Do you have a problem with that? I don't.
      • bongodongobob 30 minutes ago

        Try MS Copilot. That shit will end the conversation if anything remotely political comes up.

        • CamperBob2 23 minutes ago

          As long as it excludes politics in general, without overt partisan bias demanded by the government, what's the problem with that? If they want to focus on other subjects, they get to do that. Other models will provide answers where Copilot doesn't.

          Chinese models, conversely, are aligned with explicit, mandatory guardrails to exalt the CCP and socialism in general. Unless you count prohibitions against adult material, drugs, explosives and the like, that is simply not the case with US-based models. Whatever biases they exhibit (like the Grok example someone else posted) are there because that's what their private maintainers want.

  • resters 36 minutes ago

    I have no doubt that open source will triumph over whatever nonsense the US Government is trying to do to attack DeepSeek. Without DeepSeek, OpeanAI Pro and Claude Pro would probably cost $1000 per month each already.

    I suspect that Grok is actually DeepSeek with a bit of tuning.

  • xpe 42 minutes ago

    Please don't just read Eric Hartford's piece. Start with the key findings from the source material: "CAISI Evaluation of DeepSeek AI Models Finds Shortcomings and Risks" [1]. Here are the single-sentence summaries:

        DeepSeek performance lags behind the best U.S. reference models.
    
        DeepSeek models cost more to use than comparable U.S. models.
    
        DeepSeek models are far more susceptible to jailbreaking attacks than U.S. models.
    
        DeepSeek models advance Chinese Communist Party (CCP) narratives.
    
        Adoption of PRC models has greatly increased since DeepSeek R1 was released.
    
    [1] https://www.nist.gov/news-events/news/2025/09/caisi-evaluati...
    • evv 27 minutes ago

      It's funny how they mixed in proprietary models like GPT-5 and Anthropic with the "comparable U.S. models".

      Until they compare open-weight models, NIST is attempting a comparison between apples and airplanes.

  • JPKab 38 minutes ago

    The CCP literally revoked the visas of key DeepSeek engineers.

    That's all we need to know.

    • falcor84 33 minutes ago

      I would like to know more

      • _ache_ 9 minutes ago

        Deepseek starts out as a one-man operation. Like any company that has attracted a lot of attention, it becomes a "target" of the CCP, which then takes measures such as prohibiting key employees from leaving the country AND setting goals such as using Huawei chips instead of NVIDIA chips.

        From a Chinese political perspective, this is a good move in the long term. From Deepseek's perspective, however, this is clearly NOT the case, as it causes the company to lose some (or even most?) of its competitiveness and fall behind in the race.

      • FooBarWidget 18 minutes ago

        They revoke passports of personnel whom they deem are at risk of being negatively influenced or even kidnapped when abroad. Re influence, think school teachers. Re kidnapping, see Meng Wangzhou (Huawei CFO).

        There is a history of important Chinese personnel being kidnapped by e.g. the US when abroad. There is also a lot of talk in western countries about "banning Chinese [all presumed spies/propagandists/agents] from entering". On a good faith basis, one would think China banning people from leaving is a good thing that aligns with western desires, and should thus be applauded. So painting the policy as sinister tells me that the real desire is something entirely different.

        • UltraSane 8 minutes ago

          "There is a history of important Chinese personnel being kidnapped by e.g. the US when abroad"

          No there isn't. China revoked their passport to keep them prisoners not to keep them safe.

          "On a good faith basis, one would think China banning people from leaving is a good thing"

          Why would anyone think imprisoning someone like this is a good thing?

    • cowpig 11 minutes ago

      Source?

      And how is that "all we need to know"? I'm not even sure what your implication is.

      Is it that some CCP officials see DeepSeek engineers as adversarial somehow? Or that they are flight risks? What does it have to do with the NIST report?

    • manishsharan 25 minutes ago

      >> The CCP literally revoked the visas of key DeepSeek engineers. That's all we need to know.

      I don't follow. Why would DeepSeek engineers need visa from CCP?