Users experimenting with DeepSeek have seen the Chinese AI chatbot reply and then censor itself in real time, providing an arresting insight into its control of information and opinion.

  • observantTrapezium@lemmy.ca
    link
    fedilink
    arrow-up
    7
    ·
    1 day ago

    I downloaded the 70B model and tried politically “naughty” questions. Even without the chatbot guardrails, it mostly says things that the CCP would approve of, but you could trick it to be more honest (not super easy!). One interesting thing is that while it usually spews this <think></think> blocks, for some politically sensitive questions (“is Taiwan part of China”) it just spits the answer.

    • RedstoneValley@sh.itjust.works
      link
      fedilink
      arrow-up
      7
      ·
      1 day ago

      I experimented with a local installation as well. The censored answers were not going to through the chain-of-thought routine, but were instant answers instead. Follow-up questions however made it spill the beans rather quickly, giving out even more juicy details than I had initially asked for.

  • andrew0@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    14
    arrow-down
    6
    ·
    1 day ago

    Wait until they learn that OpenAI does the exact same thing. Try to get advice on how to crack software and see how far you get.

  • The Snark Urge@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    2
    ·
    1 day ago

    DeepSeek and all LLMs are all massively overvalued, but also, isn’t it fun to watch corpo media turn on a dime in service of shareholders to rebuild sentiment in tech stocks. Show’s over folks, back to line-goey-uppy.

    Tinfoil hat mode: Watch NVIDIA stock recover in a week like this wasn’t some manufactured rug pull. Probably by friends of that Intel exec who went on the news talking about how he bought the dip. Not being super serious, but it’d be hardly the most blatant manipulation we’ve seen.

  • bdonvrA
    link
    fedilink
    arrow-up
    5
    arrow-down
    1
    ·
    1 day ago

    Lol that’s not at all unique to DeepSeek. I remember recording my screen to see outputs on other models before they censored the message

    • Pennomi@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      edit-2
      1 day ago

      One thing that’s so weird to me is that while DeepSeek is somewhat too large to run on my PC, just it existing makes it easy to distill the reasoning functionality into other smaller models. I’ve been running a 34B distill locally and it’s been much better than any other local model I’ve tried so far.