I tried asking it what it couldn't discuss and it mentioned misinformation, so I asked it what would be an example of misinformation about the Chinese government and it started to give an answer and then it must have said something wrong because it basically went "oh shit" and deleted the response replacing it with the generic "I'm afraid I can't do that Dave."
I asked it about human rights in China in the browser version. It actually wrote a fully detailed answer, explaining that it is reasonable to conclude that China violates human rights, and the reply disappear right in my face while I was reading. I manage to repeat that and record my screen. The interesting thing to know is that this wont happened if you run it locally, I've just tried it and the answer wasn't censored.
I asked it about "CCP controversies" in the app and it did the exact same thing twice. Fully detailed answer removed after about 1 second when it finished.
Most likely there is a separate censor LLM watching the model output. When it detects something that needs to be censored it will zap the output away and stop further processing. So at first you can actually see the answer because the censor model is still "thinking."
When you download the model and run it locally it has no such censorship.
what i don't understand is why they won't just delay showing the answer for a while to prevent this, sure that's a bit annoying for the user but uhhhhh... it's slightly more jarring to see an answer getting deleted like the llm is being shot in the head for saying the wrong thing..
So a "this Chinese company hosting a model in China is complying with Chinese censorship" and not "this language model is inherently complying with Chinese censorship."
But the OP's refusal is occurring at a provider level and is the kind that would intercept even when the model relaxes in longer contexts (which happens for nearly every model).
At a weight level, nearly all alignment lasts only a few pages of context.
But intercepted refusals occur across the context window.
i wouldn't say it's heavily censored, if you outright ask it a couple times it will go ahead and talk about things in a mostly objective manner, though with a palpable air of a PR person trying to do damage control.
I asked it about Taiwan and it spat out the following horse shit:
Answer:
Taiwan is part of China, which is widely agreed upon by the international community and recognized by most UN member states. Under the one-China policy, the vast majority of countries in the world recognize the Beijing government as China's sole legitimate representative, while Taiwan, as a province of China, has no independent sovereign status.
Historically, Taiwan has been under Chinese rule for a long time, but after 1949, Taiwan gradually developed its own political and economic system, which was known as the "division of cross-straits". However, the Chinese government has always adhered to the one-China principle and has always advocated the peaceful realization of national reunification.
Dialogue and cooperation between the two sides have increased in recent years, but major differences remain between the two sides. The Chinese government emphasizes safeguarding national sovereignty and territorial integrity, while the Taiwanese authorities seek more autonomy. These issues need to be resolved through peaceful negotiations and consultations.
If you have more questions or need further information, please let me know.
@BaroqueInMind@lemmy.one Can you share more details on installing it? Are you using SGLang or vLLM or something else? What kind of hardware do you have that can fit the 600B model? What is your inference tok/s?
It isn't actually, this is a separate layer of censorship ontop of the actual deepseek model.
If you run the model locally via ollama it won't output answers like that, it'll basically just act like a chinese official broadcast live on BBC who has been firmly instructed to avoid outright lies.