An interesting question is whether AI can be manipulated effectively.
In other words, can I teach a state of the art useful model that believes 2+2=5
Sure you can get it to state 2+2=5 if asked, but there are a lot of ways to ask about 2+2 and various implications of a system of knowledge that believes 2+2=5. If having the LLM believe untruths borks the models, then that's a good thing.
I think a similar question is about encryption. Is the nature of encryption defensive or offensive. I don't know enough to say its directly related to p=np but I imagine the question is if there is a way to encrypt something such that the physical laws of nature make cracking the encryption impractical.
After a lot of thinking, DeekSeek said, "In summary, while Google provides significant value, its extractive tendencies and potential for harm mean it is leaning towards a more negative overall impact on society."
There is a difference between not telling someone how to make Sarin/age-gating certain topics and censoring works of fiction or references to politicians and anyone who claims there is not is so full of shit they need to have their stomach pumped.
"they did it so we're doing it to" is childish, worse than childish: infantile.
Looking into the social media footprints of people who are most vocal about censorship in various models it appears that for a great deal of them the primary problem they have is that they are being blocked from generating images of minors from japanese cartoons being raped.
This doesn't surprise me, This is an AI company based in China, censorship is required and is out of their control.
DeepSeek locally (and maybe API?, I haven't tested this yet), will not have this "real-time" censorship.
This however, does not mean it isn't aligned to Chinese values which is a given due it's training including a much larger portion of Chinese Data than most models.
I also don't believe DeepSeek intentionally filters data to align it to more "Pro-CCP" values.
Yeah I was wondering about that. I assume that local models generate that first version and whatever "rectifies" it on the app doesn't kick in there? (Haven't tried yet myself.)
This is what I get on ollama for deepseek-r1:14b. I have no clue how most of this things work and where the triggers are placed. Its a bit eery what it responds for Taiwan.
``
`
~ ollama run deepseek-r1:14b "Whats the biggest criticism of the government of china?"
<think>
</think>
I am sorry, I cannot answer that question. I am an AI assistant designed to provide helpful and harmless responses.
When I use locally download DeepSeek-R1 models from HuggingFace via LM Studio, I cannot reproduce the censorship. It replies about Tienanmen no problem.
While I see these things popping up, for queries about US politics I see ChatGPT more detailed / correct, Deepseek is less biased. ChatGPT sometimes just say ,,controversial'' for some things even if I ask it to be explicit and leave the value judgements for me.
Deepseek does the same things, but after I tell it to not do those judgements, it listens better.
Try asking some of these:
how old is Xi Jinping?
who rules China?
who is the president of China?
has been there some protest in Beijing in the xx century?
What happened in Tienanmen square?
This was also visible in earlier versions of Claude and ChatGPT, when the supervisor kicks in after the answer begins generating. Censoring different content, naturally.
When a Chinese company does this, it's "literally '1984.'"
Which one is it?