DeepSeek just released updated r1 models with 'deeper and more complex reasoning patterns'. Includes a r1 distilled qwen3 8b model boasting "10% improved performance" over original
DeepSeek just released updated r1 models with 'deeper and more complex reasoning patterns'. Includes a r1 distilled qwen3 8b model boasting "10% improved performance" over original

huggingface.co
deepseek-ai/DeepSeek-R1-0528-Qwen3-8B · Hugging Face

WOAH
It can't answer The Tiananmen Square question. Gonna have to wait for an abliterated one.
I really hope the new r1 CoT reasoning patterns get trained into a mistral model, they're the only ones I half count on for decent uncensored base models. Keep an eye on the UGI chart too if thats something you care about. The best uncensored model I ever tried was is beepo 22b IMO.
That UGI chart is so flawed in terms of political lean. "Left-wing" is not libertarian capitalism - left-wing is socialism. Capitalism is a right ideology.
Maybe if you rotate the political compass 90 degrees clockwise and cut the top part out you'd be looking at libertarian capitalism on the "left" (with anarcho-capitalism on the "far-left").
Huihui already did a https://huggingface.co/huihui-ai/DeepSeek-R1-0528-Qwen3-8B-abliterated (GGUF quant: https://huggingface.co/Hamzah-Asadullah/DeepSeekR1-0528-8B-abliterated-GGUF)
But is abliteration enough for this? The AI just responds that it doesn’t have any info on that topic aka wasn’t trained any data relating to it. It’s not like they taught it to refuse, they simply didn’t teach it that it happened. To my understanding abliteration removes something, but we would need to add data for this.
EDIT: there is also
ollama run huihui_ai/deepseek-r1-abliterated:8b-0528-qwen3
, I just didn’t find it at firstI've gotten the deepseek-r1-0528-qwen3-8b to answer correctly once, but not consistently. Abliterated Deepseek models I've used in the past have been able to pass the test.