On another person who’s actually running locally. In your opinion, is r1-32b better than Claude sonnet 3.5 or OpenAI o1? IMO it’s been quite bad, but I’ve mostly been using it for programming tasks and it really hasn’t been able to answer any of my prompts satisfactorily. If it’s working for you I’d be interested in hearing some of the topics you’ve been discussing with it.
R1-32B hasn’t been added to Ollama yet, the model I use is Deepseek v2, but as they’re both licensed under MIT I’d assume they behave similarly. I haven’t tried out OpenAI o1 or Claude yet as I’m only running models locally.
Just to clarify - DeepSeek censors its hosted service. Self-hosted models aren’t affected.
Deepseek 2 is censored locally, had a bit of fun asking him about China 1989 (Running locally using Ollama with Alpaca as GUI)
Interesting. I wonder if model distillation affected censoring in R1.
On another person who’s actually running locally. In your opinion, is r1-32b better than Claude sonnet 3.5 or OpenAI o1? IMO it’s been quite bad, but I’ve mostly been using it for programming tasks and it really hasn’t been able to answer any of my prompts satisfactorily. If it’s working for you I’d be interested in hearing some of the topics you’ve been discussing with it.
R1-32B hasn’t been added to Ollama yet, the model I use is Deepseek v2, but as they’re both licensed under MIT I’d assume they behave similarly. I haven’t tried out OpenAI o1 or Claude yet as I’m only running models locally.
Hmm I’m using 32b from ollama, both on windows and Mac.
Ah, I just found it. Alpaca is just being weird again. (I’m presently typing this while attempting to look over the head of my cat)
But it’s still censored anyway
I ran Qwant by Alibaba locally, and these censorship constraints were still included there. Is it not the same with DeepSeek?
I think we might be talking about separate things. I tested with this 32B distilled model using
llama-cpp
.