2,004 7 months ago

(Unsloth Dynamic Quants) A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token.

7 months ago

9ce3ae0fb95d · 140GB ·

deepseek2
·
671B
·
IQ1_S
{ "stop": [ "<|begin▁of▁sentence|>", "<|end▁of▁sentence|>",
{{- range $i, $_ := .Messages }} {{- if eq .Role "user" }}<|User|> {{- else if eq .Role "assista

Readme

Note: this model requires Ollama 0.5.5 or later.

https://unsloth.ai/blog/deepseekr1-dynamic

https://huggingface.co/deepseek-ai/DeepSeek-V3

Quants computing resources are sponsored by Fujian Xiaowei Technology Co., Ltd. / 12301.cc / Mondrian Project