383 9 months ago

QwQ is an experimental research model focused on advancing AI reasoning capabilities. i-matrix quantizations.

tools

Models

View all →

23 models

qwq-32b:latest

20GB · 32K context window · Text · 9 months ago

qwq-32b:q2_k

12GB · 32K context window · Text · 9 months ago

qwq-32b:q3_k_s

14GB · 32K context window · Text · 9 months ago

qwq-32b:q3_k_m

16GB · 32K context window · Text · 9 months ago

qwq-32b:q3_k_l

17GB · 32K context window · Text · 9 months ago

qwq-32b:q4_0

19GB · 32K context window · Text · 9 months ago

qwq-32b:q4_1

21GB · 32K context window · Text · 9 months ago

qwq-32b:q4_k_s

19GB · 32K context window · Text · 9 months ago

qwq-32b:Q4_K_M

20GB · 32K context window · Text · 9 months ago

qwq-32b:q5_0

23GB · 32K context window · Text · 9 months ago

qwq-32b:q5_1

25GB · 32K context window · Text · 9 months ago

qwq-32b:q5_k_s

23GB · 32K context window · Text · 9 months ago

qwq-32b:q5_k_m

23GB · 32K context window · Text · 9 months ago

qwq-32b:q6_k

27GB · 32K context window · Text · 9 months ago

qwq-32b:q8_0

35GB · 32K context window · Text · 9 months ago

qwq-32b:iq2_xxs

9.0GB · 32K context window · Text · 9 months ago

qwq-32b:iq2_xs

10.0GB · 32K context window · Text · 9 months ago

qwq-32b:iq2_s

10GB · 32K context window · Text · 9 months ago

qwq-32b:iq3_xxs

13GB · 32K context window · Text · 9 months ago

qwq-32b:iq3_xs

14GB · 32K context window · Text · 9 months ago

qwq-32b:iq3_s

14GB · 32K context window · Text · 9 months ago

qwq-32b:iq4_xs

18GB · 32K context window · Text · 9 months ago

qwq-32b:iq4_nl

19GB · 32K context window · Text · 9 months ago

Readme

  • Quantization from fp32
  • All models using i-matrix calibration_datav3.txt
  • iq2_s and q6_k barely works properly

QwQ is a 32B parameter experimental research model developed by the Qwen Team, focused on advancing AI reasoning capabilities.

image.png

image.png

QwQ is a 32B parameter experimental research model developed by the Qwen Team, focused on advancing AI reasoning capabilities.

QwQ demonstrates remarkable performance across these benchmarks:

  • 65.2% on GPQA, showcasing its graduate-level scientific reasoning capabilities
  • 50.0% on AIME, highlighting its strong mathematical problem-solving skills
  • 90.6% on MATH-500, demonstrating exceptional mathematical comprehension across diverse topics
  • 50.0% on LiveCodeBench, validating its robust programming abilities in real-world scenarios.

These results underscore QwQ’s significant advancement in analytical and problem-solving capabilities, particularly in technical domains requiring deep reasoning.

As a preview release, it demonstrates promising analytical abilities while having several important limitations:

  • Language Mixing and Code-Switching: The model may mix languages or switch between them unexpectedly, affecting response clarity.
  • Recursive Reasoning Loops: The model may enter circular reasoning patterns, leading to lengthy responses without a conclusive answer.
  • Safety and Ethical Considerations: The model requires enhanced safety measures to ensure reliable and secure performance, and users should exercise caution when deploying it.
  • Performance and Benchmark Limitations: The model excels in math and coding but has room for improvement in other areas, such as common sense reasoning and nuanced language understanding.