66 1 year ago

AbacusAI Dracarys2 72B

tools

1 year ago

e9b11aa77172 · 47GB ·

qwen2
·
72.7B
·
Q4_K_M
{{- if .Messages }} {{- if or .System .Tools }}<|im_start|>system {{- if .System }} {{ .System }} {{
{ "num_batch": 128, "num_ctx": 50000, "num_keep": 512, "temperature": 0.1, "top_

Readme

AbacusAI Dracarys2 72B

As featured on the Aider Leaderboard this fine tune of Qwen 2.5 72b is suited to coding tasks.

It’s highly recommended to run Ollama with K/V cache quantisation set to Q8_0 with Ollama build from the PR that adds this (https://github.com/ollama/ollama/pull/6279) to 12 the amount of vRAM used by the context.

Defaults:

num_ctx 50K To be useful for medium-larger coding tasks
num_batch 128 To reduce memory overhead of the larger context
num_keep 512 To improve context overflow for coding
temperature 0.1 To reduce hallucinations
top_p 0.8 To increase quality