eas/
codellama:34b-16k

246 1 year ago

Codelama with 16k context unlocked in modelfile

1 year ago

be95e5b84e7f · 16GB ·

llama
·
33.7B
·
Q3_K_M
{ "num_ctx": 16384, "num_gqa": 8, "rope_frequency_base": 1000000, "stop": [
[INST] <<SYS>>{{ .System }}<</SYS>> {{ .Prompt }} [/INST]

Readme

13b version is q4_K_M quantization

34b version is q3_K_M quantization