6,419 Downloads Updated 1 year ago
ollama run leeplenty/lumimaid-v0.2:8b
Updated 1 year ago
1 year ago
c86e6264e705 · 8.5GB ·
NFSW
The 8B model is based on: Meta-Llama-3.1-8B-Instruct
Wandb: https://wandb.ai/undis95/Lumi-Llama-3-1-8B?nw=nwuserundis95
The 12B model is based on: Mistral-Nemo-Instruct-2407
Wandb: https://wandb.ai/undis95/Lumi-Mistral-Nemo?nw=nwuserundis95
NOTE: As explained on Mistral-Nemo-Instruct-2407 repo, it’s recommended to use a low temperature, please experiment!
The 70B model is based on: Meta-Llama-3.1-70B-Instruct
Wandb: https://wandb.ai/undis95/Lumi-Llama-3-1-70B?nw=nwuserundis95
The 123B model is based on: Mistral-Large-Instruct
Wandb: https://wandb.ai/undis95/Lumi-Mistral-Large?nw=nwuserundis95
Lumimaid 0.1 -> 0.2 is a HUGE step up dataset wise.
As some people have told us our models are sloppy, Ikari decided to say fuck it and literally nuke all chats out with most slop.
Our dataset stayed the same since day one, we added data over time, cleaned them, and repeat. After not releasing model for a while because we were never satisfied, we think it’s time to come back!
Prompt template: Llama-3-Instruct
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
{output}<|eot_id|>
Prompt template: Mistral
<s>[INST] {input} [/INST] {output}</s>
Credits:
We sadly didn’t find the sources of the following, DM us if you recognize your set !
Datasets credits:
Others
Undi: If you want to support us, you can here.
IkariDev: Visit my retro/neocities style website please kek