52 Downloads Updated 3 months ago
Aqui-VL 24B Mistral is an advanced language model based on Mistral Small 3.1, designed to deliver exceptional performance while remaining accessible on consumer-grade hardware. This is the first open weights model from Aqui Solutions, the company behind AquiGPT. With 23.6 billion parameters, it can run efficiently on a single RTX 4090 GPU or a 32GB Mac, making cutting-edge AI capabilities available to researchers, developers, and enthusiasts.
Aqui-VL 24B Mistral demonstrates competitive performance across multiple domains:
Benchmark | Aqui-VL 24B Mistral | Mistral Small 3.1 | Llama 3.1 70B |
---|---|---|---|
IFEval (Instruction Following) | 88.3% | 82.6% | 87.5% |
MMLU (General Knowledge) | 80.9% | 80.5% | 86.0% |
GPQA (Science Q&A) | 44.7% | 44.4% | 46.7% |
HumanEval (Coding) | 92.5% | 88.9% | 80.5% |
MATH (Mathematics) | 69.3% | 69.5% | 68.0% |
MMMU (General Vision) | 64.0% | 62.5% | N/A* |
ChartQA (Chart Analysis) | 87.6% | 86.2% | N/A* |
DocVQA (Document Analysis) | 94.9% | 94.1% | N/A* |
Average Text Performance | 75.1% | 73.2% | 73.7% |
Average Vision Performance | 82.2% | 80.9% | N/A* |
*Llama 3.1 70B does not include vision capabilities
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
# Load model and tokenizer
model_name = "aquigpt/aqui-vl-24b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype=torch.float16,
device_map="auto"
)
# Generate text
prompt = "Explain quantum computing in simple terms:"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=200, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
# Pull the model
ollama pull aquiffoo/aqui-vl-24b
# Run interactive chat
ollama run aquiffoo/aqui-vl-24b
# Download quantized model (Q4_K_M, 14.4GB)
wget https://huggingface.co/aquigpt/aqui-vl-24b/resolve/main/aqui-vl-24b-q4_k_m.gguf
# Run with llama.cpp
./main -m aqui-vl-24b-q4_k_m.gguf -p "Your prompt here" -n 100
With an 88.9% score on HumanEval, Aqui-VL 24B Mistral excels at: - Writing clean, efficient code in multiple languages - Debugging and code review - Algorithm implementation - Technical documentation
Strong vision capabilities enable: - PDF document analysis and Q&A - Chart and graph interpretation - Scientific paper comprehension - Business report analysis
Aqui-VL 24B Mistral is available exclusively in Q4_K_M quantization, optimized for the best balance of performance and hardware compatibility:
Aqui-VL 24B Mistral supports: - Parameter-efficient fine-tuning (LoRA, QLoRA) - Full fine-tuning for specialized domains - Custom tokenizer training - Multi-modal fine-tuning for specific vision tasks
This model is released under the Apache 2.0 License, making it suitable for both research and commercial applications.
For questions and support regarding Aqui-VL 24B Mistral, please visit the Hugging Face repository and use the community discussions section.
Built upon the excellent foundation of Mistral Small 3.1 by Mistral AI. Special thanks to the open-source community for tools and datasets that made this model possible.
Copyright 2025 Aqui Solutions. All rights reserved