161.4K Downloads Updated 1 year ago
Nous Hermes was released by Nous Research. There are two main variants here, a 13B parameter model based on Llama, and a 7B and 13B parameter model based on Llama 2. They are all general-use models trained with the same datasets.
The model used in the example below is the Nous Hermes Llama 2 model, with 7b parameters, which is a general chat model.
ollama serve
)curl -X POST http://localhost:11434/api/generate -d '{
"model": "nous-hermes",
"prompt":"Explain the process of how a refrigerator works to keep the contents inside cold."
}'
ollama run nous-hermes
Note: The ollama run
command performs an ollama pull
if the model is not already downloaded. To download the model without running it, use ollama pull nous-hermes
If you run into issues with higher quantization levels, try using the q4 model or shut down any other programs that are using a lot of memory.
Ollama offers many variants of the Nous Hermes model that are quantized based on the official models to run well locally.
Nous Hermes Llama 2 is the original Nous Hermes model based on the original Llama model.
Example: ollama run nous-hermes
Nous Hermes Llama 1 is the original Nous Hermes model based on the original Llama model.
Example: ollama run nous-hermes:13b-q4_0
By default, Ollama uses 4-bit quantization. To try other quantization levels, please try the other tags. The number after the q represents the number of bits used for quantization (i.e. q4 means 4-bit quantization). The higher the number, the more accurate the model is, but the slower it runs, and the more memory it requires.
Aliases |
---|
latest, 7b, 7b-llama2, 7b-llama2-q4_0 |
13b, 13b-llama2, 13b-llama2-q4_0 |
Nous Hermes Llama 2 source on Ollama
7b parameters original source: Nous Research
13b parameters original source: Nous Research
Nous Hermes Llama 1 source on Ollama
13b parameters original source: Nous Research