Skip to main content
Version: dev

Monitoring Ollama Local LLM Service

HertzBeat monitors Ollama local LLM service including version info, installed models and running models.

Preparation

Ensure that Ollama is running and the API is accessible. By default, Ollama listens on port 11434.

If Ollama is running on a remote server, you may need to set the OLLAMA_HOST environment variable to 0.0.0.0 to allow external access.

Configuration Parameters

Parameter NameParameter Description
Monitoring HostThe target IPV4, IPV6 or domain name of the Ollama service. Note: without protocol header (eg: https://, http://).
Task NameThe name that identifies this monitoring task, which must be unique.
PortThe port Ollama service is listening on, default is 11434.
SSLWhether to use HTTPS to connect to the Ollama service.
API KeyTo directly access the API key of ollama.com.
CollectorConfigure which collector is used to schedule data collection for this monitoring.
Monitoring IntervalThe interval for periodically collecting data, in seconds. The minimum interval that can be set is 30 seconds.
Bound TagsTags for categorizing and managing monitoring resources.
Description/RemarksAdditional remarks to identify and describe this monitoring. Users can add notes here.

Collection Metrics

Metric Set: Version Info

Metric NameMetric UnitMetric Description
VersionNoneThe version of the Ollama service.

Metric Set: Installed Models

Metric NameMetric UnitMetric Description
Model NameNoneThe name of the installed model.
Model SizeMBThe size of the model file.
Parameter SizeNoneThe parameter scale of the model (e.g., 7B, 13B).
Quantization LevelNoneThe quantization level of the model (e.g., Q4_0, Q8_0).
Model FamilyNoneThe model family (e.g., llama, qwen).
FormatNoneThe model format (e.g., gguf).
Modified AtNoneThe last modified time of the model.

Metric Set: Running Models

Metric NameMetric UnitMetric Description
Model NameNoneThe name of the running model.
Model SizeMBThe size of the model in memory.
VRAM SizeMBThe VRAM occupied by the model.
Expires AtNoneThe time when the model will be unloaded from memory.