Retrieve a list of models that are currently running
curl http://localhost:11434/api/ps{
"models": [
{
"model": "gemma3",
"size": 6591830464,
"digest": "a2af6cc3eb7fa8be8504abaf9b04e88f17a119ec3f04a3addf55f92841195f5a",
"details": {
"parent_model": "",
"format": "gguf",
"family": "gemma3",
"families": [
"gemma3"
],
"parameter_size": "4.3B",
"quantization_level": "Q4_K_M"
},
"expires_at": "2025-10-17T16:47:07.93355-07:00",
"size_vram": 5333539264,
"context_length": 4096
}
]
}Models currently loaded into memory
Currently running models
Show child attributes
Name of the running model
Size of the model in bytes
SHA256 digest of the model
Model details such as format and family
Time when the model will be unloaded
VRAM usage in bytes
Context length for the running model
curl http://localhost:11434/api/ps{
"models": [
{
"model": "gemma3",
"size": 6591830464,
"digest": "a2af6cc3eb7fa8be8504abaf9b04e88f17a119ec3f04a3addf55f92841195f5a",
"details": {
"parent_model": "",
"format": "gguf",
"family": "gemma3",
"families": [
"gemma3"
],
"parameter_size": "4.3B",
"quantization_level": "Q4_K_M"
},
"expires_at": "2025-10-17T16:47:07.93355-07:00",
"size_vram": 5333539264,
"context_length": 4096
}
]
}