Ollama
The Ollama free and open-source software for easily getting up and running with AI models can be used through containerisation with Podman (see Podman). An example of using the software to run a model is given below.
Using Ollama to run Meta’s Llama 4 Scout Model On an NVIDIA H100 GPU
To run the Llama 4 Scout model on an NVIDIA H100 GPU on the Lovelace cluster simply start an interactive GPU session (see Accessing Compute Resources), create a folder for Ollama, start an Ollama instance in this session, and then execute commands with the session.
# Acquire a GPU session
# This is subject to demand and availability of H100 GPUs
# It may be worth running on a different resources
srun --pty -p gpu_h100 --gpus 1 bash
# Create a folder for ollama
mkdir -p "$HOME/.ollama"
# Start an Ollama instance
podman run -d -v "$HOME/.ollama:/root/.ollama" --name ollama --device nvidia.com/gpu=all docker.io/ollama/ollama
# Run the model with the Ollama instance
podman exec -it ollama ollama run llama4:scout
You will then see a prompt and will be able to ask the Llama 4 model questions. Other ways of interacting with Ollama are available but not covered here.
Once done, bring down your Ollama instance and exit your GPU session.
podman kill ollama
podman rm ollama
exit