Running Ollama on the HPC Cluster

  • Load slurm utils (sinteractive)
module load slurm/utils
sinteractive --gres=gpu:1

wait until you got resources assigned !!

  • Once you got an interactive session with gpu (a bit crowded right now), load the module ollama and start the user daemon
module load ollama/0.1.41
start-ollama
  • Now you can use the download, list, run commands of ollama (here for example : run llama3)
ollama run llama3
  • Once completed you can stop the user daemon and terminate the interactive session
stop-ollama
exit