====== Running Ollama on the HPC Cluster ====== ===== Starting an interactive session for the ollama user service ===== * Load slurm utils (sinteractive) module load slurm/utils * Start an interactive session using sinteractive, see : https://hpc.wiki.utwente.nl/slurm:sinteractive sinteractive --gres=gpu:1 wait until you got resources assigned !! ==== Starting the ollama server and interacting with it ==== * Once you got an interactive session with gpu (a bit crowded right now), load the module ollama and start the user daemon module load ollama/0.1.41 start-ollama * Now you can use the download, list, run commands of ollama (here for example : run llama3) ollama run llama3 * Once completed you can stop the user daemon and terminate the interactive session stop-ollama exit