Configurations
Using GPU
Runing Ollama models on GPU
Presenton supports GPU acceleration when using Ollama models, significantly improving performance β especially for larger models.
To enable GPU support, you need to install and configure the NVIDIA Container Toolkit.
π οΈ Step 1: Install NVIDIA Container Toolkit
Follow the official guide to install the toolkit:
π https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html
π Step 2: Run Presenton with GPU
Once installed, use the --gpus=all
flag when running the container:
- Running without environment variables
- Running with environment variables