Ollama with GPU
Based on the detailed guide from geek.sg:
-
Hardware Requirements
- A server with NVIDIA GPU (tested with RTX 3060 12GB)
- Minimum 32GB RAM recommended
- Sufficient storage space for models
-
Software Setup
- Install NVIDIA drivers
- Install NVIDIA Container Toolkit
- Configure Docker to use NVIDIA runtime
-
Coolify Configuration
- Deploy Ollama through Coolify’s one-click installer
- Modify the Docker compose configuration to include GPU support
- Add required environment variables for GPU acceleration
-
Model Management
- Pull and manage your preferred LLM models
- Monitor GPU usage and performance
- Adjust model parameters as needed
For the complete detailed guide, visit the original article.