- Added server-gpu and full-gpu Docker Compose profiles with NVIDIA CUDA support for GPU-accelerated inference - Created server-gpu.Dockerfile with PyTorch CUDA base image and GPU runtime configuration - Added compose.gpu.yaml for optional GPU allocation customization (device count, IDs, memory settings) - Documented GPU setup requirements for Linux (nvidia-container-toolkit), Windows WSL2, and macOS in docker/CLAUDE