How to run AI models locally

Local model execution commands for VLLM for dual NVidia RTX6000 Pro

How to connect Ozeki AI Gateway to Ollama

How to run local AI Models on 24 GB Vram, RTX 3090 or RTX 4090