local-llm-expert
sickn33/antigravity-awesome-skills
An expert guide for deploying and optimizing Large Language Models (LLMs) entirely on local hardware. Mastering the entire local AI ecosystem, including inference engines like Ollama, vLLM, and llama.cpp. Provides deep knowledge on quantization formats (GGUF, EXL2, AWQ), VRAM calculation, and ensuring privacy-first, offline deployment for state-of-the-art models.