local-llm-expert
sickn33/antigravity-awesome-skills
Expert guidance for private, offline local LLM inference: selecting optimal quant formats (GGUF, EXL2, AWQ), sizing VRAM/RAM, and running models with Ollama, llama.cpp, vLLM or LM Studio while keeping privacy-first deployments secure.