Local LLMs

Running language models locally, quantization, and inference