Quantized Llama models with increased speed and a reduced memory footprint October 24, 2024 by Comments