Back to Writing
Challenges Jul 15, 2024 4 min read

Weekly Challenge: Optimizing Transformer Inference

This week’s challenge was to get a 7B parameter model running at >10 tokens/sec on a standard MacBook Air (M1) without destroying perplexity.

Approach

I utilized llama.cpp for 4-bit quantization (Q4_K_M). The results were surprisingly robust. The memory footprint dropped from ~14GB (FP16) to ~4GB, fitting entirely within the unified memory of the base M1 chip.

Thanks for reading.