Return to Article Details
Bit-Width Quantization and Prompt Optimization: Achieving 90% Energy Savings in Large Language Models
Download
Download PDF