Return to Article Details Bit-Width Quantization and Prompt Optimization: Achieving 90% Energy Savings in Large Language Models Download Download PDF