Don't Waste Bits! Adaptive KV-Cache Quantization for Lightweight On-Device LLMs | ScienceToStartup