Prompt Compression in the Wild: Measuring Latency, Rate Adherence, and Quality for Faster LLM Inference | ScienceToStartup