Are Latent Reasoning Models Easily Interpretable? | ScienceToStartup