LLM Safety From Within: Detecting Harmful Content with Internal Representations | ScienceToStartup