LLM Safety From Within: Detecting Harmful Content with Internal Representations - Databubble