OpenAI has introduced a novel monitoring system for its AI reasoning models, o3 and o4-mini, designed to intercept prompts associated with biological and chemical threats. This system ensures that the models do not offer guidance for potentially harmful activities. Tailored to comprehend OpenAI's content policies, the system operates seamlessly alongside the models, detecting risk-related keywords and directing the models to refrain from offering dangerous advice.
