As per reports, Andrea Vallone, who previously served as the Head of 'Model Policy' Research at OpenAI, has taken up a position on Anthropic's Alignment Team. Her primary focus will be on establishing behavioral standards for AI systems operating in delicate contexts, notably in the realm of mental health. During her tenure at OpenAI, Vallone played a pivotal role in ensuring the secure implementation of GPT-4 and GPT-5, and was also involved in the creation of widely adopted safety training protocols. Reporting to Jan Leike, who has been a vocal critic of OpenAI's approach to safety, Vallone will persist in her exploration of the emotional safety parameters associated with AI.
