Research Assistant in Reliable AI Alignment
6 days ago
London
The position entails the development of algorithms, theoretical frameworks, and code for aligning large language models through Reinforcement Learning from Human Feedback (RLHF).This role demands expertise in leveraging recent advancements in active learning, robust optimization, and uncertainty ...