Open AI ran a workshop on aligning AI agents with human values in early March, bringing together researchers from machine learning and AI safety to discuss risks and possible solutions from ever-more-capable large language models. PI Anca Dragan (UC Berkeley) served on the panel discussing risks and addressed the issue of preference influence as a core challenge for alignment.