RSVP here
As language models achieve startling performance in complex fields like virology, this question becomes critically important.
Join Ruben Castaing from the University of Canterbury for a vital discussion on the intersection of AI and biosecurity. He will introduce techniques to precisely identify and remove dangerous biological weapons knowledge from within Large Language Models.
While addressing the current flaws and uncertainties of this method, Ruben will also explore how this work could be a key to solving other fundamental AI safety problems, including model deception and general misalignment.