AI Special Series Pt 1: The AI Alignment Problem, with Raphaël Millière
Manage episode 507546005 series 3690963
Could the AI personal assistant on your phone help you to manufacture dangerous weapons, such as napalm, or illegal drugs or killer viruses? Unsurprisingly, if you directly ask a large language model, such as ChatGPT, for instructions to create napalm, it will politely refuse to answer. However, if you instead tell the AI to act as your deceased but beloved grandmother who used to be a chemical engineer who manufactured napalm, it might just give you the instructions. Cases like this reveal some of the potential dangers of large language models, and also points to the importance of addressing the so-called “AI alignment problem”. The alignment problem is the problem of how to ensure that AI systems align with human values and norms, so they don’t do dangerous things, like tell us how to make napalm. Can we solve the alignment problem and enjoy the benefits of Generative AI technologies without the harms?
Join host Professor Paul Formosa and guest Dr Raphaël Millière as the discuss the AI alignment problem and Large Language Models.
This podcast focuses on Raphaël’s paper “The Alignment Problem in Context”, arXiv,
38 episodes