PhD Course on LLMs

Andrea T. will deliver an 8-hour lecture series for the PhD course The countdown begins! Just few days left until we kick off the Large Language Models: From BERT to GPT with Reinforcement Learning, starting on March 20, 2025.
Abstract: Research in artificial intelligence for natural language processing (NLP) has a long timeline, spanning over 70 years and characterized by several milestones, peaking with the recent pre-trained deep contextualized language models based on Transformers. Such models have gained tremendous success, bringing significant performance boosts in a wide range of tasks and benchmarks in NLP. Also, the principles underlying Transformers and their deep-learning architectural traits are used in the current state-of-the-art of Computer Vision and Speech Processing. In this context, the proposed course aims to provide an analysis of the key concepts and neural architectures that characterize the Transformer-based language models, with particular emphasis on the BERT models and on the generative Transformers, namely the GPT and InstructGPT family, which include the very recent and revolutionary ChatGPT. The course also offers a glimpse into future perspectives and new opportunities of artificial intelligence for supporting various application fields, possibly of high societal impact, such as law and healthcare.