Roberto Dessì will present an overview of deep emergent communication, covering various central topics from the present and recent past, as well as discussing current shortcomings and suggest future directions.
Översikt
- Datum:Startar 20 november 2023, 14:00Slutar 20 november 2023, 15:00
- Plats:Zoom Link: https://chalmers.zoom.us/j/64432575085 Password: mondays24
- Språk:Engelska
Large pre-trained deep networks are the standard building blocks of modern AI applications. This raises fundamental questions about how to control their behaviour and how to make them efficiently interact with each other. The field of deep net emergent communication tackles these challenges by studying how to induce communication protocols between neural network agents, and how to include humans in the communication loop. Traditionally, this research had focussed on relatively small-scale experiments where two networks have to develop a discrete code from scratch for referential communication. However, with the rise of large pre-trained language models that can work well on many tasks, the emphasis is now shifting on how to let these models interact through a language-like channel in order to engage in more complex behaviors.
In this talk, I'll present an overview of deep emergent communication, covering various central topics from the present and recent past, as well as discussing current shortcomings and suggest future directions. The talks tries to bridge the gap between researchers wanting to develop more flexible AI systems, but also to cognitive scientists and linguists interested in the evolution of communication systems.
About the Speaker:
Roberto Dessì is a last-year Ph.D. student at Universitat Pompeu Fabra in Barcelona as part of an industrial PhD with Meta AI Paris and working under the supervision of Marco Baroni. His background is at the intersection of computer engineering (B.Sc. at La Sapienza in Rome) and cognitive science (M.Sc. at the University of Trento) with a strong research focus on understanding how to teach machines to understand and generate language in more human-like ways. Specifically, during his PhD he has worked on a variety of topics such as linguistic compositionality in sequence to sequence models, emergent communication, reinforcement learning (RL) and language, and most recently on large language models that can have access to external tools. He organized several iterations of the emergent communication workshop at NeurIPS and ICLR and his research has been published in top conferences like NeurIPS, EMNLP, ACL, ICLR, and CVPR.