Forum for Artificial Intelligence

Archive


Forum for Artificial Intelligence

[ Home   |   About FAI   |   Upcoming talks   |   Past talks ]



This website is the archive for past Forum for Artificial Intelligence talks. Please click this link to navigate to the list of current talks.

FAI meets every other week (or so) to discuss scientific, philosophical, and cultural issues in artificial intelligence. Both technical research topics and broader inter-disciplinary aspects of AI are covered, and all are welcome to attend!

If you would like to be added to the FAI mailing list, subscribe here. If you have any questions or comments, please send email to Catherine Andersson.






[ Upcoming talks ]





Fri, September 24
11:00AM
Cynthia Matuszek
UMBC
Approaches to Grounded Language Acquisition from Human Interaction
Fri, October 1
11:00AM
Sam Bowman
NYU
Overclaiming Is a Problem. Underclaiming May Be Worse.
Fri, October 15
11:00AM
Pat Langley
Institute for the Study of Learning and Expertise and Center for Design Research, Stanford University
Forty Years of Machine Learning: Myths, Metaphors, Paradigms, and Challenges
Fri, October 22
11:00AM
Frank Dellaert
Georgia Tech
Factor Graphs for Perception and Action
Fri, October 29
11:00AM
Aniket Bera
UMD
Designing Emotionally-Intelligent Digital Humans that Move, Express, and Feel Like Us!
Fri, November 5
11:00AM
Matthew Walter
TTIC
Learning Better Ways to Measure and Move: Joint Optimization of an Agent's Physical Design and Computational Reasoning
Fri, November 12
11:00AM
Stefanos Nikolaidis
USC
Towards Robust HRI: A Quality Diversity Approach
Fri, November 19
11:00AM
Yejin Choi
University of Washington
David V.S. Goliath: the Art of Leaderboarding in the Era of Extreme-Scale Neural Models
Fri, December 10
11:00AM
Peter Clark
Allen Institute for AI
Systematic Reasoning and Explanation over Natural Language
Fri, February 11
11:00AM
Siddharth Srivastava
Arizona State University
Principles and Algorithms for Data-Efficient Assistive Sequential Decision Making
Fri, February 18
11:00AM
Tal Linzen
NYU
Causal analysis of the syntactic representations used by Transformers
Fri, March 4
11:00AM
Abdelrahman Mohamed
FAIR
Recent advances in speech representation learning
Fri, March 11
11:00AM
Gopal Ramchurn
University of Southampton
Trustworthy Human-AI Partnerships
Fri, March 25
11:00AM
Peter Stone
University of Texas at Austin
Outracing Champion Gran Turismo Drivers with Deep Reinforcement Learning
Fri, April 8
11:00AM
Aljosa Osep
Carnegie Mellon University
Unifying segmentation, tracking, and forecasting for 3D embodied scene understanding
Fri, April 15
11:00AM
Nigel G. Ward
UTEP
A Dimensional Model of Interaction Style Variation in Spoken Dialog
Mon, August 1
11:00AM
Charles Sutton
Google AI and University of Edinburgh
Program Synthesis, Program Semantics, and Large Language Models

Friday, September 24, 2021, 11:00AM



Approaches to Grounded Language Acquisition from Human Interaction

Cynthia Matuszek   [homepage]

UMBC

Abstract: As robots move from labs and factories into human-centric spaces, it becomes progressively harder to predetermine the environments and interactions they will need to be able to handle. Letting robots learn from end users via natural language is an intuitive, versatile approach to handling novel situations robustly. Grounded language acquisition is concerned with learning to understand language in the context of the physical world. In this presentation, I will give an overview of our work on using joint statistical models to learn the grounded semantics of natural language describing an agent's environment. I will discuss the role of speech in grounded language learning, including introducing a new dataset and results on learning directly from that speech, and will describe work on learning language in simulated environments.

About the speaker:

Bio: Cynthia Matuszek is an assistant professor of computer science and electrical engineering at the University of Maryland, Baltimore County, and the director of UMBC’s Interactive Robotics and Language lab. After working as a researcher on the Cyc project, she received her Ph.D. in computer science and engineering from the University of Washington in 2014. Her research is focused on how robots can learn grounded language from interactions with non-specialists, which includes work in not only robotics, but human-robot interactions, natural language, machine learning, machine bias, and collaborative robot learning, informed by a background in common-sense reasoning and classical artificial intelligence. Dr Matuszek has been named in the IEEE bi-annual “10 to watch in AI,” and has published in machine learning, artificial intelligence, robotics, and human-robot interaction venues.

Watch Online

Friday, October 1, 2021, 11:00AM



Overclaiming Is a Problem. Underclaiming May Be Worse.

Sam Bowman   [homepage]

NYU

Abstract: In an effort to avoid reinforcing widespread hype about the capabilities of state-of-the-art NLP systems, researchers have developed habits in framing and citation that serve to deemphasize the field's successes, even at the cost of making misleadingly strong claims about the limits of our best systems. This is a problem, though, and it may be more serious than it looks: It limits our ability to mitigate short-term harms from NLP deployments and it limits our ability to prepare for the impacts of highly effective future systems, which may be significantly greater. This paper urges researchers to be careful about these claims, and suggests some research directions that will make it easier to avoid or rebut them. This talk will draw primarily from a work-in-progress paper on this issue and secondarily from a recent NAACL position paper with George Dahl, with additional references to several other evaluation-related projects from my group.

About the speaker:

Bio: Sam Bowman has been on the faculty at NYU since 2016, when he completed PhD with Chris Manning and Chris Potts at Stanford. At NYU, he is a member of the Center for Data Science, the Department of Linguistics, and Courant Institute's Department of Computer Science. His research focuses on data, evaluation techniques, and modeling techniques for sentence and paragraph understanding in natural language processing, and on applications of machine learning to scientific questions in linguistic syntax and semantics. He is the senior organizer behind the GLUE and SuperGLUE benchmark competitions and he has received a 2015 EMNLP Best Resource Paper Award, a 2019 *SEM Best Paper Award, a 2017 Google Faculty Research Award, and a 2021 NSF CAREER award.

Watch Online

Friday, October 15, 2021, 11:00AM



Forty Years of Machine Learning: Myths, Metaphors, Paradigms, and Challenges

Pat Langley   [homepage]

Institute for the Study of Learning and Expertise and Center for Design Research, Stanford University

The machine learning community declared itself as a new branch of artificial intelligence at its first workshop in 1980. In this talk, I review the field's forty-year history, covering developments that led to its intellectual advances and its application successes. I also examine some myths that have emerged along the way, metaphors that have guided research, and paradigms that have drawn attention in different periods. In closing, I describe some open challenges that, despite recent impressive successes, suggest the field remains far from reproducing the breadth and depth of human learning.

About the speaker:

Dr. Pat Langley serves as Director of the Institute for the Study of Learning and Expertise and as Research Scientist at Stanford University's Center for Design Research. He has contributed to artificial intelligence and cognitive science for more than 40 years, having published over 300 papers and five books on these topics. Dr. Langley developed some of the first computational approaches to scientific knowledge discovery, and he was an early champion of experimental studies of machine learning and its application to real-world problems. He is the founding editor of two journals, Machine Learning in 1986 and Advances in Cognitive Systems in 2012, and he is a Fellow of both AAAI and the Cognitive Science Society. Dr. Langley's current research focuses on architectures for embodied agents, abductive methods for plan understanding, and learning procedures from written instructions.

Watch Online

Friday, October 22, 2021, 11:00AM



Factor Graphs for Perception and Action

Frank Dellaert   [homepage]

Georgia Tech

Factor graphs have been very successful in providing a lingua franca in which to phrase robotics perception and navigation problems. In this talk I will re-visit some of those successes, also discussed in depth in [a recent review article](https://t.co/Xc0RMXyeYY). However, I will focus on our more recent work in the talk, centered on using factor graphs for *action​*. In particular, I will discuss our efforts in motion planning, trajectory optimization, optimal control, and model-predictive control, highlighting in each how factor graphs provide an intuitive and natural framework in which to think about these problems and generate state of the art solutions.

About the speaker:

Bio: Frank Dellaert is a Professor at Georgia Tech's School of Interactive Computing, and a research scientist at Google AI. He has previously done stints at Skydio, a drone startup, and Facebook Reality Labs. His work is on sensor fusion and the use of large-scale graphical models for robotics sensing, thinking, and acting. With his students and collaborators, he created the popular sensor fusion/SLAM library GTSAM, see gtsam.org.

Watch Online

Friday, October 29, 2021, 11:00AM



Designing Emotionally-Intelligent Digital Humans that Move, Express, and Feel Like Us!

Aniket Bera   [homepage]

UMD

The creation of intelligent virtual agents (IVAs) or digital humans is vital for many virtual and augmented reality systems. As the world increasingly uses digital and virtual platforms for everyday communication and interactions, there is a heightened need to create human-like virtual avatars and agents endowed with social and emotional intelligence. Interactions between humans and virtual agents are being used in different areas including, VR, games and story-telling, computer-aided design, social robotics, and healthcare. Designing and building intelligent agents that can communicate and connect with people is necessary but not sufficient. Researchers must also consider how these IVAs will inspire trust and desire among humans. Knowing the perceived affective states and social-psychological constructs (such as behavior, emotions, psychology, motivations, and beliefs) of humans in such scenarios allows the agents to make more informed decisions and navigate and interact in a socially intelligent manner. In this talk, I will give an overview of our recent work on simulating intelligent, interactive, and immersive human-like agents who can also learn, understand and be sentient to the world around them using a combination of emotive gestures, gaits, and expressions. Finally, I will also talk about our many ongoing projects which use our AI-driven IVAs, including intelligent digital humans for urban simulation, crowd simulation, mental health and therapy applications, and social robotics

About the speaker:

Bio: Aniket Bera is an Assistant Research Professor at the Department of Computer Science. His core research interests are in Affective Computing, Computer Graphics (AR/VR, Augmented Intelligence, Multi-Agent Simulation), Autonomous Agents, Cognitive modeling and planning for intelligent characters. His work has won multiple awards at top Graphics/VR conferences. He has previously worked in many research labs, including Disney Research, Intel, and the Centre for Development of Advanced Computing. Aniket's research has been featured on CBS, WIRED, Forbes, FastCompany, Times of India, etc.

Watch Online

Friday, November 5, 2021, 11:00AM



Learning Better Ways to Measure and Move: Joint Optimization of an Agent's Physical Design and Computational Reasoning

Matthew Walter   [homepage]

TTIC

The recent surge of progress in machine learning foreshadows the advent of sophisticated intelligent devices and agents capable of rich interactions with the physical world. Many of these these advances focus on building better computational methods for inference and control---computational reasoning methods trained to discover and exploit the statistical structure and relationships in their problem domain. However, the design of physical interfaces through which a machine senses and acts in its environment is as critical to its success as the efficacy of its computational reasoning. Perception problems become easier when sensors provide measurements that are more informative towards the quantities to be inferred. Control policies become more effective when an agent's physical design permits greater robustness and dexterity in its actions. Thus, the problems of physical design and computational reasoning are coupled, and the answer to what combination is optimal naturally depends on the environment the machine operates in and the task before it. I will present learning-based methods that perform automated, data-driven optimization over sensor measurement strategies and physical configurations jointly with computational inference and control. I will first describe a framework that reasons over the configuration of sensor networks in conjunction with the corresponding algorithm that infers spatial phenomena from noisy sensor readings. Key to the framework is encoding sensor network design as a differential neural layer that interfaces with a neural network for inference, allowing for joint optimization using standard techniques for training neural networks. Next, I will present a method that draws on the success of data-driven approaches to continuous control to jointly optimize the physical structure of legged robots and the control policy that enables them to locomote. The method maintains a distribution over designs and uses reinforcement learning to optimize a shared control policy to maximize the expected reward over the design distribution. I will then describe recent work that extends this approach to the coupled design and control of physically realizable soft robots. If time permits, I will conclude with a discussion of ongoing work that seeks to improve test-time generalization of the learned policies.

About the speaker:

Matthew R. Walter is an assistant professor at the Toyota Technological Institute at Chicago. His interests revolve around the realization of intelligent, perceptually aware robots that are able to act robustly and effectively in unstructured environments, particularly with and alongside people. His research focuses on machine learning-based solutions that allow robots to learn to understand and interact with the people, places, and objects in their surroundings. Matthew has investigated these areas in the context of various robotic platforms, including autonomous underwater vehicles, self-driving cars, voice-commandable wheelchairs, mobile manipulators, and autonomous cars for (rubber) ducks. Matthew obtained his Ph.D. from the Massachusetts Institute of Technology and the Woods Hole Oceanographic Institution, where his thesis focused on improving the efficiency of inference for simultaneous localization and mapping.

Watch Online

Friday, November 12, 2021, 11:00AM



Towards Robust HRI: A Quality Diversity Approach

Stefanos Nikolaidis   [homepage]

USC

The growth of scale and complexity of interactions between humans and robots highlights the need for new computational methods to automatically evaluate novel algorithms and applications. Exploring the diverse scenarios of interaction between humans and robots in simulation can improve understanding of complex HRI systems and avoid potentially costly failures in real-world settings. In this talk, I propose formulating the problem of automatic scenario generation in HRI as a quality diversity problem, where the goal is not to find a single global optimum, but a diverse range of failure scenarios that explore both environments and human actions. I show how standard quality diversity algorithms can discover surprising and unexpected failure cases in the shared autonomy domain. I then discuss the development of a new class of quality diversity algorithms that significantly improve the search of the scenario space and the integration of these algorithms with generative models, which enables the generation of complex and realistic scenarios. Finally, I discuss applications in procedural content generation and human preference learning.

About the speaker:

Stefanos Nikolaidis is an Assistant Professor of Computer Science at the University of Southern California and leads the Interactive and Collaborative Autonomous Robotics Systems (ICAROS) lab. His research focuses on stochastic optimization approaches for learning and evaluation of human-robot interactions. His work leads to end-to-end solutions that enable deployed robotic systems to act optimally when interacting with people in practical, real-world applications. Stefanos completed his PhD at Carnegie Mellon's Robotics Institute and received an MS from MIT, a MEng from the University of Tokyo and a BS from the National Technical University of Athens. His research has been recognized with an oral presentation at NeurIPS and best paper awards and nominations from the IEEE/ACM International Conference on Human-Robot Interaction, the International Conference on Intelligent Robots and Systems, and the International Symposium on Robotics.

Watch Online

Friday, November 19, 2021, 11:00AM



David V.S. Goliath: the Art of Leaderboarding in the Era of Extreme-Scale Neural Models

Yejin Choi   [homepage]

University of Washington

Scale appears to be the winning recipe in today's leaderboards. And yet, extreme-scale neural models are still brittle to make errors that are often nonsensical and even counterintuitive. In this talk, I will argue for the importance of knowledge, especially commonsense knowledge, and demonstrate how smaller models developed in academia can still have an edge over larger industry-scale models, if powered with knowledge. First, I will introduce "symbolic knowledge distillation", a new framework to distill larger neural language models into smaller commonsense models, which leads to a machine-authored KB that wins, for the first time, over a human-authored KB in all criteria: scale, accuracy, and diversity. Next, I will introduce a new conceptual framework for language-based commonsense moral reasoning, and discuss how we can teach neural language models about complex social norms and human values, so that the machine can reason that “helping a friend” is generally a good thing to do, but “helping a friend spread fake news” is not. Finally, I will discuss an approach to multimodal script knowledge, which leads to new SOTA performances on a dozen leaderboards that require grounded, temporal, and causal commonsense reasoning.

About the speaker:

Bio: Yejin Choi is Brett Helsel Professor at the Paul G. Allen School of Computer Science & Engineering at the University of Washington and also a senior research manager at AI2 overseeing the project Mosaic. Her research focuses on commonsense knowledge and reasoning, language grounding with vision and perception, and AI for social good. She is a co-recepient of the ACL Test of Time award in 2021, the CVPR Longuet-Higgins Prize (test of time award) in 2021, the AAAI Outstanding Paper Award (best paper award) in 2020, the Borg Early Career Award (BECA) in 2018, the inaugural Alexa Prize Challenge in 2017, IEEE AI's 10 to Watch in 2016, and the Marr Prize (best paper award) at ICCV 2013. She received her Ph.D. in Computer Science at Cornell University and BS in Computer Science and Engineering at Seoul National University in Korea.

Watch Online

Friday, December 10, 2021, 11:00AM



Systematic Reasoning and Explanation over Natural Language

Peter Clark   [homepage]

Allen Institute for AI

Recent work has shown that transformers can be trained to reason *systematically* with natural language (NL) statements, answering questions with answers implied by a set of provided facts and rules, and even generating proofs for those conclusions. However, these systems required all the knowledge to be provided explicitly as input. In this talk, I will describe our current work on generalizing this to real NL problems, where the system produces faithful, entailment-based proofs for its answers, including materializing its own latent knowledge as needed for those proofs. The resulting reasoning-supported answers can then be inspected, debugged, and corrected by the user, offering new opportunities for interactive problem-solving dialogs, and taking a step towards "teachable systems" that can learn from such dialogs over time.

About the speaker:

Peter Clark (peterc@allenai.org) is a Senior Research Manager at the Allen Institute for AI (AI2) and leads the Aristo Project. His work focuses on natural language processing, machine reasoning, and world knowledge, and the interplay between these three areas.

Watch Online

Friday, February 11, 2022, 11:00AM



Principles and Algorithms for Data-Efficient Assistive Sequential Decision Making

Siddharth Srivastava   [homepage]

Arizona State University

Can we balance efficiency and reliability while designing assistive AI systems? What would such AI systems need to provide? In this talk I will present some of our recent work addressing these questions. In particular, I will show that a few fundamental principles of abstraction are remarkably effective in designing efficient and reliable AI systems that can learn, plan and act over extended horizons. Our results show that abstraction mechanisms are invaluable not only in improving the data efficiency of sequential decision making, but also in developing AI systems that can explain their own behavior to non-experts, and in computing user-interpretable assessments of the limits and capabilities of Black-Box AI systems. I will also present some of our work on learning the requisite abstractions in a bottom-up fashion. Throughout the talk I will highlight the theoretical guarantees that our methods provide along with results from empirical evaluations featuring AI systems such as decision support systems and physical robots.

About the speaker:

Siddharth Srivastava is an Assistant Professor of Computer Science in the School of Computing and Augmented Intelligence at Arizona State University. Prof. Srivastava was a Staff Scientist at the United Technologies Research Center in Berkeley. Prior to that, he was a postdoctoral researcher in the RUGS group at the University of California Berkeley. He received his PhD in Computer Science from the University of Massachusetts Amherst. His research interests include robotics and AI, with a focus on reasoning, planning, and acting under uncertainty. His work on integrated task and motion planning for household robotics has received coverage from international news media. He is a recipient of the NSF CAREER award, a Best Paper award at the International Conference on Automated Planning and Scheduling (ICAPS) and an Outstanding Dissertation award from the Department of Computer Science at UMass Amherst. He served as conference chair for ICAPS 2019 and currently serves as an Associate Editor for the Journal of AI Research.

Watch Online

Friday, February 18, 2022, 11:00AM



Causal analysis of the syntactic representations used by Transformers

Tal Linzen   [homepage]

NYU

The success of artificial neural networks in language processing tasks has underscored the need to understand how they accomplish their behavior, and, in particular, how their internal vector representations support that behavior. The probing paradigm, which has often been invoked to address this question, relies on the (typically implicit) assumption that if a classifier can decode a particular piece of information from the model's intermediate representation, then that information plays a role in shaping the model's behavior. This assumption is not necessarily justified. Using the test case of everyone's favorite syntactic phenomenon - English subject-verb number agreement - I will present an approach that provides much stronger evidence for the *causal* role of the encoding of a particular linguistic feature in the model's behavior. This approach, which we refer to as AlterRep, modifies the internal representation in question such that it encodes the opposite value of that feature; e.g., if BERT originally encoded a particular word as occurring inside a relative clause, we modify the representation to encode that it is not inside the relative clause. I will show that the conclusions of this method diverge from those of the probing method. Finally, if time permits, I will present a method based on causal mediation analysis that makes it possible to draw causal conclusions by applying counterfactual interventions to the *inputs*, contrasting with AlterRep which intervenes on the model's internal representations.

About the speaker:

Tal Linzen is an Assistant Professor of Linguistics and Data Science at New York University and a Research Scientist at Google. Before moving to NYU in 2020, he was a faculty member at Johns Hopkins University, a postdoctoral researcher at the École Normale Supérieure in Paris, and a PhD student at NYU. At NYU, Tal directs the Computational Psycholinguistics Lab, which develops computational models of human language comprehension and acquisition, as well as methods for interpreting and evaluating neural network models for language technologies.

Watch Online

Friday, March 4, 2022, 11:00AM



Recent advances in speech representation learning

Abdelrahman Mohamed   [homepage]

FAIR

Self-supervised representation learning methods recently achieved great successes in NLP and computer vision domains, reaching new performance levels while reducing required labels for many downstream scenarios. Speech representation learning is experiencing similar progress. This talk will present our recent work on self-supervised speech representation learning from audio and video (AV-HuBERT). Learning such high-quality speech representations also enabled our research on the Generative Spoken Language Modeling (GSLM) tasks. I’ll share our recent advances on that front.

About the speaker:

Abdelrahman Mohamed is a research scientist at Facebook AI Research (FAIR). Before FAIR, he was a principal scientist/manager in Amazon Alexa and a researcher in Microsoft Research. Abdelrahman received his Ph.D. from the University of Toronto, working with Geoffrey Hinton and Gerald Penn as part of the team that started the Deep Learning revolution in Spoken Language Processing in 2009. He is the recipient of the IEEE Signal Processing Society Best Journal Paper Award for 2016. His research interests span Deep Learning, Spoken Language Processing, and Natural Language Understanding. Abdelrahman has been focusing lately on improving, using, and benchmarking learned speech representations, e.g. HuBERT, Wav2vec 2.0, TextlessNLP, and SUPERB.

Watch Online

Friday, March 11, 2022, 11:00AM



Trustworthy Human-AI Partnerships

Gopal Ramchurn   [homepage]

University of Southampton

Recent advances in AI, Machine learning and Robotics have significantly enhanced the capabilities of machines. Machine intelligence is now able to support human decision making, augment human capabilities, and, in some cases, take over control from humans and act fully autonomously. Machines are becoming more tightly embedded into systems alongside humans, interacting and influencing each other in a number of ways. Such human-AI partnerships are a new form of socio-technical system in which the potential synergies between humans and machines are much more fully utilised. Designing, building, and deploying human-AI partnerships present a number of new challenges as we begin to understand their impact on our physical and mental well-being, our personal freedoms, and those of the wider society. In this talk I will focus on the challenges in designing trustworthy human-AI partnerships. I will explore the multiple elements of trust in human-AI partnerships and discuss the associated research challenges.I will also aim to identify the risks associated with human-AI partnerships and therefore determine the associated measures to mitigate these risks. I will conclude by giving a brief overview of the UKRI Trustworthy Autonomous Systems Programme (www.tas.ac.uk), a £33m programme launched in 2020 involving over 20 universities, 100+ industry partners, and over 200 researchers.

About the speaker:

Prof. Sarvapali Ramchurn is a Professor of Artificial Intelligence, Turing Fellow, and Fellow of the Institution of Engineering and Technology. He is the Director of the UKRI Trustworthy Autonomous Systems hub (www.tas.ac.uk) and Co-Director of the Shell-Southampton Centre for Maritime Futures. He is also a Co-CEO of Empati Ltd, an AI startup working on decentralised green hydrogen technologies. His research is about the design of Responsible Artificial Intelligence for socio-technical applications including energy systems and disaster management. He has won multiple best paper awards for his research and is a winner of the AXA Research Fund Award (2018) for his work on Responsible Artificial Intelligence. He has pioneered the development of AI-based disaster response systems using multi-UAV systems, AI-driven large-scale battery management for smart grids, and an AI bot that outperformed more than 5M human players (top 0.7%) in the English Premier League Fantasy Football Tournament. His papers have been cited more than 9000 times (according to Google scholar). He is originally from Mauritius and is interested in promoting applications of AI for environmental sustainability.

Watch Online

Friday, March 25, 2022, 11:00AM



Outracing Champion Gran Turismo Drivers with Deep Reinforcement Learning

Peter Stone   [homepage]

University of Texas at Austin

Many potential applications of artificial intelligence involve making real-time decisions in physical systems while interacting with humans. Automobile racing represents an extreme example of these conditions; drivers must execute complex tactical manoeuvres to pass or block opponents while operating their vehicles at their traction limits. Racing simulations, such as the PlayStation game Gran Turismo, faithfully reproduce the non-linear control challenges of real race cars while also encapsulating the complex multi-agent interactions. Here we describe how we trained agents for Gran Turismo that can compete with the world's best e-sports drivers. We combine state-of-the-art, model-free, deep reinforcement learning algorithms with mixed-scenario training to learn an integrated control policy that combines exceptional speed with impressive tactics. In addition, we construct a reward function that enables the agent to be competitive while adhering to racing's important, but under-specified, sportsmanship rules. We demonstrate the capabilities of our agent, Gran Turismo Sophy, by winning a head-to-head competition against four of the world's best Gran Turismo drivers. By describing how we trained championship-level racers, we demonstrate the possibilities and challenges of using these techniques to control complex dynamical systems in domains where agents must respect imprecisely defined human norms.

About the speaker:

Dr. Peter Stone is the David Bruton, Jr. Centennial Professor and Associate Chair of Computer Science, as well as Director of Texas Robotics, at the University of Texas at Austin. In 2013 he was awarded the University of Texas System Regents' Outstanding Teaching Award and in 2014 he was inducted into the UT Austin Academy of Distinguished Teachers, earning him the title of University Distinguished Teaching Professor. Professor Stone's research interests in Artificial Intelligence include machine learning (especially reinforcement learning), multiagent systems, and robotics. Professor Stone received his Ph.D in Computer Science in 1998 from Carnegie Mellon University. From 1999 to 2002 he was a Senior Technical Staff Member in the Artificial Intelligence Principles Research Department at AT&T Labs - Research. He is an Alfred P. Sloan Research Fellow, Guggenheim Fellow, AAAI Fellow, IEEE Fellow, AAAS Fellow, ACM Fellow, Fulbright Scholar, and 2004 ONR Young Investigator. In 2007 he received the prestigious IJCAI Computers and Thought Award, given biannually to the top AI researcher under the age of 35, and in 2016 he was awarded the ACM/SIGAI Autonomous Agents Research Award. Professor Stone co-founded Cogitai, Inc., a startup company focused on continual learning, in 2015, and currently serves as Executive Director of Sony AI America.

Watch Online

Friday, April 8, 2022, 11:00AM



Unifying segmentation, tracking, and forecasting for 3D embodied scene understanding

Aljosa Osep   [homepage]

Carnegie Mellon University

The future of AI is embodied. Think Alexa but with a robot body that can actually help out at home, or robotaxis to easily take you wherever, whenever, and however. But how do we get there? From the perception point of view, we first need to answer questions such as where I am or where I need to go, what is around me, and how it moves. I will start with the question of what is around me and present our recent efforts toward the unification of temporal 3D semantic and instance segmentation for streaming lidar perception. I will then zoom into the temporal aspect of this task, tackling the question of how does it move and talk about how far we can get in 3D object tracking by purely relying on modeling geometric relations using graph neural networks. I will continue with how it could move and discuss how end-to-end forecasting can be posed as future object detection. Finally, I will conclude my talk with the question of where I need to go and talk about city-scale position localization based on textural descriptions of the visual surroundings.

About the speaker:

Aljosa Osep is currently working as a postdoctoral fellow at the Robotics Institute of Carnegie Mellon University in Pittsburgh and Dynamic Vision and Learning Group at the Technical University in Munich. When not exploring the world, he is working on cool research problems that lie on the intersection of computer vision, robotics, and machine learning. His current focus is on scaling object detection, segmentation, and tracking methods to the open world, in which future robots will need to operate.

Watch Online

Friday, April 15, 2022, 11:00AM



A Dimensional Model of Interaction Style Variation in Spoken Dialog

Nigel G. Ward   [homepage]

UTEP

In spoken dialog people vary their interaction styles, and dialog systems should also be able to do so. Previous work has elucidated many aspects of style variation and adaptation, but a general model has been lacking. We here describe a dimensional model of the space of interaction styles, derived by applying Principal Component Analysis to a large sampling of American English conversations, each represented by 84 novel features that encode the frequencies of diverse interaction-related prosodic behaviors. The top 8 dimensions were meaningfully interpretable, and include aspects previously noted in the literature, such as engaged vs. uninterested, but also new ones, such as positive assessment vs. negative feeling and factual vs. thoughtful. Further, regarding individual differences in interaction style, we find that individual style tendencies were surprisingly weak, with a predictive model based on individual tendencies only barely outperforming a speaker-independent model. There are interesting implications for dialog systems design, tuning, and adaptation.

About the speaker:

Nigel G. Ward is Professor of Computer Science at the University of Texas at El Paso. He received his Ph.D. from the University of California at Berkeley. On the faculty of the University of Tokyo for ten years before joining UTEP, in 2015-2016 he was a Fulbright Scholar and Visiting Professor at Kyoto University. He is chair of ISCA’s Speech Prosody SIG and the author of Prosodic Patterns in English Conversation. He is organizing the Special Session on Speaking Styles and Interaction Styles at Interspeech 2022. Ward's research areas are at the intersection of spoken language and human-computer interaction. Current topics include the subtle non-lexical and prosodic signals that enable inference of a dialog partner's needs, intentions, and feelings at the sub-second level; and ways to model and exploit these phenomena to improve dialog system responsiveness, information retrieval from audio, and language assessment and teaching. These projects apply multiple methods: linguistic, corpus-based modeling, systems building, and experimental.

Watch Online

Monday, August 1, 2022, 11:00AM



Program Synthesis, Program Semantics, and Large Language Models

Charles Sutton   [homepage]

Google AI and University of Edinburgh

I will describe our experience with two generations of large language models for code at Google. These models show a range of abilities, including generating small programs from natural language descriptions and engaging in dialog about code, incorporating human feedback to improve solutions. However, in a deeper sense, these models seem not to understand the code that they write, in the sense that they are generally unable to predict the output of a program given a specific input. I will discuss our subsequent efforts to improve the "code understanding" abilities of LMs, by asking them to emit intermediate computation steps as tokens onto a "scratchpad". These same models are able to perform complex multi-step computations when asked to perform the operation "step by step", showing the results of intermediate computations, even operations that the LM could not perform directly.

About the speaker:

Charles Sutton is a Research Scientist at Google Brain and a Reader (equivalent to Associate Professor: http://bit.ly/1W9UhqT) in Machine Learning at the University of Edinburgh. He has published over 75 papers in probabilistic machine learning and deep learning, motivated by the demands of a broad range of applications, including natural language processing (NLP), analysis of computer systems, sustainable energy, data analysis, and software engineering. His work in machine learning for software engineering has won two ACM Distinguished Paper Awards. He has served as Director of the EPSRC Centre for Doctoral Training in Data Science at the University of Edinburgh, and has previously been a Fellow of the Alan Turing Institute, the UK’s national research institute for artificial intelligence and data science.

[ FAI Archives ]

Fall 2022 - Spring 2023

Fall 2021 - Spring 2022

Fall 2020 - Spring 2021

Fall 2019 - Spring 2020

Fall 2018 - Spring 2019

Fall 2017 - Spring 2018

Fall 2016 - Spring 2017

Fall 2015 - Spring 2016

Fall 2014 - Spring 2015

Fall 2013 - Spring 2014

Fall 2012 - Spring 2013

Fall 2011 - Spring 2012

Fall 2010 - Spring 2011

Fall 2009 - Spring 2010

Fall 2008 - Spring 2009

Fall 2007 - Spring 2008

Fall 2006 - Spring 2007

Fall 2005 - Spring 2006

Spring 2005

Fall 2004

Spring 2004

Fall 2003

Spring 2003

Fall 2002

Spring 2002

Fall 2001

Spring 2001

Fall 2000

Spring 2000