Forum for Artificial Intelligence

Archive


Forum for Artificial Intelligence

[ Home   |   About FAI   |   Upcoming talks   |   Past talks ]



This website is the archive for past Forum for Artificial Intelligence talks. Please click this link to navigate to the list of current talks.

FAI meets every other week (or so) to discuss scientific, philosophical, and cultural issues in artificial intelligence. Both technical research topics and broader inter-disciplinary aspects of AI are covered, and all are welcome to attend!

If you would like to be added to the FAI mailing list, subscribe here. If you have any questions or comments, please send email to Catherine Andersson.






[ Upcoming talks ]





Fri, September 19
11:00AM
Jivko Sinapov
University of Texas at Austin
Behavior-Grounded Multisensory Object Perception and Exploration by a Humanoid Robot
Fri, October 3
11:00AM
George Konidaris
Duke University
Robots, Skills, and Symbols
Fri, November 7
11:00AM
Byron Wallace
School of Information, University of Texas at Austin
Automating evidence synthesis via machine learning and natural language processing
Fri, November 21
11:00AM
Marc Levoy
Stanford University and Google
What Google Glass means for the future of photography
Fri, January 23
11:00AM
Shimon Whiteson
University of Amsterdam
Relative Upper Confidence Bound for the K-Armed Dueling Bandit Problem
Fri, January 23
2:00PM
Rina Dechter
University of California, Irvine
Modern Exact and Approximate Combinatorial Optimization Algorithms for Graphical Models
Fri, January 30
12:00PM
Jeffrey Bigham
Carnegie Mellon University
Deeply Integrating Human and Machine Intelligence To Power Deployable Systems
Mon, February 2
11:00AM
Wei Xu
University of Pennsylvania
Modeling Lexically Divergent Paraphrases in Twitter (and Shakespeare!)
Fri, February 13
11:00AM
Cornelia Caragea
University of North Texas
Keyphrase Extraction in Citation Networks: How Do Citation Contexts Help?
Fri, February 20
11:00AM
Stefanie Tellex
Brown University
Human-Robot Collaboration
Fri, April 3
11:00AM
Noah Smith
Carnegie Mellon University
Machine Learning about People from their Language
Fri, April 10
11:00AM
Emma Brunskill
Carnegie Mellon University
Faster Learning for Better Decisions

Friday, September 19, 2014, 11:00AM



Behavior-Grounded Multisensory Object Perception and Exploration by a Humanoid Robot

Jivko Sinapov   [homepage]

University of Texas at Austin

Infants use exploratory behaviors to learn about the objects around them. Psychologists have theorized that behaviors such as touching, pressing, lifting, and dropping enable infants to form grounded object representations. For example, scratching an object can provide information about its roughness, while lifting it can provide information about its weight. In a sense, the exploratory behavior acts as a ``question'' to the object, which is subsequently ``answered" by the sensory stimuli produced during the execution of the behavior. In contrast, most object representations used by robots today rely solely on computer vision or laser scan data, gathered through passive observation. Such disembodied approaches to robotic perception may be useful for recognizing an object using a 3D model database, but nevertheless, will fail to infer object properties that cannot be detected using vision alone.

To bridge this gap, my research has pursued a framework for object perception and exploration in which the robot's representation of objects is grounded in its own sensorimotor experience with them. In this framework, an object is represented by sensorimotor contingencies that span a diverse set of exploratory behaviors and sensory modalities. Results from several large-scale experimental studies show that the behavior-grounded object representation enables a robot to solve a wide variety of tasks including recognition of objects based on the stimuli that they produce, object grouping and sorting, and learning category labels that describe objects and their properties.

About the speaker:

Jivko Sinapov received the Ph.D. degree in computer science and human-computer interaction from Iowa State University (ISU) in 2013. While working towards his Ph.D. at ISU’s Developmental Robotics Lab, he developed novel methods for behavioral object exploration and multi-modal perception. He is currently a Postdoctoral Fellow working with Peter Stone at the Artificial Intelligence lab. His research interests include developmental robotics, computational perception, autonomous manipulation, and human-robot interaction.

Friday, October 3, 2014, 11:00AM



Robots, Skills, and Symbols

George Konidaris   [homepage]

Duke University

Robots are increasingly becoming a part of our daily lives, from the automated vacuum cleaners in our homes to the rovers exploring Mars. However, while recent years have seen dramatic progress in the development of affordable, general-purpose robot hardware, the capabilities of that hardware far exceed our ability to write software to adequately control.

The key challenge here is one of abstraction. Generally capable behavior requires high-level reasoning and planning, but perception and actuation must ultimately be performed using noisy, high-bandwidth, low-level sensors and effectors. My research uses methods from hierarchical reinforcement learning as a basis for constructing robot control hierarchies through the use of learned motor controllers, or skills.

The first part of my talk will present work on autonomous robot skill acquisition. I will demonstrate a robot system that learns to complete a task, and then extracts components of its solution as reusable skills, which it deploys to quickly solve a second task. The second part will briefly focus on practical methods for acquiring skill control policies, through the use human demonstration and active learning. Finally, I will present my recent work on establishing a link between the skills available to a robot and the abstract representations it should use to plan with them. I will discuss the implications of these results for building true action hierarchies for reinforcement learning problems.

About the speaker:

George Konidaris is an Assistant Professor of Computer Science and Electrical and Computer Engineering at Duke University. He holds a BScHons from the University of the Witwatersrand, an MSc from the University of Edinburgh, and a PhD from the University of Massachusetts Amherst, having completed his thesis under the supervision of Professor Andy Barto. Prior to joining Duke, he was a postdoctoral researcher at MIT with Professors Leslie Kaelbling and Tomas Lozano-Perez.

Friday, November 7, 2014, 11:00AM



Automating evidence synthesis via machine learning and natural language processing

Byron Wallace   [homepage]

School of Information, University of Texas at Austin

Evidence-based medicine (EBM) looks to inform patient care with the totality of available relevant evidence. Systematic reviews are the cornerstone of EBM and are critical to modern healthcare, informing everything from national health policy to bedside decision-making. But conducting systematic reviews is extremely laborious (and hence expensive): producing a single review requires thousands of person-hours. Moreover, the exponential expansion of the biomedical literature base has imposed an unprecedented burden on reviewers, thus multiplying costs. Researchers can no longer keep up with the primary literature, and this hinders the practice of evidence-based care.

To mitigate this issue, I will discuss past and recent advances in machine learning and natural language processing methods that look to optimize the practice of EBM. These include methods for semi-automating evidence identification (i.e., citation screening) and more recent work on automating the extraction of structured data from full-text published articles describing clinical trials. As I will discuss, these problems pose challenging problems from a machine learning vantage point, and hence motivate the development of novel approaches. I will present evaluations of these methods in the context of EBM.

About the speaker:

Byron Wallace is an assistant professor in the School of Information at the University of Texas at Austin. He holds a PhD in Computer Science from Tufts University, where he was advised by Carla Brodley. Prior to joining UT, he was research faculty at Brown University, where he was part of the Center for Evidence-Based Medicine and also affiliated with the Brown Laboratory for Linguistic Information Processing. His primary research is in applications of machine learning and natural language processing to problems in health -- particularly in evidence-based medicine.

Wallace's work is supported by grants from the NSF and the ARO. He was recognized as The Runner Up for the 2013 ACM Special Interest Group on Knowledge Discovery and Data Mining (SIG KDD) for his thesis work.

Friday, November 21, 2014, 11:00AM



What Google Glass means for the future of photography

Marc Levoy   [homepage]

Stanford University and Google

Although head-mounted cameras (and displays) are not new, Google Glass has the potential to make these devices commonplace. This has implications for the practice, art, and uses of photography. So what's different about doing photography with Glass? First, Glass doesn't work like a conventional camera; it's hands-free, point-of-view, always available, and instantly triggerable. Second, Glass facilitates different uses than a conventional camera: recording documents, making visual todo lists, logging your life, and swapping eyes with other Glass users. Third, Glass will be an open platform, unlike most cameras. This is not easy, because Glass is a heterogeneous computing platform, with multiple processors having different performance, efficiency, and programmability. The challenge is to invent software abstractions that allow control over the camera as well as access to these specialized processors. Finally, devices like Glass that are head-mounted and perform computational photography in real time have the potential to give wearers "superhero vision", like seeing in the dark, or magnifying subtle motion or changes. If such devices can also perform computer vision in real time and are connected to the cloud, then they can do face recognition, live language translation, and information recall. The hard part is not imagining these capabilities, but deciding which ones are feasible, useful, and socially acceptable.

About the speaker:

Marc Levoy is the VMware Founders Professor of Computer Science and Electrical Engineering, Emeritus, at Stanford University. He received a Bachelor's and Master's in Architecture from Cornell University in 1976 and 1978, and a PhD in Computer Science from the University of North Carolina at Chapel Hill in 1989. In the 1970's Levoy worked on computer animation, developing a cartoon animation system that was used by Hanna-Barbera Productions to make The Flintstones, Scooby Doo, and other shows. In the 1980's Levoy worked on volume rendering, a technique for displaying three-dimensional functions such as computed tomography (CT) or magnetic resonance (MR) data. In the 1990's he worked on 3D laser scanning, culminating in the Digital Michelangelo Project, in which he and his students spent a year in Italy digitizing the statues of Michelangelo. In the 2000's he worked on computational photography and microscopy, including light field imaging as commercialized by Lytro and other companies. At Stanford he taught computer graphics and the science of art, and still teaches digital photography. Outside of academia, Levoy co-designed the Google book scanner, launched Google's Street View project, and currently leads a team in GoogleX that has worked on Project Glass and the Nexus 6 HDR+ mode. Awards: Charles Goodwin Sands Medal for best undergraduate thesis (1976), National Science Foundation Presidential Young Investigator (1991), ACM SIGGRAPH Computer Graphics Achievement Award (1996), ACM Fellow (2007).

Friday, January 23, 2015, 11:00AM



Relative Upper Confidence Bound for the K-Armed Dueling Bandit Problem

Shimon Whiteson   [homepage]

University of Amsterdam

In this talk, I will propose a new method for the K-armed dueling bandit problem, a variation on the regular K-armed bandit problem that offers only relative feedback about pairs of arms. Our approach extends the Upper Confidence Bound algorithm to the relative setting by using estimates of the pairwise probabilities to select a promising arm and applying Upper Confidence Bound with the winner as a benchmark. We prove a sharp finite-time regret bound of order O(K log T) on a very general class of dueling bandit problems that matches a lower bound proven by Yue et al. In addition, our empirical results using real data from an information retrieval application show that it greatly outperforms the state of the art.

About the speaker:

Shimon Whiteson is an Associate Professor and the head of the Autonomous Agents section at the Informatics Institute of the University of Amsterdam. He was recently awarded an ERC Starting Grant from the European Research Council as well as a VIDI grant for mid-career researchers from the Dutch national science foundation. He received his PhD in 2007 from the University of Texas at Austin, under the supervision of Peter Stone. His research focuses on decision-theoretic planning and learning, with applications to robotics, multi-camera tracking systems, and information retrieval.

Friday, January 23, 2015, 2:00PM



Modern Exact and Approximate Combinatorial Optimization Algorithms for Graphical Models

Rina Dechter   [homepage]

University of California, Irvine

In this talk I will present several principles behind state of the art algorithms for solving combinatorial optimization tasks defined over graphical models (Bayesian networks, Markov networks, constraint networks, satisfiability) and demonstrate their performance on some benchmarks.

Specifically I will present branch and bound search algorithms which explore the AND/OR search space over graphical models and thus exploit problem's decomposition (using AND nodes), equivalence (by caching) and pruning irrelevant subspaces via the power of bounding heuristics. In particular I will show how the two ideas of mini-bucket partitioning which relaxes the input problem using node duplication only, combined with linear programming relaxations ideas which optimize cost-shifting/re-parameterization schemes, can yield tight bounding heuristic information within systematic, anytime, search.

Notably, a solver for finding the most probable explanation (MPE or MAP), embedding these principles has won first place in all time categories in the 2012 PASCAL2 approximate inference challenge, and first or second place in the UAI-2014 competitions. Recent work on parallel/distributed schemes and on m-best anytime solutions may be mentioned, as time permits.

Parts of this work were done jointly with: Lars Otten, Alex Ihler, Radu Marinescu, Natasha Flerova, Kalev Kask

About the speaker:

Rina Dechter is a professor of Computer Science at the University of California, Irvine. She received her PhD in Computer Science at UCLA in 1985, an MS degree in Applied Mathematic from the Weizmann Institute and a B.S in Mathematics and Statistics from the Hebrew University, Jerusalem. Her research centers on computational aspects of automated reasoning and knowledge representation including search, constraint processing and probabilistic reasoning. Professor Dechter is an author of Constraint Processing published by Morgan Kaufmann, 2003, has authored over 150 research papers, and has served on the editorial boards of: Artificial Intelligence, the Constraint Journal, Journal of Artificial Intelligence Research and journal of Machine Learning (JLMR). She was awarded the Presidential Young investigator award in 1991, is a fellow of the American association of Artificial Intelligence since 1994, was a Radcliffe Fellowship 2005-2006, received the 2007 Association of Constraint Programming (ACP) research excellence award and is a 2013 Fellow of the ACM. She has been Co-Editor-in-Chief of Artificial Intelligence, since 2011.

Friday, January 30, 2015, 12:00PM



Deeply Integrating Human and Machine Intelligence To Power Deployable Systems

Jeffrey Bigham   [homepage]

Carnegie Mellon University

Over the past few years, I have been developing and deploying interactive crowd-powered systems that solve characteristic ?hard? problems to help people get things done in their everyday lives. For instance, VizWiz answers visual questions for blind people in seconds, Legion drives robots in response to natural language commands, Chorus holds helpful general conversations with human partners, and Scribe converts streaming speech to text in less than five seconds.

My research envisions a future in which the intelligent systems that we have dreamed about for decades, which have inspired generations of computer scientists from its beginning, are brought about for the benefit of people. My work illustrates a path for achieving this vision by leveraging the on-demand labor of people to fill in for components that we cannot currently automate, by building frameworks that allow groups to do together what even expert individuals cannot do alone, and by gradually allowing machines to take over in a data-driven way. A crowd-powered world may seem counter to the goals of computer science, but I believe that it is precisely by creating and deploying the systems of our dreams that will learn how to advance computer science to create the machines that will someday realize them.

[Talk hosted by UT School of Information; Talk info here]

About the speaker:

Jeffrey P. Bigham is an Associate Professor in the Human-Computer Interaction and Language Technologies Institutes in the School of Computer Science at Carnegie Mellon University. He uses clever combinations of crowds and computation to build truly intelligent systems, often with a focus on systems supporting people with disabilities. Dr. Bigham received his B.S.E degree in Computer Science from Princeton University in 2003, and received his Ph.D. in Computer Science and Engineering from the University of Washington in 2009. From 2009 to 2013, he was an Assistant Professor at the University of Rochester, where he founded the ROC HCI human-computer interaction research group. He has been a Visiting Researcher at MIT CSAIL, Microsoft Research, and Google[x]. He has received a number of awards for his work, including the MIT Technology Review Top 35 Innovators Under 35 Award, the Alfred P. Sloan Fellowship, and the National Science Foundation CAREER Award.

Monday, February 2, 2015, 11:00AM



Modeling Lexically Divergent Paraphrases in Twitter (and Shakespeare!)

Wei Xu   [homepage]

University of Pennsylvania

Paraphrases are alternative linguistic expressions of the same meaning. Identifying paraphrases is fundamental to many natural language processing tasks and has been extensively studied for the standard contemporary English. In this talk I will present MULTIP (Multi-instance Learning Paraphrase Model), a joint word-sentence alignment model suited to identify paraphrases within the noisy user-generated texts on Twitter. The model infers latent word-level paraphrase anchors from only sentence-level annotations during learning. This is a major departure from previous approaches that rely on lexical or distributional similarities over sentence pairs. By reducing the dependence on word overlap as evidence of paraphrase, our approach identifies more lexically divergent expressions with equivalent meaning. For experiments, we constructed a Twitter Paraphrase Corpus using a novel and efficient crowdsourcing methodology. Our new approach improves the state-of-the-art performance of a method that combines a latent space model with a feature-based supervised classifier. I will also present findings on paraphrasing between standard English and Shakespearean styles.

Joint work with Chris Callison-Burch (UPenn), Bill Dolan (MSR), Alan Ritter (OSU), Yangfeng Ji (GaTech), Colin Cherry (NRC) and Ralph Grishman (NYU).

About the speaker:

Wei Xu is a postdoc in Computer and Information Science Department at University of Pennsylvania, working with Chris Callison-Burch. Her research focuses on paraphrases, social media and information extraction. She received her PhD in Computer Science from New York University. She is organizing the SemEval-2015 shared task on Paraphrase and Semantic Similarity in Twitter, and the Workshop on Noisy User-generated Text at ACL-2015 (http://noisy-text.github.io/). During her PhD, she visited University of Washington for two years and interned at Microsoft Research, ETS and Amazon.com.

Friday, February 13, 2015, 11:00AM



Keyphrase Extraction in Citation Networks: How Do Citation Contexts Help?

Cornelia Caragea   [homepage]

University of North Texas

Keyphrase extraction is defined as the problem of automatically extracting descriptive phrases or concepts from documents. Keyphrases for a document act as a concise summary of the document and have been successfully used in many applications such as query formulation, document clustering, classification, recommendation, indexing, and summarization. Previous approaches to keyphrase extraction generally used the textual content of a target document or a local neighborhood that consists of textually-similar documents. We posit that, in a scholarly domain, in addition to a document’s textual content and textually-similar neighbors, other informative neighborhoods exist that have the potential to improve keyphrase extraction. In a scholarly domain, research papers are not isolated. Rather, they are highly inter-connected in giant citation networks, in which papers cite or are cited by other papers in appropriate citation contexts, i.e., short text segments surrounding a citation’s mention. These contexts are not arbitrary, but they serve as brief summaries of a cited paper. We effectively exploit citation context information for keyphrase extraction and show remarkable improvements in performance over strong baselines in both supervised and unsupervised settings.

About the speaker:

Cornelia Caragea is an Assistant Professor at the University of North Texas, where she directs the Machine Learning group. Her research interests lie at the intersection of machine learning, information retrieval, and natural language processing, with applications to scholarly digital libraries. She has published research papers in prestigious venues such as AAAI, IJCAI, WWW, and ICDM. Cornelia reviewed for many journals including Nature, ACM Transactions on Intelligent Systems and Technology, and IEEE Transactions on Knowledge and Data Engineering, served on several NSF panels, and was a program committee member for top conferences such as ACL, IJCAI, Coling, and CIKM. She also helped organize several workshops on scholarly big data in conferences such as AAAI, IEEE BigData, and CIKM.

Cornelia earned a Bachelor of Science degree in Computer Science and Mathematics from the University of Bucharest, and a Ph.D. in Computer Science from the Iowa State University. Prior to joining UNT in Fall 2012, she was a post-doctoral researcher at the Pennsylvania State University. Her appointment at UNT marks one of the University’s unique approaches to faculty hires: she is part of the Knowledge Discovery from Digital Information research cluster.

Friday, February 20, 2015, 11:00AM



Human-Robot Collaboration

Stefanie Tellex   [homepage]

Brown University

In order for robots to collaborate with humans, they must infer helpful actions in the physical world by observing the human's language, gesture, and actions. A particular challenge for robots that operate under uncertainty is identifying and recovering from failures in perception, actuation, and language interpretation. I will describe our approaches to automatic failure recovery using language and probabilistic methods. First we describe how a robot can use a probabilistic language grounding framework to employ information-theoretic dialog strategies, asking targeted questions to reduce uncertainty about different parts of a natural language command. Second, I will show how to invert a model for interpreting language to generate targeted natural language requests for help from a human partner, enabling a robot team to actively solicit help from a person when they encounter problems. And third, I will describe steps toward incremental interpretation of language and gesture as an enabling technology for making robots that use coordination actions to establish common ground with their human partner. This approach points the way toward more general models of human-robot collaboration building world models from both linguistic and non-linguistic input, following complex grounded natural language commands, and engaging in fluid, flexible collaboration with their human partners.

About the speaker:

Stefanie Tellex is an Assistant Professor of Computer Science and Assistant Professor of Engineering at Brown University. Her group, the Humans To Robots Lab, creates robots that seamlessly collaborate with people to meet their needs using language, gesture, and probabilistic inference, aiming to empower every person with a collaborative robot. She completed her Ph.D. at the MIT Media Lab in 2010, where she developed models for the meanings of spatial prepositions and motion verbs. Her postdoctoral work at MIT CSAIL focused on creating robots that understand natural language. She has published at SIGIR, HRI, RSS, AAAI, IROS, ICAPs and ICMI, winning Best Student Paper at SIGIR and ICMI, and Best Paper at RSS. She was named one of IEEE Spectrum's AI's 10 to Watch and won the Richard B. Salomon Faculty Research Award at Brown University. Her research interests include probabilistic graphical models, human-robot interaction, and grounded language understanding.

Friday, April 3, 2015, 11:00AM



Machine Learning about People from their Language

Noah Smith   [homepage]

Carnegie Mellon University

This talk describes new analysis algorithms for text data aimed at understanding the social world from which the data emerged. The political world offers some excellent questions to explore: Do US presidential candidates "move to the political center" after winning a primary election? Are Supreme Court justices swayed by amicus curiae briefs, documents crafted at great expense? I'll show how our computational models capture theoretical commitments and uncertainty, offering new tools for exploring these kinds of questions and more. Time permitting, we'll close with an analysis of a quarter million biographies, discussing what can be discovered about human lives as well as those who write about them.

The primary collaborators on this research are my Ph.D. students David Bamman and Yanchuan Sim; collaborators from the Political Science Department at UNC Chapel Hill, Brice Acree, and Justin Gross; and Bryan Routledge from the Tepper School of Business at CMU.

About the speaker:

Noah Smith is Associate Professor of Language Technologies and Machine Learning in the School of Computer Science at Carnegie Mellon University. In fall 2015, he will join the University of Washington as Associate Professor of Computer Science & Engineering. He received his Ph.D. in Computer Science from Johns Hopkins University in 2006 and his B.S. in Computer Science and B.A. in Linguistics from the University of Maryland in 2001. His research interests include statistical natural language processing, especially unsupervised methods, machine learning for structured data, and applications of natural language processing. His book, Linguistic Structure Prediction, covers many of these topics. He has served on the editorial board of the journals Computational Linguistics (2009–2011), Journal of Artificial Intelligence Research (2011–present), and Transactions of the Association for Computational Linguistics (2012–present), and as the secretary-treasurer of SIGDAT (2012–present). His research group, Noah's ARK, is currently supported by the NSF, DARPA, IARPA, ARO, and gifts from Amazon and Google. Smith's work has been recognized with a Finmeccanica career development chair at CMU (2011–2014), an NSF CAREER award (2011–2016), a Hertz Foundation graduate fellowship (2001–2006), numerous best paper nominations and awards, and coverage by NPR, BBC, CBC, New York Times, Washington Post, and Time.

Friday, April 10, 2015, 11:00AM



Faster Learning for Better Decisions

Emma Brunskill   [homepage]

Carnegie Mellon University

A fundamental goal of artificial intelligence is to create agents that learn to make good decisions as they interact with a stochastic environment. Some of the most exciting and valuable potential applications involve systems that interact directly with humans, such as intelligent tutoring systems or medical interfaces. In these cases, sample efficiency is highly important, as each decision, good or bad, is impacting a real person. I will describe our research on tackling this challenge, including transfer learning across sequential decision making tasks, as well as its relevance to improving educational tools.

About the speaker:

Emma Brunskill is an assistant professor in the computer science department at Carnegie Mellon University. She is also affiliated with the machine learning department at CMU. She works on reinforcement learning, focusing on applications that involve artificial agents interacting with people, such as intelligent tutoring systems. She is a Rhodes Scholar, Microsoft Faculty Fellow and NSF CAREER award recipient, and her work has received best paper nominations in Education Data Mining (2012, 2013) and CHI (2014).

[ FAI Archives ]

Fall 2022 - Spring 2023

Fall 2021 - Spring 2022

Fall 2020 - Spring 2021

Fall 2019 - Spring 2020

Fall 2018 - Spring 2019

Fall 2017 - Spring 2018

Fall 2016 - Spring 2017

Fall 2015 - Spring 2016

Fall 2014 - Spring 2015

Fall 2013 - Spring 2014

Fall 2012 - Spring 2013

Fall 2011 - Spring 2012

Fall 2010 - Spring 2011

Fall 2009 - Spring 2010

Fall 2008 - Spring 2009

Fall 2007 - Spring 2008

Fall 2006 - Spring 2007

Fall 2005 - Spring 2006

Spring 2005

Fall 2004

Spring 2004

Fall 2003

Spring 2003

Fall 2002

Spring 2002

Fall 2001

Spring 2001

Fall 2000

Spring 2000