Faraz Torabi
faraztrb [at] cs.utexas.edu

I recently defended my PhD from Learning Agents Research Group in the Computer Science Department at the University of Texas at Austin, where I worked with Prof. Peter Stone. My PhD research focused on imitation learning where, in general, the goal is for an autonomous agent to learn a task by observing another, more expert agent performs that task. I was also a part of the UT Austin Villa RoboCup 3D simulation team where I helped improving the passing strategy and worked on improving the kick and walk skills of the agents.

I am currently on the job market for research positions in industry. In general, I am interested in machine learning, reinforcement learning, and imitation learning.

CV  /  Google Scholar  /  Dissertation

Publications Preprints
DEALIO: Data-Efficient Adversarial Learning for Imitation from Observation
DEALIO: Data-Efficient Adversarial Learning for Imitation from Observation
DEALIO: Data-Efficient Adversarial Learning for Imitation from Observation
Faraz Torabi, Garrett Warnell, Peter Stone
[pdf] [bib] [video]

we propose an imitation from observation (IfO) algorithm, Data-Efficient Adversarial Learning for Imitation from Observation (DEALIO), that incorporates ideas from model-based reinforcement learning with adversarial methods for IfO in order to increase the data efficiency of these methods without sacrificing performance. Specifically, we consider time-varying linear Gaussian policies, and propose a method that integrates the linear-quadratic regulator with path integral policy improvement into an existing adversarial IfO framework. The result is a more data-efficient IfO algorithm with better performance, which we show empirically in four simulation domains: using far fewer interactions with the environment, the proposed method exhibits similar or better performance than the existing technique.

Skeletal Feature Compensation for Imitation Learning with Embodiment Mismatch
Skeletal Feature Compensation for Imitation Learning with Embodiment Mismatch
Skeletal Feature Compensation for Imitation Learning with Embodiment Mismatch
Eddy Hudson, Garrett Warnell, Faraz Torabi, Peter Stone
[pdf] [bib] [video]

We propose a new imitation learning algorithm called SILEM (Skeletal feature compensation for Imitation Learning with Embodiment Mismatch) that can help reliably train a learner to imitate an expert with a different body. Central to SILEM is a learned affine transform that compensates for differences in the skeletal features (e.g. joint angles, height of head, etc.) derived from the expert and learner. We also provide qualitative and quantitative results, teaching simulated humanoid agents, including Atlas from Boston Dynamics, to walk by observing human demonstrations.

Refereed Conferences
RIDM: Reinforced Inverse Dynamics Modeling for Learning from a Single Observed Demonstration
RIDM: Reinforced Inverse Dynamics Modeling for Learning from a Single Observed Demonstration
RIDM: Reinforced Inverse Dynamics Modeling for Learning from a Single Observed Demonstration
Brahma Pavse, Faraz Torabi, Josiah Hanna, Garrett Warnell, Peter Stone
[pdf] [bib] [slides] [video]

We propose a new technique for performing integrated IL and RL called reinforced inverse dynamics modeling (RIDM). RIDM uses recent ideas from model-based imitation from observation (IfO) to find a controller that is able to both mimic the demonstrated single state trajectory and also maximize an external environment reward signal. We experimentally compare our technique to a reasonable baseline algorithm in relevant scenarios on six MuJoCo domains, finding that it produces state-of-the-art results. We also apply our algorithm to simulated robot soccer skill learning and a UR5 physical robot arm.

Imitation Learning from Video by Leveraging Proprioception
Imitation Learning from Video by Leveraging Proprioception
Imitation Learning from Video by Leveraging Proprioception
Faraz Torabi, Garrett Warnell, Peter Stone
Interantional Joint Conference on Artificial intelligence (IJCAI), 2019
[pdf] [bib] [slides]

We build upon our previous work, GAIfO, proposing an algorithm that uses a GAN-like architecture to learn tasks perform IfO directly from videos. Unlike our prior work, however, our method also uses proprioceptive information from the imitating agent during the learning process. We show that the addition of such information will improve both learning speed and the final performance of the imitator in several standard simulation domains.

Recent Advances in Imitation Learning from Observation
Recent Advances in Imitation Learning from Observation
Recent Advances in Imitation Learning from Observation
Faraz Torabi, Garrett Warnell, Peter Stone
Interantional Joint Conference on Artificial intelligence (IJCAI), 2019
[pdf] [bib] [slides]

Conventionally, in imitation learning the imitator has access to both state and action information generated by an expert performing the task. However, requiring the action information prevents imitation learning from a large number of existing valuable learning resources such as online videos of humans performing tasks. To overcome this issue, the specific problem of imitation from observation (IfO) has recently garnered a great deal of attention, in which the imitator only has access to the state information (e.g., video frames) generated by the expert. In this paper, we provide a literature review of methods developed for IfO, and then point out some open research problems and potential future work.

Leveraging Human Guidance for Deep Reinforcement Learning Tasks
Leveraging Human Guidance for Deep Reinforcement Learning Tasks
Leveraging Human Guidance for Deep Reinforcement Learning Tasks
Ruohan Zhang, Faraz Torabi, Lin Guan, Dana H. Ballard, Peter Stone
Interantional Joint Conference on Artificial intelligence (IJCAI), 2019
[pdf] [bib]

Reinforcement learning agents can learn to solve sequential decision tasks by interacting with the environment. Human knowledge of how to solve these tasks can be incorporated using imitation learning, where the agent learns to imitate human demonstrated decisions. However, human guidance is not limited to the demonstrations. Other types of guidance could be more suitable for certain tasks and require less human effort. In this paper we provide a high-level overview of five recent learning frameworks that primarily rely on human guidance other than conventional, step-by-step action demonstrations. We review the motivation, assumption, and implementation of each framework.

Behavioral Cloning from Observation
Behavioral Cloning from Observation
Behavioral Cloning from Observation
Faraz Torabi, Garrett Warnell, Peter Stone
Interantional Joint Conference on Artificial intelligence (IJCAI), 2018
[pdf] [bib]

We propose a two-phase, autonomous imitation from observation technique called behavioral cloning from observation (BCO). We allow the agent to acquire experience in a self-supervised fashion. This experience is used to develop a model which is then utilized to learn a particular task by observing an expert perform that task without the knowledge of the specific actions taken.

Workshops, Symposia, Extended Abstracts
Generative Adversarial Imitation from Observation
Faraz Torabi, Sean Geiger, Garrett Warnell, Peter Stone
International Conference on Machine Learning Workshop on Imitation, Intent, and Interaction (I3), 2019
[pdf] [bib] [slides]

Adversarial Imitation Learning from State-only Demonstrations
Faraz Torabi, Garrett Warnell, Peter Stone
Autonomous Agents and Multi-Agent Systems (AAMAS) Extended Abstract, 2019
[pdf] [bib]

Sample-efficient Adversarial Imitation Learning from Observation
Faraz Torabi, Sean Geiger, Garrett Warnell, Peter Stone
International Conference on Machine Learning Workshop on Imitation, Intent, and Interaction (I3), 2019
[pdf] [bib]

Imitation Learning from Observation
Faraz Torabi
Association for the Advancement of Artificial Intelligence (AAAI) Doctoral Consortium, 2019
[pdf] [bib]

Book Chapters
UT Austin Villa: RoboCup 2019 3D Simulation League Competition and Technical Challenge Champions
Patrick MacAlpine, Faraz Torabi, Brahma Pavse, Peter Stone
Robot Soccer World Cup. Springer, 2019
[pdf] [bib]

UT Austin Villa: RoboCup 2018 3D Simulation League Champions
Patrick MacAlpine, Faraz Torabi, Brahma Pavse, John Sigmon, Peter Stone
Robot Soccer World Cup. Springer, 2018
[pdf] [bib]


Website template credits.