David Zuckerman

David Zuckerman, Professor

Visiting Harvard for the Fall 2022 semester.

Curriculum Vitae

Prospective students and interns,
please read FAQ before emailing me.

Background image is the atrium of the Gates Dell building, which houses the Department of Computer Science at the University of Texas at Austin.

Mailing Information Administrative Assistant
Department of Computer Science
The University of Texas at Austin
2317 Speedway, D9500
Austin, TX 78712
GDC 4.508
Phone: (512)471-9729
Fax: (512)471-8885
Email: diz at cs.utexas.edu
Monica Aguilar
GDC 3.502
Email: maguilar at cs.utexas.edu
Phone: (512)471-9595

100 second talk about randomness on The Academic Minute, produced by an NPR affiliate.

David Zuckerman holds an Endowed Professorship in the Computer Science Department at the University of Texas at Austin. He received an A.B. in Mathematics from Harvard University, where he was a Putnam Fellow, in 1987, and a Ph.D. in Computer Science from U.C. Berkeley in 1991. He was a postdoctoral fellow at MIT from 1991-1993 and at Hebrew University in the Fall of 1993. He has been with the University of Texas since then, visiting U.C. Berkeley from 1999-2000, Harvard University from 2004-2005, and the Institute for Advanced Study from 2011-12. In Spring 2017, he visited U.C. Berkeley, co-organizing the Simons Program on Pseudorandomness.

His research focuses primarily on pseudorandomness and the role of randomness in computing. He is best known for his work on randomness extractors and their applications. His other research interests include coding theory, distributed computing, cryptography, inapproximability, and other areas of complexity theory. His research awards include a 30-Year Test of Time Award at FOCS 2021, a Simons Investigator Award, a Best Paper Award at STOC 2016, ACM Fellow, a Guggenheim Fellowship, a Packard Fellowship for Science and Engineering, a Sloan Research Fellowship, and an NSF Young Investigator Award.

For more details, see his C.V.

The primary goal of David Zuckerman's research is to understand whether access to random numbers can make computation more efficient. An important advance in computer science was the development of efficient randomized algorithms to solve computational problems not known to have efficient deterministic algorithms. Such problems include number-theoretic problems, such as factoring polynomials, as well as a variety of approximation problems.

In practice, however, computers do not have access to truly random numbers. Instead, they use the outputs of pseudorandom number generators or physical devices. It is therefore natural and important to ask whether we can still make use of the efficient randomized algorithms even if we have no source of randomness, or perhaps a defective one. Zuckerman has shown how to convert randomized algorithms into robust randomized algorithms that work even if the source of randomness is defective. For certain natural classes of algorithms, he has been able to convert the randomized algorithms into deterministic ones.

The key tool for much of this work is the randomness extractor. An extractor is an efficient algorithm which extracts high-quality randomness from a low-quality random source, using a few auxiliary high-quality random bits. Building on Zuckerman's earlier work, he and Nisan defined and constructed the first good extractor, and he subsequently constructed several others. Moreover, Zuckerman showed how extractors can be used to construct various types of pseudorandom generators, results which do not directly involve low-quality random sources. More surprisingly, he showed that extractors have important applications in seemingly unrelated areas, including non-blocking network and expander constructions, coding theory, and the difficulty of approximating certain problems.

In 2016, Zuckerman and his then PhD student Eshan Chattopadhyay solved a longstanding open problem by introducing an algorithm that combines two independent low-quality random sources to create one high-quality random source. Previous attempts needed at least one of the two input sources to be of moderately high-quality. The new algorithm, called a two-source extractor, also gives a major improvement to an important mathematical problem in Ramsey Theory.

Most of Zuckerman's work is joint with others; see Papers and Talks for details.

Site Design by Gigi Taylor, PhD
Site Development by Monica Aguilar