全球超过130所大学在使用他编写的教材《Introduction to Computer Systems》
Web search engines have become fixtures in our society, but few people realize that they are actually publicly accessible, massive computing systems, where a single query can unleash the power of several hundred processors operating on a data set of over 200 terabytes. With Internet search, computing has risen to entirely new levels of scale, especially in terms of the sizes of the data sets involved. Google and its competitors have created a new class of large-scale computer systems, which we label "Data-Intensive Scalable Computer" (DISC) systems. DISC systems differ from conventional supercomputers in their focus is on data: they acquire and maintain continually changing data sets, in addition to performing large-scale computations over the data. DISC points the way to new ways of organizing large-scale computing systems to be more robust, scalable, and cost effective than are current high-performance computing systems.
Programs for DISC systems must be written in ways that allows them to be executed in a loosely-coupled asynchronous environment, such as the Map/Reduce framework pioneered by Google. Although Map/Reduce has surprisingly broad applicability, a richer set of programming languages and models is required to realize the full potential of DISC.