Biography:Uzi Vishkin

From HandWiki
Uzi Vishkin
עוזי וישקין
Born1953
Tel Aviv, Israel
Alma materHebrew University
Technion
Scientific career
Fieldsparallel algorithms
InstitutionsIBM Thomas J. Watson Research Center
New York University
Tel Aviv University
University of Maryland, College Park
Doctoral advisorYossi Shiloach

Uzi Vishkin (Hebrew: עוזי וישקין; born 1953) is a computer scientist at the University of Maryland, College Park, where he is Professor of Electrical and Computer Engineering at the University of Maryland Institute for Advanced Computer Studies (UMIACS). Uzi Vishkin is known for his work in the field of parallel computing. In 1996, he was inducted as a Fellow of the Association for Computing Machinery, with the following citation: "One of the pioneers of parallel algorithms research, Dr. Vishkin's seminal contributions played a leading role in forming and shaping what thinking in parallel has come to mean in the fundamental theory of Computer Science."[1] In 2025, he won the ACM SPAA Parallel Computing Award with the following citation: “A founder of PRAM algorithmic theory, Uzi Vishkin contributed the formative work-depth think-parallel framework, many fundamental parallel algorithms and the PRAM-on-chip system framework and its commitment to silicon.” [2]

Biography

Uzi Vishkin was born in Tel Aviv, Israel. He completed his B.Sc. (1974) and M.Sc. in Mathematics (advised by Robert Aumann) at the Hebrew University, before earning his D.Sc. in Computer Science at the Technion (1981). He then spent a year working at the IBM Thomas J. Watson Research Center in Yorktown Heights, New York. From 1982 to 1984, he worked at the department of computer science at New York University and remained affiliated with it till 1988. From 1984 until 1997 he worked in the computer science department of Tel Aviv University, serving as its chair from 1987 to 1988. Since 1988 he is with the University of Maryland, College Park.

PRAM-on-chip

A notable rudimentary abstraction—that any single instruction available for execution in a serial program executes immediately—made serial computing simple. A consequence of this abstraction is a step-by-step (inductive) explication of the instruction available next for execution. The rudimentary parallel abstraction behind the PRAM-on-chip concept, dubbed Immediate Concurrent Execution (ICE) in (Vishkin 2011), is that indefinitely many instructions available for concurrent execution execute immediately. A consequence of ICE is a step-by-step (inductive) explication (also known as lock-step) of the instructions available next for concurrent execution. Moving beyond the serial von Neumann computer (the only successful general purpose platform to date), the aspiration of the PRAM-on-chip concept is that computer science will again be able to augment mathematical induction with a simple one-line computing abstraction. A chronological overview of the evolution of the PRAM-on-chip concept and its hardware and software prototyping follow. In the 1980s and 1990s, Uzi Vishkin co-authored several articles that helped building a theory of parallel algorithms in a mathematical model called parallel random access machine (PRAM), which is a generalization for parallel computing of the standard serial computing model random-access machine (RAM). The parallel machines needed for implementing the PRAM model have not yet been built at the time, and quite a few challenged the ability to ever build such machines. Concluding in 1997[3] that the transistor count on chip as implied by Moore's Law will allow building a powerful parallel computer on a single silicon chip within a decade, he developed a PRAM-On-Chip vision that called for building a parallel computer on a single chip that allows programmers to develop their algorithms for the PRAM model. He went on to invent the explicit multi-threaded (XMT) computer architecture that enables implementation of this PRAM theory, and led his research team to completing in January 2007 a 64-processor computer[4] named Paraleap,[5] that demonstrates the overall concept. The XMT concept was presented in (Vishkin Dascal), (Naishlos Nuzman), the XMT 64-processor computer in (Wen Vishkin), in (Vishkin 2011) and most recently in (Ghanim Vishkin), where it was shown that lock-step parallel programming (using ICE) can achieve the same performance as the fastest hand-tuned multi-threaded code on XMT systems. Such inductive lock-step approach stands in contrast to multi-threaded programming approaches of other many core systems that are known for challenging programmers. The demonstration of XMT comprised several hardware and software components, as well as teaching PRAM algorithms in order to program the XMT Paraleap, using a language called XMTC. Since making parallel programming easy is one of the biggest challenges facing computer science today, the demonstration also sought to include teaching the basics of PRAM algorithms and XMTC programming to students ranging from high-school to graduate school.

Following his XMT related inventions, Uzi Vishkin was named 2024 Fellow of the National Academy of Inventors (NAI). A University of Maryland announcement [6] noted: "Two of Vishkin's 2005 patents integrating parallel processing accelerators into the CPU, or "brain” of the computer, led computer design into a new era. The best-known example is CPUs coupled with integrated graphics processing units, present in well over a billion devices including desktop and laptop computers built since the 2010s".

Parallel algorithms

In the field of parallel algorithms, Uzi Vishkin co-authored the paper (Shiloach Vishkin) that contributed the work-time (WT) (sometimes called work-depth) framework for conceptualizing and describing parallel algorithms. The WT framework was adopted as the basic presentation framework in the parallel algorithms books (JaJa 1992) and (Keller Kessler), as well as in the class notes (Vishkin 2009). In the WT framework, a parallel algorithm is first described in terms of parallel rounds. For each round, the operations to be performed are characterized, but several issues can be suppressed. For example, the number of operations at each round need not be clear, processors need not be mentioned and any information that may help with the assignment of processors to jobs need not be accounted for. Second, the suppressed information is provided. The inclusion of the suppressed information is, in fact, guided by the proof of a scheduling theorem due to (Brent 1974). The WT framework is useful since while it can greatly simplify the initial description of a parallel algorithm, inserting the details suppressed by that initial description is often not very difficult. Similarly, first casting an algorithm in the WT framework can be very helpful for programming it in XMTC. (Vishkin 2011) explains the simple connection between the WT framework and the more rudimentary ICE abstraction noted above.

In the field of parallel and distributed algorithms, one of the seminal papers co-authored by Uzi Vishkin is (Cole Vishkin). This work introduced an efficient parallel technique for graph coloring. The Cole–Vishkin algorithm finds a vertex colouring in an n-cycle in O(log* n) synchronous communication rounds. This algorithm is nowadays presented in many textbooks, including Introduction to Algorithms by Cormen et al.,[7] and it forms the basis of many other distributed algorithms for graph colouring.[8]

Other contributions by Uzi Vishkin and various co-authors include parallel algorithms for list ranking, lowest common ancestor, spanning trees, and biconnected components.

Selected publications

Notes

  1. ACM: Fellows Award / Uzi Vishkin.
  2. SPAA Parallel Computing Award.
  3. Vishkin, Uzi. Spawn-join instruction set architecture for providing explicit multithreading. U.S. Patent 6,463,527. See also (Vishkin Dascal).
  4. University of Maryland, press release, June 26, 2007: "Maryland Professor Creates Desktop Supercomputer" .
  5. University of Maryland, A. James Clark School of Engineering, press release, November 28, 2007: "Next Big "Leap" in Computing Technology Gets a Name".
  6. Maryland Today, December 10, 2024: "2 Maryland Engineers Named to National Academy of Inventors: Tao, Vishkin Honored for ‘Creating Tangible Impacts’".
  7. 1st ed., Section 30.5.
  8. See, e.g., (Goldberg Plotkin).

References

  • Baase, Sara; Van Gelder, Allen (2000), Computer Algorithms Introduction to Design and Analysis (Third ed.), Addison-Wesley, ISBN 978-0-201-61244-8 
  • "The parallel evaluation of general arithmetic expressions", Journal of the ACM 21 (2): 201–208, 1974, doi:10.1145/321812.321815 .
  • Cormen, Thomas H. (1990), Introduction to Algorithms (First ed.), MIT Press and McGraw-Hill, ISBN 978-0-262-03141-7 
  • "Parallel algorithmic techniques for combinatorial computation", Annu. Rev. Comput. Sci. 3: 233–283, 1988, doi:10.1146/annurev.cs.03.060188.001313 

This survey paper cites 16 papers co-authored by Vishkin

Cites 36 papers co-authored by Vishkin

  • "A Survey of Parallel Algorithms for Shared-Memory Machines", University of California, Berkeley, Department of EECS, Tech. Rep. UCB/CSD-88-408, 1988 

This survey paper cites 20 papers co-authored by Vishkin

  • Keller, Jorg; Kessler, Cristoph W.; Traeff, Jesper L. (2001), Practical PRAM Programming, Wiley-Interscience, ISBN 978-0-471-35351-5 

Cites 19 papers co-authored by Vishkin