Biography:Stuart Geman

From HandWiki
Revision as of 07:37, 7 February 2024 by LinuxGuru (talk | contribs) (change)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Short description: American mathematician
Stuart A. Geman
Stuart Geman.jpg
Geman lecturing on the Gibbs sampler
Born (1949-03-23) March 23, 1949 (age 75)
Chicago , Illinois, USA
NationalityAmerican
Alma materUniversity of Michigan B.S. (1971)
Dartmouth College M.S. (1973)
Massachusetts Institute of Technology Ph.D. (1977)
Scientific career
FieldsMathematics
InstitutionsBrown University
ThesisStochastic Differential Equations with Smooth Mixing Processes (1977)
Doctoral advisorHerman Chernoff
Frank Kozin
Doctoral studentsBarry R. Davis
Websitewww.dam.brown.edu/people/geman/

Stuart Alan Geman (born March 23, 1949) is an American mathematician, known for influential contributions to computer vision, statistics, probability theory, machine learning, and the neurosciences.[1][2][3][4] He and his brother, Donald Geman, are well known for proposing the Gibbs sampler, and for the first proof of convergence of the simulated annealing algorithm.[5][6]

Biography

Geman was born and raised in Chicago. He was educated at the University of Michigan (B.S., Physics, 1971), Dartmouth Medical College (MS, Neurophysiology, 1973), and the Massachusetts Institute of Technology (Ph.D, Applied Mathematics, 1977).

Since 1977, he has been a member of the faculty at Brown University, where he has worked in the Pattern Theory group, and is currently the James Manning Professor of Applied Mathematics. He has received many honors and awards, including selection as a Presidential Young Investigator and as an ISI Highly Cited researcher. He is an elected member of the International Statistical Institute, and a fellow of the Institute of Mathematical Statistics and of the American Mathematical Society.[7] He was elected to the US National Academy of Sciences in 2011.

Work

Geman's scientific contributions span work in probabilistic and statistical approaches to artificial intelligence, Markov random fields, Markov chain Monte Carlo (MCMC) methods, nonparametric inference, random matrices, random dynamical systems, neural networks, neurophysiology, financial markets, and natural image statistics. Particularly notable works include: the development of the Gibbs sampler, proof of convergence of simulated annealing,[8][9] foundational contributions to the Markov random field ("graphical model") approach to inference in vision and machine learning,[3][10] and work on the compositional foundations of vision and cognition.[11][12]

Notes

  1. Thomas P. Ryan; William H. Woodall (2005). "The Most-Cited Statistical Papers". Journal of Applied Statistics 32 (5): 461–474. doi:10.1080/02664760500079373. 
  2. S. Kotz; N.L. Johnson (1997). Breakthroughs in Statistics, Volume III. New York, NY: Springer Verlag. 
  3. 3.0 3.1 [Wikipedia] List of important publications in computer science.
  4. Sharon Bertsch Mcgrayne (2011). The theory that would not die. New York and London: Yale University Press. https://archive.org/details/theorythatwouldn0000mcgr. 
  5. S. Geman; D. Geman (1984). "Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images". IEEE Transactions on Pattern Analysis and Machine Intelligence 6 (6): 721–741. doi:10.1109/TPAMI.1984.4767596. PMID 22499653. 
  6. Google Scholar: Stochastic Relaxation, Gibbs Distributions and the Bayesian Restoration.
  7. List of Fellows of the American Mathematical Society, retrieved 2013-08-27.
  8. P.J. van Laarhoven; E.H. Aarts (1987). Simulated annealing: Theory and applications. Netherlands: Kluwer. Bibcode1987sata.book.....L. 
  9. P. Salamon; P. Sibani; R. Frost (2002). Facts, Conjectures, and Improvements for Simulated Annealing. Philadelphia, PA: Society for Industrial and Applied Mathematics. 
  10. C. Bishop (2006). Pattern recognition and machine learning. New York: Springer. 
  11. N. Chater; J.B. Tenenbaum; A. Yuille (2005). "Probabilistic models of cognition: Conceptual foundations". Trends in Cognitive Sciences 10 (7): 287–291. doi:10.1016/j.tics.2006.05.007. PMID 16807064. https://escholarship.org/content/qt1g84199d/qt1g84199d.pdf?t=lrgwg6. 
  12. B. Ommer; J.M. Buhmann (2010). "Learning the compositional structure of visual object categories for recognition". IEEE Transactions on Pattern Analysis and Machine Intelligence 32 (3): 501–516. doi:10.1109/tpami.2009.22. PMID 20075474.