David A. McAllester (born May 30, 1956) is an American computer scientist who is Professor and former chief academic officer at the Toyota Technological Institute at Chicago. He received his B.S., M.S. and Ph.D. degrees from the Massachusetts Institute of Technology in 1978, 1979 and 1987 respectively. His PhD was supervised by Gerald Sussman. He was on the faculty of Cornell University for the academic year 1987–1988 and on the faculty of MIT from 1988 to 1995. He was a member of technical staff at AT&T Labs-Research from 1995 to 2002. He has been a fellow of the American Association of Artificial Intelligence since 1997.[3] He has written over 100 refereed publications.

David A. McAllester
Born (1956-05-30) May 30, 1956 (age 68)
United States
Alma materMassachusetts Institute of Technology
Known forArtificial intelligence
AwardsAAAI Classic Paper Award (2010)[1]
International Conference on Logic Programming Test of Time award (2014)[2]
Scientific career
FieldsComputer Science, Artificial Intelligence, Machine Learning
InstitutionsMassachusetts Institute of Technology
Toyota Technological Institute at Chicago
Doctoral advisorGerald Sussman

McAllester's research areas include machine learning theory, the theory of programming languages, automated reasoning, AI planning, computer game playing (computer chess) and computational linguistics. A 1991 paper on AI planning[4] proved to be one of the most influential papers of the decade in that area.[5] A 1993 paper on computer game algorithms[6] influenced the design of the algorithms used in the Deep Blue chess system that defeated Garry Kasparov.[7] A 1998 paper on machine learning theory[8] introduced PAC-Bayesian theorems which combine Bayesian and non-Bayesian methods.

Opinions on artificial intelligence

edit

McAllester has voiced concerns about the potential dangers of artificial intelligence, writing in an article to the Pittsburgh Tribune-Review that it is inevitable that fully automated intelligent machines will be able to design and build smarter, better versions of themselves, an event known as the singularity. The singularity would enable machines to become infinitely intelligent, and would pose an "incredibly dangerous scenario". McAllester estimates a 10 percent probability of the Singularity occurring within 25 years, and a 90 percent probability of it occurring within 75 years.[9] He appeared on the AAAI Presidential Panel on Long-Term AI Futures in 2009:,[10] and considers the dangers of superintelligent AI worth taking seriously:

I am uncomfortable saying that we are ninety-nine per cent certain that we are safe for fifty years... That feels like hubris to me.[11]

He was later described as discussing the singularity at the panel in terms of two major milestones in artificial intelligence:

1) Operational Sentience: We can easily converse with computers. 2) The AI Chain Reaction: A computer that boot straps itself to a better self. Repeat.[12]

McAllester has also written on friendly artificial intelligence on his blog. He says that before machines become capable of programming themselves (potentially leading to the singularity), there should be a period where they are moderately intelligent in which it should be possible to test out giving them a purpose or mission that should render them safe to humans:

I personally believe that it is likely that within a decade agents will be capable of compelling conversation about the everyday events that are the topics of non-technical dinner conversations. I think this will happen long before machines can program themselves leading to an intelligence explosion. The early stages of artificial general intelligence (AGI) will be safe. However, the early stages of AGI will provide an excellent test bed for the servant mission or other approaches to friendly AI ... If there is a coming era of safe (not too intelligent) AGI then we will have time to think further about later more dangerous eras.[13]

References

edit
  1. ^ "AAAI Classic Paper Award". AAAI. 2016. Retrieved 19 August 2016.
  2. ^ "Pascal's paper stands the test of time". Australian National University. 23 April 2014. Retrieved 19 August 2016.
  3. ^ "David McAllester biography". Toyota Technological Institute at Chicago. Retrieved 19 August 2016.
  4. ^ McAllester, David; Rosenblitt, David (December 1991). "Systematic Nonlinear Planning" (PDF). Proceedings AAAI-91. AAAI: 634–639. Retrieved 19 August 2016.
  5. ^ "Google Scholar Citations". Google Scholar. 2016. Retrieved 19 August 2016.
  6. ^ McAllester, David; Yuret, Deniz (20 October 1993). "Alpha-Beta-Conspiracy Search". ICGA Journal. Draft. CiteSeerX 10.1.1.44.6969.
  7. ^ Campbell, Murray S.; Joseph Hoane, Jr., A.; Hsu, Feng-hsiung (1999). "Search Control Methods in Deep Blue" (PDF). AAAI Technical Report SS-99-07. AAAI: 19–23. Archived from the original (PDF) on 14 September 2016. Retrieved 16 August 2016. To the best of our knowledge, the idea of separating the white and black depth computation was first suggested by David McAllester. A later paper (McAllester and Yuret 1993) derived an algorithm, ABC, from conspiracy theory (McAllester 1988).
  8. ^ McAllester, David (1998). "Some PAC-Bayesian theorems". Proceedings of the eleventh annual conference on Computational learning theory - COLT' 98. Association for Computing Machinery. pp. 230–234. CiteSeerX 10.1.1.21.1745. doi:10.1145/279943.279989. ISBN 978-1581130577. S2CID 53234792. Retrieved 19 August 2016.
  9. ^ Cronin, Mike (2 November 2009). "Futurists' report reviews dangers of smart robots". Pittsburgh Tribune-Review. Retrieved 20 August 2016.
  10. ^ "Asilomar Meeting on Long-Term AI Futures". Microsoft Research. 2009. Retrieved 20 August 2016.
  11. ^ Khatchadourian, Raffi (23 November 2015). "The Doomsday Invention: Will artificial intelligence bring us utopia or destruction?". The New Yorker. Retrieved 23 August 2016.
  12. ^ Fortnow, Lance (31 July 2009). "The Singularity". Computational Complexity. Retrieved 20 August 2016.
  13. ^ McAllester, David (10 August 2014). "Friendly AI and the Servant Mission". Machine Thoughts. WordPress. Retrieved 20 August 2016.
edit