Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Eliezer Yudkowsky

From Wikipedia, the free encyclopedia
American AI researcher and writer (born 1979)

Eliezer Yudkowsky
Yudkowsky atStanford University in 2006
Born
Eliezer Shlomo (or Solomon) Yudkowsky

(1979-09-11)September 11, 1979 (age 46)
Chicago, Illinois
OrganizationMachine Intelligence Research Institute
Known forCoining the termfriendly artificial intelligence
Research onAI safety
Rationality writing
Founder ofLessWrong
Websitewww.yudkowsky.net

Eliezer S. Yudkowsky (/ˌɛliˈɛzərjʊdˈkski/EL-ee-EH-zər yuud-KOW-skee;[1] born September 11, 1979) is an American artificial intelligence researcher[2][3][4][5] and writer ondecision theory andethics, best known for popularizing ideas related tofriendly artificial intelligence.[6][7] He is the founder of and a research fellow at theMachine Intelligence Research Institute (MIRI), a private research nonprofit based inBerkeley, California.[8] His work on the prospect of a runawayintelligence explosion influenced philosopherNick Bostrom's 2014 bookSuperintelligence: Paths, Dangers, Strategies.[9]

Work in artificial intelligence safety

[edit]
See also:Machine Intelligence Research Institute

Goal learning and incentives in software systems

[edit]

Yudkowsky's views on the safety challenges future generations of AI systems pose are discussed inStuart Russell's andPeter Norvig's undergraduate textbookArtificial Intelligence: A Modern Approach. Noting the difficulty of formally specifying general-purpose goals by hand, Russell and Norvig cite Yudkowsky's proposal that autonomous and adaptive systems be designed to learn correct behavior over time:

Yudkowsky (2008)[10] goes into more detail about how to design aFriendly AI. He asserts that friendliness (a desire not to harm humans) should be designed in from the start, but that the designers should recognize both that their own designs may be flawed, and that the robot will learn and evolve over time. Thus the challenge is one of mechanism design—to design a mechanism for evolving AI under a system of checks and balances, and to give the systems utility functions that will remain friendly in the face of such changes.[6]

In response to theinstrumental convergence concern, which implies that autonomous decision-making systems with poorly designed goals would have default incentives to mistreat humans, Yudkowsky and other MIRI researchers have recommended that work be done to specify software agents that converge on safe default behaviors even when their goals are misspecified.[11][7] Yudkowsky also proposed in 2004 a theoreticalAI alignment framework calledcoherent extrapolated volition, which involves designing AIs to pursue what people would desire under idealepistemic and moral conditions.[12]

Capabilities forecasting

[edit]

In theintelligence explosion scenario hypothesized byI. J. Good, recursively self-improving AI systems quickly transition from subhuman general intelligence tosuperintelligence.Nick Bostrom's 2014 bookSuperintelligence: Paths, Dangers, Strategies sketches out Good's argument in detail, while citing Yudkowsky on the risk thatanthropomorphizing advanced AI systems will cause people to misunderstand the nature of an intelligence explosion. "AI might make anapparently sharp jump in intelligence purely as the result of anthropomorphism, the human tendency to think of 'village idiot' and 'Einstein' as the extreme ends of the intelligence scale, instead of nearly indistinguishable points on the scale of minds-in-general."[6][10][12]

InArtificial Intelligence: A Modern Approach, Russell and Norvig raise the objection that there are known limits to intelligent problem-solving fromcomputational complexity theory; if there are strong limits on how efficiently algorithms can solve various tasks, an intelligence explosion may not be possible.[6]

Time op-ed

[edit]

In a 2023 op-ed forTime magazine, Yudkowsky discussed the risk of artificial intelligence and advocated for international agreements to limit it, including a total halt on the development of AI.[13][14] He suggested that participating countries should be willing to take military action, such as "destroy[ing] a rogue datacenter by airstrike", to enforce such a moratorium.[5] The article helped introduce the debate about AI alignment to the mainstream, leading a reporter to ask PresidentJoe Biden a question about AI safety at a press briefing.[2]

If Anyone Builds It, Everyone Dies

[edit]

Together withNate Soares, Yudkowsky wroteIf Anyone Builds It, Everyone Dies, which was published byLittle, Brown and Company on September 16, 2025.[15]

Rationality writing

[edit]

Between 2006 and 2009, Yudkowsky andRobin Hanson were the principal contributors toOvercoming Bias, a cognitive and social science blog sponsored by theFuture of Humanity Institute of Oxford University. In February 2009, Yudkowsky foundedLessWrong, a "community blog devoted to refining the art of human rationality".[16][17]Overcoming Bias has since functioned as Hanson's personal blog.

Over 300 blog posts by Yudkowsky on philosophy and science (originally written onLessWrong andOvercoming Bias) were released as an ebook,Rationality: From AI to Zombies, by MIRI in 2015.[18] This book is also referred to as The Sequences.[19] MIRI has also publishedInadequate Equilibria, Yudkowsky's 2017 ebook on societal inefficiencies.[20]

Yudkowsky has also written several works of fiction. Hisfanfiction novelHarry Potter and the Methods of Rationality uses plot elements fromJ. K. Rowling'sHarry Potter series to illustrate topics in science and rationality.[16][21][22]

Personal life

[edit]

Yudkowsky is anautodidact[23] and did not attend high school or college.[24] He is Jewish and was raised as aModern Orthodox Jew, but is now secular.[25][26]

Academic publications

[edit]

See also

[edit]

References

[edit]
  1. ^"Eliezer Yudkowsky on “Three Major Singularity Schools”" onYouTube. February 16, 2012. Timestamp 1:18.
  2. ^abSilver, Nate (April 10, 2023)."How Concerned Are Americans About The Pitfalls Of AI?".FiveThirtyEight.Archived from the original on April 17, 2023. RetrievedApril 17, 2023.
  3. ^Ocampo, Rodolfo (April 4, 2023)."I used to work at Google and now I'm an AI researcher. Here's why slowing down AI development is wise".The Conversation.Archived from the original on April 11, 2023. RetrievedJune 19, 2023.
  4. ^Gault, Matthew (March 31, 2023)."AI Theorist Says Nuclear War Preferable to Developing Advanced AI".Vice.Archived from the original on May 15, 2023. RetrievedJune 19, 2023.
  5. ^abHutson, Matthew (May 16, 2023)."Can We Stop Runaway A.I.?".The New Yorker.ISSN 0028-792X.Archived from the original on May 19, 2023. RetrievedMay 19, 2023.Eliezer Yudkowsky, a researcher at the Machine Intelligence Research Institute, in the Bay Area, has likened A.I.-safety recommendations to a fire-alarm system. A classic experiment found that, when smoky mist began filling a room containing multiple people, most didn't report it. They saw others remaining stoic and downplayed the danger. An official alarm may signal that it's legitimate to take action. But, in A.I., there's no one with the clear authority to sound such an alarm, and people will always disagree about which advances count as evidence of a conflagration. "There will be no fire alarm that is not an actual running AGI," Yudkowsky has written. Even if everyone agrees on the threat, no company or country will want to pause on its own, for fear of being passed by competitors. ... That may require quitting A.I. cold turkey before we feel it's time to stop, rather than getting closer and closer to the edge, tempting fate. But shutting it all down would call for draconian measures—perhaps even steps as extreme as those espoused by Yudkowsky, who recently wrote, in an editorial forTime, that we should "be willing to destroy a rogue datacenter by airstrike," even at the risk of sparking "a full nuclear exchange."
  6. ^abcdRussell, Stuart;Norvig, Peter (2009).Artificial Intelligence: A Modern Approach. Prentice Hall.ISBN 978-0-13-604259-4.
  7. ^abLeighton, Jonathan (2011).The Battle for Compassion: Ethics in an Apathetic Universe. Algora.ISBN 978-0-87586-870-7.
  8. ^Kurzweil, Ray (2005).The Singularity Is Near. New York City: Viking Penguin.ISBN 978-0-670-03384-3.
  9. ^Ford, Paul (February 11, 2015)."Our Fear of Artificial Intelligence".MIT Technology Review.Archived from the original on March 30, 2019. RetrievedApril 9, 2019.
  10. ^abYudkowsky, Eliezer (2008)."Artificial Intelligence as a Positive and Negative Factor in Global Risk"(PDF). InBostrom, Nick; Ćirković, Milan (eds.).Global Catastrophic Risks. Oxford University Press.ISBN 978-0199606504.Archived(PDF) from the original on March 2, 2013. RetrievedOctober 16, 2015.
  11. ^Soares, Nate; Fallenstein, Benja;Yudkowsky, Eliezer (2015)."Corrigibility".AAAI Workshops: Workshops at the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, January 25–26, 2015. AAAI Publications.Archived from the original on January 15, 2016. RetrievedOctober 16, 2015.
  12. ^abBostrom, Nick (2014).Superintelligence: Paths, Dangers, Strategies. Oxford University Press.ISBN 978-0199678112.
  13. ^Moss, Sebastian (March 30, 2023).""Be willing to destroy a rogue data center by airstrike" - leading AI alignment researcher pens Time piece calling for ban on large GPU clusters".Data Center Dynamics.Archived from the original on April 17, 2023. RetrievedApril 17, 2023.
  14. ^Ferguson, Niall (April 9, 2023)."The Aliens Have Landed, and We Created Them".Bloomberg.Archived from the original on April 9, 2023. RetrievedApril 17, 2023.
  15. ^Yudkowsky, Eliezer;Soares, Nate (September 16, 2025).If Anyone Builds It, Everyone Dies.Little, Brown and Company.ISBN 978-0-316-59564-3.
  16. ^abMiller, James (2012).Singularity Rising. BenBella Books, Inc.ISBN 978-1936661657.
  17. ^Miller, James (July 28, 2011)."You Can Learn How To Become More Rational".Business Insider.Archived from the original on August 10, 2018. RetrievedMarch 25, 2014.
  18. ^Miller, James D."Rifts in Rationality – New Rambler Review".newramblerreview.com.Archived from the original on July 28, 2018. RetrievedJuly 28, 2018.
  19. ^Metz, Cade (August 4, 2025)."The Rise of Silicon Valley's Techno-Religion". RetrievedAugust 5, 2025.The Sequences, the ur-text that gave rise to the [Rationalist] movement.
  20. ^Machine Intelligence Research Institute."Inadequate Equilibria: Where and How Civilizations Get Stuck".Archived from the original on September 21, 2020. RetrievedMay 13, 2020.
  21. ^Snyder, Daniel D. (July 18, 2011)."'Harry Potter' and the Key to Immortality".The Atlantic.Archived from the original on December 23, 2015. RetrievedJune 13, 2022.
  22. ^Packer, George (2011)."No Death, No Taxes: The Libertarian Futurism of a Silicon Valley Billionaire".The New Yorker. p. 54.Archived from the original on December 14, 2016. RetrievedOctober 12, 2015.
  23. ^Matthews, Dylan; Pinkerton, Byrd (June 19, 2019)."He co-founded Skype. Now he's spending his fortune on stopping dangerous AI".Vox.Archived from the original on March 6, 2020. RetrievedMarch 22, 2020.
  24. ^Saperstein, Gregory (August 9, 2012)."5 Minutes With a Visionary: Eliezer Yudkowsky".CNBC.Archived from the original on August 1, 2017. RetrievedSeptember 9, 2017.
  25. ^Elia-Shalev, Asaf (December 1, 2022)."Synagogues are joining an 'effective altruism' initiative. Will the Sam Bankman-Fried scandal stop them?".Jewish Telegraphic Agency. RetrievedDecember 4, 2023.
  26. ^Yudkowsky, Eliezer (October 4, 2007)."Avoiding your belief's real weak points".LessWrong.Archived from the original on May 2, 2021. RetrievedApril 30, 2021.

External links

[edit]
Wikimedia Commons has media related toEliezer Yudkowsky.
Wikiquote has quotations related toEliezer Yudkowsky.
People
Organizations
Works
Concepts
Concepts
Organizations
People
Other
Concepts
Key figures
Organizations
Focus areas
Literature
Events
Overviews
Currents
Organizations
People
International
National
Academics
Artists
Other
Retrieved from "https://en.wikipedia.org/w/index.php?title=Eliezer_Yudkowsky&oldid=1318315570"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp