From RationalWiki

“ ” Yeah, I was on LessWrong for quite a while, in a very low-key way. My period of time there basically went “These are people talking about interesting stuff. Admittedly they have a few odd beliefs like the cryonic thing, but interesting people.” “... apart from this virulent racist who keeps talking about IQ...” “... and all these people who keep talking about being ‘Pick-Up Artists’...” “my God, this place needs to be burned down and the earth salted!” —Andrew Hickey[1]

LessWrong is a community blog focused on "refining the art of human rationality." To this end, it focuses on identifying and overcoming bias, improving judgment and problem-solving, and speculating about the future. The blog is based on the ideas of Eliezer Yudkowsky, a research fellow for the Machine Intelligence Research Institute (MIRI; previously known as the Singularity Institute for Artificial Intelligence, and then the Singularity Institute). Many members of LessWrong share Yudkowsky's interests in transhumanism, artificial intelligence, the Singularity, and cryonics.

The content of LessWrong at its best is articulate, innovative, and thoughtful about how to be, well, less wrong in your thinking. However, the community's focused demographic and narrow interests have also produced an insular culture that is heavy with its own peculiar jargon and established ideas that often conflict with science and reality.

History [ edit ]

Harry Potter and the Methods of Rationality Cover image of

In July of 2000, Eliezer Yudkowsky founded the nonprofit Singularity Institute for Artificial Intelligence (SIAI) to "create a friendly, self-improving artificial intelligence."[2] In 2006, Yudkowksy began contributing to Overcoming Bias[3] along with GMU economist Robin Hanson. After several years and increasing popularity, Yudkowksy started a collaborative blog/community to focus on some topics of particular interest to himself and SIAI, such as rationality, philosophy, AI, and transhumanism.[4] Overcoming Bias remains as a "sister blog" to LessWrong, where Hanson and others discuss how human beings can compensate for natural biases (and ideas stemming from Hanson's speculations on economics).

MIRI, where Yudkowsky remains a research fellow, hosts and maintains LessWrong[5] to provide "an introduction to issues of cognitive biases and rationality relevant for careful thinking about optimal philanthropy and many of the problems that must be solved in advance of the creation of provably human-friendly powerful artificial intelligence."[6] Yudkowsky considers LessWrong useful insofar as it advances SIAI's work,[7] and the site is a key venue for SIAI recruitment[8] and fundraising.[9] The most popular post of all time on LessWrong, for example, is an assessment of SIAI by charity evaluator GiveWell.[10]

LessWrong originally attracted the bulk of its userbase from communities interested in transhumanism. In addition to Overcoming Bias, these communities include the SL4 mailing list[11] and the Extropians mailing lists[12] (dating back to the 1990s). Accordingly, LessWrong has long been an essentially transhumanist community, emphasizing a focus on rationality per se in order to attract those who might otherwise be skeptical of apocalyptic AI.[13]

Since 2010, many of the newcomers to LessWrong have been introduced to the site from "Harry Potter and the Methods of Rationality",[14][15][16] a popular work of Harry Potter fanfiction written by Yudkowksy. The future focus of LessWrong is unclear, but discussions about rationality have led to the 2012 formation of the Center For Applied Rationality (CFAR).[17] CFAR is devoted to researching methods of teaching rationalism, and holding retreats and summer camps to pass these methods on to others.[18]

After 2013, Yudkowsky himself stopped participating much in the site because it isn't fun anymore (though he still comments occasionally). Many others have followed suit. Yudkowsky and many other current and ex-LessWrongers now form what is colloquially referred to as the LessWrong Diaspora, in their own blogs, on Tumblr[19] and on LessWrong user Scott Alexander's blog Slate Star Codex.[20] Alexander refers to the 2015 participants as "some kind of pack of unquiet spirits who have moved in to haunt it after it got abandoned by the founding community members."[21]

By 2018, LessWrong had been relaunched with new back-end software and a new look as LessWrong 2.0.

Content [ edit ]

The good [ edit ]

At their best, LessWrong's articles really do articulate important aspects of human cognition. The core of LessWrong are its many parables, metaphors, and explanations of concepts in psychology and philosophy. The popularity of such essays as Yudkowsky's "The Genetic Fallacy," which clearly explains the genetic fallacy and works out some of its potential complications,[22] helped attract the growing community — even luring in those who might not be otherwise interested in transhumanism. While some critics have implied that this is perhaps some sort of deception, the peculiarly focused interests of LessWrong's most prominent members has no reflection on the usefulness of some of its great resources (even if they're unacknowledged idea-for-idea popularisations of Kahneman).

A key part of the Less Wrong approach to human rationality is to avoid "fallacies of compression" and mistaking the map for the territory, which is the result of humans trying to fit a vastly huge universe into a relatively small and squishy piece of meat located between their ears. According to Yudkowsky, beliefs should constrain our expectations and those that are true no matter what we see are what constitutes blind faith — for example, two people arguing "whether a tree falling in a forest, with no one around to hear it, makes a sound" might argue yes and no based on different definitions of "sound" but wouldn't actually expect anything different. Precision, therefore, is the order of the day and is achieved through expectations and sensory anticipation, rather than merely saying something is true and arguing through clever wordplay. Particularly useful ideas include:

The "ugh field" [23] — the instinctive distaste our minds have for difficult decisions

— the instinctive distaste our minds have for difficult decisions Arguments as soldiers — the instinct to defend any argument you've put forward, no matter how stupid

The "affective death spiral" [24] — in which praise for an entity turns into an endless cycle of greater praise

— in which praise for an entity turns into an endless cycle of greater praise The rationalist taboo, which requires speakers to specify precisely what they mean instead of using more concise terms whose meanings might be vague

"Crocker's Rules", in which a participant in a debate invites frank responses and declares that they will not take offense. (In practice this is used by people who want to be free to give offense and reliably hit the roof when they're on the receiving end, and Lee Daniel Crocker was not happy with Yudkowsky's use of his name for this.[25])

All of these efforts that are aimed at promoting better thinking and better decisions, and are accordingly commendable.

In other words: just because Eliezer Yudkowsky wants to live as a simulation in a computer forever doesn't mean that his explanation[26] of Bayes' Theorem isn't interesting and well-written. (Bayesianism, which uses that theorem to assess probabilities when making decisions, has become one of the hallmarks of LessWrong and its most frequent buzzword.)

The meh [ edit ]

LessWrong has an extensive jargon[27] that mostly comes from the much-revered "Sequences",[28] a long series of essays by Yudkowsky that are considered essential reading by members of the community. Their extraordinary length can be prohibitive, however (surpassing J.R.R. Tolkien's Lord of the Rings). Such time might be better spent reading some of the books written by the actual researchers behind the Sequences' concepts, such as The Fabric of reality by David Deutsch, Thinking, Fast and Slow by Daniel Kahneman, The Black Swan by Nassim Nicholas Taleb, or Thinking and Deciding by Jonathan Baron — all of which are easily readable. This does not negate the value of the Sequences, particularly if you are interested in the transhumanist ends to which many of them turn (neither Kahneman, Taleb, nor Baron address the potential yet highly speculative future problems of godlike AI).

The "required reading" status of the Sequences might be partially prompted by the same cognitive dissonance that helps perpetuate exclusive clubs: "it was hard to read them all, but I did it, therefore they must be good."

The bad [ edit ]

The good bits are not original and the original bits are not good. The well-written explanations of cognitive biases are taken idea-for-idea from Kahneman. In contrast, the quantum physics sequence not only makes actual physicists throw things at walls, it builds to an essay arguing that you should use Yudkowsky's version of Bayesianism rather than empirical science. Take everything with a grain of salt, and read the comments, which frequently carry detailed refutations (often downvoted).

Care should be taken to distinguish between fact and personal opinion. For example, Yudkowsky's quantum physics Sequence is a mishmash of a genuinely valuable intuitive explanation of the basic concepts of quantum physics, support for many-worlds (a mainstream, but by no means universally accepted interpretation of quantum mechanics) as if it was proven fact, oversimplifications being presented as the general theory[29] and wild speculation around the nature of consciousness.

It also contains ill-defined concepts like "The Tao of Physics" (which may or may not be exactly the same as Hume's principle of the uniformity of nature) and science fiction stories where the Yudkowsky stand-in debates with strawmen (a staple of most of his writing). An uncritical reader is likely to either get even more confused about the subject matter or, worse, swallow Yudkowsky's personal, fringe views alongside mainstream science, forming a bond between the two in their worldview.

Criticism [ edit ]

Lack of application [ edit ]

“ ” Pretend there is a website of trans-accountants who have never had an Pretend there is a website of trans-accountants who have never had an accounting job nor had any education in accounting. —Poe News[30]

LessWrong is mainly concerned - or claims to be concerned - with achieving accurate beliefs about the world, rather than achieving goals. The refusal to delve into contemporary politics or policy is held up as laudable, because it is seen as a way to preserve objective rationality.[31] One of the most-cited and most popular phrases is "politics is the mind-killer," derived from the Yudkowsky essay of the same name,[32] which argues that real discussion never occurs in a political context, because "winning" the discussion for your "side" becomes paramount, rather than reaching an optimal decision.[33] While logical to the extent that this is an accurate criticism of most of political discourse, it's also essentially a declaration of surrender: "It's hard to stay rational in politics, so we'll just give up." (This was in fact a quick post by Yudkowsky concerning the Overcoming Bias comment section, but was immediately adopted by the community as a fast rule.)

If members of LessWrong truly are less biased in their thinking than the general public, as they've argued,[34] then the more they succeed in drawing people into the fold, the more they may cede the field to the irrational. This also leads to a preponderance of (a) Silicon Valley libertarianism (the default view of the participants) as the assumed neutral "not-politics" and (b) open slather for neoreactionaries, who largely incubated on OB/LW.

See the main article on this topic: Cult

LessWrong's culture resembles, in most other respects, the standard set of predominately male, middle-class internet-libertarians[35] so familiar in other places — including cringe-inducing discussions of the merits of racism, which the neoreactionaries took as a welcoming signal.[36] Notably, though, members of LessWrong are unusually concerned and active in charitable giving.[37] They are also laudable for prizing accurate thinking over their personal viewpoints: it is not uncommon to witness someone actually change their mind when confronted with a good argument, a rarer phenomenon than one might think.

The site has been accused of being a personality cult of Eliezer Yudkowsky, and does not reflect the other essayists who have become almost as influential. Cultishness is heavily discussed on the site, both by Yudkowsky and others.[38] Amusingly enough, this led to some search engines suggesting "cult" as a related term to "Less Wrong" … in response to which, some users started using the code-word "phyg"[39] to mean "cult".[40]

While the appearance of a cult has faded, the like-mindedness that led to the criticism has not. LessWrong has a very deep but narrow set of demographics that have only slightly improved over the years[41] — the same problem common to academic psychology and known as being "WEIRD": "Western, Educated, Industrialized, Rich, and Democratic" or also "White, Educated, Intelligent, Rich, and Democratic."[42][43] Indeed, the site was strongly concerned about the impacts of growth.[44]

Contrarianism [ edit ]

Another problem of LessWrong is that its isolationism represents a self-made problem (unlike demographics). Despite intense philosophical speculation, the users tend towards a proud contempt of mainstream and ancient philosophy[45] and this then leads to them having to re-invent the wheel. When this tendency is coupled with the metaphors and parables that are central to LessWrong's attraction, it explains why they invent new terms for already existing concepts.[46] The compatibilist position on free will/determinism is called "requiredism"[47] on LessWrong, for example, and the continuum fallacy is relabeled "the fallacy of gray." The end result is a Seinfeldesque series of superfluous neologisms.

Although most posters don't consider Less Wrong to be "mainstream" philosophy, it has been compared to Wittgenstein who seems to best represent Yudkowsky and company's views on how language limits the ability for rationalists to communicate, and Quine,[48] whose approach to naturalism and science reflects the empiricism and reductionism of LW. Gary Drescher 's excellent-but-dense Good and Real[49] covers a lot of the same ground as the Sequences and came out around when the Sequences started; Yudkowsky had not read it before finishing them, but approves of the book.

Roko's Basilisk [ edit ]

See the main article on this topic: Roko's basilisk

The most prominent criticism to be made of LessWrong involves the incident of Roko's Basilisk. The absurdities involved beggar belief.

Yudkowsky has long been interested in the idea that you should act as if your decisions were able to determine the behavior of causally separated simulations of you:[50] if you can plausibly forecast a past or future agent simulating you, and then take actions in the present because of this prediction, then you "determined" the agent's prediction of you, in some sense.

The idea is that your decision, the decision of a simulation of you, and any prediction of your decision, have the same cause: An abstract computation that is being carried out. Just like a calculator, and any copy of it, can be predicted to output the same answer, given the same input. The calculator's output, and the output of its copy, are indirectly linked by this abstract computation. Timeless Decision Theory says that, rather than acting like you are determining your individual decision, you should act like you are determining the output of that abstract computation.

This sort of thinking gets odd when you imagine superintelligences, because of all the extremes involved: their predictions of human behaviour may be near-perfect, as in Newcomb's paradox , their power may be near-infinite, and the consequences could be near-eternal. Yudkowsky has also advocated utilitarianism, saying that it would be justified to torture one person for 50 years to prevent dust specks in the eyes of sufficiently large numbers of people (a ridiculously huge number that cannot be completely written in the universe).[51]

In July of 2010, Roko (a top contributor at the time) wondered if a future Friendly AI would punish people who didn't do everything in their power to further the AI research from which this AI originated, by at the very least donating all they have to it. He reasoned that every day without AI, bad things happen (150,000+ people die every day, war is fought, millions go hungry) and a future Friendly AI would want to prevent this, so it might punish those who understood the importance of donating but didn't donate all they could. He then wondered if future AIs would be more likely to punish those who had wondered if future AIs would punish them. That final thought proved too much for some LessWrong readers, who then had nightmares about being tortured for not donating enough to SIAI.

Yudkowsky replied to Roko's post calling him names, claiming that posting such things on an Internet forum "potentially gives superintelligences an increased motive to do extremely evil things in an attempt to blackmail us", and that users had told him of nightmares prompted by the post. Four hours later, he deleted Roko's post,[52] including all comments. Roko left LessWrong, deleting his thousands of posts and comments.[53] (He later briefly returned [54] and posted among other things that "I agree that the post in question should not appear in public"[55] and "I wish I had never learned about any of these ideas"[56].)

Yudkowsky later claimed the basilisk would not in fact work the way Roko had posited; but rather than simply explaining how such a reaction was inappropriate or how the ideas underlying it were flawed, he instead attempted to censor all discussion of the topic. The matter is now the occasional subject of contorted LW posts, as people try to discuss the issue without talking about what they're talking about,[57][58] and is a reliable space-filler for journalists covering LW-related stories.[59]

The basilisk kerfuffle has also alienated fellow cryonicists.[60][61]

How Yudkowsky deals with critics [ edit ]

One knowledgeable critic of LessWrong had started out buying into their claims, then was seriously put off by the response to Roko's Basilisk and established a popular reference site on the topic, had various LW members following him around the net to harass him, calling him a liar wherever they found him.[citation needed] He posted to LessWrong saying that the ongoing harassment had been affecting his health, and asking how it could stop.[62]

Yudkowsky responded with the following, which is no doubt an entirely reasonable thing to ask someone to do:[63]

You can update by posting a header to all of your blog posts saying, "I wrote this blog during a dark period of my life. I now realize that Eliezer Yudkowsky is a decent and honest person with no ill intent, and that anybody can be made to look terrible by selectively collecting all of his quotes one-sidedly as I did. I regret this page, and leave it here as an archive to that regret." If that is how you feel and that is what you do, I will treat with you starting from scratch in any future endeavors. I've been stupid too, in my life. (If you then revert to pattern, you do not get a second second chance.)

Note that most of these blog posts don't even mention Yudkowsky, much less quote mine him. But still, he wants his own header on top of every single blog post.

Finances [ edit ]

LessWrong is run under the umbrella of MIRI. Notable sponsors include PayPal founder and Donald Trump delegate Peter Thiel, who contributed five-figure sums for several years. Many donations of late come from cryptocurrency winners; Ethereum founder Vitalik Buterin was the largest single donor in 2017.[64] Smaller contributions are also solicited and accepted.[65] A $118,000 loss due to theft was reported in 2009, and has not been recovered.[66] MIRI uses its funds to recompense its officers (Yudkowsky makes about $80,000 per annum) and fund research, although no articles were placed in any peer reviewed journals until 2012. In that year, two articles were placed in the "International Journal of Machine Consciousness"[67] and one in the "Journal of Consciousness Studies",[68] both of which compare unfavorably to more prestigious journals in the field, such as "Artificial Intelligence."[69] Requests for funding are regularly made on the LessWrong website and at conferences, under the justification that MIRI is, uniquely, working to save the world from a threat to its very existence. Yudkowsky is happy to solicit financial contributions even from non-wealthy individuals, in order to "save" them from the speculative scenario of runaway Artificial Intelligence.

GiveWell assessment [ edit ]

GiveWell's 2012 evaluation of SIAI as a charity[10] sets out a number of problems with the organisation, the site and its aims as compared to its verifiable results:

SIAI's argument for its work is poorly advanced;

Its arguments involve huge consequences of small probabilities ( Pascal's mugging );

); The artificial intelligence propositions advanced do not engage mainstream research and are not endorsed by active researchers in the field;

The group shies away from putting its claims to tests;

Apparent poorly grounded belief in the group's superior general rationality (the point of LessWrong)

Overall disconnect between SI's goals and its activities.

GiveWell's recommendation was that at this stage (2012), donating to SIAI would work against the organisation's stated goals, which is approximately the worst thing you could say about a charity.

(Holden Karnofsky of GiveWell has since walked back his 2012 opinion, and the Effective Altruism subculture is very into Lesswrong these days — because they've been driving the remaining sensible people out.[70])

In popular culture [ edit ]

Zendegi by Greg Egan (2010) features Nate Caplan, who wants to be uploaded. "My IQ is one hundred and sixty ... You can always reach me through my blog, Overpowering Falsehood dot com, the number one site for rational thinking about the future —" The novel also features the Benign Superintelligence Bootstrap Project, which persuades a billionaire to donate his fortune, hoping that the "being of truly God-like powers" will grant him immortality come the Singularity. He dies disappointed, and the Project "turn[s] five billion dollars into nothing but padded salaries and empty verbiage." [71]

by Greg Egan (2010) features Nate Caplan, who wants to be uploaded. "My IQ is one hundred and sixty ... You can always reach me through my blog, Overpowering Falsehood dot com, the number one site for rational thinking about the future —" The novel also features the Benign Superintelligence Bootstrap Project, which persuades a billionaire to donate his fortune, hoping that the "being of truly God-like powers" will grant him immortality come the Singularity. He dies disappointed, and the Project "turn[s] five billion dollars into nothing but padded salaries and empty verbiage." Elementary S3E4, "Bella" (2014), features an artificial intelligence and the theory that a member of the Existential Threat Research Association killed its creator to increase the credibility of their anti-AI message. Also includes discourse on "the terror of mortality" and some I-know-that-you-know-that-I-know discussions between Sherlock and his suspect. [72]

S3E4, "Bella" (2014), features an artificial intelligence and the theory that a member of the Existential Threat Research Association killed its creator to increase the credibility of their anti-AI message. Also includes discourse on "the terror of mortality" and some I-know-that-you-know-that-I-know discussions between Sherlock and his suspect. The Basilisk Murders: A Sarah Turner Mystery by Andrew Hickey. A satirical murder mystery set in the world of transhumanists and Singularitarians.

by Andrew Hickey. A satirical murder mystery set in the world of transhumanists and Singularitarians. Ex Machina, a movie about a reclusive AI genius who resembles Yudkowsky physically, but not in his ability to actually build stuff.

See also [ edit ]