From RationalWiki

Scott Alexander (1984–) is the pen-name of a LessWrong-rationalist blogger and psychiatrist. After graduating with a bachelor’s degree magna cum laude in Philosophy,[1] he gained an MD, and then completed a residency as a psychiatrist-in-training.

He began writing on Less Wrong under the name Yvain, and then branched out into his own blog, Slate Star Codex (a near-anagram of "Scott Alexander"). SSC has become one of the top-tier blogs for LessWrong-style rationalists, this and his related Tumblr being linchpins of the LessWrong Diaspora.[2]

As is customary in the writing of psychiatrists and psychologists, he mashes up details of different patients when he writes about them, so as to fictionalize the accounts and avoid his patients being identified, as well as using a pseudonym himself.

He blogs on many subjects other than psychiatry. He is particularly fond of eugenics, which he expressly advocates a positive version of.

SSC posts tend to range from long to extremely long. Alexander uses Twitter[3] and Tumblr[4] to post short/frivolous posts and puns.

Notable internet publications include his giant anti-neoreactionary FAQ,[5] his giant anti-libertarian FAQ,[6] his map of the rationalist blogosphere,[7] and a long collection of quotations from actual computer scientists on the subject of why we should take AI risk seriously.[8] Additionally, he posted a lengthy and famed criticism[9] of feminism, which had been spurred by a feminist backlash[10][11] against a blog comment[12] by MIT professor Scott Aaronson.

Alexander is a frequent visitor of local LessWrong-rationalist meetups[13] in the US, and organizes some of them himself.

The Slate Star Codex blog was taken down on 23 June 2020, on the apparent basis that a New York Times article on the Slate Star Codex subculture by Cade Metz was going to use Alexander's real name, and he feared for his safety as he had been harassed previously at work over his blogging. (This just happened to coincide with the reporter starting to get in touch with SSC critics, and not just cheerleaders.[14]) Alexander recommended readers move to the Reddit forums /r/slatestarcodex and /r/themotte.[15]

Political and social views

He does not always censor racist and sexist opinions in his comments section, which some of his fellow LessWrong-style rationalists have a problem with[16] (except on open threads, where race and gender discussions are always banned).

Iranian secularist Kaveh Mousavi, while agreeing with Alexander that the intellectually bankrupt sections of the social justice community should be heavily critiqued, has nonetheless criticized Alexander himself for having an Americentric view of social issues and of creating a false equivalence between social justice advocates and social conservatives, as well as of downplaying discrimination against women and minorities in Western countries.[17] It is worth noting, though, that Alexander has been willing to defend parts of social justice he views as worthwhile, such as uses for trigger warnings[18] and acknowledging discrimination still exists and has massive economic costs.[19]

Neoreaction

Alexander is critical of neoreactionaries, having written what is generally regarded as the definitive takedown of neoreaction,[5] though, per the header, he later took back some of the points he made in it. His blogroll is full of neoreactionaries and his comment section contains a lot of neoreactionary discussion, because he knows a pile of them personally, and he keeps discussing their ideas in his blog and for example, he considers Mencius Moldbug's Unqualified Reservations blog an obvious go-to reference his readers will immediately understand when he's talking about gay relationship counselling.[20]

Feminism

Alexander does not identify as a feminist or an anti-feminist,[21] but feels like he has been unfairly associated with both.

He talked of "the sane 30%-or-so of feminists"[22][note 1] and described some essays as “blurring the already thin line between feminism and literally Voldemort”.[24] He apparently regrets the popularity of this phrase, saying "NO NEED TO TAKE THIS ONE SENTENCE OUT OF CONTEXT AND TRY TO SPREAD IT ALL OVER THE INTERNET", though it really doesn't improve at all with context.[note 2]

In the post meant to clarify his position on feminism and feminist issues "SSC on Feminism", he described his negative attitudes towards the movement as:

“ ” I think there’s a whole corner of Internet feminism – the Jezebel, Gawker, and Modal Tumblr User faction – which is really scary. …

I think there’s a whole corner of Internet feminism – the Jezebel, Gawker, and Modal Tumblr User faction – which is really scary. … This strain is absolutely not the entirety of the movement – but it has become a big enough piece of the movement, and sufficiently dangerous to anybody who doesn’t share their views, that I think it really needs talking about and can’t be dismissed as “a few bad apples”. …

I will sometimes complain about “feminists” in a way that doesn’t necessarily mean the millions of feminists who follow good discussion norms and treat other people with respect. I’m trying to generalize less now and be much more precise about how I mean only a certain strain, but I have left the older posts untouched. I will sometimes complain about “feminists” in a way that doesn’t necessarily mean the millions of feminists who follow good discussion norms and treat other people with respect. I’m trying to generalize less now and be much more precise about how I mean only a certain strain, but I have left the older posts untouched.

Alexander enthusiastically supported James Damore's Google's Ideological Echo Chamber: How Bias Clouds Our Thinking About Diversity and Inclusion which suggested that the gender imbalance in tech fields was at least partly due to a greater proportion of men than women having the kinds of interests, inclinations, and talents that drew them to tech jobs, and suggested that changing Google's requirements for tech jobs might be needed to attract more women. This was after Alexander's fans on Reddit /r/slatestarcodex had spotted[25] that it was largely a restatement of Alexander's arguments in his post "Gender Imbalances Are Mostly Not Due To Offensive Attitudes".[26] Alexander wrote:[27]

“ ” And if you're reading this — sorry, huge respect for what you're trying, but it's pretty doomed. The best hope is a Fabian strategy of making sure enough there's enough of an underground of people who know what's up that they can quietly self-sort, form bubbles of liveability, and curb the worst excesses without forming a clear target for anybody. If you actually go riding in on a white horse waving a paper marked "ANTI-DIVERSITY MANIFESTO", you're just providing justification for the next round of purges.

Libertarianism

“ ” I feel pretty okay about both being sort of a libertarian and writing an essay arguing against libertarianism, because the world generally isn't libertarian enough but the sorts of people who read long online political essays generally are way more libertarian than can possibly be healthy.[28] I feel pretty okay about both being sort of a libertarian and writing an essay arguing against libertarianism, because the world generally isn't libertarian enough but the sorts of people who read long online political essays generally are way more libertarian than can possibly be healthy.

Communism

He is highly critical of communism, and has more generally been persistently critical of what he views as millenarian ideologies, i.e., a catastrophe will destroy the current system, handwave, a new Golden Age will arise from the ashes.[29] Much as with neoreaction, this hasn't stopped him from writing long book reports and getting very interested, for example, in the details of central planning in the USSR.[30]

Existential risks

Alexander believes that the risks of superintelligent AIs (e.g. the risk of them misconstruing our goals and turning us all into paperclips) have been repeatedly misrepresented and downplayed by the media, that while immediate disaster is unlikely, the threat is worth taking seriously, and now is a good time to research it.[31]

However, Alexander, who echoes the views of Machine Intelligence Research Institute (MIRI), Stephen Hawking, Elon Musk and Nick Bostrom on this, is not an AI researcher, nor a computer scientist (and the same goes for most of the "researchers" at MIRI, including Eliezer Yudkowsky). An actual AI researcher, Richard Loosemore, has criticized the assumptions behind many of the MIRI-style superintelligent AI doomsday scenarios, pointing out that an AI that thought it could correctly interpret the core goals of humanity but got them so hideously wrong would not in fact be worthy of the name "intelligent" at all, and that this is not merely a naming issue but a basic design issue for AIs.[32] Alexander has compiled a list of renowned and accomplished AI researchers expressing concerns about AI.[8]

Effective altruism

Alexander finds the logic of effective altruism difficult to accept intellectually, having come up with a very counterintuitive thought experiment about it, but is inclined to offer effective altruism his moral support anyway.[citation needed] Alexander is a big supporter of charity on similar grounds and often gives speeches on efficient charity,[33] and currently supports the Giving What We Can project which attempts to separate effective charities from inefficient ones.[34]

Race and IQ

Alexander identifies with the 'hereditarian left',[35] and considers The Bell Curve co-author Charles Murray to be a close ideological ally.[36][37] He has also expressed support for Gregory Cochran and Henry Harpending's hypothesis that the frequency of congenital diseases among Ashkenazi Jews (of which Alexander is one) is caused by selection for intelligence,[38] as opposed to the multiple bottlenecks and founder effects for which there is actual evidence.[39] There is almost nothing he won't try to apply human biodiversity to, e.g. Harry Potter.[40]

The Slate Star Codex comments section and the /r/slatestarcodex subreddit are even more extreme on this issue, with Cochran, Steve Sailer and Emil O. W. Kirkegaard all having taken part in the discussion.

/r/slatestarcodex

As usual, you can make anything worse by adding Reddit. /r/slatestarcodex is an unofficial fan forum for the blog. Scott comments occasionally and is a moderator. The culture wars (a regular weekly thread, until it was recently branched off to the Scott-endorsed /r/themotte) and pseudoscientific racialism of "human biodiversity" are regular and upvoted topics (literally advocating the Fourteen Words will get you 40+ upvotes[41] and admiring replies). Of course, much more offensive than the racism is objecting to the racism, which gets you a day's ban.[42] According to one moderator, "A belief in HBD doesn’t automatically equate to racism", somehow.

The moderators have a partial registry of bans.[43]

After pressure from his friends, Alexander banned "culture war" discussions from /r/slatestarcodex and moved them to a new subreddit, /r/themotte — which is now a haven for race realism and for the sort of white nationalism that thinks it's erudite. Alexander simultaneously disclaimed /r/themotte and kept recommending it.

In popular culture

Dark Enlightenment philosopher Nick Land's 2014 psychological horror novella Phyl-Undhu includes a technological cult reminiscent of LessWrong, and a character called "Alex Scott" expressing some of Scott's ideas on the Doomsday Hypothesis, with an intelligence at the end of time you can communicate with, and a cultist pushed out of the cult who "wants to have not thought certain things."

Notes

↑ [23] The evil enemy in question being He specified in the comments: The word “sane” in that context should not be taken to mean “stupid” or even “holds stupid views”, but rather “willing to hold rational discussions about their views with someone they are tempted to consider an evil enemy, based on the Principle of Charity”The evil enemy in question being neoreactionaries ↑ I dunno, you fuck one pony.