51
submitted 8 months ago* (last edited 8 months ago) by dgerard@awful.systems to c/sneerclub@awful.systems

content warning: Zack Davis. so of course this is merely the intro to Zack's unquenchable outrage at Yudkowsky using the pronouns that someone wants to be called by

top 50 comments
sorted by: hot top controversial new old
[-] Amoeba_Girl@awful.systems 25 points 8 months ago

"It was bad that the New York Times called Scott a racist, because he's a racist but in a way that makes it correct to be racist."

[-] Amoeba_Girl@awful.systems 27 points 8 months ago* (last edited 8 months ago)

oh holy shit I was only a handful of paragraphs in but he literally says that!!!

So The New York Times implicitly accuses us of being racists, like Charles Murray, and instead of pointing out that being a racist like Charles Murray is the obviously correct position that sensible people will tend to reach in the course of being sensible, we disingenuously deny everything.

one point for (pseudo)intellectual honesty i guess!

[-] sc_griffith@awful.systems 14 points 8 months ago

I nominate this accidental called shot for Sneer Of The Week

[-] slopjockey@awful.systems 10 points 8 months ago

I wanted to see what kind of person would defend Scott, but disparage Murray so..

ctrl-f "Murray" read quote related ctrl-w

I almost don't want to know what motte the author has set up to make Murray's hereditarianism seem "obviously correct" and "sensible". Are we supposed to ignore the cross burning and constant race science posting on twitter?

load more comments (3 replies)
[-] swlabr@awful.systems 8 points 8 months ago

This circuitous, n-tuple negative tenor pervades the whole piece. It's first and foremost frustrating, but after that just sad.

[-] Soyweiser@awful.systems 8 points 8 months ago

Congrats at succeeding at an intellectual Turing test. And they say we strawman them.

[-] 200fifty@awful.systems 19 points 8 months ago

I love this unhinged Yudkowsky quote buried in here:

This is a filter affecting your evidence; it has not to my own knowledge filtered out a giant valid counterargument that invalidates this whole post. I would have kept silent in that case, for to speak then would have been dishonest.

Personally, I'm used to operating without the cognitive support of a civilization in controversial domains, and have some confidence in my own ability to independently invent everything important that would be on the other side of the filter and check it myself before speaking. So you know, from having read this, that I checked all the speakable and unspeakable arguments I had thought of, and concluded that this speakable argument would be good on net to publish[...]

Zack is actually correct that this is a pretty wild thing to say... "Rest assured that I considered all possible counterarguments against my position which I was able to generate with my mega super brain. No, I haven't actually looked at the arguments against my position, but I'm confident in my ability to think of everything that people who disagree with me would say."

It so happens that Yudkowsky is on the 'right side' politically in this particular case, but man, this is real sloppy for someone who claims to be on the side of capital-T truth.

The problem is... well, Zack correctly recognizes Yudkowsky is maybe not as world-changingly smart as he presents himself, and may be engaging in motivated reasoning rather than disinterested truth-seeking, but then his solution (a) doesn't involve questioning his belief in the rest of the robot apocalypse mythos, and (b) does involve running crying directly into the arms of Moldbug and a bunch of TERFs, which like, dude. Maybe consider critically interrogating those people's arguments too??

load more comments (1 replies)
[-] swlabr@awful.systems 17 points 8 months ago

Epistemic status: I have made it through this swamp. I did not read the footnotes. I glossed over the comments. Also as far as I can tell I am a cishet male, which is probably relevant to mention in reviewing this.

So here is where I am at. This post fractal sucks. Davis clearly doesn’t identify with the gender essentialist notion of male, but they also want to hold onto the idea of a biological sex binary. It sucks. I don’t think this is a unique situation- there are plenty of stories about trans kids in conservative/religious households. This particular conservative religion is the robot cult. The oppressive parents are the ghost of 2007!Yud and Scoot.

[-] swlabr@awful.systems 16 points 8 months ago

A standout, self-contained example of the general badness of this post is this anecdote (CW: transphobia/misgendering):

A trans woman I follow on Twitter complained that a receptionist at her workplace said she looked like some male celebrity. "I'm so mad," she fumed. "I look like this right now"—there was a photo attached to the Tweet—"how could anyone ever think that was an okay thing to say?" It is genuinely sad that the author of those Tweets didn't get perceived in the way she would prefer! But the thing I want her to understand, a thing I think any sane adult (on Earth, and not just dath ilan) should understand— It was a compliment! That receptionist was almost certainly thinking of someone like David Bowie or Eddie Izzard, rather than being hateful. The author should have graciously accepted the compliment and done something to pass better next time.[15] The horror of trans culture is that it's impossible to imagine any of these people doing that—noticing that they're behaving like a TERF's hostile stereotype of a narcissistic, gaslighting trans-identified man and snapping out of it.

It is baffling that they choose to speculate about/rationalise the offending comment to cast aspersions on the protagonist and “trans culture” as a whole.

[-] sc_griffith@awful.systems 16 points 8 months ago

"eyy the lady shoulda taken it as a compliment!" is so classically misogynist that this is wrapping back around to gender affirmation

[-] Amoeba_Girl@awful.systems 13 points 8 months ago* (last edited 8 months ago)

it's stupid and i wouldn't have brought it up if zack wasn't such a fucking dick, but that's fucking rich coming from someone who's so high on their own completely misguided biological essentialism, at the exclusion of any social reality, that they think an apt and plausible way to pass as a woman in public is to use a full head silicone mask

alas, too much of a transphobe to know better

[-] sc_griffith@awful.systems 11 points 8 months ago

Until I came across a comics table hawking TransCat, the "first" (self-aware scare quotes included) transgender superhero. I had to stop and look: just the catchphrase promised an exemplar of everything I'm fighting—not out of hatred, but out of a shared love that I think I have the more faithful interpretation of. I opened the cover of one of the displayed issues to peek inside. The art quality was ... not good. "There's so much I could say that doesn't fit in this context," I said to the table's proprietor, whose appearance I will not describe.

a vile thing to write at any time, but to do so within this post in particular is just incredible

[-] swlabr@awful.systems 10 points 8 months ago* (last edited 8 months ago)

??? Zack is a “I think you should leave” character

[-] dgerard@awful.systems 9 points 8 months ago

jesus fucking christ

[-] MBM@lemmings.world 8 points 8 months ago

This is wild, there's like multiple essays worth of text and subtext in there. It almost makes me sad for Zack and his severe brain worms

load more comments (3 replies)
[-] scruiser@awful.systems 14 points 8 months ago* (last edited 8 months ago)

ghost of 2007!Yud

This part gets me the most. The current day Yud isn't transphobic (enough? idk) so Zack has to piece together his older writings on semantics and epistemology to get a more transphobic gender essentialist version of past Yud.

[-] dgerard@awful.systems 13 points 8 months ago

This particular conservative religion is the robot cult.

and remember that the robot cult is considerably less transphobic than Zack, and that's half the problem Zack has.

[-] sailor_sega_saturn@awful.systems 17 points 8 months ago* (last edited 8 months ago)

Another meandering manifesto from Zach hoping that trans women would stop existing ~~so that we stop tempting him with the prospect of actually maybe being happy for once~~.

To clear this up once and for all 'she' is a very solid category with a fixed and unchanging logical definition which can be derived via straightforward Bayesian reasoning and should only be used to refer to the following and anyone who disagrees HATES MATH:

  1. Boats
  2. Cars
  3. A swedish chiptune artist (obligatory music link)
  4. She-Slimes from the dragon quest series of videogames (fun fact: male She-slimes have the ability to undergo slimification and form a King she-slime when eight gather together!)
  5. Ladies (a lady is of course a title that can be used by anyone who kneads bread, per etymology online)
load more comments (5 replies)
[-] swlabr@awful.systems 14 points 8 months ago* (last edited 8 months ago)

post begins with Atlas Shrugged excerpt

ah fuck guess I'm reading this one.

69 minute read

at least there's that.

[-] swlabr@awful.systems 15 points 8 months ago

AH FUCK it's this person again. I choose not to remember rat names if I can help it but guess I gotta commit this one to memory.

[-] dgerard@awful.systems 9 points 8 months ago

i did warn you!

[-] mosiacmango@lemm.ee 13 points 8 months ago* (last edited 8 months ago)

Begins with atlas shrug quotes, is rambling insanity that runs far too long.

Yeah, that tracks.

[-] scruiser@awful.systems 12 points 8 months ago* (last edited 8 months ago)

So, I was morbidly curious about what Zack has to say about the Brennan emails (as I think they've been under-discussed, if not outright deliberately ignored, in lesswrong discussion), I found to my horror I actually agree with a side point of Zack's. From the footnotes:

It seems notable (though I didn't note it at the time of my comment) that Brennan didn't break any promises. In Brennan's account, Alexander "did not first say 'can I tell you something in confidence?' or anything like that." Scott unilaterally said in the email, "I will appreciate if you NEVER TELL ANYONE I SAID THIS, not even in confidence. And by 'appreciate', I mean that if you ever do, I'll probably either leave the Internet forever or seek some sort of horrible revenge", but we have no evidence that Topher agreed.

To see why the lack of a promise is potentially significant, imagine if someone were guilty of a serious crime (like murder or stealing billions of dollars of their customers' money) and unilaterally confessed to an acquaintance but added, "Never tell anyone I said this, or I'll seek some sort of horrible revenge." In that case, I think more people's moral intuitions would side with the reporter.

Of course, Zack's ultimate conclusion on this subject is the exact opposite of the correct one I think:

I think that to people who have read and understood Alexander's work, there is nothing surprising or scandalous about the contents of the email.

I think the main reason someone would consider the email a scandalous revelation is if they hadn't read Slate Star Codex that deeply—if their picture of Scott Alexander as a political writer was "that guy who's so committed to charitable discourse

Gee Zack, I wonder why so many people misread Scott? ...Its almost like he is intentionally misleading about his true views in order to subtly shift the Overton window of rationalist discourse and intentionally presents himself as simply committed to charitable discourse while actually having a hidden agenda! And the bloated length of Scott's writing doesn't help with clarity either. Of course Zack, who writes tens of thousands of words to indirectly complain about perceived hypocrisy of Eliezer's in order to indirectly push gender essentialist views, probably finds Scott's writings a perfectly reasonable length.

Edit: oh and a added bonus on the Brennan Emails... Seeing them brought up again I connected some dots I had missed. I had seen (and sneered at) this Yud quote before:

I feel like it should have been obvious to anyone at this point that anybody who openly hates on this community generally or me personally is probably also a bad person inside and has no ethics and will hurt you if you trust them, but in case it wasn't obvious consider the point made explicitly.

But somehow I had missed or didn't realize the subtext was the emails that laid clear Scott's racism:

(Subtext: Topher Brennan. Do not provide any link in comments to Topher's publication of private emails, explicitly marked as private, from Scott Alexander.)

Hmm... I'm not sure to update (usage of rationalist lingo is deliberate and ironic) in the direction of "Eliezer is stubbornly naive on Scott's racism" or "Eliezer is deliberately covering for Scott's racism". Since I'm not a rationalist my probabilities don't have to sum to 1, so I'm gonna go with both.

[-] dgerard@awful.systems 11 points 8 months ago* (last edited 8 months ago)

(I think this is the best and most important post in the sequence; I suspect that many readers who didn't and shouldn't bother with the previous three posts, may benefit from this one.)

that's ... wonderful, Zack

also, I think this is the first mention of the Brennan email on LW?

load more comments (2 replies)
[-] Soyweiser@awful.systems 11 points 8 months ago* (last edited 8 months ago)

69 minute read.

That is not nice.

Reminds me of the poetic reaction to the epic poem (1400+ lines) about May by Herman Gorter. In reaction to this Hendrik de Vries wrote: (Loosely translated, with the ABBBA rhyme scheme destroyed, im sorry)

"Gorter, Gorter!

I wanted to read your Maycanto

But soon I was afraid

That I, before I died, would not be able to finish it.

Shorter! Shorter! Shorter!"

OG in Dutch: "Gorter, Gorter!

'k Heb uw Meizang willen lezen

Maar begon al gauw te vrezen

Dat het, voor mijn dood, niet uit zou wezen.

Korter! Korter! Korter!"

Not that relevant, but I always thought it was a fun story, and it is a nice piece of poetic shade thrown.

Edit: formatting eurgh.

[-] swlabr@awful.systems 10 points 8 months ago

I don’t recognise this ABBA song, must be a deep cut (jk)

[-] dgerard@awful.systems 11 points 8 months ago* (last edited 8 months ago)

b-side of "Gimme Gimme Gimme 17,000 Words After Midnight"

load more comments (1 replies)
[-] gerikson@awful.systems 11 points 8 months ago

I don't believe this is a correct use of a diaeresis: "preöccupied" (and not just because it looks ridiculous if you know Swedish)

[-] carlitoscohones@awful.systems 11 points 8 months ago

Eschewing the mainstream use of language and formatting is a sign of genius.

[-] gerikson@awful.systems 10 points 8 months ago

Bring back röckdöts!

(Funniest example of that is the band Tröjan, which literally means "the sweater" in Swedish)

[-] AllNewTypeFace@leminal.space 8 points 8 months ago

Whomst’ve even come up with such a preposterous usage

load more comments (1 replies)
[-] carlitoscohones@awful.systems 10 points 8 months ago
[-] Collectivist@awful.systems 8 points 8 months ago

While the writer is wrong, the post itself is actually quite interesting and made me think more about epistemic luck. I think Zack does correctly point out cases where I would say rationalists got epistemically lucky, although his views on the matter seem entirely different. I think this quote is a good microcosm of this post:

The Times's insinuation that Scott Alexander is a racist like Charles Murray seems like a "Gettier attack": the charge is essentially correct, even though the evidence used to prosecute the charge before a jury of distracted New York Times readers is completely bogus.

A "Gettier attack" is a very interesting concept I will keep in my back pocket, but he clearly doesn't know what a Gettier problem is. With a Gettier case a belief is both true and justified, but still not knowledge because the usually solid justification fails unexpectedly. The classic example is looking at your watch and seeing it's 7:00, believing it's 7:00, and it actually is 7:00, but it isn't knowledge because the usually solid justification of "my watch tells the time" failed unexpectedly when your watch broke when it reached 7:00 the last time and has been stuck on 7:00 ever since. You got epistemically lucky.

So while this isn't a "Gettier attack" Zack did get at least a partial dose of epistemic luck. He believes it isn't justified and therefore a Gettier attack, but in fact, you need justification for a Gettier attack, and it is justified, so he got some epistemic luck writing about epistemic luck. This is what a good chunk of this post feels like.

[-] YouKnowWhoTheFuckIAM@awful.systems 9 points 8 months ago* (last edited 8 months ago)

This “Gettier” attack seems to me to have no more interesting content than a “stopped clock”. To use an extremely similar, extremely common phrase, the New York Times would have been “right for the wrong reasons” to call Scott Alexander a racist. And this would be conceptually identical to pointing out that, I dunno, crazed conspiracy theorists suggested before he was caught that Jeffrey Epstein was part of an extensive paedophile network.

But we see this happen all the time, in fact it’s such a key building block of our daily experience that we have at least two cliches devoted to capturing it.

Perhaps it would be interesting if we were to pick out authentic Gettier cases which are also accusations of some kind, but it seems likely that in any case (i.e. all cases) where an accusation is levelled with complex evidence, the character of justification fails to be the very kind which would generate a Gettier case. Gettier cases cease to function like Gettier cases when there is a swathe of evidence to be assessed, because already our sense of justification is partial and difficult to target with the precision characteristic of unexpected failure - such cases turn out to be just “stopped clocks”. The sense of counter-intuitivity here seems mostly to be generated by the convoluted grammar of your summarising assessment, but this is just an example of bare recursivity, since you’re applying the language of the post to the post itself.

load more comments (10 replies)
load more comments (1 replies)
load more comments
view more: next ›
this post was submitted on 04 Mar 2024
51 points (100.0% liked)

SneerClub

984 readers
21 users here now

Hurling ordure at the TREACLES, especially those closely related to LessWrong.

AI-Industrial-Complex grift is fine as long as it sufficiently relates to the AI doom from the TREACLES. (Though TechTakes may be more suitable.)

This is sneer club, not debate club. Unless it's amusing debate.

[Especially don't debate the race scientists, if any sneak in - we ban and delete them as unsuitable for the server.]

founded 1 year ago
MODERATORS