• Sign in
  • Sign up
Elektrine
EN
Log in Register
Modes
Overview Chat Timeline Communities Gallery Lists Friends Email Vault DNS VPN
Back to Timeline !community @gerikson
In reply to 4 earlier posts
@Architeuthis@awful.systems on awful.systems Open parent
Storytime with Rationalist Rabbi Scott Alexander
This was posted on catholic easter sunday on the ssc subreddit. It’s a posted-on-April 1st-for-plausible-deniability siskind post from back in 2018, where he outlines a kind of argument about how an all-powerfull entity that’s God in all but name (and obviously emanated from a culture discovering AGI) is actually “logically necessary”. He calls the whole thing “The Hour I First Believed”. I think it’s notable for being a bit of a treasure trove of rationalist weird accepted truths, such as: All copies of a consciousness share a self, because consciousness is like an equation, or something: But if consciousness is a mathematical object, it might be that two copies of the same consciousness are impossible. If you create a second copy, you just have the consciousness having the same single stream of conscious experience on two different physical substrates. Which is both the original transhumanist cope to enable so-called consciousness upload so it’s not just copying a simulacrum of your personality to a computer while you continue to rot away, and also what makes the basilisk torturing you possible. And it’s corollary, Simulation Capture: This means that an AI can actually “capture” you, piece by piece, into its simulation. First your consciousness is just in the real world. Then your consciousness is distributed across one real-world copy and a million simulated copies. Then the AI makes the simulated copies slightly different, and 99.9999% of you is in the simulation. which is a kind of nuts I hadn’t happened upon before. There’s also a bunch of rationalist decision theory stuff which I think make obvious how they were concocted to serve this type of narrative in the first place, instead for being broadly useful, Yud posing as a decision theory trailblazer notwithstanding.
Open parent Original URL
0
0
0
@YourNetworkIsHaunted@awful.systems on awful.systems Open parent
The decision theory stuff itself ought to be called out more for playing pretty fast and loose with reality to begin with. “If you have a supercomputer that perfectly simulates blah blah blah” is such a fundamentally bad premise because once you presume such a thing exists you’re committing to the same basic metaphysical problems that you would if you replaced the computer with God. In particular I think it commits you to hard determinism at which point there’s no sense arguing about what the right action is because the answer was set in stone not just before you entered the room but when the initial state of the universe was set up. Like, there’s a version of this where the question is meaningful in which case the premise is impossible, and a version where we accept the premise as given and render the question pointless. Why are you doing decision theory in a hypothetical world where nobody really makes decisions? Or we could acknowledge that yudkowskian decision theory is just singularity apologetics and accept the impossible elements of the premise on faith.
Open parent Original URL
0
0
0
@Architeuthis@awful.systems on awful.systems Open parent
Luckily we should be getting trickle down free will, since all universes are (of course) able to develop technology to perfectly simulate universes of lesser complexity, which seems to imply the existence of a special universe of ultimate complexity where all others emanate from, possibly in line with ain soph or equivalent mystical concept. I don’t know how that squares with that blabbing about the tegmarkian multiverse that supposedly posits that mathematically simple universes “exist ‘more’”, which siskind probably just included to reinforce his consciousness as a non-physical, mathematical object premise.
Open parent Original URL
0
0
0
@it_wasnt_arson@awful.systems on awful.systems Open parent
I continue to be endlessly fascinated by Anathem, by virtue of enjoying it as a kid for the wacky speculative metaphysics, enjoying it as an adult for the case study it presents in how Neal Stephenson can get you nodding along to a set of faux-lectures strung together by road tripping until he gets you to an obviously false conclusion, and now the fact that The Wick is apparently what rationalists actually believe in, just substituting simulations and reality-hacking for quantum woo and nukes? The Incanter Basilisk can entrap your consciousness by manipulating which timelines your brain is quantum-entangled with coexisting copies of your psyche exist in the multimetaverse and selecting among them to give you quantum immortality 51% attack you into the Matrix, I guess.
Open parent Original URL
0
0
0
0
gerikson
gerikson in !community
@gerikson@awful.systems · 2d
Been a long time since I read Anathem, and I’m kinda surprised Stephenson hasn’t been outed as an out-and-out Nazi. I mean it’s good if he’s not, just that he seems the type. Although I found the terrorist character in REAMDE weirdly well written.
View on awful.systems
0
0
0
Sign in to interact

Loading comments...

313k7r1n3

Company

  • About
  • Contact
  • FAQ

Legal

  • Terms of Service
  • Privacy Policy
  • VPN Policy

Email Settings

IMAP: mail.elektrine.com:993

POP3: pop3.elektrine.com:995

SMTP: mail.elektrine.com:465

SSL/TLS required

Support

  • support@elektrine.com
  • Report Security Issue

Connect

Tor Hidden Service

khav7sdajxu6om3arvglevskg2vwuy7luyjcwfwg6xnkd7qtskr2vhad.onion
© 2026 Elektrine. All rights reserved. • Server: 14:54:34 UTC