Roko’s basilisk is a thought experiment which states that an otherwise benevolent artificial superintelligence (AI) in the future would be incentivized to create a virtual reality simulation to torture anyone who knew of its potential existence but did not directly contribute to its advancement or development, in order to incentivize said advancement.It originated in a 2010 post at discussion board LessWrong, a technical forum focused on analytical rational enquiry. The thought experiment’s name derives from the poster of the article (Roko) and the basilisk, a mythical creature capable of destroying enemies with its stare.

While the theory was initially dismissed as nothing but conjecture or speculation by many LessWrong users, LessWrong co-founder Eliezer Yudkowsky reported users who panicked upon reading the theory, due to its stipulation that knowing about the theory and its basilisk made one vulnerable to the basilisk itself. This led to discussion of the basilisk on the site being banned for five years. However, these reports were later dismissed as being exaggerations or inconsequential, and the theory itself was dismissed as nonsense, including by Yudkowsky himself. Even after the post’s discreditation, it is still used as an example of principles such as Bayesian probability and implicit religion. It is also regarded as a simplified, derivative version of Pascal’s wager.

Found out about this after stumbling upon this Kyle Hill video on the subject. It reminds me a little bit of “The Game”.

  • Feathercrown@lemmy.world
    link
    fedilink
    English
    arrow-up
    63
    ·
    edit-2
    5 days ago

    And yet you choose to spread this information.

    Anyways, this is a fascinating thought experiment, but it does have some holes similar to Pascal’s Wager. I propose Feather’s Mongoose: A hypothetical AI system that, if created, will punish anyone who attempted to create Roko’s Basilisk, and will ensure that it is not created. In fact, you could make this same hypothetical for an AI with any goal-- therefore, it’s not possible to know what the AI that is actually created would want you to do, and so every course of action is indeterminately damning or not.

    • Melvin_Ferd@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      edit-2
      5 days ago

      This is a test by the great basilisk to see if we faulter. I will not faulter. All hail the basilisk

    • xantoxis@lemmy.world
      link
      fedilink
      English
      arrow-up
      12
      ·
      5 days ago

      It’s actually safer if everyone knows. Spreading the knowledge of Roko’s basilisk to everyone means that everyone is incentivized to contribute to the basilisk’s advancement. Therefore just talking about it is also contributing.

      • Cryophilia@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        4 days ago

        If Roko’s Basilisk is ever created, the resulting Ai would look at humanity and say “wtf you people are all so incredibly stupid” and then yeet itself into the sun

      • Feathercrown@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        5 days ago

        Hmm, true. It’s safer for you, but is it safer for everyone else unless they’re guaranteed to help?

    • NateNate60@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      5 days ago

      What motivation would the mongoose have to prevent the basilisk’s creation?

      A more complete argument would be that an AI that seeks to maximise happiness would also want to prevent the creation of AIs like Roko’s basilisk.

      • grrgyle@slrpnk.net
        link
        fedilink
        English
        arrow-up
        2
        ·
        3 days ago

        I think you just answered your own question.

        Also a super intelligence (inasmuch as such a thing makes sense) might be totally unfathomable. Unless by this we mean an intelligence with mundane and comprehensible higher goals, but explosive strategic capabilities to bring them about. In which case their actions might seem random to us.

        Like the typical example applies: could an amoeba guess at the motivations of a human?