this time in open letter format! that’ll sure do it!

there are “risks”, which they are definite about - the risks are not hypothetical, the risks are real! it’s totes even had some acknowledgement in other places! totes real defs for sure this time guize

    • o7___o7@awful.systems
      link
      fedilink
      English
      arrow-up
      4
      ·
      6 months ago

      Rad as heck!

      btw, (sorry if this is prying!) considering your line of work, is all of this acausal robot god stuff especially weird and off-putting for you? Do your coworkers seem to be resistant to it?

      • BigMuffin69@awful.systems
        link
        fedilink
        English
        arrow-up
        6
        ·
        edit-2
        5 months ago

        Not prying! Thankful to say, none of my coworkers have ever brought up ye olde basilisk, the closest anyone has ever gotten has been jokes about the LLMs taking over, but never too seriously.

        No, I don’t find the acasual robot god stuff too weird b.c. we already had Pascal’s wager. But holy shit, people actually full throat believing it to the point that they are having panic attacks wtf. Like:

        1. Full human body simulation -> my brother-in-law is a computational chemist, they spend huge amounts of compute modeling simple few atom systems. To build a complete human simulation, you’d be computing every force interaction for approx ~ 10^28 atoms, like this is ludicrous.

        2. The chuckle fucks who are posing this are suggesting ok, once the robot god can sim you (which again, doubt), it’s going to be able to use that simulation of you to model your decisions and optimize against you.

        So we have an optimization problem like:

        min_{x,y} f(x) s.t. y in argmin{ g(x,y),(x,y) in X*Y}

        where x and f(x) would be the decision variables and obj function 🐍 is trying to minimize, and y and g(x,y) is the objective of me, the simulated human who has its own goals, (don’t get turned to paperclips).

        This is a bilevel optimization problem, and it’s very, very nasty to solve. Even in the nicest case possible, that somehow g,f, are convex functions and X,Y are all convex sets, (which is an insane ask considering y and g entails a complete human sim), this problem is provably NP-hard.

        Basically, to build the acasual god, first you need a computer larger than the known universe, and this probably isn’t sufficient.

        Weird note: while I was in academia, I actually did do some work on training ANN to model the constraint that y is a minimizer of a follower problem by using an ANN to act as a proxy for g(x,*), and then encoding a representation of the trained network into a single level optimization problem… we got some nice results for some special low dim problems where we had lots of data🦍 🦍 🦍 🦍 🦍