I remember role playing cops and robbers as a kid. I could point my finger and shout “bang bang I got you” but if my friend didn’t pretend to be mortally wounded and instead just kept running around there’s really nothing I could do.
I didn’t read this but I confident it can be summarized as “how many hostile AGIs can we confine to the head of a pin?”
Word Count needs to be added to the crackpot index.
td;lr
No control method exists to safely contain the global feedback effects of self-sufficient learning machinery. What if this control problem turns out to be an unsolvable problem?
While I agree this article is TL and I DR it, this is not an abstract. This is a redundant lede and attempted clickbait at that.
Oh wait I just noticed the L and D are swapped. Feel free not to tell me whether that’s a typo or some smarmy lesswrongism.
too dong, leedn’t read?
too damn lesswrong
Nobody tell these guys that the control problem is just the halting problem and first year CS students already know the answer.
remembering how Thiel paid Buterin to drop out of his comp sci course so he spent all of 2018 trying to implement plans for Ethereum that only required that P=NP
it’s kind of amazing how many of the “I’ll never use CS theory in my career ever” folks end up trying to implement a fast SAT solver without realizing
@dgerard @sneerclub It’s a reasonably close approximation, at least…
On a similar note, Yud’s decision theory that hinges on an AI (presumably a Turing Machine) predicting what a human (Turing-Complete at the least) does with 100% accuracy.
…huh. somehow among all the many things wrong with TDT, I never cottoned to the fact that it just reduces to the halting problem
are rats just convinced that Alan Turing never considered what if computer but more complex? cause there’s a whole branch of math dedicated to computability regardless of the complexity of the computation substrate, and Alan helped invent it. of course they don’t know about this because they ignore the parts of computer science that disagree with their stupid ideas
Actually I might have done goofed with that one; now that I think of it, if you assume some jackoff amount of computing power then a human brain (assuming nothing uncomputable happens there, so sad Penrose noises) could be simulated from first principles for a limited amount of time, no actual proof of possible future outcomes needed. This still leaves the problem of how exactly do you get all the data for that (and I think any uncertainity would require an exponential increase in paths you have to simulate), especially without killing the human in question.
:chefkiss:
@sue_me_please Don’t think this reply will properly show up on awful.systems, but I can’t resist to sneer.
It amuses me that for a while the LW people saw Musk as a great example, and he just went ‘I would solve the control problem by making them human friendly and making the robots have low grip strength. Easy peasy’ Amazed that wasn’t a crack ping moment for a lot of them.
the sneer looks good to me!