I mean, I heard this level of fear around nukes, and I'm glad they have regulations and I think regs should be on any weapon, but the situation you lay out just seems like sensationalism and I don't see how it's different than other weapons of mass destruction. And we haven't had total war like you're describing since WWII - and at the time it couldn't be handled, no one knew what to do about nukes, or chain guns, or tanks. War is a never ending series of escalations that we're never ready for and our lack of preparedness pushes us to new escalations. I get that it's scary but it seems like fear mongering. But, if it's not, what's the suggestion to stop the world from ending?
Lol exactly. He can't "grock" the difference between speculation and fact. He doesn't seem to understand social norms or be able to understand basic social cues. He either doesn't read comments closely or has a a reading disability or doesn't know the recent history of some very basic A.I. (self driving cars). He often seems unable to distinguish between criticism of his manner and the material he is presenting (as if the material and his behavior are the same thing. Taking criticisms of his behavior as disagreement with the matetial). He is a super strange person who seem to either have a screw loose or is lingering on the far edge of the functional spectrum. Hubski's A.I. Pope is super bizarre.
There are a handful of hubskiers I know of, myself included, who have done/are doing real AI work. I've been trying to ignore this dude's threads, but I'll point out that everyone else I know of who knows their shit hasn't bothered to comment in them either. Real-world AI is mostly boring, because things you can automate are always boring. PROFOUND/SCARY AI THING articles are always more science fiction than science.
You know it's quite hilarious -- my facebook friends (and I only have a few hundred, my actual friends) include Oren Etzioni and Ben Goertzal and Roman Yampolskiy and Toby Walsh and Sebastian Thrun and Rob Enderle and many others -- it's you who have your heads on backwards. I humbly suggest you rethink your position -- it's really not going to work any more.
Elaborating now that I'm not half-asleep, AI algorithms are either automating logical inference or automating statistical inference or a mixture. You can call a spam filter AI if you want and no one will call you on it, but it's just boring old hypothesis testing if you do what it does with pencil and paper. Likewise self-driving cars aren't different in kind from missile guidance systems, they can just do more because we have better computers. AI is more who you studied with, how you want to think about the problems you're solving and, less benignly, how the marketing people want to talk about the problems you're solving than a distinct kind of technology.
I'm not entirely sure how to answer that, but if you wikipedia Kalman Filter and follow along you're a couple of months from following Sebastien Thrun's book and the rest depends on which side of the "DARPA got us this far" and "google has hookers and blow" you fall on.
I use grock correctly in a sentence (with quotation marks because it's silly) and you have to act like I'm some kind if big dummy letting me know it a "real" word used by Heinlein. For some reason you just have to assume that every one is an ignoramus even when all the evidence is to the contrary. I know my Heinlein, I even remember how to drive the tractor.
That's precisely it -- it really is fundamentally different, and when you truly grok that about AI and AGI, you'll be the same as me; horrified beyond belief. It's funny, because I used to be exactly like you, skepical of off-the-wall theories, and very confident we could do something, even if it would not be known right now. But about 8 years ago I realized that AI was totally different, a whole new threat, and this time it was terribly real, like nothing that has ever been seen before, ever. This was a fundamental advance, and it WILL be incredibly beneficial and useful without a doubt, but the downside is equally horrific -- literally the end of this world in our lifetime. I reasearched it for many years before I came to this conclusion -- it was not a half-assed remark or sassy conclusion by any means, and I really feel like you better research it too, more closely. The programmers lazy arm in ISIS, typing out instructions for the end of this world. It's 100000 times different and easier than a soldiers iron specification, and it will be the death of us. There IS NO SOLUTION -- so we have to think differently, and maybe -- maybe -- we can come up with something meaningful.
Hahaha... I understand; it takes awhile to really understand so don't worry. And BTW Bostrom is talking about control, not actually really sticking to a paperclip maximizer... You know I actually majored in Philosophy and Math at Berkeley, although Math was always my specialty. I was quite good :-)
You sure do like moving the goalposts, don't you? So you grok this all so well but can't offer a solution? Not really making a good case for why we should listen to you on this issue. Earlier, you said: From reading your argument, you seem to have totally internalized its underlying assumptions, but skip over the part where you support them. Why will it be military AI? How will ISIS or some other such group have access to it? Why will their version be able to defeat one developed in the US? Why would an AGI do what they tell it to (or us for that matter)? Also, you still haven't answered my earlier comment asking why you think you've said anything many of us haven't already thought about.when you truly grok that about AI and AGI, you'll be the same as me; horrified beyond belief.
There IS NO SOLUTION -- so we have to think differently, and maybe -- maybe -- we can come up with something meaningful.
It's really ANI for the military case, not AGI, but that does not mean it can kill us very effectively. It can and it will.