We have a reasonably clear sense of what "good" is, but it's not perfect. Suffering is bad, pleasure is good, more people living enjoyable lives is good, yes, but tradeoffs are hard. How much worse is it to go blind than to lose your leg? [1] How do we compare the death of someone at eighty to the death of someone at twelve? If you wanted to build some automated system that would go from data about the world to a number representing how well it's doing, where you would prefer any world that scored higher to any world scoring lower, that would be very difficult.
Say, however, that you've built a metric that you think matches your values well and you put some powerful optimizer to work maximizing that metric. This optimizer might do many things you think are great, but it might be that the easiest ways to maximize the metric are the ones that pull it apart from your values. Perhaps after it's in place it turns out your metric included many things that only strongly correlated with what you cared about, where the correlation breaks down under maximization.
What confuses me is that the people who warn about this scenario with respect to AI are often the same people in favor of futarchy. They both involve trying to define your values and then setting an indifferent optimizer to work on them. If you think AI would be very dangerous but futarchy would be very good, why?
I also posted this on my blog.
[1] This is a question people working in public health try to answer with Disability Weights for DALYs.
Not sure how you missed it... but I speak about how people should be able to choose where their taxes go. Maybe you missed it because I get swamped with downvotes?
Right now the government engages in activities that some people consider to be immoral. For example, pacifists consider war to be immoral. You think that there's absolutely nothing wrong with pacifists being forced to fund war. Instead of worrying about how pacifists currently have to give war a leg to stand on... you want to worry about how we're going to prevent robots from being immoral.
When evilness, like beauty, is in the eye of the beholder... it's just as futile to try and prevent AIs from being immoral as it is to try and prevent humans from being immoral. What isn't futile however is to fight for people's freedom not to invest in immorality.
Any case you worry about is a case where an AI that you consider to be immoral ends up with too many resources at its disposal. Because you're really not going to worry about...
So you worry about a case where an immoral AI ends up with too many resources at its disposal. But that's exactly the same thing that I worry about with humans. And if it's exactly the same thing that I worry about with humans... then it's a given that my worry is the same regardless of whether the immoral individual is human, AI, alien or other.
In other words, you have this bizarre double standard for humans and AI. You want to prevent immoral AIs from coming into existence yet you think nothing of forcing humans to give immoral humans a leg to stand on.
Oh gods, you're doing that again. "How dare you be talking about something other than my pet issue! That proves you're on the wrong side of my pet issue, which proves you're inconsistent and insincere!"
There is a reason why you keep getting "swamped with downvotes". That reason is that you are wasting other people's time and attention, and appear not to care. As long as you continue to behave in this obnoxious and antisocial fashion, you will continue to get swamped with downvotes. And, not coincidentally, your rudeness and obtuseness will incline people to think less favourably of your proposal. If someone else more reasonable comes along with an economic proposal like yours, the first reaction of people who've interacted with you here is likely to be that bit more negative because they'll associate the idea with rudeness and obtuseness.
Please consider whether that is really what you want.