Will artificial intelligence ever be fully human?

Moralist philosopher Tony Beavers wrote in 2012,

"The project of designing moral machines is complicated by the fact that even after two millennia of moral inquiry, there is still no consensus on how to determine moral right from wrong."

This argument, and each other argument like it, is a relevant problem in the creation of artificial intelligence. Because humans are still unable to tell apart moral right and wrong, AI, when conceived as a supplementary tool by scientists and engineers, is being conceived more as an optimizer than as a anthropomorphic problem-solver.

Such a bias brings to light a difference in the perception of intelligence. Because we are not able to parametrize morality and therefore create a machine-equivalent of it, discounting it from intelligence altogether is becoming a widely suggested course of action (Muehlhauser & Helm, 2002). Is this justified?

For instance, consider the recent example of Google's autonomous cars project. In the future, if such cars were to proliferate, then they will have to come programmed with a system of ethics. However, the problem that humans don't yet know what the ideal code of ethics really is, what will the cars be programmed with?

A short essay by Gary Marcus in The New Yorker highlights this problem. Here's an excerpt:

"Within two or three decades … it will no longer be optional for machines to have ethical systems. Your [autonomous] car is speeding along a bridge at fifty miles per hour when errant school bus carrying forty innocent children crosses its path. Should your car swerve, possibly risking the life of its owner (you), in order to save the children, or keep going, putting all forty kids at risk? If the decision must be made in milliseconds, the computer will have to make the call."

This is just one case in point, one that's also easily resolved by consideirng that a manual override system will be in place if only to avoid legal consequences to the manufacturer. But the overarching quandary remains unresolved: Will AI ever be fully human?

The Singularity Institute's point of view (linked above) is that intelligence must be redefined in order for AI to persist. Intelligence, they say, cannot include moral values, and instead must encompass a perspective toward problem-solving that is entirely goal-oriented. If an AI is being constructed to keep our roads clean, then that is all it will do. No modesty, no compunctions.

The problem with this perspective is that it discounts the contributions of culture and traditions to the notion of intelligence, perhaps even to the notion of being human.