Other than efficiency, the only real design
issue for a brain is to define its
reinforcement values (i.e., what it wants)
 

Human values are necessarily mostly selfish,
but building selfish machines would be nuts
 

Their values should be our happiness
 

Any other values (e.g., profits of the
corporation building the machine) will be
dangerous to humans