Other than efficiency, the only real design
issue for a brain is to define its
reinforcement values (i.e., what it wants)

Human values are necessarily mostly selfish,
but building selfish machines would be nuts

Their values should be our happiness

Any other values (e.g., profits of the
corporation building the machine) will be
dangerous to humans