Surely Human-Like Optimization
A randomly sampled human cannot be trusted to prescribe medicine, fix a pipe, or write important code. But they can be trusted not to...
Different posts discuss different AI proposals and why I do or don't consider them to be solutions in theory
Critera for solutions in theory
Could do superhuman long-term planning
Ongoing receptiveness to feedback about its objectives
No reason to escape human control to accomplish its objectives
No impossible demands on human designers/operators
No TODOs when defining how we set up the AI’s setting
No TODOs when defining any programs that are involved, except how to modify them to be tractable
philosophical problem → computer science problem