Okay. So going back to this thesis that you’re saying, that AI alignment cannot be top-down, what if we just throw in some top-down instructions to AI, for example, like, try to not make humanity go extinct? Isn’t that a pretty good top-down directive?

Keyboard shortcuts

j previous speech k next speech