With all due respect, one sentence explaining how you think the mind works isn't really worth much. It doesn't amount to much more than "the brain tries to explain reality." Yes, ok, but how do you translate that into some algorithm? How does it relate to gradient descent methods on neural networks?
Surprise can be defined as an unexpected difference or anomaly. One Class Classification or anomaly detection can be a very good trainer for a generic AI.
Et cetera, et cetera. The length of something does not necessarily imply that an idea is weak, maybe the idea is really deep? Dismissing an idea based on length is idiotic.