"Any intelligent fool can make things bigger and more complex... It takes a touch of genius - and a lot of courage to move in the opposite direction."I agree. Most likely the solution to General Artificial Intelligence is simple and elegant. The current models seem ... complex.
Ray Kurzweil talks about how our intelligence is built from layers of simple mechanics. Each part of the layer works identically, but it gathers info from the prior layer. I like this idea, it sounds "simple". The difficulty is determining what the "rules" are for the smallest element in the layer.
Monica Anderson talks about a non-model based approach. Specifically one in which "intuition" plays the largest role in "getting to" intelligence. She is definitely on to something. Her ideas that the world as we see it is not built on specific models (forumula's) is right on target.
My primary concern with this field is that there is no clear agreed upon definition of when "intelligence" has been reached. Like wise there is no test for intelligence, at least for a digital avatar. We have the Turing Test but that comes up short on a few fronts, in my opinion.
Recall that the goal of General or Strong Artificial Intelligence is to create a machine that can successfully perform any intellectual task that a human being can.
In simple words, it has to be able to learn anything.
So we need simple rules that can enable a system to learn anything.
From this thought my mind jumps through the following sequence of thoughts:
- Humans have many senses, AI would need to have many types of input.
- Wait, an infant, in his mom's belly has very few senses.
- Rather, he has all senses, but the input coming in is very small.
- AI with multiple inputs should be able to "learn" something" with very few inputs
- As we grow new senses are introduced one at a time. We grow fingers before ears, ears before eyes, and so on.
- Perhaps the method by which the AI can learn, needs to be adaptable to many different types of inputs
- So all inputs regardless of source must resolve to the same "signals"
- I think all the input needs to be in a binary state
- Or all inputs need to be on a sigmoid curve (value between 0 and 1
Now I know that I have Sigmoid coming into my input layer...what rules can I build around this?
Another interesting item, at a certain point in our learning, we do not need a lot of exposure to new input to "learn" a new concept. For example, the effort to learn how to add is significantly harder than the effort to learn how to subtract, once we have learned how to add.
Put another way, when placed in a new environment, but the environment is "similar" to something we know, we learn quickly. The more the environment is different the slower we learn.
What this means is that once our AI has learned a specific subject area, learning in that area should be quick. For example, a maze solving AI would solve mazes quickly with few "learning" cycles, but that same AI placed in a Turing test would take "longer" to learn how to interact within that system.