Why Machines will become Hyper-Intelligent before Humans do     <home>

These notes detail some of the reasons why it will be much easier to build hyper-intelligent non-organic machines, than to dramatically enhance individual human intelligence – why SI (super intelligences) will appear before radical human IA (intelligence augmentation).

In this context, super/ hyper intelligence refers to general intelligence far exceeding current top human abilities in (at least) the areas of abstract thinking, innovation, and practical problem solving; Systems obviously clever enough to understand human language and affairs, science and technology, and – crucially – to be able to radically enhance their own design.

Preamble - Setting the context: (for a wider overview see: Advanced Intelligence: SI, IA, and the Global Brain)

Intelligence, versus knowledge and tools (embedded knowledge/ ability) – Intelligence is a measure of what can be accomplished utilizing given knowledge and tools. It includes the ability to develop novel ideas and tools from an existing base. Abstract, inductive thought and creativity are crucial aspects of this.

Individual versus collective smarts/ Maximum versus aggregate intelligence – Individual, maximum intelligence is an overall limiting factor. It imposes limits on the maximum complexity of problems that can be effectively tackled. Quantity can ultimately not make up for quality: A million monkeys can not solve problems that individual humans can - A million porn or Pokemon messages will not cure cancer (see note1). This is not to say that group interaction and collaboration add nothing - they clearly do increase overall smarts.

Rates of change in individual human intelligence, machine 'intelligence', and overall human ability -

Will machines remain extensions of our capabilities – tools, under our control? It seems unlikely. Illustration: Imagine that we engineered or bred animals for maximum intelligence. To the extent that we succeeded, they would likely disagree with some of our goals and agendas. For example, one can easily imagine an 'Animal Liberation' movement.

Would SIs at least be supportive of our goals? If intelligence in machines does increase much faster than in humans, then their relationship to us (positive, negative, or neutral) will determine whether we will eventually be able to catch-up/ integrate. However, there would be some lag (though not necessarily very long) between the time machines become smarter than us, and when they become smart enough to uplift us – dramatically upgrading brains is a harder problem than achieving nanotech or curing ageing.

This leads to the questions: Is it in fact easier to dramatically increase machine or human intelligence? Can technological advances in intelligence be integrated into humans as fast as they can be applied to machines? Naturally, underlying this discussion is the assumption that SI is possible

Advantages of working with designed instead of evolved systems - machines, rather than brains:

Engineered solutions are much easier to understand, modify, enhance, and debug. Also, we don't have to limit ourselves to the single solution to intelligence created by a blind, unconscious Watchmaker with his own agenda (survival) -

Advantages of working with artificial systems instead of humans - hardware/ software, rather than wetware:

Artificial systems development offers more flexibility, ease of design, speed, scalability, and better financial return -

Capitalizing on a massive hardware overhang:

There are several ways in which current and near-term processing capacity can be leveraged much more effectively (millions of times), thus mitigating apparent performance limitations. Capitalizing on this overhang will benefit machine development much more than human augmentation efforts (scalability, self-improvement, etc) -

Illustration: Imagine that a widely accepted, workable theory/ design for AI existed, and that there was a particularly urgent need for it (avert imminent meteor strike): I am sure that we could 'find' millions of times more effective processing power than is currently employed for AI (by the means listed above). I predict that much better theories will soon emerge, and that commercial forces alone will have much of the driving force as imminent disaster.

Some additional problems with radical human intelligence enhancements/ upgrades:

Notes -

1) - A similar constraint applies to machines: Maximum intelligence of individual designs will ultimately determine the level of overall smarts in a network. Furthermore, tightly coupled machine clusters/ units are likely to be most intelligent because of the importance of coordinated, high-bandwidth meta-cognition and sub-process communication.

2) - Who is working on 'real' AI? Of all the people working in the field called 'AI'...
80% don't believe in the concept of General Intelligence (but instead, in a large collection of specific skills & knowledge)
of those that do, 80% don't believe that (super) human-level intelligence is possible - either ever, or for long, long time
of those that do, 80% work on domain specific AI projects for commercial or academic-political reasons (results are a lot quicker)
of those left, 80% have a poor conceptual framework....

Peter Voss, June 2001