Will our human intuition be a match for true artificial intelligence?

The year 1997 saw the ultimate man versus machine tournament, with chess grandmaster Garry Kasparov losing to a machine called Deep Blue.

Earlier this year, in what was hailed as another breakthrough in artificial intelligence (AI) research, Google’s AlphaGo defeated a professional Go player.

Go is an ancient Chinese board game that has hitherto been difficult for a computer to play at a high level due to its deceptively complex gameplay. Where chess is played on a board of 8 x 8 squares, Go is typically played on a board of 19 x 19 squares.

These are all worthy engineering achievements, but what does it mean for research into genuine machine intelligence and the predicted artificial intelligence that will surpass human intelligence?

Arguably, not much. To understand why, we need to delve a little more into the complexity of the games and the differences between how machines and humans play.

It is estimated the number of possible games of chess is 10120 while the lowest limit of games for Go is 10360. These numbers are big, even for a computer. If you’re not quite convinced of this, the estimated number of atoms in the observable universe is merely 1079 – minuscule in comparison.

Game-playing AI still cannot foresee every possible game play and, just like us, has to consider the options and make a decision on what move to make. For brevity, we’ll mainly stick with chess as it’s more widely known. Let’s look at how a computer plays first.

The machine

Most chess programs operate via brute-force search, which means they look through as many future positions as they can before before making a choice.

This results in a tree of possible combinations called the search tree. Here’s an example:

An example of a search tree for a particular game. David Ireland, Author provided
An example of a search tree for a particular game. David Ireland, Author provided

The search tree starts with a root that represents current game play. And the branches are all the possible game plays. Each level of the tree is called a ply, which is a single move by a player.

Deep Blue’s specialised hardware allowed it to search future game play at a staggering 200 million chess positions per second. Even today, most chess AI programs only compute about 5 million positions per second.

Not only does the AI have to search through a large collection of chess positions, but at some stage, it must evaluate them for their potential worth. This is done by a so-called evaluation function.

Deep Blue’s evaluation function was developed by a team of programmers and chess grandmasters who distilled their knowledge of chess into a function that evaluates piece strength, king safety, control of the centre, piece mobility, pawn structure and many other characteristics — everything a novice is taught.

This allows a particular board position to be scored with a single number. Think of the evaluation function as something like this:

How a chess-playing AI evaluates the chessboard. David Ireland, Author provided
How a chess-playing AI evaluates the chessboard. David Ireland, Author provided

The higher the number, the better the position is for the machine. The machine seeks to maximise this function in its favour, and minimise it for its opponent.

The human

A person, in stark contrast, only considers three to five positions per second, at best. How, then, did Kasparov give Deep Blue a run for its money?

This question has fascinated cognitive scientists who have yet to agree on a computational theory on how even an amateur plays chess.

Nevertheless, there’s been extensive psychological research into the cognitive processes involved in how players of various strengths perceive the chessboard and how they go about selecting a move.

Studies conducted on eye movements of expert players as they select a move showed little consistency with searching a tree of possible moves. People, it seems, pay more attention to squares that contain active attacking and defending pieces and perceive the pieces on the board as groups or chunks rather than as individual pieces.

In an even more revealing experiment, novice and expert players were shown a chess position taken from a game for five seconds. They were then asked to reproduce the board from memory. Expert players were able to reconstruct the board much more accurately than novice players.

Curiously, when they were asked to reconstruct a board that had the pieces randomly distributed, experts did no better than novices.

It is believed that through constant play, a player accumulates a large number of chunks that could be thought of as a language of chess. These chunks were not present with the randomly distributed board and, as such, the experts’ perception was no better than that of the novice.

This language encodes positions, threats, blocks, defences, attacks, forks and the many other complex combinations that arise. It allows players to determine and prioritise pressures on the board and reveal opportunities and dangers.

The language of chess is a higher level of perception of the chessboard that still alludes AI and cognitive science researchers.

Let’s take a look at an interesting position.

What is white’s winning strategy?

Two kings are on either side of a pawn blockade. White has an opportunity to promote the pawn on F6 to a stronger piece. But that square is being guarded by the black king.

An example problem where human intuition often triumphs over AI. David Ireland, Author provided
An example problem where human intuition often triumphs over AI. David Ireland, Author provided

For white to win, the white king must move around the blockade via column A and force the black king away. Defeat for black is then inevitable.

Simple enough? Not entirely for a chess AI, which has more difficulty perceiving white’s advantage. This is because it would need to search to a depth of 20 ply to find white’s advantage. In this position, at 15 plies there are 10,142,484,904,590 possible positions (we tried computing to 20 plies but after one week of computation, we gave up).

Most computer chess programs won’t see the winning strategy. Instead, they will move the white king to the centre of the board which is the common strategy when there are only a few pieces on the board.

Human intuition is still a powerful force.

Higher level of perception

A famous AI researcher, Professor Douglas Hofstader, believes analogy is the core of cognition. We humans, certainly bring our own analogies to the game: gambits, sacrifices, and blockades, among other things.

Alas, research into the field of cognitive science has waned over the past decade in favour of more practical and profitable direct AI approaches as seen in Watson and AlphaGo.

Nevertheless, there has been sporadic research output on so-called cognitive architectures (CHREST) that model human perception, learning, memory, and problem solving.

Some play chess (CHUMP) not by searching for a plethora of combinations but by perceiving patterns and relationships between pieces and squares. And just like most humans, they play mediocre chess.

It’s worth pondering: if true artificial intelligence is established, will it begin with an explosion of intelligence or something smaller and imperceptible?The Conversation

David Ireland, Electronic Engineer and Research Scientist at the Australian E-Health Research Centre, CSIRO

This article was originally published on The Conversation. Read the original article.


  1. The test position you posted is absolutely wrong. Many engines find the winning strat in less than 100ms.

    info depth 1 score cp 313 time 0 nodes 6 nps 6000 hashfull 0 pv f6f7
    info depth 2 score cp 285 time 0 nodes 56 nps 56000 hashfull 0 pv e1d2
    info depth 3 score cp 283 time 0 nodes 139 nps 139000 hashfull 0 pv e1d2
    info depth 4 score cp 246 time 0 nodes 243 nps 243000 hashfull 0 pv e1d2
    info depth 5 score cp 232 time 0 nodes 510 nps 510000 hashfull 0 pv e1d2
    info depth 6 score cp 246 time 0 nodes 870 nps 870000 hashfull 0 pv e1d2
    info depth 7 score cp 271 time 0 nodes 1464 nps 1464000 hashfull 0 pv e1d2
    info depth 8 score cp 234 time 16 nodes 2684 nps 157882 hashfull 0 pv e1d2
    info depth 9 score cp 283 time 16 nodes 4260 nps 250588 hashfull 0 pv e1d2
    info depth 10 score cp 249 time 16 nodes 6639 nps 390529 hashfull 0 pv e1d2
    info depth 11 score cp 259 time 16 nodes 8323 nps 489588 hashfull 0 pv e1d2
    info depth 12 score cp 246 time 16 nodes 12298 nps 723411 hashfull 0 pv e1d2
    info depth 13 score cp 259 time 16 nodes 14744 nps 867294 hashfull 0 pv e1d2
    info depth 14 score cp 246 time 32 nodes 19725 nps 597727 hashfull 0 pv e1d2
    info depth 15 score cp 271 time 32 nodes 23692 nps 717939 hashfull 0 pv e1d2
    info depth 16 score cp 246 time 32 nodes 27612 nps 836727 hashfull 0 pv e1d2
    info depth 17 score cp 271 time 32 nodes 34144 nps 1034666 hashfull 0 pv e1d2
    info depth 18 score cp 451 time 47 nodes 44809 nps 933520 hashfull 0 pv e1d2

  2. I only observe that the human response to advances in artificial intelligence is all-too-human. For many years, chess was held up as a benchmark of human intelligence, but when Deep Blue became unbeatable, we reconsidered. Go became a new standard, for we saw that it was a substantially more intuitive and supposedly less mechanistic game. But now inevitably Go too as succumbed to automation. But no matter, we’re going to shift the goal posts again.
    Now, I don’t say any of this cynically, but rather to highlight the very human ability to change the rules.
    I don’t know how humans do it, but it can’t be algorithmic can it? How could an algorithmic intelligence ever truly change its own rules? If we programmed a program to change its own rules, it could only do so within a framework that was programmed into it. That is, it would need to have a set of rules for breaking the rules.
    But the rules for breaking rules couldn’t be broken.
    I like it that the author David Ireland referenced Douglas Hofstadter, who also believes that recursion and “strange” loops are inherent to consciousness. Maybe the rules for breaking the rules for breaking the rules is not in fact an infinite regression but instead bottoms out after a few loops, and can therefore be programmed.
    Steve Wilson, Constellation Research.

  3. The discussion is predicated on a very poor mechanistic question.

  4. It depends on the definition of true artificial (digital) intelligence. There are numerous animals which exhibit degrees of intelligence more superior, more responsive than machines capable of pattern recognition, game playing and counting. However, digital intelligence is much less, perhaps even unrestricted, as to what it can achieve. Having abilities to comprehend written text and to appreciate this, with an ability to respond accordingly, are well within the digital capacity of current moderate size machines.

Commenting on this post has been disabled.