After writing about Ray Kurzweil’s ambitious plan to create a super-intelligent personal assistant in his new job at Google (see “Ray Kurzweil Plans to Create a Mind at Google—and Have it Serve You”), I sent a note to Boris Katz, a researcher in MIT’s Computer Scientist and Artificial Intelligence Lab who’s spent decades trying to give machines the ability to parse the information conveyed through language, to ask him what he makes of the endeavor.

Here’s what Katz has to say about Kurzweil’s new project:

I certainly agree with Ray that understanding intelligence is a very important project, but I don’t believe that at this point we know enough about how the brain works to be able to build the kind of understanding he says he is interested in into a product.

A cross section showing the somatosensory cortex of a mouse. Neurons, at the bottom, and dendrites, reaching up, have been colored by green fluorescent protein from jellyfish (CC BY-SA 2.0).

After writing about Ray Kurzweil’s ambitious plan to create a super-intelligent personal assistant in his new job at Google (see “Ray Kurzweil Plans to Create a Mind at Google—and Have it Serve You”), I sent a note to Boris Katz, a researcher in MIT’s Computer Scientist and Artificial Intelligence Lab who’s spent decades trying to give machines the ability to parse the information conveyed through language, to ask him what he makes of the endeavor.

Here’s what Katz has to say about Kurzweil’s new project:

I certainly agree with Ray that understanding intelligence is a very important project, but I don’t believe that at this point we know enough about how the brain works to be able to build the kind of understanding he says he is interested in into a product.

I previously interviewed Katz for an article about Apple’s Siri (see “Social Intelligence”). He explained that constructing meaning from language goes well beyond learning vocabulary and grammar, often relying on a lifetime of experience with the world. This is why Siri is only capable of responding to a fairly narrow set of questions or commands, even if Apple’s designers have done a clever job of making Siri seem as if it’s understanding goes much deeper.

Kurzweil believes he can build something approaching human intelligence by constructing a model of a brain based on simple principles and then having that model gorge itself on an enormous quantities of information—everything Google indexes from the Web and beyond.

There are reasons to believe this type of approach might just work. Google’s own language translation technology has made remarkable strides simply by ingesting vast quantities of documents already translated by hand and then applying statistical learning techniques to figure out what translations work best. Likewise, IBM’s Watson demonstrated a remarkable ability to answer Jeopardy questions by applying similar statistical techniques to information gathered from sources including the website Wikipedia (see “How IBM Plans to Win Jeopardy!”). But this is very different from the way humans develop an understanding of the world and of language.

If it does not provide an accurate representation of how the brain works, the question is whether Kurzweil’s approach will hit a wall in terms of simply mimicking that understanding by producing really useful responses to very sophisticated questions.

To read more, click here.