I liked the paper but I disagree with the primary thesis:
I suggest in the context of question-answering that what
matters when it comes to the science of AI is not a
good semblance of intelligent behaviour at all, but
the behaviour itself, what it depends on, and how
it can be achieved
I believe what matters when it comes to the science of AI is a good semblance of intelligent behaviour, regardless of how it is arrived at. As I have said before, what I strive for is the “illusion of intelligence”. If I can use some “party trick” to accomplish the task easily, so much the better.
Until we have a computational environment that more directly emulates the human brain, researchers need to take shortcuts or use other algorithms to accomplish our goals.
The ability to correctly resolve Winograd Schemas and anaphora is a nice goal, but so is being able to emulate a human in a free flowing conversation.
I would submit that resolving anaphora is not that important in a conversational AI. At worse, the AI can guess at the resolution, which gives it a 50/50 chance of being correct. If wrong, the human often corrects the bot and that resolves the issue within 1-2 volleys.
But, in a conversation, there are often clues that can give a bot a much better chance at resolving the anaphora correctly. Even if these do not exist up to the point in the conversation that the statement is made, it could be left unresolved until ultimately it can be resolved, guessed or ignored.
Leaving it unresolved (or assigning both items the attribute until resolution) is similar in concept to using particle swarms with 2 particles.
A particle consists of an assignment of a value to a set of variables and an associated weight. The probability of a proposition, given some evidence, is proportional to the weighted proportion of the weights of the particles in which the proposition is true. A set of particles is a population.
The author also points out another “cheap trick” that lets an mature AI guess at the correct statement:
If you can find no evidence for the existence of something, assume that it does not exist.
This leads the AI to increase the number of correct responses. And, all it takes is 1 correction to add to the AI’s knowledge base. What percentage of correct responses would be good enough. Would humans of different ages perform better/worse on a similar test?
We want multiple-choice questions that people can answer
easily. But we also want to avoid as much as possible questions that can be answered using cheap tricks (aka heuristics).
Consider for example, the question posed earlier:
Could a crocodile run a steeplechase?
yes
no
The intent here is clear. The question can be answered by
thinking it through: a crocodile has short legs; the hedges in a
steeplechase would be too tall for the crocodile to jump over;
so no, a crocodile cannot run a steeplechase.
Skynet-AI responded:
AI LOG:
USER:Could a crocodile run a steeplechase?
AI: Could a crocodile run a steeplechase? Why do you ask?
“Why do you ask?” may be a correct response to a nonsensical question. But, what if I also told you that the AI brought up a wiki page which contained a place called “Steeplechase Park” that included a thrill ride called “Crocodile Run”. Would the answer then be yes?
I think the real question is related to the “bot grounding” problem. There are 2 environments that AIs live in, open and closed worlds. In a closed world, everything is known. The objects, properties and relationships are all well defined. Some might say this is an expert system.
In an open world, some are known, others are unknown. Some things may be learned during the course of a conversation and some things must be calculated. The “computational linguistics” may include things like math, comparisons, reasoning and logic. Humans have a rich environment and are nurtured from birth to learn about an increasingly rich set of things in their environment. No such support system is there for an AI. Why should we expect a 4 year old AI to communicate and reason at the same level as an adult human?
The process of understanding what is known and unknown, how to represent it efficiently, how to do the comparisons and calculations that humans take for granted, and the learning methodology that integrates all, makes AI a varied and interesting field.