I’ve put together some notes on my experience of this year’s Loebner Prize final.
I arrived at Bletchley Park at around 9:45 am and after signing in, went to the Education Centre at Block B where this year’s contest as to be held.
The 4 AI laptops were up and running and after saying hello to Hugh and the organisers, I set up Tutor and Mitsuku on their allocated machines. The installation process went extremely well and I had both bots installed and configured to run within 10 minutes.
This year, the AISB were installing some extra node.js software on the machines and I believe that this was related to the live webcast, as well as removing the network delay that plagued the competition in 2014. Unfortunately, this software proved to be causing lots of problems, as I was unable to test that Mitsuku and Tutor were working and had no network delays when talking via the judge program until around 1:20pm, the contest was due to start at 1:30pm.
Eventually, all the problems with it were resolved and the contest started around 1:50pm, 20 minutes after the scheduled start time.
This year’s judges were:
Joanne Pransky: World’s First Robot Psychiatrist - http://www.robot.md
David Boyle: Author of “Alan Turing: Unlocking the Enigma”
Jow Hewitt: Brand Strategist. Landor - http://landor.com
Tom Cheshire: Technology Correspondent with Sky News - http://news.sky.com/technology
The human confederates were:
Lisa Guthrie - Kingston University
Emily Donovan - 2MUL
Memo Akten - Goldsmiths
Prashant Aparajeya - Goldsmiths
The details of the judges and confederates were unknown, as the last page of the handout was deliberately unavailable until after the contest. I suggested this last year, as it would be easy for a judge to do something like:
Judge: What is your name?
Entity: I am called Maria
Judge: Well according to the handout, none of the confederates are called Maria and so you must be a bot.
Round 1 got underway with no hitches. However about 10 minutes into the round, the organisers noticed that Rose had a “debug error” message displayed on the screen but she looked to be responding to the judge’s messages and was displaying them back to the judge, so they decided to leave it running.
In round 2, I noticed from watching the output on the webcast that Arckon was producing lots of dduupplliiccaattee letters. I can only assume this was due to the program not clearing the program folder quick enough after processing each character. I mentioned it to the organisers but there was little they could do, as it seemed to be the program itself. Unfortunately, this meant it was producing nonsense and didn’t stand much of a chance this year.
Rose had the same error message on screen as round 1 but still seemed to be working.
Round 3 saw the judge talking to Rose leave the room about 5 minutes into the round, saying he had something he needed to do. I voiced my opinion that I thought this was unfair towards Rose and we should wait for him to return to restart the round but it appears that Rose wasn’t responding at all at this point. Dr Keedwell tried to contact Bruce and I posted a message on chatbots.org and Twitter mentioning this but I don’t believe Bruce saw it.
Round 4 saw Mitsuku fail to respond to any of the judge’s input. I checked the program folder and could see the judge was talking to her but Mitsuku refused to respond. Fortunately, as I was on site, I could soon remedy the fault, which was down to a config file on the organiser’s node.js setup which needed amending between rounds and she started working again after about 5 minutes.
The results saw three of the judges score Mitsuku in top place and one of the judges ranked Rose as the top bot. This meant that Mitsuku won with a score of 1.25 - The best score possible is 1 (lower is better)
The final placings were:
1 - Mitsuku
2 - Tutor
3 - Rose
4 - Arckon
Sky News were filming for the event and broadcast several times during the day on their TV channel. A company from Channel 4 television here in the UK were also filming for a documentary. The event was well attended by the public who seemed to be enjoying the exhibits of a NAO robot, software where they could make their own collages from internet photos and even the opportunity to take a Turing Test with a chatbot for themselves.
Suggestions for next year:
1 - Do a mock set up of the computers before the contest day. This will allow plenty of time to sort out any issues, as it looked like the contest was going to be delayed by quite a while this year. Luckily, it was only 20 minutes.
2 - The programmers should be available on standby to assist the organisers in case of difficulties or at least have a representative at the contest. Had I not been on site to fix Mitsuku, I would have missed an entire round. Technical issues with the programs meant that in reality, it was a contest between Mitsuku and Tutor this year and although naturally, I am overjoyed to have won. It would have been good to win with all 4 bots showing their full potential.