Google's Geoffrey Hinton, an artificial intelligence pioneer, has outlined an advance in the technology that improves the rate at which computers correctly identify images and with reliance on less data.
Hinton, an academic whose previous work on artificial neural networks is considered foundational to the commercialisation of machine learning, detailed the approach, known as capsule networks, in two research papers posted anonymously on academic websites last week.
The approach could mean computers learn to identify a photograph of a face taken from a different angle from those it had in its bank of known images. It could also be applied to speech and video recognition.
"This is a much more robust way of identifying objects," Hinton told attendees at the Go North technology summit hosted by Alphabet’s Google, detailing proof of a thesis he had first theorised in 1979.
In the work with Google researchers Sara Sabour and Nicholas Frost, individual capsules - small groups of virtual neurons - were instructed to identify parts of a larger whole and the fixed relationships between them.
The system then confirmed whether those same features were present in images the system had never seen before.
Artificial neural networks mimic the behavior of neurons to enable computers to operate more like the human brain.
Hinton said early testing of the technique had come up with half the errors of current image recognition techniques.
The bundling of neurons working together to determine both whether a feature is present and its characteristics also means the system should require less data to make its predictions.
"The hope is that maybe we might require less data to learn good classifiers of objects, because they have this ability of generalising to unseen perspectives or configurations of images," said Hugo Larochelle, who heads Google Brain's research efforts in Montreal.
"That's a big problem right now that machine learning and deep learning needs to address, these methods right now require a lot of data to work," he said.
Hinton likened the advance to work two of his students developed in 2009 on speech recognition using neural networks that improved on existing technology and was incorporated into the Android operating system in 2012.
Still, he cautioned it was early days.
"This is just a theory," he said. "It worked quite impressively on a small dataset" but now needs to be tested on larger datasets, he added.
Peer review of the findings is expected in December.
(Reporting by Alastair Sharp; Editing by Andrew Hay)
- APAC CIOs ahead of peers on AI, IoT, chatbots: Gartner survey
- Why Alibaba Cloud faces a steep challenge outside China
- Google and Salesforce ink global cloud integration deal
- Microsoft delves deeper into AI with new kit bag of tools
- Serko inks memorandum of understanding with ATPI Group for AI kit
- HPE introduces AI recommendation engine to InfoSight
- Kiwi AI start-up secures significant investment to boost shopper experience