The Independent's journalism is supported by our readers. When you purchase through links on our site, we may earn commission. 

Electronic tongue reveals ‘inner thoughts’ of AI

Researchers say device gives glimpse into mysterious decision-making process of artificial intelligence

Anthony Cuthbertson
Thursday 10 October 2024 20:11 BST
Comments
The electronic tongue uses graphene and an artificial neural network to detect different tastes
The electronic tongue uses graphene and an artificial neural network to detect different tastes (Das Lab)

Your support helps us to tell the story

This election is still a dead heat, according to most polls. In a fight with such wafer-thin margins, we need reporters on the ground talking to the people Trump and Harris are courting. Your support allows us to keep sending journalists to the story.

The Independent is trusted by 27 million Americans from across the entire political spectrum every month. Unlike many other quality news outlets, we choose not to lock you out of our reporting and analysis with paywalls. But quality journalism must still be paid for.

Help us keep bring these critical stories to light. Your support makes all the difference.

Scientists have developed an electronic tongue capable of discerning between different tastes better than a human.

A team from Penn State in the US claims the graphene-based device holds the potential to “revolutionise” the detection of chemical and environmental changes, which could be used in everything from medical diagnostics, to spotting when food has been spoiled.

The new technology also offers unique insight into the “inner thoughts” of artificial intelligence, which has largely been an opaque field until now due to something referred to as the black box problem.

The team were able to do this by reverse engineering how the neural network came to its final determination when identifying differences between various types of milk, coffee and fizzy drinks.

This process gave the researchers “a glimpse into the neural network’s decision-making process”, which they claimed could lead to better AI safety and development.

“We’re trying to make an artificial tongue, but the process of how we experience different foods involves more than just the tongue,” said Saptarshi Das, a professor of engineering science and mechanics at Penn State.

“We have the tongue itself, consisting of taste receptors that interact with food species and send their information to the gustatory cortex — a biological neural network.”

The neural network used by the electronic tongue was able to reach a tasting accuracy of more than 95 per cent when put against human-selected parameters.

Using a method called Shapley additive explanations, the researchers could delve into the decision-making process of the neural network.

Rather than assessing individual human-assigned parameters, the neural network considered the data it determined was most important for assessing different tastes.

“We found that the network looked at more subtle characteristics in the data – things we, as humans, struggle to define properly,” Professor Das said.

“And because the neural network considers the sensor characteristics holistically, it mitigates variations that might occur day-to-day. In terms of the milk, the neural network can determine the varying water content of the milk and, in that context, determine if any indicators of degradation are meaningful enough to be considered a food safety issue.”

The research was detailed in a study, titled ‘Robust chemical analysis with graphene chemosensors and machine learning’, published in the journal Nature.

Join our commenting forum

Join thought-provoking conversations, follow other Independent readers and see their replies

Comments

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in