Why asking an AI to explain itself can make things worse – MIT Technology Review
Posted: January 29, 2020 at 5:48 pm
Upol Ehsan once took a test ride in an Uber self-driving car. Instead of fretting about the empty drivers seat, anxious passengers were encouraged to watch a pacifier screen that showed a cars-eye view of the road: hazards picked out in orange and red, safe zones in cool blue.
For Ehsan, who studies the way humans interact with AI at the Georgia Institute of Technology in Atlanta, the intended message was clear: Dont get freaked outthis is why the car is doing what its doing. But something about the alien-looking street scene highlighted the strangeness of the experience rather than reassuring. It got Ehsan thinking: what if the self-driving car could really explain itself?
The success of deep learning is due to tinkering: the best neural networks are tweaked and adapted to make better ones, and practical results have outpaced theoretical understanding. As a result, the details of how a trained model works are typically unknown. We have come to think of them as black boxes.
A lot of the time were okay with that when it comes to things like playing Go or translating text or picking the next Netflix show to binge on. But if AI is to be used to help make decisions in law enforcement, medical diagnosis, and driverless cars, then we need to understand how it reaches those decisionsand know when they are wrong.
People need the power to disagree with or reject an automated decision, says Iris Howley, a computer scientist at Williams College in Williamstown, Massachusetts. Without this, people will push back against the technology. You can see this playing out right now with the public response to facial recognition systems, she says.
Ehsan is part of a small but growing group of researchers trying to make AIs better at explaining themselves, to help us look inside the black box. The aim of so-called interpretable or explainable AI (XAI) is to help people understand what features in the data a neural network is actually learningand thus whether the resulting model is accurate and unbiased.
One solution is to build machine-learning systems that show their workings: so-called glassboxas opposed to black-boxAI. Glassbox models are typically much-simplified versions of a neural network in which it is easier to track how different pieces of data affect the model.
There are people in the community who advocate for the use of glassbox models in any high-stakes setting, says Jennifer Wortman Vaughan, a computer scientist at Microsoft Research. I largely agree. Simple glassbox models can perform as well as more complicated neural networks on certain types of structured data, such as tables of statistics. For some applications that's all you need.
But it depends on the domain. If we want to learn from messy data like images or text, were stuck with deepand thus opaqueneural networks. The ability of these networks to draw meaningful connections between very large numbers of disparate features is bound up with their complexity.
Even here, glassbox machine learning could help. One solution is to take two passes at the data, training an imperfect glassbox model as a debugging step to uncover potential errors that you might want to correct. Once the data has been cleaned up, a more accurate black-box model can be trained.
It's a tricky balance, however. Too much transparency can lead to information overload. In a 2018 study looking at how professional users interact with machine-learning tools, Vaughan found that transparent models can actually make it harder to detect and correct the models mistakes.
Another approach is to include visualizations that show a few key properties of the model and its underlying data. The idea is that you can see serious problems at a glance. For example, the model could be relying too much on certain features, which could signal bias.
These visualization tools have proved incredibly popular in the short time theyve been around. But do they really help? In the first study of its kind, Vaughan and her team have tried to find outand exposed some serious issues.
The team took two popular interpretability tools that give an overview of a model via charts and data plots, highlighting things that the neural network picked up on most in training. Six machine-learning professionals were recruited from within Microsoft, all different in education, job roles, and experience. They took part in a mock interaction with a deep neural network trained on a national income data set taken from the 1994 US census. The experiment was designed specifically to mimic the way data scientists use interpretability tools in the kinds of tasks they face routinely.
What the team found was striking. Sure, the tools sometimes helped people spot missing values in the data. But this usefulness was overshadowed by a tendency to over-trust and misread the visualizations. In some cases, users couldnt even describe what the visualizations were showing. This led to incorrect assumptions about the data set, the models, and the interpretability tools themselves. And it instilled a false confidence about the tools that made participants more gung-ho about deploying the models, even when they felt something wasnt quite right. Worryingly, this was true even when the output had been manipulated to show explanations that made no sense.
To back up the findings from their small user study, the researchers then conducted an online survey of around 200 machine-learning professionals recruited via mailing lists and social media. They found similar confusion and misplaced confidence.
Worse, many participants were happy to use the visualizations to make decisions about deploying the model despite admitting that they did not understand the math behind them. It was particularly surprising to see people justify oddities in the data by creating narratives that explained them, says Harmanpreet Kaur at the University of Michigan, a coauthor on the study. The automation bias was a very important factor that we had not considered.
Ah, the automation bias. In other words, people are primed to trust computers. Its not a new phenomenon. When it comes to automated systems from aircraft autopilots to spell checkers, studies have shown that humans often accept the choices they make even when they are obviously wrong. But when this happens with tools designed to help us avoid this very phenomenon, we have an even bigger problem.
What can we do about it? For some, part of the trouble with the first wave of XAI is that it is dominated by machine-learning researchers, most of whom are expert users of AI systems. Says Tim Miller of the University of Melbourne, who studies how humans use AI systems: The inmates are running the asylum.
This is what Ehsan realized sitting in the back of the driverless Uber. It is easier to understand what an automated system is doingand see when it is making a mistakeif it gives reasons for its actions the way a human would. Ehsan and his colleague Mark Riedl are developing a machine-learning system that automatically generates such rationales in natural language. In an early prototype, the pair took a neural network that had learned how to play the classic 1980s video game Frogger and trained it to provide a reason every time it made a move.
Upol Ehsan
To do this, they showed the system many examples of humans playing the game while talking out loud about what they were doing. They then took a neural network for translating between two natural languages and adapted it to translate instead between actions in the game and natural-language rationales for those actions. Now, when the neural network sees an action in the game, it translates it into an explanation. The result is a Frogger-playing AI that says things like Im moving left to stay behind the blue truck every time it moves.
Ehsan and Riedls work is just a start. For one thing, it is not clear whether a machine-learning system will always be able to provide a natural-language rationale for its actions. Take DeepMinds board-game-playing AI AlphaZero. One of the most striking features of the software is its ability to make winning moves that most human players would not think to try at that point in a game. If AlphaZero were able to explain its moves, would they always make sense?
Reasons help whether we understand them or not, says Ehsan: The goal of human-centered XAI is not just to make the user agree to what the AI is sayingit is also to provoke reflection. Riedl recalls watching the livestream of the tournament match between DeepMind's AI and Korean Go champion Lee Sedol. The commentators were talking about what AlphaGo was seeing and thinking. "That wasnt how AlphaGo worked," says Riedl. "But I felt that the commentary was essential to understanding what was happening."
What this new wave of XAI researchers agree on is that if AI systems are to be used by more people, those people must be part of the design from the startand different people need different kinds of explanations. (This is backed up by a new study from Howley and her colleagues, in which they show that peoples ability to understand an interactive or static visualization depends on their education levels.) Think of a cancer-diagnosing AI, says Ehsan. Youd want the explanation it gives to an oncologist to be very different from the explanation it gives to the patient.
Ultimately, we want AIs to explain themselves not only to data scientists and doctors but to police officers using face recognition technology, teachers using analytics software in their classrooms, students trying to make sense of their social-media feedsand anyone sitting in the backseat of a self-driving car. Weve always known that people over-trust technology, and thats especially true with AI systems, says Riedl. The more you say its smart, the more people are convinced that its smarter than they are.
Explanations that anyone can understand should help pop that bubble.
Original post:
Why asking an AI to explain itself can make things worse - MIT Technology Review
- Facebooks Hanabi-playing AI achieves state-of-the-art results - VentureBeat [Last Updated On: December 11th, 2019] [Originally Added On: December 11th, 2019]
- Biggest scientific discoveries of the 2010s decade: photos - Business Insider [Last Updated On: December 11th, 2019] [Originally Added On: December 11th, 2019]
- DeepMind co-founder moves to Google as the AI lab positions itself for the future - The Verge [Last Updated On: December 11th, 2019] [Originally Added On: December 11th, 2019]
- AlphaGo - Wikipedia [Last Updated On: December 11th, 2019] [Originally Added On: December 11th, 2019]
- DeepMind Vs Google: The Inner Feud Between Two Tech Behemoths - Analytics India Magazine [Last Updated On: December 18th, 2019] [Originally Added On: December 18th, 2019]
- AI is dangerous, but not for the reasons you think. - OUPblog [Last Updated On: December 18th, 2019] [Originally Added On: December 18th, 2019]
- The Perils and Promise of Artificial Conscientiousness - WIRED [Last Updated On: December 18th, 2019] [Originally Added On: December 18th, 2019]
- AI has bested chess and Go, but it struggles to find a diamond in Minecraft - The Verge [Last Updated On: December 18th, 2019] [Originally Added On: December 18th, 2019]
- What is AlphaGo? - Definition from WhatIs.com [Last Updated On: December 22nd, 2019] [Originally Added On: December 22nd, 2019]
- What are neural-symbolic AI methods and why will they dominate 2020? - The Next Web [Last Updated On: January 18th, 2020] [Originally Added On: January 18th, 2020]
- AlphaZero beat humans at Chess and StarCraft, now it's working with quantum computers - The Next Web [Last Updated On: January 18th, 2020] [Originally Added On: January 18th, 2020]
- Why The Race For AI Dominance Is More Global Than You Think - Forbes [Last Updated On: February 10th, 2020] [Originally Added On: February 10th, 2020]
- AI on steroids: Much bigger neural nets to come with new hardware, say Bengio, Hinton, and LeCun - ZDNet [Last Updated On: February 10th, 2020] [Originally Added On: February 10th, 2020]
- I think, therefore I am said the machine to the stunned humans - Innovation Excellence [Last Updated On: February 10th, 2020] [Originally Added On: February 10th, 2020]
- From Deception to Attrition: AI and the Changing Face of Warfare - War on the Rocks [Last Updated On: February 20th, 2020] [Originally Added On: February 20th, 2020]
- Levels And Limits Of AI - Forbes [Last Updated On: February 20th, 2020] [Originally Added On: February 20th, 2020]
- How to overcome the limitations of AI - TechTarget [Last Updated On: February 20th, 2020] [Originally Added On: February 20th, 2020]
- The top 5 technologies that will change health care over the next decade - MarketWatch [Last Updated On: February 25th, 2020] [Originally Added On: February 25th, 2020]
- Chess grandmaster Gary Kasparov predicts AI will disrupt 96 percent of all jobs - The Next Web [Last Updated On: February 25th, 2020] [Originally Added On: February 25th, 2020]
- Enterprise AI Books to Read This Spring - DevOps.com [Last Updated On: March 14th, 2020] [Originally Added On: March 14th, 2020]
- The New ABCs: Artificial Intelligence, Blockchain And How Each Complements The Other - JD Supra [Last Updated On: March 14th, 2020] [Originally Added On: March 14th, 2020]
- The Turing Test is Dead. Long Live The Lovelace Test - Walter Bradley Center for Natural and Artificial Intelligence [Last Updated On: April 8th, 2020] [Originally Added On: April 8th, 2020]
- QuickBooks is still the gold standard for small business accounting. Learn how it's done now. - The Next Web [Last Updated On: April 19th, 2020] [Originally Added On: April 19th, 2020]
- This A.I. makes up gibberish words and definitions that sound astonishingly real - Digital Trends [Last Updated On: May 17th, 2020] [Originally Added On: May 17th, 2020]
- The Hardware in Microsofts OpenAI Supercomputer Is Insane - ENGINEERING.com [Last Updated On: June 5th, 2020] [Originally Added On: June 5th, 2020]
- Why the buzz around DeepMind is dissipating as it transitions from games to science - CNBC [Last Updated On: June 5th, 2020] [Originally Added On: June 5th, 2020]
- AlphaGo (2017) - Rotten Tomatoes [Last Updated On: June 5th, 2020] [Originally Added On: June 5th, 2020]
- AlphaGo - Top Documentary Films [Last Updated On: June 5th, 2020] [Originally Added On: June 5th, 2020]
- Enterprise hits and misses - contactless payments on the rise, equality on the corporate agenda, and Zoom and Slack in review - Diginomica [Last Updated On: June 8th, 2020] [Originally Added On: June 8th, 2020]
- Is Dystopian Future Inevitable with Unprecedented Advancements in AI? - Analytics Insight [Last Updated On: June 26th, 2020] [Originally Added On: June 26th, 2020]
- Test your Python skills with these 10 projects - Best gaming pro [Last Updated On: October 3rd, 2020] [Originally Added On: October 3rd, 2020]
- In the Know - UCI News [Last Updated On: October 3rd, 2020] [Originally Added On: October 3rd, 2020]
- How to Understand if AI is Swapping Civilization - Analytics Insight [Last Updated On: October 3rd, 2020] [Originally Added On: October 3rd, 2020]
- Investing in Artificial Intelligence (AI) - Everything You Need to Know - Securities.io [Last Updated On: November 2nd, 2020] [Originally Added On: November 2nd, 2020]
- What the hell is reinforcement learning and how does it work? - The Next Web [Last Updated On: November 2nd, 2020] [Originally Added On: November 2nd, 2020]
- An AI winter may be inevitable. What we should fear more: an AI ice age - ITProPortal [Last Updated On: December 4th, 2020] [Originally Added On: December 4th, 2020]
- Are Computers That Win at Chess Smarter Than Geniuses? - Walter Bradley Center for Natural and Artificial Intelligence [Last Updated On: December 4th, 2020] [Originally Added On: December 4th, 2020]
- What are proteins and why do they fold? - DW (English) [Last Updated On: December 12th, 2020] [Originally Added On: December 12th, 2020]
- Are we ready for bots with feelings? Life Hacks by Charles Assisi - Hindustan Times [Last Updated On: December 12th, 2020] [Originally Added On: December 12th, 2020]
- Examining the world through signals and systems - MIT News [Last Updated On: February 10th, 2021] [Originally Added On: February 10th, 2021]
- How AI is being used for COVID-19 vaccine creation and distribution - TechRepublic [Last Updated On: April 24th, 2021] [Originally Added On: April 24th, 2021]
- The 13 Best Deep Learning Courses and Online Training for 2021 - Solutions Review [Last Updated On: April 24th, 2021] [Originally Added On: April 24th, 2021]
- Why AI That Teaches Itself to Achieve a Goal Is the Next Big Thing - Harvard Business Review [Last Updated On: April 24th, 2021] [Originally Added On: April 24th, 2021]
- The Alpha of 'Go'. What is AlphaGo? | by Christopher Golizio | Apr, 2021 | Medium - Medium [Last Updated On: April 24th, 2021] [Originally Added On: April 24th, 2021]
- How will Edge Artificial Intelligence (AI) Chips Take IoT Devices to the Next Level - Enterprise Apps Today [Last Updated On: July 6th, 2022] [Originally Added On: July 6th, 2022]
- Machines with Minds? The Lovelace Test vs. the Turing Test - Walter Bradley Center for Natural and Artificial Intelligence [Last Updated On: July 6th, 2022] [Originally Added On: July 6th, 2022]
- For AI to Be Creative, Here's What It Would Take - Discovery Institute [Last Updated On: July 6th, 2022] [Originally Added On: July 6th, 2022]
- What is my chatbot thinking? Nothing. Here's why the Google sentient bot debate is flawed - Diginomica [Last Updated On: August 7th, 2022] [Originally Added On: August 7th, 2022]
- Incoherent, creepy and gorgeous: we asked six leading artists to make work using AI and here are the results - The Guardian [Last Updated On: December 4th, 2022] [Originally Added On: December 4th, 2022]
- Top 5 Applications of Reinforcement Learning in Real-Life - Analytics Insight [Last Updated On: December 4th, 2022] [Originally Added On: December 4th, 2022]
- OpenAI tweaks ChatGPT to avoid dangerous AI information - The Register [Last Updated On: December 4th, 2022] [Originally Added On: December 4th, 2022]
- Go champion who faced off against Google's AlphaGo says the rise of AI strips the games of artistry - DIGITIMES [Last Updated On: April 4th, 2024] [Originally Added On: April 4th, 2024]