AI May Be Catching up With Human Reasoning

AI May Be Catching up With Human Reasoning

AI May Be Catching up With Human Reasoning GA S REGULAR Menu Lifewire Tech for Humans Newsletter! Search Close GO News > Smart & Connected Life

AI May Be Catching up With Human Reasoning

Matching wits against algorithms

By Sascha Brodsky Sascha Brodsky Senior Tech Reporter Macalester College Columbia University Sascha Brodsky is a freelance journalist based in New York City. His writing has appeared in The Atlantic, the Guardian, the Los Angeles Times and many other publications. lifewire's editorial guidelines Published on April 12, 2022 10:13AM EDT Fact checked by Jerri Ledford Fact checked by Jerri Ledford Western Kentucky University Gulf Coast Community College Jerri L. Ledford has been writing, editing, and fact-checking tech stories since 1994. Her work has appeared in Computerworld, PC Magazine, Information Today, and many others. lifewire's fact checking process Tweet Share Email Tweet Share Email Smart & Connected Life Mobile Phones Internet & Security Computers & Tablets Smart Life Home Theater & Entertainment Software & Apps Social Media Streaming Gaming Researchers have created techniques that let users rank the results of a machine-learning model's behavior. Experts say the method shows that machines are catching up to humans' thinking abilities. Advances in AI could speed up the development of computers' ability to understand language and revolutionize the way AI and humans interact. KanawatTH / Getty Images A new technique that measures the reasoning power of artificial intelligence (AI) shows that machines are catching up to humans in their abilities to think, experts say. Researchers at MIT and IBM Research have created a method that enables a user to rank the results of a machine-learning model's behavior. Their technique, called Shared Interest, incorporates metrics that compare how well a model's thinking matches people's. "Today, AI is capable of reaching (and, in some cases, exceeding) human performance in specific tasks, including image recognition and language understanding," Pieter Buteneers, director of engineering in machine learning and AI at the communications company Sinch, told Lifewire in an email interview. "With natural language processing (NLP), AI systems can interpret, write and speak languages as well as humans, and the AI can even adjust its dialect and tone to align with its human peers."

Artificial Smarts

AI often produces results without explaining why those decisions are correct. And tools that help experts make sense of a model’s reasoning often only provide insights, only one example at a time. AI is usually trained using millions of data inputs, making it hard for a human to evaluate enough decisions to identify patterns. In a recent paper, the researchers said that Shared Interest could help a user uncover trends in a model’s decision-making. And these insights could allow the user to decide whether a model is ready to be deployed. “In developing Shared Interest, our goal is to be able to scale up this analysis process so that you could understand on a more global level what your model’s behavior is,” Angie Boggust, a co-author of the paper, said in the news release. Shared Interest uses a technique that shows how a machine-learning model made a particular decision, known as saliency methods. If the model is classifying images, saliency methods highlight areas of an image that are important to the model when it makes its decision. Shared Interest works by comparing saliency methods to human-generated annotations. Researchers used Shared Interest to help a dermatologist determine if he should trust a machine-learning model designed to help diagnose cancer from photos of skin lesions. Shared Interest enabled the dermatologist to quickly see examples of the model’s correct and incorrect predictions. The dermatologist decided he could not trust the model because it made too many predictions based on image artifacts rather than actual lesions. “The value here is that using Shared Interest, we are able to see these patterns emerge in our model’s behavior. In about half an hour, the dermatologist was able to decide whether or not to trust the model and whether or not to deploy it,” Boggust said. The reasoning behind a model’s decision is important to both machine learning researcher and the decision-maker.

Measuring Progress

The work by MIT researchers could be a significant step forward for AI’s progress toward human-level intelligence, Ben Hagag, head of research at Darrow, a company that uses machine learning algorithms, said that told Lifewire in an email interview. “The reasoning behind a model’s decision is important to both machine learning researcher and the decision-maker,” Hagag said. “The former wants to understand how good the model is and how it can be improved, whereas the latter wants to develop a sense of confidence in the model, so they need to understand why that output was predicted.” But Hagag cautioned that the MIT research is based on the assumption that we understand or can annotate human understanding or human reasoning. “However, there is a possibility that this might not be accurate, so more work on understanding human decision-making is necessary,” Hagag added. metamorworks / Getty Images Advances in AI could speed up the development of computers’ ability to understand language and revolutionize the way AI and humans interact, Buteneers said. Chatbots can understand hundreds of languages at a time, and AI assistants can scan bodies of text for answers to questions or irregularities. “Some algorithms can even identify when messages are fraudulent, which can help businesses and consumers alike to weed out spam messages,” Buteneers added. But, said Buteneers, AI still makes some mistakes that humans never would. “While some worry that AI will replace human jobs, the reality is we’ll always need people working alongside AI bots to help keep them in check and keep these mistakes at bay while maintaining a human touch in business,” he added. Was this page helpful? Thanks for letting us know! Get the Latest Tech News Delivered Every Day Subscribe Tell us why! Other Not enough details Hard to understand Submit More from Lifewire Your Car's Driving Assistance Tech Isn't Meant to Be Used Alone—Here's Why Your Next Flight Might Be More On-Time Thanks to AI What Is Artificial Intelligence? What Is a Neural Network? Artificial Intelligence vs. Machine Learning Why AI Needs to Be Regulated The Four Types of Artificial Intelligence AI Might Not Be Your Best Source for Advice Just Yet How AI Could Monitor Its Dangerous Offspring Artificial Intelligence Isn't Taking Over Anytime Soon, Right? Experts Wonder if AI Is Creating Its Own Language AI Discoveries Could Soon Power Your Car New Tech Could Let Gadgets Understand Your Conversations How Satellite Images Could Improve Lives AI Could Give 3D Printers New Capabilities AI Could Stop Snooping By Predicting What You’ll Say Newsletter Sign Up Newsletter Sign Up Newsletter Sign Up Newsletter Sign Up Newsletter Sign Up By clicking “Accept All Cookies”, you agree to the storing of cookies on your device to enhance site navigation, analyze site usage, and assist in our marketing efforts. Cookies Settings Accept All Cookies
Share:
0 comments

Comments (0)

Leave a Comment

Minimum 10 characters required

* All fields are required. Comments are moderated before appearing.

No comments yet. Be the first to comment!

AI May Be Catching up With Human Reasoning | Trend Now | Trend Now