What AI can and can’t do (yet) for your business

Here is a brief excerpt from an article written by Michael Chui, James Manyika, and Mehdi Miremadi for the McKinsey Quarterly, published by McKinsey & Company. To read the complete article, check out other resources, learn more about the firm, obtain subscription information, and register to receive email alerts, please click here.

To learn more about the McKinsey Quarterly, please click here.

*     *     *

Artificial intelligence is a moving target. Here’s how to take better aim.

Artificial intelligence (AI) seems to be everywhere. We experience it at home and on our phones. Before we know it—if entrepreneurs and business innovators are to be believed—AI will be in just about every product and service we buy and use. In addition, its application to business problem solving is growing in leaps and bounds. And at the same time, concerns about AI’s implications are rising: we worry about the impact of AI-enabled automation on the workplace, employment, and society.

A reality sometimes lost amid both the fears and the headline triumphs, such as Alexa, Siri, and AlphaGo, is that the AI technologies themselves—namely, machine learning and its subset, deep learning—have plenty of limitations that will still require considerable effort to overcome. This is an article about those limitations, aimed at helping executives better understand what may be holding back their AI efforts. Along the way, we will also highlight promising advances that are poised to address some of the limitations and create a new wave of opportunities.

Our perspectives rest on a combination of work at the front lines—researching, analyzing, and assessing hundreds of real-world use cases—and our collaborations with some of the thought leaders, pioneering scientists, and engineers working at the frontiers of AI. We’ve sought to distill this experience to help executives who often, in our experience, are exposed only to their own initiatives and not well calibrated as to where the frontier is or what the pace setters are already doing with AI.

Simply put, AI’s challenges and limitations are creating a “moving target” problem for leaders: It is hard to reach a leading edge that’s always advancing. It is also disappointing when AI efforts run into real-world barriers, which can lessen the appetite for further investment or encourage a wait-and-see attitude, while others charge ahead. As recent McKinsey Global Institute research indicates, there’s a yawning divide between leaders and laggards in the application of AI both across and within sectors (Exhibit 1).

Leaders in the adoption of AI also intend to invest more in the near future compared with laggards.

Executives hoping to narrow the gap must be able to address AI in an informed way. In other words, they need to understand not just where AI can boost innovation, insight, and decision making; lead to revenue growth; and capture of efficiencies—but also where AI can’t yet provide value. What’s more, they must appreciate the relationship and distinctions between technical constraints and organizational ones, such as cultural barriers; a dearth of personnel capable of building business-ready, AI-powered applications; and the “last mile” challenge of embedding AI in products and processes. If you want to become a leader who understands some of the critical technical challenges slowing AI’s advance and is prepared to exploit promising developments that could overcome those limitations and potentially bend the trajectory of AI—read on.

Challenges, limitations, and opportunities

A useful starting point is to understand recent advances in deep-learning techniques. Arguably the most exciting developments in AI, these advances are delivering jumps in the accuracy of classification and prediction, and are doing so without the usual “feature engineering” associated with traditional supervised learning. Deep learning uses large-scale neural networks that can contain millions of simulated “neurons” structured in layers. The most common networks are called convolutional neural networks (CNNs) and recurrent neural networks (RNNs). These neural networks learn through the use of training data and backpropagation algorithms.

While much progress has been made, more still needs to be done.1 A critical step is to fit the AI approach to the problem and the availability of data. Since these systems are “trained” rather than programmed, the various processes often require huge amounts of labeled data to perform complex tasks accurately. Obtaining large data sets can be difficult. In some domains, they may simply not be available, but even when available, the labeling efforts can require enormous human resources.

Further, it can be difficult to discern how a mathematical model trained by deep learning arrives at a particular prediction, recommendation, or decision. A black box, even one that does what it’s supposed to, may have limited utility, especially where the predictions or decisions impact society and hold ramifications that can affect individual well-being. In such cases, users sometimes need to know the “whys” behind the workings, such as why an algorithm reached its recommendations—from making factual findings with legal repercussions to arriving at business decisions, such as lending, that have regulatory repercussions—and why certain factors (and not others) were so critical in a given instance.

Let’s explore five interconnected ways in which these limitations, and the solutions emerging to address them, are starting to play out.

[Here’s the first.]

Limitation 1: Data labeling

Most current AI models are trained through “supervised learning.” This means that humans must label and categorize the underlying data, which can be a sizable and error-prone chore. For example, companies developing self-driving-car technologies are hiring hundreds of people to manually annotate hours of video feeds from prototype vehicles to help train these systems. At the same time, promising new techniques are emerging, such as in-stream supervision (demonstrated by Eric Horvitz and his colleagues at Microsoft Research), in which data can be labeled in the course of natural usage.2 Unsupervised or semisupervised approaches reduce the need for large, labeled data sets. Two promising techniques are reinforcement learning and generative adversarial networks.

Reinforcement learning. This unsupervised technique allows algorithms to learn tasks simply by trial and error. The methodology hearkens to a “carrot and stick” approach: for every attempt an algorithm makes at performing a task, it receives a “reward” (such as a higher score) if the behavior is successful or a “punishment” if it isn’t. With repetition, performance improves, in many cases surpassing human capabilities—so long as the learning environment is representative of the real world.

Reinforcement learning has famously been used in training computers to play games—most recently, in conjunction with deep-learning techniques. In May 2017, for example, it helped the AI system AlphaGo to defeat world champion Ke Jie in the game of Go. In another example, Microsoft has fielded decision services that draw on reinforcement learning and adapt to user preferences. The potential application of reinforcement learning cuts across many business arenas. Possibilities include an AI-driven trading portfolio that acquires or loses points for gains or losses in value, respectively; a product-recommendation engine that receives points for every recommendation-driven sale; and truck-routing software that receives a reward for on-time deliveries or reducing fuel consumption.

Reinforcement learning can also help AI transcend the natural and social limitations of human labeling by developing previously unimagined solutions and strategies that even seasoned practitioners might never have considered. Recently, for example, the system AlphaGo Zero, using a novel form of reinforcement learning, defeated its predecessor AlphaGo after learning to play Go from scratch. That meant starting with completely random play against itself rather than training on Go games played by and with humans.3

Generative adversarial networks (GANs). In this semisupervised learning method, two networks compete against each other to improve and refine their understanding of a concept. To recognize what birds look like, for example, one network attempts to distinguish between genuine and fake images of birds, and its opposing network attempts to trick it by producing what look very much like images of birds, but aren’t. As the two networks square off, each model’s representation of a bird becomes more accurate.

The ability of GANs to generate increasingly believable examples of data can significantly reduce the need for data sets labeled by humans. Training an algorithm to identify different types of tumors from medical images, for example, would typically require millions of human-labeled images with the type or stage of a given tumor. By using a GAN trained to generate increasingly realistic images of different types of tumors, researchers could train a tumor-detection algorithm that combines a much smaller human-labeled data set with the GAN’s output.

While the application of GANs in precise disease diagnoses is still a way off, researchers have begun using GANs in increasingly sophisticated contexts. These include understanding and producing artwork in the style of a particular artist and using satellite imagery, along with an understanding of geographical features, to create up-to-date maps of rapidly developing areas.

*     *     *

Here is a direct link to the complete article.

Michael Chui is a partner of the McKinsey Global Institute (MGI) and is based in McKinsey’s San Francisco office; James Manyika is the chairman of MGI and a senior partner in the San Francisco office; and Mehdi Miremadi is a partner in the Chicago office.The authors wish to thank Jack Clark at OpenAI, Jeffrey Dean at Google Brain, Professor Barbara Grosz at Harvard University, Demis Hassabis at DeepMind, Eric Horvitz at Microsoft Research, and Martin Wicke for their insights on the ideas in this article. They also wish to thank their McKinsey colleagues Steven Adler, Ali Akhtar, Adib Ayay, Ira Chadha, Rita Chung, Nicolaus Henke, Sankalp Malhotra, and Pieter Nel for their contributions to this article.

 

Posted in

Leave a Comment





This site uses Akismet to reduce spam. Learn how your comment data is processed.