GENERATIVA AI IN YOUR BUSINESS

0:00 TODD GLICKMAN: But now, I'm very pleased to introduce Dr. George 0:08 Westerman, Principal Research Scientist at the MIT Sloan School of Management. Dr. Westerman is a world-recognized thought leader 0:16 in leading transformation and competitive advantage through technological innovation. 0:22 He works at the really exciting intersection of executive leadership and digital innovation 0:28 to provide continuous, valuable insights to leaders in all industries. George, welcome. 0:33 GEORGE WESTERMAN: Thanks, Todd. 0:39 It's really great to be here. And it's especially great because at home, we 0:44 got 5 inches of snow yesterday. And there was no snow on the ground here. So that's a good thing. 0:50 So yeah, I teach in the management school. And so what I focus on-- he said technology, innovation, 0:56 and leadership. I try to take technologies and make them-- to demystify them, to help leaders understand what's happening there 1:03 and really think about how do you design your organization and how do you lead transformation? So I did some of the earliest research 1:10 on digital transformation 15 years ago. To tell you how long ago that was, nobody used the word digital transformation back then. 1:18 And so I wrote some book on that. I had done a work with IT organizations before that. I've done work in skills and careers. 1:26 And now, of course, AI, all of those things come together. AI is affecting all of that. 1:32 And so what I want to do today is talk to you a little bit about how do you think about generative AI and how do you work that in your strategy? 1:40 Now, Eric has been doing AI for almost as long as I've been alive. 1:46 And I want to try to tell you about AI from a management 1:52 standpoint. And that means it will-- I'll try to make it crystal clear for you in trying to make 1:57 decisions in your organization. But it also means that if Eric and I disagree, he's correct. 2:05 So that's where we want to go. My job is to be as clear as possible, help you figure out the questions you can ask, 2:11 the instincts to develop, as you are making decisions, because all of you, probably, are having to make decisions 2:17 with AI. How can AI help you understand that a little bit better? Or if I could say it a different way, 2:23 how can I make it a little bit less scary and make you feel more competent in dealing with this stuff? So are you ready to go? 2:31 Three pieces here-- number one is, what is AI? Now, Eric already talked about what is AI. 2:36 I'm going to give you my version of it. And think about these elements of the different kinds of AI 2:42 and how to think about it. Number two is GenAI because that's really where we're spending a lot of time now, 2:47 a lot of attention on ChatGPT, and Gemini, and Midjourney, and all these, Sora, and all these others. 2:52 How do you think about making that work in your organization? Because it's tremendously powerful, but also very risky. 2:59 And how do you think about that? And then the last is, what does that mean? How are companies innovating with this right now? 3:06 So first, what is AI? And like I said, it's a little bit scary, after having followed an AI professor. 3:12 But like I said, it's mostly correct and hopefully very clear when we talk about it. What is AI? 3:17 Well, the thing about what is AI is, it's really hard. The definitions Eric gave were from 2003. 3:24 Now, Google does not track Google Trends back in 2003. They start in 2004. 3:30 And look at generative AI. Nobody was talking about it because it didn't exist back 3:35 then. But then deep learning, he talked about that, and you could see how that came on 10 years ago or something 3:42 like that. That was really cool. We had a thing called the Work of the Future Initiative. 3:48 And the question there was, with deep learning doing what it is doing, what jobs will be left 3:54 after the robots eat the jobs? And we found that, actually, they don't eat the jobs. They eat pieces of the jobs. 4:01 And also, there's a lot of extra effort that goes in. It's not just the technology. But that was the scariest, most interesting AI. 4:09 And now, what do we call that? We call that traditional AI. That's how fast these things move. 4:14 Go back a little bit farther, this is machine learning. And you can see, why is it bigger? 4:22 Well, it's actually a superset of these other two and then artificial intelligence. And the interesting thing about this, you see, 4:28 artificial intelligence was more common. And now, it's less common because artificial intelligence 4:34 back then was a different kind of thing than it is now, just like Eric was talking about. So with these elements, how do you make sense out of this mess 4:43 when it keeps moving so quickly? I'm going to really first give you my important law 4:49 that if you remember nothing else, remember this. You know Moore's law, Moore's law and the exponential. 4:57 These things keep growing at an exponential rate. This is Westerman's law. 5:04 Westerman's law is different. Technology changes quickly. Organizations change much more slowly. 5:11 I know this to be true. I should not have to say this is true. But I know it to be true. And you know it to be true. 5:16 But our technology people often forget this. And your technology vendors often forget this. 5:23 The hard part is not adopting the technology. It's changing the way you do business. 5:29 It's not the digital. It's the transformation, is the hard part. And so what does that mean? 5:35 Technology is not the problem. Transformation is. So it's not really a technical problem. It's also very much a leadership problem. 5:44 We need to get both of those things in there. And so, for example, in talking about AI, 5:49 we've talked to two-- we've talked to a lot of executives. But here's two that we can quote. Matthew Evans, who worked on AI for the manufacturing 5:57 process at Airbus, he says, "Strictly speaking, we don't invest in AI or natural language processing or image. 6:04 We are always investing in a business problem." Or if we go on to the Home Depot, Fahim Siddiqui-- 6:11 another thing I do is I run the MIT Sloan CIO Leadership Award every year. 6:16 And Fahim was one of our finalists last year. And Home Depot is a home goods store. 6:22 If you need hardware, or saws, or wood, or anything, you go to Home Depot. 6:27 I'm renovating a house right now, so I spent a lot of time in Home Depot. And Fahim and his team have created a really good experience 6:34 for a really, really complicated store. He says, we should always be looking for extraordinary experiences We want to bring joy to the user. 6:43 We want to delight. The technology-- that's secondary. And once again, we forget that all the time. 6:50 But we can't forget that because technology provides zero value to your company. 6:56 What you do with the technology is what creates that value, how you change your business or your products to make it work better. 7:04 So what's the first thing from a management side? What's the first thing you need to know about AI? 7:12 Artificial intelligence is not intelligent. And Eric mentioned this, and we should continue that. 7:22 Basically, it's a program that executes, but it doesn't have that context knowledge. 7:28 And we have to remember that. It's basically executing a formula. And that formula is what you've programmed it to do, 7:34 what it's learned how to do. But it may not be the right thing. 7:40 Aude Oliva, who has created AI that can actually read what you are thinking, with brain scans, 7:47 one of the smartest people I know, she says, artificial intelligence should be artificial idiots. 7:54 That's one way to think about this. It's not intelligent. But the thing is, it can act very intelligently. 8:00 And that's very useful, as long as we are careful, as long as we do it in the right way. 8:07 So the digital transformation research-- this is an update. We started this in 2010. 8:13 This is an update from 2021. Where can you look for opportunities not from AI, not just from AI, but from mobile, and from Internet of Things, 8:22 and from other technologies you're facing? Four good areas you can look-- one is creating an emotionally engaging, targeted, personalized 8:33 customer experience, number one. Number two is operations, not just automation, 8:39 but the ability to adapt and adjust as things go on. 8:45 That's industry 4.0. Business models-- not just being the Alibaba or the Amazon 8:51 of your industry, but what can you do with information to make more low-hanging fruit, turning 8:57 your products into services, these kinds of things? And what we rediscovered-- we should have noted before-- 9:02 employee experience. Employee experience is a critical part. We know that satisfied employees lead to satisfied customers. 9:10 We also know that if you have a bad employee experience, it's a very good indication that your systems and your processes 9:18 are not running the way they should or your incentives. So when you're looking for what to do with AI, 9:23 it's the same way we think about any other digital transformation work. You can look in these areas, and not only in these areas, 9:31 but the opportunity to go across, like Home Depot has done, like Airbus has done. You'll see also that this rides on top 9:38 of your systems, your data, your systems. And if that's a mess, it's hard to do this. 9:44 AI actually helps a little bit with that messy data. But the better your data is, the cleaner your processes are, 9:51 the better you will be. So I really think of AI as the next stage 9:57 of digital transformation. The same principles apply. But there's more that you can do. 10:02 And there are more powerful opportunities. And so we still need to lead this, but we need to think more and more about what we can do. 10:09 So here are some things we could do with generative AI. I have designed courses over time. 10:16 We're doing one in the Philippines right now, a set of courses. And instead of me standing up, and putting a suit, on and going 10:23 in front of a camera, now we just type words in, and she will say it instead. And if you don't like how she looks, 10:29 somebody else that looks different will do it. And they can do it in a different language. And frankly, I can do it with a deepfake very easily. 10:36 I don't need to go to a studio anymore. Also, any corporate literature you have 10:43 can be in any language you want, instantly, with a push of a button. 10:48 Over here, certainly coding-- many people are doing coding right now. How many of you are your programmers? 10:54 Are they using Copilots and things like that? Well, if they are not, it's time to start. 11:00 This really does help. Now, there are questions about whether it's better for the senior people or the junior people, 11:06 but there's a learning process. And it is better. It's not only better for coding, but also enforcing standards 11:12 and making documentation. People hate doing that. The computer can help them. 11:17 What else is there? Well, Cresta. Cresta is a call center tool, especially for sales. 11:23 And you heard some really good startups. In a randomized trial at MIT, they found that everybody using this tool got better. 11:33 The most senior people got 14% better. The most junior people got 34% better. 11:41 If you have the senior and the junior people, they raise the bar for everybody, and it moves up. What does it do? 11:47 It listens to you talk and it gives you hints while you are talking. That person is getting confused. 11:54 Explain the product better. That person is getting angry. Try this to calm them down. 12:01 And then at the end, it comes back and says, looking across what you've been doing, you don't close fast enough. 12:07 Try to get to the sale faster, these kinds of things. It becomes what a good supervisor would be. 12:12 But it's with you all the time. And I'm working on a project right now with the media lab, where we are creating a personalized 12:18 tutor for the programming class, the first Python class, for minority institutions, minority learning institutions. 12:31 The idea is, if we can get people through that first class, with an individualized tutor, they are on a set for a career 12:38 that they never would have had possible. But people drop out of that first class because they do not have a tutor. So we're developing that right now. 12:44 And then last but not least, those of you who are using any products, whether it's SAP, or Workday, or Adobe, this is being integrated 12:51 into all of your products. So this is happening all over the place. But once again, it's not the technology, 12:57 it's what do you do with the technology that makes it better. The other thing that's really important here is we talk about generative AI solutions. 13:05 But most of the best solutions are a combination of generative AI, traditional AI, boring old IT, 13:14 and also people and process in there. So Lemonade, they have a very specific niche 13:21 in the insurance market. They write 98% of their policies, 98% of first claim notices, and they handle 50% of claims automatically. 13:30 Through this combination of different AIs and also traditional systems, what about that 50%? 13:36 Well, the easy stuff, the computer does it. And the hard stuff, they give it to a person to figure out what to do. 13:42 That's how their model works. 13:47 And this is Sysco. This is not the technology company Cisco. This is Sysco, the food service delivery. 13:54 They have one of the largest truck fleets in the world. Restaurants around the world get supplied by these. 14:01 And here are all the different ways that Sysco is applying AI in what they do. 14:06 Now, this is not a high tech company. They deliver cans of stuff to restaurants. But you can see all the different things 14:12 they have on the customer experience side over there and also on the back office. 14:17 And when Tom Peck, when I was talking with him about what he does, all those little brains, 14:24 those are opportunities where generative AI can help. We think about generative AI in helping a salesperson figure 14:33 out their call planning. But how about generative AI in helping you route things around a warehouse or a-- 14:42 we don't have this thing they want. We don't have this kind of mushroom. But this kind of mushroom will work, so all kinds 14:49 of opportunities there. So if you ask three AI experts, what are the categories of AI, 14:58 you will get five different answers. Nobody can really agree. So I'm going to give you my four categories. 15:05 And I'm not going to tell you this is right. I'm going to tell you it's mine, if that helps, 15:11 so four categories. And this should look familiar because Eric talked about a lot of this. And what I want to do is walk you through these, 15:18 just to give you a little bit more instinct on what to do about these things. So number one, rule-based systems-- Eric called it expert systems. 15:25 Put a lot of if/then statements together. 1984, when I first started programming, 15:30 one of my first jobs was to create an engine for a rule-based system. And like Eric said, these things don't work beyond the simplest 15:38 problems. But now, rules engines are better and if you use them in the right environments, prescriptions, 15:45 loan making these kinds of things. Within the limits in which they are good, they are useful. 15:51 But don't try to get outside that context. Then there's econometrics, which is statistics. 15:58 Most of us learned in school deep learning, generative AI. Let's walk through these things. So expert systems, we talked about this. 16:05 How do you program them? You talk to an expert. And the expert tries to figure it out. 16:11 But there's a nice paradox, which is that we know more than we can tell. Or as Eric said, your two-year-old 16:17 understands physics but couldn't tell you how that works. And we have that problem all the time. 16:23 But you have to talk to an expert. You don't need any data, the expert. The good thing is, they will give you a precise answer, 16:31 and they will give you the same answer every time. But they do not adapt. 16:38 You add a rule, things get a little tricky. So that's one thing. Econometrics, that's the next thing. 16:43 That's statistics. Many of you have probably learned statistics. You've probably forgotten much of it. 16:48 But we still use this all the time. If you have structured data, meaning data that you can put into a spreadsheet, usually 16:56 numeric data, these things work. And they work really well. They're pretty cheap to program and especially 17:04 the two-click things. How do we create a formula for those two dimensions 17:09 that will help us to figure out, is it a blue thing or a green thing? And they tend to be-- there are false positives. 17:16 There are false negatives. But they tend to be pretty good. The other is looking at the trend line, looking at the regression 17:22 These things actually work remarkably well with numeric data and also things 17:27 you can turn into numeric data, like is he going to buy my product based on what I do to him? 17:33 You can also do multiple dimensions. Finding that trend line with two dimensions is easy. 17:39 Try it with 100 dimensions. I have a team right now currently looking 17:44 at 100 million CVs, resumes, and tracking people's careers 17:50 through that 100 million resumes to try to figure out what is a more or less productive path. 17:56 You're not going to do that by just looking at two dimensions. And then, of course, you don't have to just look at lines. 18:03 But you do have to have an idea of what the functional form is that you're looking at. So it's often numeric. 18:08 There are some things, but you can do this. It's relatively cheap. It works really well. You get precise answers. 18:13 You get the same answer every time. But it has to be numeric data. Then you have deep learning. 18:19 And Eric talked a lot about deep learning. This was the coolest thing ever 15 years ago. And it's still pretty darn cool. 18:26 Now, the thing is, Eric also showed you there are 50 different versions of this. So I'm just going to give you one archetype. 18:33 This picture-- do you know what that picture is? It's a neural net. But what this picture mostly is? 18:40 It's a thing that people use to scare you. It's a complicated-looking thing. 18:45 And there's a little bit of magic involved in there. So Eric explained this, but let me explain it also a little bit more. 18:50 You're taking inputs. You're running them through a set of weighted averages and coming out with a prediction of multiple different states 18:58 at the end. And it sounds very easy, but it gets beyond human capability 19:06 really fast. You train it with labeled data. You need to know the truth. 19:11 That's why every time you log in, it says prove you are human. Tell us where there is a car. 19:17 Tell us where there's a bicycle. You are labeling data to train these things. Every time you get on social media and they do the de-aging 19:23 challenge-- show us you and show us you 20 years ago-- you're creating labeled data of old and young people. 19:31 You need that data. And then you do the-- the outputs are repeatable. But they are in no way explainable. 19:37 Our smartest people are just starting to figure out how to make this explainable. And it's going to be a little while. 19:42 But I'm going to give you an example now. At MIT-- I'm going to get a little complicated on you, 19:49 but I promise, not too complicated. The good thing is, I'm going to try to take this scary picture and make it a little bit more real. 19:55 Eric started. I'll try it. And after the two of us, hopefully, it will make sense. 20:01 What are those? Anybody? They're numbers. 20:07 Now, try to tell me, if we were to say features, loops 20:13 and lines and these kinds of things, tell me the formula to look at the number 2. 20:20 How would you create the rules to decide what is a number 2? 20:27 Well, it has a little loop at the top. And it has a loop down here. And then the number goes down. 20:32 That's how a two looks, except no loop there. 20:38 And there the line is going across instead of going down. If you were to try to program this with a bunch of features, 20:46 you would probably not get very far. But 35 lines of code, 25 to 35 lines of code 20:51 can do this no problem at all by doing this kind of thing. What do you do? 20:56 Well, first of all, you take these two-dimensional objects and you turn them into a one-dimensional set of numbers. 21:03 So 28 by 28 image becomes 784 pixels. Now, what you have just done is you've 21:09 thrown away all the relationships between what is on top of each other. You've thrown it away. And the algorithm doesn't care. 21:16 And then you put the numbers in the left hand side, and you have random numbers here. And it spits out 10 estimates, 0 through 9, 21:25 how likely is it that that is the number. And so you come through, and you get this. 21:31 Well, we think maybe it's a 2, a 1. We think maybe it's a 2. 21:41 Maybe it's a 7 and some stronger or less strong estimates. And then you go back and say, hey, it's really a 9. 21:48 And you go back the other way, and you turn all the knobs to make it a little bit better. 21:57 Now, if that bothers you, how much do you turn each knob, it should bother you because nobody knows 22:05 how much you turn each knob. But we do know, if you do this 10,000 times, the knobs end up in the right place. 22:12 So if it sounds like magic to you, it is magic. It's the magic of large numbers and good algorithms. 22:18 And so if you're wildly uncomfortable, then you are like the rest of us. We are all wildly uncomfortable with this. 22:24 But it does work. So you do that 10,000 times, and it gets there. 22:29 Now, what does that mean? Well, that means you've got to have labeled data. You need to know what that number really is. 22:35 You need to do the reinforcement over and over again. And you need to back chain to adjust those numbers. 22:42 If your data is biased-- let's say we have only data on men, but not women. 22:48 Well, then what happens when a woman comes in? Amazon found this out. They were reviewing resumes to try 22:54 to figure out what resumes look like their best engineers. And they were routinely rejecting women. 23:03 Why was that? None of their engineers were women. And women talk differently. 23:09 By the way, if you were the captain of the women's swimming team, you would never get a job. None of the men that they trained it on 23:16 were ever captain of the women's swimming team and also different use. So they had to go and fix that bias, same thing 23:23 if you train on men versus women, children versus adults, different parts of the world. 23:28 So you got to get rid of that. You got to test that bias. You also need to make sure you actually have some kind of accuracy in your data. 23:35 So that's it, no more on the neural nets. Do you understand that picture? I could test you, and you would know how to say that now? 23:45 Well, at least nobody can scare you anymore. So then generative AI continues. And as Eric said, what they do now, 23:52 not only they're using a lot of these similar algorithms, but they're also inventing. And they're saying, given what you told me and everything 23:59 I know, what's the next best word or pair of words to say? And then it says, once I have that, 24:06 given everything you told me and everything I know, plus everything I just invented, what's the next best word? 24:12 What's the next best word? And it can get things very right, as you've seen. It can also get them very wrong. 24:19 I first did my bio with ChatGPT. And the bio was really good, expert on digital transformation 24:25 and innovation strategy, author of four award winning books, blah, blah, blah. I loved it. 24:31 I've only written three books. I don't know what that fourth book is. 24:37 And it wouldn't tell me. So you want to be careful. But the interesting thing about that is-- well, 24:42 I'll come back to that in a minute. It randomly generates next. So why do you get a different answer every time? 24:48 Because it is random. It's meant to be that way. And you can use that. 24:53 It creates new things. It doesn't just classify. It creates new. It can be used for good as bad. 25:00 You can get some boring stuff. You can also get creative stuff. But you also get this, the hallucinations. 25:06 My favorite is this. This lawyer prepared all of his court documents using ChatGPT. 25:12 And unfortunately, some of the cases that he cited as precedent for what he's doing were not real. 25:20 That made the judge very angry. So not only did he get in trouble with the judge, 25:25 but he ended up in all the business press as being the laughingstock. Don't do that. But here's the thing. 25:31 We are so worried about these hallucinations. But how many of us know a perfect employee? 25:40 People make mistakes too. And so the idea is not to expect a perfect answer, maybe, except if your car is driving. 25:47 But the answer is to get these things to work well and put the right controls in, in case they are wrong, 25:54 the same thing we do with people. So why do we expect the computers to be better than people in every case? 26:00 Why don't we just put the right processes in place to account for the fact that these are sometimes not right? 26:05 The other things, of course, huge training data this thing 26:11 needs and huge amounts of energy. And so these are other topics we can talk about over time. 26:18 So the problem is, you got to start with the problem 26:23 and figure out the right technique. So one way to think about it is there are questions, you can ask. How accurate do I have to be? 26:29 And what's the cost of being wrong? Getting in a traffic accident or making a wrong medical decision, 26:36 there's a big cost. Sending the wrong marketing message to somebody, no cost at all. 26:43 Do you need the answer to be explainable? Because if it does, then those first two are probably useful. 26:49 Those last two, they're not explainable. How do you do that? Do you need the answers to be the same every time? 26:57 You're not going to get that with generative AI. And confidentiality-- and confidentiality is less of a problem right now. 27:03 You can get these generative things that are going to keep you, or at least the vendors promise they will keep you safe. 27:09 But these are things to think about, right? Then on the other side is the data. Do you have a source of truth, and how true is that data? 27:18 And number two, is it generalizable? Or are you going to reject all women that 27:23 apply for your company? So as you're thinking about how to do it, these are questions you can ask about this problem I'm 27:31 trying to solve right now. And the questions, either you can make the decision 27:37 or you can ask these questions of your technical people. And first of all, they will now be a little bit afraid of you. 27:42 But second of all, you will have a better conversation because you are asking these questions. So first bit, what is AI? 27:52 Four things. Here's how to think about them. Here, if you like traffic lights, here's a nice little chart you can use 27:58 to think about these things. Where are they good and bad? But that's only part of the problem. 28:06 That's the part saying, I have a specific problem I want to solve. How do I think about what the right technique is? 28:12 The real problem is how do I make this work in my organization? 28:19 So I have more questions for you, maybe some more answers. Are you ready? 28:28 How do I make it work in the organization? Remember, come back to this technology is not the problem. 28:34 Transformation is. How do we put the processes, the policies, 28:39 and the capabilities in place to do this the right way? Because doing it once is hard. 28:45 But you need to do it over and over, and you need to make it mix in with the rest of the stuff that you do. 28:50 Here are three challenges to think about when you try to make this stuff work in your organization. 28:57 And I'd like to say, this is only GenAI, but this is basically anything. How do we work this in? 29:03 But with generative AI, how do we prioritize what to do and what not to do? What do we do first? 29:10 What do we do second? And what do we never do? Number two, risk management. 29:16 What if we are wrong? What if we have a privacy problem? These kinds of things. And last is the capabilities, so four sets of questions 29:23 you can ask for that. And then after this, I'll have one more little section for you. 29:30 How do you ensure the safety and the value of what you're doing? And one of the big other ones is learning. 29:37 How do you not only learn about this innovation, but how do you learn across innovations so you can choose the best way to do things, 29:44 and you don't end up doing the same thing 10 times? 29:50 One of the big questions to think about here is, how does your governance process look? 29:58 And what we've seen in generative AI is multiple approaches, but they tend to fall into top down, 30:05 this is a very risky, very expensive thing. Let's control everything. 30:10 And the good thing about that is, we're not going to make mistakes. We're not going to throw money away. 30:15 We're probably not going to kill a lot of people. But we're also not going to innovate very well 30:21 because the central part of large organizations, it doesn't always know the best opportunities. 30:27 The other side is the decentralized way. Hey, go out and innovate. 30:32 Do whatever you want. Now, here are some rules. Don't steal money. 30:37 Don't be bad to people. But here are some rules. Innovate within these rules, and tell us about what you're doing. 30:45 It's a tremendously good way to discover really interesting ideas at the edges. But it's a great way to waste money also. 30:52 And also, depending on how careful they're being, you can also break some laws when you do that because where the central might try to be very 30:59 careful, the edges might not. And so what we saw in generative AI is the very top down, safe, but slow, or the very decentralized, 31:10 very fast, but more risky and more costly. We saw both of those coming out. So at Societe Generale, one of the big French banks, 31:19 they did it in a very centralized way. But they asked everybody to say, if you want to do this, 31:26 what would you do? And they got 700 use cases. And what they decided is they did a few of those right away. 31:34 But the rest they said, wait, because all these cases, an intelligent agent is important. 31:40 A chatbot to talk to people is important. Programming is important. Let's do this the right way, and we will build everything on top. 31:47 So they did a few things fast. They told other things to wait until they 31:53 had these other capabilities set up. They tried to bridge the centralized and decentralized. 31:58 Sysco that I told you about, they did a different way. They said you know what? This is just another technology. 32:06 We have a way we think about this. And the first thing we think about is, can we buy this instead of building it? 32:14 And if we can, we will do that. The second thing they say is, if we think it needs to be programmed, can we do it cheaper and easier 32:21 with expert systems and statistics? Or do we do the fancy, complicated stuff? 32:27 And only when it answers all those questions do they do the leading edge, the generative AI. 32:32 So you see these governance rules, they fit in there. You got to figure out what works with your company, 32:38 and how does it fit the risk profile that you're thinking about? 32:43 Some other questions you want to ask, though, are also just as important. Is our culture ready for this? 32:51 What happens when the computer is just as smart as our expert person? 32:57 15 years ago, this started happening in banks, where the computers were as good at judging a high quality or a low quality loan 33:05 than people were-- or insurance. The computers were just as good at underwriting some things 33:11 as people. The people felt very threatened by that. And there was a whole process. 33:17 It's happening in all kinds of jobs. Now, if you've seen the movie Moneyball, have you seen-- 33:22 I don't know whether-- In America, baseball is a big deal. I don't know if it is here. But there's a big question. 33:28 How do I tell who is a good potentially new employee for this company? 33:34 Maybe the computer is better, and people don't like that. So do we have the humility to work with these? 33:40 Or do we fight against them? Do we have ethics to do the right thing and not 33:46 the wrong thing? And also in the culture, how good are we 33:51 at experimenting and trying things and failing fast, rather than trying to have the first answer, 33:57 the right answer, before we start? The culture's got to be ready for these things. 34:04 And the way you do that, you have to drive that. You don't just say plug in the model. You got to help people get comfortable, 34:09 with innovation units, other things. And another one also with people being comfortable, 34:15 do we have the skills? But what does this mean for careers? 34:22 A friend of mine, Daniel Rock, looked at GPTs, and he has calculated through a very rigorous process 34:28 that 46% of all jobs are likely to have 50% of their tasks replaced by AI over time. 34:38 Now, some jobs, it'll be 100%. These are just-- 46% of jobs will have half of their tasks 34:46 go. What if you do that job? How do you feel about that? And am I going to-- am I going to go along with this? 34:53 Or am I going to reject this because I think it's going to hurt my job? So what do you what do you do about this? 34:59 Certainly, I've done some writing on this. And certainly after lunch, you're going to see some really interesting stuff on that. 35:06 But AI doesn't have to replace you. How can it make your job easier, reduce your cognitive load? 35:12 How can it do-- one of the things I do for all the time, I'm a pretty good writer. But I don't do those sexy, creative titles. 35:19 So I let AI do it now. Give me five potential titles for this. 35:25 Or I can give it my text and say, make this more exciting. And the other thing we did not expect 35:32 is the ability to learn with AI. AI is a tremendous teaching tool. 35:37 And I'm finding I'm learning all the time, just by having it help me think, having it ask questions about what I'm doing. 35:44 And certainly, I talked about this tutor we are doing. And many places are doing these tutors, including at Cresta 35:50 and at one of the startups we talked about. So it does not have to replace us. 35:55 Are you having that conversation with your people about how this-- you're not just saying, we're not going to hurt you because nobody believes that. 36:02 But can you have the conversation about how we can help you introduce this? 36:08 So, for example, Dentsu Creative, an advertising firm, that is a firm that you might think should be a little bit nervous about these things. 36:15 And what they're finding is that these creative people love this. 36:23 But they're being very systematic in the company about how they introduced this. First of all, they are saying all the boring stuff. 36:29 Your planning, your proposals, those kinds of things, it will do that for you, take a few words 36:36 and turn it into a proposal. That's great. What they didn't expect and that they love now is, let's draw a picture. 36:43 And instead of saying, go away, and in a week, we will have a design for you, now they 36:48 say, go get a cup of coffee, and in five minutes, we'll have a design for you. And now, we can iterate that and iterate it. 36:54 You can do it with the client. It's better for both. They didn't expect that. The other thing they're doing, though, is they're not saying, 37:01 you use it. You use it. What they're saying is, everybody use it. We will help you. And let's get together and share your ideas. 37:08 What is working for you? What's not working for you? These office hours where people share their tricks. So people are investing as a group in getting better. 37:15 And you see what happens. It frees up the time that creatives need to be creative. 37:22 All the stuff they hate, they don't have to do. They can't see themselves going back to the old way. 37:28 So what they did, they were very careful about helping people understand how this can be good for them, not just bad for them. 37:34 They involved them in that process. And that's how they avoided the rejection, and they get better with this. 37:41 In coding, in your marketing people, in many places, with customer service, you will have this thing going on, 37:48 a lot of fear. We've done a thing called the Global Opportunity Forum, where 37:54 we are getting companies, talking to companies about career questions. How are you developing your people? 38:01 What are the skills that you need? What's the best way that you can do to help your people grow with you, instead 38:06 of being afraid as you grow? And so if you're interested, contact me about how we are getting companies together 38:12 to think through this challenge. So that's that. You ready for the last part of this talk? 38:20 We talked first of all about types of AI, four types. And if you have a particular problem, 38:26 how do you decide on the right type? And I gave you questions you can ask. Then we talked about having your company have the capability 38:35 to use this well to transform in a better way and questions you can ask there about culture and helping 38:41 your people be ready. So we just completed a study-- it's coming out tomorrow 38:47 in Sloan Management Review-- about how our company is looking at transformation with generative AI. 38:53 And we were looking for these giant transformations. Let's change the entire underwriting process in a firm. 39:00 Or let's turn over all of our sales and customer service to a computer. 39:06 We didn't find those. But we did find a lot of smaller transformations happening 39:13 that people are starting to get a lot of value from this and setting themselves up for the bigger transformation 39:19 over time. So what we do is, if you think about the big transformations as being transformation with a big capital T, 39:27 what we found is a company are doing a lot of transformation with a little t, smaller transformations. 39:33 But they're doing it in a systematic way that is helping them get ready. And it looks like this. 39:41 Level 1 is, most companies already are thinking about individual productivity. 39:48 Here is a version of the LLM that is safe to use. 39:53 Or here's our own private version. McKinsey, for example, now has an LLM 39:58 that looks across all of their slide decks. And it's not perfect. But if you need to know about energy generation in Southeast 40:06 Asia, McKinsey probably did this once before. And it can help to answer some questions for you. 40:12 So not only the public things, but the private things they're starting to do from an individual productivity 40:17 standpoint, very low risk. They're just informing people. It's a good way to get started. 40:24 Number two is the specialized roles and tasks. Let's start to transform that. So we're seeing it in call centers a lot. 40:30 We're seeing it in coding. We're seeing it in some other things. Typically, a human still stays in the loop. 40:36 But sometimes for low risk things, they're letting the computer take over, like at Lemonade. 40:42 Lemonade is 50% of things get processed automatically, 50%. The more risky things they take on. 40:49 And so that's starting to happen. And then what we're not seeing as much of, except in technology companies, is actually 40:56 doing the direct customer impact. So how do we think about this? In the first level, individual productivity, 41:03 summarize these documents. Many banks I know that the minute a company releases its financials, a minute later, 41:12 five minutes later, all the spreadsheets are updated. You had to have a lot of interns that did that before. 41:19 Now, it just happens. These kinds of things can happen. Sum-- well, sorry. That's the next piece. But summarizing documents, hey, what 41:25 just happened at my meeting? And in the company specific LLMs, the opportunities in the tasks are like that one. 41:32 Can we make sense of the latest quarterly report for financials? Can we think about customer support? 41:38 That's human-in-the-loop. They tend to be lower risk. And you keep the human in the loop 41:44 to help mediate that risk even more. And last but not least, we're seeing this farther thing, 41:50 the direct engagement. And we're seeing this in online right now. The people that make Coach and Kate Spade things, 41:58 things that only this big, but they cost a lot of money, they're personalizing a very conversational approach 42:06 to help you. Just as if you went in a store and that person would make you buy that handbag that you charge 42:11 too much money for, now online, they can start to do the same thing. Relatively low risk, those kinds of things 42:18 are starting to happen and also the idea of doing the first or second tier of customer service. 42:24 And these are happening. The product companies are already doing it because they have to. 42:29 What we're seeing less of is this bottom thing, transformation of large processes. 42:36 What will happen, from what we're hearing, is that you're not going to get GenAI replacing the whole. 42:41 But combinations will do it. GenAI will have a part. So many companies, we already have the forms that you give. 42:49 Generative AI will translate that into data that the rest of the processes can use. 42:55 And then it will translate it back into conversational later. You'll see these combinations happening more and more. 43:02 So you want to think about the risks and the capabilities. What's happening here, we call this the risk slope, 43:08 how as you grow, you grow your risk management capability, but you also grow your capability 43:14 to do more and more bigger things over time. Here's the reason why this is so important. 43:23 The proofs of concept, it's easy to do them. But making it work in a large group of users or customers, 43:29 getting from the lab to reality is really hard. And in fact, in a large bank, she said, 43:35 "The more stuff you do, the more stuff you find to do." She did not mean more features we can build. She meant more errors we have to solve. 43:41 And one way to think about it is H&M, the fashion company, 43:48 "It's almost like putting a tire on a car." How many of you have changed a tire on your car? 43:54 Not many people. Well, if you've ever changed the tire, you know that you do not put a bolt on really hard and another one, because you will bend the tire. 44:01 You'll bend the rim. Put a little bit, a little bit, a little bit, a little bit. That's how you want to think about your AI capability 44:08 in your companies. Do it here. Learn from there for the next thing, the next thing 44:13 So what's the conclusion? I just gave you a lot of stuff. 44:19 But that's OK. This is MIT. You're smart people AI can seem to be intelligent, but be 44:28 intelligent in how you use it. But remember, just because it's not perfect, that doesn't mean it's bad. 44:36 Your people also make errors. Let's put the right processes, right controls, to deal with the fact that people are sometimes wrong. 44:44 Number two, start with the problem, not the technology. And many times, the answer will be combinations, not just 44:50 one of these. You look at the task that needs to be done. Number three, get started now because if you're 44:57 going to put that tire on your car a little bit, you work up the growth slope, the risk slope. 45:02 And every time you do something, you learn more to be better at it. 45:08 You got to help your people be ready, because if they're not ready, they will fight you. 45:13 And they will fight you either very actively, or more likely, they'll just say, oh, that's hard. I don't know what to do. 45:19 Those of you who have teenagers, you know they do both. And last but not least, continuously improve. 45:27 What can you do with small t transformations that will help the large T transformations later on? 45:33 So I have exactly ten seconds left. So I'm going to ask for questions, 45:39 even though I'm not allowed to. So let's take a few questions. But I won't go too far over. Can we do some questions? 45:46 So by the way, lots of good reading if you want it, including the new article, so thank you very much. 45:53 What should we do with questions? They're up here. 46:00 I don't get to choose the question. This is scary. Our agency is pushing everyone to use AI, even the blue collar 46:07 workers. Is this the right move? Is it necessary for everybody? The interesting question thing that you said in there 46:12 is, they're pushing everyone. How can you encourage everyone? 46:17 Do you see the difference? If I say, you have to do this, half the people are going to say, you can't tell me what to do. 46:25 If you say, hey, this is a fun set of tools, let's try to help you think through it, they might be a little bit more interested. 46:31 So what I suggest is, encourage everybody to do it, and you might learn something. 46:36 And if you encourage everybody to do it, they might actually suggest some innovations and fight this a little bit more. 46:43 What's another question? There we go. Many of our works revolve around achieving efficiency and optimization. 46:52 When AI can do that better, what would be the focus of human's work? So certainly, we had the Work of the Future Initiative. 46:59 Ben's going to talk about this more also after lunch, some really fascinating things. But I'm convinced there will be plenty for people to do. 47:08 And in an ideal world, you take all the boring stuff, let the computers do that, and you 47:14 let the people do more interesting stuff. Now, that's an ideal world. But can we get closer? 47:20 If you think of the way automation has worked all over time, it has always taken the routine work. 47:26 But there's still been a role for humans. And when you look at the way advanced manufacturing works, you still have people in there. 47:32 They're just not doing the really boring stuff anymore. So I think there will be a role for humans. 47:37 And the tricky part is trying to figure out how to introduce the right things in the right way 47:42 and to engage your people in the process. The Global Opportunity Forum we're talking about 47:50 is another opportunity there. Next up, how do you balance creativity with and without AI? 47:57 What are the advantages, disadvantages of flexing ideas? 48:02 I don't see any disadvantage. Actually, I do see a disadvantage. I do it all the time. 48:07 What are some creative ways to think about this? You can always ignore it. But what would I do without the computer? 48:13 I would go to a couple of friends. What's a creative way to think about this? So I don't see a problem with helping 48:20 the computer be more helping. The only challenge is, sometimes you start down a route, 48:26 and you get stuck in that route. So just as if you're talking to people, when you're talking with the computer, 48:33 be ready to back out of that route and try a different thing. You can get uncreative because you get stuck in a certain path, 48:42 but absolutely, flex the ideas all you can. You don't have to accept those ideas. How can we prepare young people for a job 48:49 15 years from now that might not exist due to AI anymore? Hopefully, if you are training-- oh, sorry. 48:54 Let me say that differently. Certain people you train to do a job. We have vocational training all over the place. 49:01 You train them for that job because nobody should expect that you start a job now, and it will 49:06 be the same job in 15 years. Now, there are some-- my mom was a school teacher. 49:13 15 years later, she was still a school teacher. But she was teaching differently. My uncle was a fireman. 49:19 15 years later, he was still a fireman, but there were new tools he had to learn to use. 49:25 So you train people for the job, if you're training them for a job. But they should also get the ability to learn. 49:31 They should gain the comfort, the growth mindset to get things. And you should always train that. 49:36 As far as training people for what will the good things be, creativity, working with people, critical thinking. 49:47 These don't seem to be going out of style. This generative AI stuff-- those don't seem to be going out of style for a while. 49:53 Actually, writing those emails, maybe. But knowing how to tune those emails to be the right kind of interaction with people, AI will help, 50:00 but those things will happen. So what I call these human skills-- and we have a framework we call the MIT-- 50:07 when I was in Open Learning, the human skills framework. In that framework, we think about how 50:12 I think, how I work with others, how I manage myself, how I lead others. 50:19 Those skills are probably not going to go away. And so how can you get those in college? How can you get those in high school? 50:25 Teamwork, writing and communicating, leading things, even when you're 12 years old. 50:33 I think that's the end of my time. So I want to thank you. And I'll be around at lunch if you want to talk. Bye. 50:38 [APPLAUSE] [MUSIC PLAYING]

Comments