âSam Altman on What Makes Him âSuper Nervousâ About AI: The OpenAI Co-Founder Thinks Tools like GPT-4 Will Be Revolutionary. But Heâs Wary of Downsidesâ, 2023-03-23 (; backlinks)â :
Kara Swisher: [Origins of scaling] âŚYou had started it [OpenAI] as a nonprofit.
Sam Altman: I am not a natural fit for a CEO; being an investor, I think, suits me very well. I got convinced that AGI was going to happen and be the most important thing I could ever work on. I think it is going to transform our society in many ways, and I wonât pretend that as soon as we started OpenAI, I was sure it was going to work, but it became clear over the intervening years, and certainly by 2018â2019, that we had a real chance here.
K. Swisher: What was it that made you think that?
S. Altman: A number of things. Itâs hard to point to just a single one, but by the time we made GPT-2, which was still weak in a lot of ways, you could look at the scaling laws and see what was going to happen. I was like, âHmm. This can go very, very far.â I got super-excited about it. Iâve never stopped being super-excited about it.
Swisher: Was there something you saw that scaled or what was the âŚ
Altman: It was looking at the data of how predictably better we could make the system with more compute, with more data.
K S: There had already been a lot of stuff going on at Google with DeepMind. They had bought that earlier, right around then [2014].
S A: Yeah. There had been a bunch of stuff, but somehow it wasnât quite the trajectory that has turned out to be the one that really works.
S: âŚWhen people critiqued ChatGPT, you wish that you said, âWait for GPT-4.â Now that itâs out, has it met expectations?
A: A lot of people seem really happy with it. Thereâs plenty of things itâs still bad at.
S: I meant your expectations.
A: Yeah. Iâm proud of it. Again, a very long way to go, but as a step forward, Iâm proud of it.
S: What are you proudest of?
A: Well, I enjoy using it, but more than that, itâs very gratifying to just go search for GPT-4 on Twitter and read what people are doing with it, the amazing discoveries people make of how to use it to be more productive, more effective, more creative, whatever they need. Itâs nice. Itâs nice to build something thatâs useful for people.
S: You tweeted that at first glance, GPT-4 seems âmore impressive than it actually is.â Why is that?
A: Well, I think thatâs been an issue with every version of these systems, not particularly GPT-4. You find these flashes of brilliance before you find the problems. And so, a thing that someone used to say about GPT-4 that has really stuck with me is it is the worldâs greatest demo creator. Because you can tolerate a lot of mistakes there, but if you need a lot of reliability for a production system, itâs not as good at that. GPT-4 makes fewer mistakes. Itâs more reliable, itâs more robust, but thereâs still a long way to go.
A: âŚThese systems are trained to do something, which is to predict the next word in a sequence. And so, itâs trying to just complete a pattern, and given its training set, this is the most likely completion. That said, the decrease [in loss] from GPT-3 â 3.5 â 4, I think is very promising. We track this internally, and every week weâre able to get the number lower and lower and lower. I think itâll require combinations of model scale, new ideasâ
S: More data.
A: A lot of user feedback.
S: Model scale is more data.
A: Not necessarily more data, but more compute thrown at the problem. Human feedbackâpeople flagging the errors for us, developing new techniques of the modelâcan tell when itâs about to go off the rails.
S: Real people just saying âThis is a mistake.â
A: âŚYeah. We pay experts to flag, to go through and label the data for usâŚNot just bounties, but we employ people. We have contractors; we work with external firms. We say we need experts in this area to help us go through and improve things. You donât just want to rely totally on random users doing whatever, trying to troll you, or anything like that.
S: So humans, more compute. What else?
A: I think that there is going to be a big new algorithmic idea, a different way that we train or use or tweak these models, different architecture perhaps. So I think weâll find that at some point.
S: Meaning what, for the non-techy?
A: Well, it could be a lot of things. You could say a different algorithm, but just some different idea of the way that we create or use these models that encourages, during training or inference time when youâre using it, that encourages the models to really ground themselves in truth, be able to cite sources. Microsoft has done some good work there. Weâre working on some things.
S: [Openness] Talk about the next steps. How does this move forward?
A: I think weâre on this very long-term exponential, and I donât mean that just for AI, although AI tooâI mean that as cumulative, human, technological progressâand itâs very hard to calibrate that, and we keep adjusting our expectations. I think if we told you 5 years ago weâd have GPT-4 today, youâd maybe be impressed. But if we told you 4 months ago after you used ChatGPT that weâd have GPT-4 today, probably not that impressed. Yet itâs the same continued exponential, so maybe where we get to a year from now, youâre like, âMeh. Itâs better, but the new iPhoneâs always a little better too.â But if you look at where weâll be in 10 years, then I think youâd be pretty impressed.
A: âŚWell, for us itâs the opposite. I mean, what weâve said all alongâand this is different than what most other AGI efforts have thoughtâis everybody needs to know about this. AGI should not be built in a secret lab with only the people who are privileged and smart enough to understand it. Part of the reason that we deploy this is, I think, we need the input of the world, and the world needs familiarity with what is in the process of happening, the ability to weigh in, to shape this together. We want that. We need that input, and people deserve it. So I think weâre not the secretive company. Weâre quite the opposite. We put the most advanced AI in the world in an API that anybody can use. I donât think that if we hadnât started doing that a few years ago, Google or anybody else would be doing it now. They would just be using it secretly to make Google search better.
S: [Funding] But you are in competition. And let me go back to someone who was one of your original funders, Elon Musk. Heâs been openly critical of OpenAI, especially as itâs gone to profits: âOpenAI was created as an open source (which is why I named itâOpenâ AI), nonprofit company to serve as a counterweight to Google, but now it has become a closed source, maximum-profit company effectively controlled by Microsoft. Not what I intended at all.â Weâre talking about open source versus closed, but what about his critique that youâre too close to the big guys?
A: I mean, most of that is not true. And Elon knows that. Weâre not controlled by Microsoft. Microsoft doesnât even have a board seat on us, we are an independent company. We have an unusual structure where we can make very different decisions than what most companies do. I think a fair part of that is we donât open-source everything anymore. Weâve been clear about why we think we were wrong there originally. We still do open-source a lot of stuff. Open sourcing CLIP was something that kicked off this whole generative image world. We recently open-sourced Whisper, we open-sourced tools, weâll open-source more stuff in the future. But I donât think it would be good right now for us to open-source GPT-4, for example. I think that would cause some degree of havoc in the world, or at least thereâs a chance of thatâwe canât be certain that it wouldnât. And by putting it out behind an API, we are able to get many, not all, but many of the benefits we want of broad access to this society being able to understand the update and think about it. But when we find some of the scarier downsides, weâre able to then fix them, and we are going to.
S: How do you respond when heâs saying youâre a closed-source maximum-profit company? Iâll leave out the control by Microsoft, but in a strong partnership with Microsoft. Which was against what he said. I remember years ago, this was something he talked about a lot and wasâ
A: Was what part?
S: âOh, we donât want these big companies to run it. If they run it, weâre doomed.â He was much more dramatic than most people.
A: So weâre a capped-profit company. We invented this new thing where we started as a nonprofitâ
S: Explain that. Explain what a capped profit is.
A: Our shareholders, who are our employees and our investors, can make a certain return. Their shares have a certain price that they can get to. But if OpenAI goes and becomes a multi-trillion-dollar company, almost all of that flows to the nonprofit that controls us.
S: What is the cap?
A: It continues to vary as we have to raise more money, but itâs much, much, much, and will remain much, smaller than anyââŚIn terms of a number, I truly donât know off the top of my head.
S: But itâs not substantial. The nonprofit gets a substantial chunk of the revenue.
A: Well, no, it gets everything over a certain amount. So if weâre not very successful, the nonprofit gets a little bit along the way, but it wonât get any appreciable amount. The goal of the cap profit is in the world where we do succeed at making AGI and we have a substantial lead over everybody else, it could become much more valuable, I think, than maybe any company out there today. Thatâs when you want almost all of it to flow to a nonprofit, I think.
S: I want to get back to what Elon was talking about. He was very adamant at the time and, again, overly dramatic, that Google and Microsoft and Amazon were going to kill us. I think he had those kinds of words, that there needed to be an alternative. What changed, in your estimation?
A: Of?
S: To change from that idea.
A: Oh, it was very simple. When we realized the level of capital we were going to need to do this, scaling turned out to be far more important than we thought, and we even thought it was going to be important then. And we tried for a while to find a path to that level of capital as a nonprofit. There was no one that was willing to do it. So we didnât want to become a fully for-profit company. We wanted to find something that would let us get the access to and the power of capitalism to finance what we needed to do, but still be able to fulfill and be governed by the nonprofit mission. So having this nonprofit that governs the capped-profit LLC, given the playing field that we saw at the time, and I still think that we see now, was the way to get to the best of all worlds. In a really well-functioning society, I think this wouldâve been a government project.
S: Thatâs correct. I was just going to make that point. The government wouldâve been your funder.
A: We talked to them. They not just would have been our funder, but they wouldâve started the project. Weâve done things like this before in this country. But the answer is not to just say, âOh well, the government doesnât do stuff like this anymore, so weâre just going to sit around and let other countries run by us and get an AGI and do whatever they want to us.â Weâre going to look at whatâs possible on this playing field.
S: Right. So Elon used to be the co-chair, and you have a lot of respect for him.
A: I do.
S: Iâm sure you thought deeply about his critiques. Have you spoken to him directly? Was there a break, or what? You two were very close, as I recall.
A: Weâve spoken directly recently.
S: And what do you make of the critiques? When you hear them from him, I mean, he can be quite in your face about things.
A: Heâs got his style.
S: Yeah.
A: To say a positive thing about Elonâ
S: Yeah, Iâd like you to.
A: âŚI think he really does care about a good future with AGI.
S: He does.
A: AndâŚI mean, heâs a jerk, whatever else you want to say about him. He has a style that is not a style that Iâd want to have for myself.
S: Heâs changed.
A: But I think he does really care, and he is feeling very stressed about what the futureâs going to look likeâ
S: For humanity.
A: For humanity.
S: When we did an interview at Tesla, he was like, âIf this doesnât work, weâre all doomed.â Which was sort of centered on his car, but nonetheless, he was correct. And this was something he talked about almost incessantly, the idea of either AI taking over and killing us, or maybe it doesnât really care. Then he decided it was like anthills; do you remember that example?
A: I donât remember the anthills part.
S: He said, âYou know how, when weâre building a highway, anthills are there and we just go over them without thinking about it?â And then he said, âWeâre like a cat, and maybe theyâll feed us and bell us, but they donât really care about us.â It went on and on; it changed and iterated over time. But I think the critique that I would most agree with him on is that big companies would control this and there couldnât be innovation in the space.
A: Well, I would say weâre evidence against that.
S: Except Microsoft, and thatâs why I thinkâ
A: Theyâre a big investor, but again, not even a board member. Like true, full independence from them.
S: So you think you are a startup in comparison with a giant partner?
A: Yeah, I mean, weâre a big start-up at this point.
S: And thereâs no way to be a nonprofit that would work?
A: If someone wants to give us tens of billions of dollars of nonprofit capital, we can go make that work.
S: Yeah. Or the government, which theyâre not.
A: We tried.
S: [GPT-3 political bias] Greg Brockman, your co-founder, said you guys made a mistake by creating AI with a quote, âLeft-leaning political bias.â What do you think of the substance of those critiques?
A: Yeah. I think the reinforcement learning from human feedback on our first version of ChatGPT was pretty left-biased, but that is now no longer true. Itâs just become an internet meme. There are some people who are intellectually honest about this. If you go look at GPT-4 and test it onâŚItâs relatively neutral. Not to say we donât have more work to do. The main thing, though, is I donât think you ever get two people agreeing that any one system is unbiased on every topic. And so giving users more control and also teaching people about how these systems work, that there is some randomness in a response, that the worst screenshot you see on Twitter is not representative of what these things do, I think is important.
S: So when you said it had a left-leaning bias, what did that mean to you? And of course theyâll run with thatâtheyâll run with that quite far.
A: People would give it these tests that score you on the political spectrum in America or whatever. And one would be all the way on the right, 10 would be all the way on the left. It would get like a 10 on all of those tests, the first version.
S: Why?
A: A number of reasons, but largely because of the reinforcement learning from human feedback stuff.
S: [Competitors] What do you think the most viable threat to OpenAI is? I hear youâre watching Claude very carefully. This is the bot from Anthropic, a company thatâs founded by former OpenAI folks and backed by Google. Is that it? Weâre recording this on Tuesday. Bard launched today; Iâm sure youâve been discussing it internally. Talk about those two to start.
A: I try to pay some attention to whatâs happening with all these other things. Itâs going to be an unbelievably competitive space. I think this is the first new technological platform in a long period of time. The thing I worry about the most is not any of those, because I think thereâs room for a lot of people, and also I think weâll just continue to offer the best product. The thing I worry about the most is that weâre somehow missing a better approach. Everyoneâs chasing us right now on large language models, kind of trained in the same way. I donât worry about them, I worry about the person who has some very different idea about how to make a more useful system.
S: But is there one that youâre watching more carefully?
A: Not especially.
S: Really? I kind of donât believe you, but really?
A: The things that I pay the most attention to are not, like, language model, start-up number 217. Itâs when I hear, âThese are 3 smart people in a garage with some very different theory of how to build AGI.â And thatâs when I pay attention.
S: Is there one that youâre paying attention to now?
A: There is one; I donât want to say.
S: You really donât want to say?
A: I really donât want to say.
A: âŚWhat I am personally most excited about is helping us greatly expand our scientific knowledge. I am a believer that a lot of our forward progress comes from increasing scientific discovery over a long period of time.
S: In any area?
A: All of the areas. I think thatâs just whatâs driven humanity forward. And if these systems can help us in many different ways, to greatly increase the rate of scientific understanding, curing diseases is an obvious example. Thereâs so many other things we can do withâ
S: AI has already moved in that directionâfolding proteins [eg. AlphaFold2] and things like that.
A: So thatâs the one that Iâm personally most excited about. But there will be many other wonderful things too. You asked me what my one was andâ
S: Is there one unusual thing that you think will be great, that youâve seen already that youâre like, âThatâs pretty cool?â
A: Using some of these new AI tutor-like applications is like, âI wish I had this when I was growing up. I could have learned so much, and so much better and faster.â And when I think about what kids today will be like by the time theyâre finished with their formal education and how much smarter and more capable and better educated they can be than us today, Iâm excited for that.
A: âŚ[Regulation] I do think we need regulation and we need industry norms about this. We spent many, many monthsâand actually really the years that itâs taken us to get good at making these modelsâgetting them ready before we put them out. It obviously became somewhat of an open secret in Silicon Valley that we had GPT-4 done for a long time and there were a lot of people who were like, âYou have to release this now; youâre holding this back from society. This is your closed AI, whatever.â But we just wanted to take the time to get it right. Thereâs a lot to learn here, and itâs hard, and in fact, we try to release things to help people get it right, even competitors. I am nervous about the shortcuts that other companies now seem like they want to take.
S: âŚYes. So this is what happens a lot of the time, even in well-regulated areas, which banks are compared to the internet. What sort of regulations does AI need in America? Lay them out. I know youâve been meeting with regulators and lawmakers.
A: Yeah, I did a 3-day trip to Washington DC earlier this year.
S: You did. So tell me what you think the regulations were and what are you telling them, and do you find them savvy as a group? I think theyâre savvier than people think.
A: Some of them are quite, quite exceptional. I think the thing that I would like to see happen immediately is just much more insight into what companies like ours are doing, companies that are training above a certain level of capability at a minimum. A thing that I think could happen now is the government should just have insight into the capabilities of our latest stuff, released or not, what our internal audit procedures and external audits we use look like, how we collect our data, how weâre red-teaming these systems, what we expect to happen, which we may be totally wrong about. We could hit a wall anytime, but our internal road-map documents, when we start a big training run, I think there could be government insight into that. And then if that can start nowâŚI do think good regulation takes a long time to develop. Itâs a real process. They can figure out how they want to have oversight.
S: Reid Hoffman has suggested a blue-ribbon panel so they learn, they learn up on this stuff, whichâ
A: Panels are fine. We could do that too, but what I mean is government auditors sitting in our buildings.
S: Congressman Ted Lieu said there needs to be an agency dedicated specifically to regulating AI. Is that a good idea?
A: I think thereâs two things you want to do. This is way out of my area of expertise, but youâre asking, so Iâll try. I think people like us who are creating these very powerful systems that could become something properly called AGI at some pointâ
S: Explain what that is.
A: Artificial general intelligence, but what people mean is just above some threshold where itâs really good. Those efforts probably do need a new regulatory effort, and I think it needs to be a global regulatory body. And then people who are using AI, like we talked about, as a medical adviser, I think the FDA can give probably very great medical regulation, but theyâll have to update it for the inclusion of AI. But I would say creation of the systems and having something like an IAEA that regulates that is one thing, and then having existing industry regulators still do their regulationâ
S: People do react badly to that, because the information bureaus, thatâs always been a real problem in Washington. Who should head that agency in the U.S.?
A: I donât know.
S: Okay. So one of the things thatâs going to happen, though, is the less intelligent ones, of which there are many, are going to seize on things like theyâve done with TikTok, possibly deservedly, but other things. Like Snap released a chatbot powered by GPT that allegedly told a 15-year-old how to mask the smell of weed and alcohol, and a 13-year-old how to set the mood for sex with an adult. Theyâre going to seize on this stuff. And the question is, whoâs liable if this is true, when a teen uses those instructions? And §230 doesnât seem to cover generative AI. Is that a problem?
A: I think we will need a new law for use of this stuff, and I think the liability will need to have a few different frameworks. If someone is tweaking the models themselves, I think itâs going to have to be the last person who touches it has the liability, and thatâsâ
S: But itâs not full immunity that the platformâs gettingâ
A: I donât think we should have full immunity. Now, that said, I understand why you want limits on it, why you do want companies to be able to experiment with this, you want users to be able to get the experience they want, but the idea of no one having any limits for generative AI, for AI in general, that feels super-wrong.