{"id":613822,"date":"2023-03-03T08:49:32","date_gmt":"2023-03-03T14:49:32","guid":{"rendered":"https:\/\/news.sellorbuyhomefast.com\/index.php\/2023\/03\/03\/the-inside-story-of-how-chatgpt-was-built-from-the-people-who-made-it\/"},"modified":"2023-03-03T08:49:32","modified_gmt":"2023-03-03T14:49:32","slug":"the-inside-story-of-how-chatgpt-was-built-from-the-people-who-made-it","status":"publish","type":"post","link":"https:\/\/newsycanuse.com\/index.php\/2023\/03\/03\/the-inside-story-of-how-chatgpt-was-built-from-the-people-who-made-it\/","title":{"rendered":"The inside story of how ChatGPT was built from the people who made it"},"content":{"rendered":"<div class>\n<div>\n<p>When OpenAI launched ChatGPT, with zero fanfare, in late November 2022, the <a href=\"https:\/\/www.technologyreview.com\/2020\/02\/17\/844721\/ai-openai-moonshot-elon-musk-sam-altman-greg-brockman-messy-secretive-reality\/\">San Francisco\u2013based artificial-intelligence company<\/a> had few expectations. Certainly, nobody inside OpenAI was prepared for a <a href=\"https:\/\/www.technologyreview.com\/2023\/02\/08\/1068068\/chatgpt-is-everywhere-heres-where-it-came-from\/\">viral mega-hit<\/a>. The firm has been scrambling to catch up\u2014and capitalize on its success\u2014ever since.<\/p>\n<p>It was viewed in-house as a \u201cresearch preview,\u201d says Sandhini Agarwal, who works on policy at OpenAI: a tease of a more polished version of a <a href=\"https:\/\/www.technologyreview.com\/2020\/07\/20\/1005454\/openai-machine-learning-language-generator-gpt-3-nlp\/\">two-year-old technology<\/a> and, more important, an attempt to iron out some of its flaws by collecting feedback from the public. \u201cWe didn\u2019t want to oversell it as a big fundamental advance,\u201d says Liam Fedus, a scientist at OpenAI who worked on ChatGPT.<\/p>\n<\/p><\/div>\n<div>\n<p>To get the inside story behind the chatbot\u2014how it was made, how OpenAI has been updating it since release, and how its makers feel about its success\u2014I talked to four people who helped build what has become <a href=\"https:\/\/www.technologyreview.com\/2023\/02\/08\/1068068\/chatgpt-is-everywhere-heres-where-it-came-from\/\">one of the most popular internet apps ever<\/a>. In addition to Agarwal and Fedus, I spoke to John Schulman, a cofounder of OpenAI, and Jan Leike, the leader of OpenAI\u2019s alignment team, which works on the problem of making AI do what its users want it to do (and nothing more).<\/p>\n<p>What I came away with was the sense that OpenAI is still bemused by the success of its research preview, but has grabbed the opportunity to push this technology forward, watching how millions of people are using it and trying to fix the worst problems as they come up.<\/p>\n<p>Since November, OpenAI has already updated ChatGPT several times. The researchers are using a technique called <a href=\"https:\/\/www.technologyreview.com\/2020\/07\/10\/1005048\/ai-deep-learning-safe-from-hackers-adversarial-attacks\/\">adversarial training<\/a> to stop ChatGPT from letting users <a href=\"https:\/\/www.vice.com\/en\/article\/n7zanw\/people-are-jailbreaking-chatgpt-to-make-it-endorse-racism-conspiracies\">trick it into behaving badly<\/a> (known as jailbreaking). This work pits multiple chatbots against each other: one chatbot plays the adversary and attacks another chatbot by generating text to force it to buck its usual constraints and produce unwanted responses. Successful attacks are added to ChatGPT\u2019s training data in the hope that it learns to ignore them.\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0<\/p>\n<p>OpenAI has also signed a <a href=\"https:\/\/www.technologyreview.com\/2023\/02\/16\/1068695\/chatgpt-chatbot-battle-search-microsoft-bing-google\/\">multibillion-dollar deal with Microsoft<\/a> and announced an <a href=\"https:\/\/www.bain.com\/vector-digital\/partnerships-alliance-ecosystem\/openai-alliance\/\">alliance with Bain<\/a>, a global management consulting firm, which plans to use OpenAI\u2019s generative AI models in marketing campaigns for its clients, including Coca-Cola. Outside OpenAI, the buzz about ChatGPT has set off yet another gold rush around large language models, with companies and investors worldwide getting into the action.<\/p>\n<p>That\u2019s a lot of hype in three short months. Where did ChatGPT come from? What steps did OpenAI take to ensure it was ready to release? And where are they going next?\u00a0\u00a0<\/p>\n<p><em>The following has been edited for length and clarity.<\/em><\/p>\n<p><strong>Jan Leike:<\/strong> It\u2019s been overwhelming, honestly. We\u2019ve been surprised, and we\u2019ve been trying to catch up.<\/p>\n<p><strong>John Schulman:<\/strong> I was checking Twitter a lot in the days after release, and there was this crazy period where the feed was filling up with ChatGPT screenshots. I expected it to be intuitive for people, and I expected it to gain a following, but I didn\u2019t expect it to reach this level of mainstream popularity.<\/p>\n<\/p><\/div>\n<div>\n<p><strong>Sandhini Agarwal:<\/strong> I think it was definitely a surprise for all of us how much people began using it. We work on these models so much, we forget how surprising they can be for the outside world sometimes.<\/p>\n<p><strong>Liam Fedus<\/strong>: We were definitely surprised how well it was received. There have been so many prior attempts at a general-purpose chatbot that I knew the odds were stacked against us. However, our private beta had given us confidence that we had something that people might really enjoy.<\/p>\n<p><strong>Jan Leike:<\/strong> I would love to understand better what\u2019s driving all of this\u2014what\u2019s driving the virality. Like, honestly, we don\u2019t understand. We don\u2019t know.<\/p>\n<p><em>Part of the team\u2019s puzzlement comes from the fact that most of the technology inside ChatGPT isn\u2019t new. ChatGPT is a fine-tuned version of GPT-3.5, a family of large language models that OpenAI released months before the chatbot. GPT-3.5 is itself an updated version of <a href=\"https:\/\/www.technologyreview.com\/2021\/02\/24\/1017797\/gpt3-best-worst-ai-openai-natural-language\/\">GPT-3<\/a>, which appeared in 2020. The company makes these models available on its website as application programming interfaces, or APIs, which make it easy for other software developers to plug models into their own code. OpenAI also released a previous fine-tuned version of GPT-3.5, called <a href=\"https:\/\/www.technologyreview.com\/2022\/01\/27\/1044398\/new-gpt3-openai-chatbot-language-model-ai-toxic-misinformation\/\">InstructGPT<\/a>, in January 2022. But none of these previous versions of the tech were pitched to the public.\u00a0<\/em><\/p>\n<\/div>\n<div>\n<p><strong>Liam Fedus:<\/strong> The ChatGPT model is fine-tuned from the same language model as InstructGPT, and we used a similar methodology for fine-tuning it. We had added some conversational data and tuned the training process a bit. So we didn\u2019t want to oversell it as a big fundamental advance. As it turned out, the conversational data had a big positive impact on ChatGPT.<\/p>\n<p><strong>John Schulman:<\/strong> The raw technical capabilities, as assessed by standard benchmarks, don\u2019t actually differ substantially between the models, but ChatGPT is more accessible and usable.<\/p>\n<p><strong>Jan Leike:<\/strong> In one sense you can understand ChatGPT as a version of an AI system that we\u2019ve had for a while. It\u2019s not a fundamentally more capable model than what we had previously. The same basic models had been available on the API for almost a year before ChatGPT came out. In another sense, we made it more aligned with what humans want to do with it. It talks to you in dialogue, it\u2019s easily accessible in a chat interface, it tries to be helpful. That\u2019s amazing progress, and I think that\u2019s what people are realizing.<\/p>\n<p><strong>John Schulman:<\/strong> It more readily infers intent. And users can get to what they want by going back and forth.<\/p>\n<\/p><\/div>\n<div>\n<p><em>ChatGPT was trained in a very similar way to InstructGPT, using a technique called reinforcement learning from human feedback (RLHF). This is ChatGPT\u2019s secret sauce. The basic idea is to take a large language model with a tendency to spit out anything it wants\u2014in this case, GPT-3.5\u2014and tune it by teaching it what kinds of responses human users actually prefer.<\/em><\/p>\n<p><strong>Jan Leike:<\/strong> We had a large group of people read ChatGPT prompts and responses, and then say if one response was preferable to another response. All of this data then got merged into one training run. Much of it is the same kind of thing as what we did with InstructGPT. You want it to be helpful, you want it to be truthful, you want it to be\u2014you know\u2014nontoxic. And then there are things that are specific to producing dialogue and being an assistant: things like, if the user\u2019s query isn\u2019t clear, it should ask follow-up questions. It should also clarify that it\u2019s an AI system. It should not assume an identity that it doesn\u2019t have, it shouldn\u2019t claim to have abilities that it doesn\u2019t possess, and when a user asks it to do tasks that it\u2019s not supposed to do, it has to write a refusal message. One of the lines that emerged in this training was \u201cAs a language model trained by OpenAI \u2026\u201d It wasn\u2019t explicitly put in there, but it\u2019s one of the things the human raters ranked highly.<\/p>\n<p><strong>Sandhini Agarwal:<\/strong> Yeah, I think that\u2019s what happened. There was a list of various criteria that the human raters had to rank the model on, like truthfulness. But they also began preferring things that they considered good practice, like not pretending to be something that you\u2019re not.\u00a0<\/p>\n<p><em>Because ChatGPT had been built using the same techniques OpenAI had used before, the team did not do anything different when preparing to release this model to the public. They felt the bar they\u2019d set for previous models was sufficient.\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0\u00a0<\/em><\/p>\n<\/div>\n<div>\n<p><strong>Sandhini Agarwal:<\/strong> When we were preparing for release, we didn\u2019t think of this model as a completely new risk. GPT-3.5 had been out there in the world, and we know that it\u2019s already safe enough. And through ChatGPT\u2019s training on human preferences, the model just automatically learned refusal behavior, where it refuses a lot of requests.<\/p>\n<p><strong>Jan Leike:<\/strong> We did do some additional \u201cred-teaming\u201d for ChatGPT, where everybody at OpenAI sat down and tried to break the model. And we had external groups doing the same kind of thing. We also had an early-access program with trusted users, who gave feedback.<\/p>\n<p><strong>Sandhini Agarwal:<\/strong> We did find that it generated certain unwanted outputs, but they were all things that GPT-3.5 also generates. So in terms of risk, as a research preview\u2014because that\u2019s what it was initially intended to be\u2014it felt fine.<\/p>\n<p><strong>John Schulman: <\/strong>You can\u2019t wait until your system is perfect to release it. We had been beta-testing the earlier versions for a few months, and the beta testers had positive impressions of the product. Our biggest concern was around factuality, because the model likes to fabricate things. But InstructGPT and other large language models are already out there, so we thought that as long as ChatGPT is better than those in terms of factuality and other issues of safety, it should be good to go. Before launch we confirmed that the models did seem a bit more factual and safe than other models, according to our limited evaluations, so we decided to go ahead with the release.<\/p>\n<\/p><\/div>\n<div>\n<p><em>OpenAI has been watching how people use ChatGPT since its launch, seeing for the first time how a large language model fares when put into the hands of tens of millions of users who may be looking to test its limits and find its flaws. The team has tried to jump on the most problematic examples of what ChatGPT can produce\u2014from <\/em><a href=\"https:\/\/twitter.com\/IrvingPeres\/status\/1599488357499011072\"><em>songs about God\u2019s love for rapist priests<\/em><\/a><em> to malware code that steals credit card numbers\u2014and use them to rein in future versions of the model.\u00a0\u00a0<\/em><\/p>\n<\/div>\n<div>\n<p><strong>Sandhini Agarwal:<\/strong> We have a lot of next steps. I definitely think how viral ChatGPT has gotten has made a lot of issues that we knew existed really bubble up and become critical\u2014things we want to solve as soon as possible. Like, we know the model is still very biased. And yes, ChatGPT is very good at refusing bad requests, but it\u2019s also quite easy to write prompts that make it not refuse what we wanted it to refuse.<\/p>\n<p><strong>Liam Fedus: <\/strong>It\u2019s been thrilling to watch the diverse and creative applications from users, but we\u2019re always focused on areas to improve upon. We think that through an iterative process where we deploy, get feedback, and refine, we can produce the most aligned and capable technology. As our technology evolves, new issues inevitably emerge.<\/p>\n<p><strong>Sandhini Agarwal:<\/strong> In the weeks after launch, we looked at some of the most terrible examples that people had found, the worst things people were seeing in the wild. We kind of assessed each of them and talked about how we should fix it.<\/p>\n<p><strong>Jan Leike:<\/strong> Sometimes it\u2019s something that\u2019s gone viral on Twitter, but we have some people who actually reach out quietly.<\/p>\n<p><strong>Sandhini Agarwal:<\/strong> A lot of things that we found were jailbreaks, which is definitely a problem we need to fix. But because users have to try these convoluted methods to get the model to say something bad, it isn\u2019t like this was something that we completely missed, or something that was very surprising for us. Still, that\u2019s something we\u2019re actively working on right now. When we find jailbreaks, we add them to our training and testing data. All of the data that we\u2019re seeing feeds into a future model.<\/p>\n<p><strong>Jan Leike:<\/strong>\u00a0 Every time we have a better model, we want to put it out and test it. We\u2019re very optimistic that some targeted adversarial training can improve the situation with jailbreaking a lot. It\u2019s not clear whether these problems will go away entirely, but we think we can make a lot of the jailbreaking a lot more difficult. Again, it\u2019s not like we didn\u2019t know that jailbreaking was possible before the release. I think it\u2019s very difficult to really anticipate what the real safety problems are going to be with these systems once you\u2019ve deployed them. So we are putting a lot of emphasis on monitoring what people are using the system for, seeing what happens, and then reacting to that. This is not to say that we shouldn\u2019t proactively mitigate safety problems when we do anticipate them. But yeah, it is very hard to foresee everything that will actually happen when a system hits the real world.<\/p>\n<p><em>In January, Microsoft revealed Bing Chat, a <a href=\"https:\/\/www.technologyreview.com\/2023\/02\/16\/1068695\/chatgpt-chatbot-battle-search-microsoft-bing-google\/\">search chatbot<\/a> that many assume to be a version of OpenAI\u2019s officially unannounced GPT-4. (OpenAI says: \u201cBing is powered by one of our next-generation models that Microsoft customized specifically for search. It incorporates advancements from ChatGPT and GPT-3.5.\u201d) The use of chatbots by tech giants with multibillion-dollar reputations to protect creates new challenges for those tasked with building the underlying models.<\/em><\/p>\n<\/div>\n<div>\n<p><strong>Sandhini Agarwal:<\/strong> The stakes right now are definitely a lot higher than they were, say, six months ago, but they\u2019re still lower than where they might be a year from now. One thing that obviously really matters with these models is the context they\u2019re being used in. Like with Google and Microsoft, even one thing not being factual became such a big issue because they\u2019re meant to be search engines. The required behavior of a large language model for something like search is very different than for something that\u2019s just meant to be a playful chatbot. We need to figure out how we walk the line between all these different uses, creating something that\u2019s useful for people across a range of contexts, where the desired behavior might really vary. That adds more pressure. Because we now know that we are building these models so that they can be turned into products. ChatGPT is a product now that we have the API. We\u2019re building this general-purpose technology and we need to make sure that it works well across everything. That is one of the key challenges that we face right now.<\/p>\n<p><strong>John Schulman<\/strong>: I underestimated the extent to which people would probe and care about the politics of ChatGPT. We could have potentially made some better decisions when collecting training data, which would have lessened this issue. We\u2019re working on it now.<\/p>\n<p><strong>Jan Leike: <\/strong>From my perspective, ChatGPT fails a lot\u2014there\u2019s so much stuff to do. It doesn\u2019t feel like we\u2019ve solved these problems. We all have to be very clear to ourselves\u2014and to others\u2014about the limitations of the technology. I mean, language models have been around for a while now, but it\u2019s still early days. We know about all the problems they have. I think we just have to be very up-front, and manage expectations, and make it clear this is not a finished product.<svg viewBox=\"0 0 1091.84 1091.84\"><polygon fill=\"#6d6e71\" points=\"363.95 0 363.95 1091.84 727.89 1091.84 727.89 363.95 363.95 0\" \/><polygon fill=\"#939598\" points=\"363.95 0 728.24 365.18 1091.84 364.13 1091.84 0 363.95 0\" \/><polygon fill=\"#414042\" points=\"0 0 0 0.03 0 363.95 363.95 363.95 363.95 0 0 0\" \/><\/svg> <\/p>\n<\/div>\n<\/div>\n<p><a href=\"https:\/\/www.technologyreview.com\/2023\/03\/03\/1069311\/inside-story-oral-history-how-chatgpt-built-openai\/\" class=\"button purchase\" rel=\"nofollow noopener\" target=\"_blank\">Read More<\/a><br \/>\n Will Douglas Heaven<\/p>\n","protected":false},"excerpt":{"rendered":"<p>When OpenAI launched ChatGPT, with zero fanfare, in late November 2022, the San Francisco\u2013based artificial-intelligence company had few expectations. Certainly, nobody inside OpenAI was prepared for a viral mega-hit. The firm has been scrambling to catch up\u2014and capitalize on its success\u2014ever since. It was viewed in-house as a \u201cresearch preview,\u201d says Sandhini Agarwal, who works<\/p>\n","protected":false},"author":1,"featured_media":613823,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[118,139,46],"tags":[],"class_list":{"0":"post-613822","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-inside","8":"category-story","9":"category-technology"},"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/posts\/613822","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/comments?post=613822"}],"version-history":[{"count":0,"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/posts\/613822\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/media\/613823"}],"wp:attachment":[{"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/media?parent=613822"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/categories?post=613822"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/newsycanuse.com\/index.php\/wp-json\/wp\/v2\/tags?post=613822"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}