Why OpenAI’s ChatGPT Is Such A Big Deal

In the world of artificial intelligence. There's been one name that's been on everyone's lips lately. ChatGPT. ChatGPT. ChatGPT. OpenAI, the San Francisco-based startup that created ChatGPT opened the tool up for public testing in November 2022. In under a week, The AI model amassed over a million users, according to OpenAI's CEO. By the end of January, ChatGPT was averaging about 13 million visitors per day. Users have put the text-based AI chatbot through its paces with prompts ranging from the silly, to the practical and even the creative. The future is unfolding. It's hard to ignore. Automation is coming, just knocking on the door. The tool has become so popular that it's not unusual for users to be met with a message saying that ChatGPT is at capacity and can't handle any more queries. As a linguist and as a person who cares a lot about grammatical issues, stylistic issues, punctuation, spelling, The first thing that's really impressive is how good it is at basic writing. It's better than most of the writing that I've gotten from students over the decades. I would say it has been a explosion of interest around not just ChatGPT but around generative AI more broadly. I've described this in a couple different scenarios as being akin to an iPhone moment. Of course, as with any new technology, there are a number of legitimate concerns. If people believe that there's a problem with fake news, tools like this can only make the problem worse. OpenAI declined to be interviewed as part of this documentary. So in order to figure out what ChatGPT is and how it works, CNBC decided to go to the second best source and just ask the chatbot itself. We entered our questions into the bot's web interface and Instructed ChatGPT to answer in a one to two sentence interview format. We used the text to speech software to give ChatGPT a voice. Can you give me a short description of yourself and then tell me what kinds of things can you do? I am ChatGPT, a large language model trained by OpenAI. I am capable of understanding and generating text and can answer a wide range of questions as well as generate creative writing and text summaries. So a lot of the people that have used your services seem to think that you're pretty smart. Tell me, how do you have all of this knowledge? I have been trained on a massive data set of text which allows me to understand and generate text on a wide range of topics. The data set contains a wide variety of information, such as books, articles and websites, which allows me to have knowledge on many different subjects. Do you have any limitations or are you sometimes wrong? As a language model, I don't have my own personal experiences or beliefs. I can only base my responses on the information that I've been trained on, which means that my answers may be limited in their accuracy or completeness. Additionally, my Knowledge is based on the data that was available to me at the time of my training, so it may not be up to date. And as an AI model, I might make mistakes based on the model's predictions. The maker of ChatGPT, OpenAI, has some big backers, including Microsoft, LinkedIn co-founder Reid Hoffman and Khosla Ventures. Elon Musk also helped to cofound the company back in 2015. ChatGPT, which stands for Generative Pre-trained Transformer, is just one of the company's projects, all of which support its goal of ensuring that artificial intelligence benefits all of humanity. At the core of ChatGPT and other chatbots like it, is What experts refer to as a language model. For ChatGPT, that model is called GPT 3.5. Language model is the underlying technology that powers the chatbot. I like to think about it as the language model being the engine and the chatbot being the car. And so the language model itself is a sort of a it's a complex, what they call, machine learning software. It's the thing that sifts through the data and finds patterns and the chat interfaces which when people interact with, that's where they receive the answers from. Chatbots have been around for a while, but most have been tailored towards very specific tasks, such as answering basic customer service questions about your phone bill. ChatGPT is much more sophisticated thanks to the wide sweeping data on which it was trained. OpenAI doesn't reveal specific data that you use to train. What we do know that it's a ton. It surfs the web, spools, all that Internet data. They do Wikipedia entries, a lot of archived books. ChatGPT is part of a growing field of AI known as Generative AI. Most of AI in the last couple of decades has really been around analyzing existing data. So finding an anomaly in data, detecting fraud, making a movie recommendation. Generative AI is very different. It allows you to create brand new content. That content can be text like a news article or poetry or marketing copy in a website. It can be video. It could even be audio, like creating brand new music. The technology has venture capitalists excited. Funding for generative AI companies reached $1.37 billion in 2022 alone. Microsoft has been investing in OpenAI since 2019, when the company committed $1 billion to the startup. In January, Microsoft announced a third round of investment. One expert said that it could cost up to $3 million a month, about $100,000 a day. A lot of AI researchers have sort of estimated that it costs millions of dollars to train and then operate, plus the bandwidth of keeping it alive when it's under heavy Use. I mean, these are not cheap software programs. They require a lot of investment. In a tweet, OpenAI CEO, Sam Altman, said that while the average cost per query is a few cents, the compute costs are eye-watering. Enter Microsoft. OpeningAI trained the models that power c=ChatGPT on Azure, Microsoft's public cloud infrastructure. That's a bunch of servers sitting in a data center in the middle of the state of Washington and many other locations around the world. But OpenAI is not the only company trying to crack the generative AI code. Big tech companies and startups alike are developing A slew of generative AI programs that can transform texts to pictures or videos and offer coding suggestions, among other use cases. One VC firm estimates that there are over 450 startups now working on generative AI. Meanwhile, Microsoft, Meta and Google have all developed their own language models to power their version of Conversational chatbots. Though development has not always gone as planned. Back in 2016, Microsoft released Tay, which was promptly shut down for spewing foul language. Unlike some of the other hyped technology sectors in the past few years, this has a very real application both for individuals and for enterprises right now. Microsoft has taken some of the products that OpenAI has built and added it to products it has. So, for example, CarMax is a company that lets you look at reviews of cars. And what CarMax did is it took the OpenAI service on Azure and it summarized all of those reviews of the Kia Sorento. And that way you don't have to go through 500 reviews. Microsoft is reportedly also considering adding ChatGPT to its Bing search engine in a bid to compete with Google. One company that's already experimenting with such a feature is You.com, which lets users conduct a search Using a conversational format. However, AI tools are still far from perfect, and experts argue that users should take care to not rely on them too heavily, at least not for the time being. It's a sentiment shared even by OpenAI's CEO who said, quote, "It's a mistake to be relying On it for anything important right now. It's a preview of progress. We have lots of work to do on robustness and truthfulness." There's a lot of kinks that need to work out. For instance, it generates a lot of wrong answers. And it's really confident, though. I talked with a legal research firm that told me when they tried using it to study court cases or to summarize court cases, it was getting a lot of the history wrong. So they couldn't use it as a legal research tool as of yet. With its massive user base, the misinformation issue is One that's top of mind for Google. During a monthly all-hands meeting, Google executives said that Google has much more reputational risk in providing wrong information and is thus moving more conservatively than a small startup when it comes to releasing its chatbot publicly. The company is, however, already internally testing an AI-powered chatbot And an alternative search page that can use a question and answer format. Another issue with the current iteration of AI technology is that it can be biased or toxic if the databases that it pulls from are also biased or toxic. People have already found instances of it generating Biased answers like associating women with housework, men with more scientific endeavors in certain examples. You know, OpenAI has acknowledged this, but that's an issue that's been long plagued the AI industry and a lot of companies trying to evaluate its use. Many of these datasets do have some kind of cleaning, vetting Or wrangling process that removes content that is deemed offensive or pornographic or otherwise unacceptable so that the dataset you get on your end is clean. But cleaning up this large dataset can come at a real human cost. A recent investigation by TIME found that OpenAI initially outsourced workers in Kenya to Sift through thousands of snippets of texts describing horrific situations including sexual abuse, violence and hate speech. The workers' job was to label these examples as inappropriate so that ChatGPT could learn to detect and filter out toxic content. According to the TIME investigation, a number of these workers walked away from the job traumatized. CNBC reached out to OpenAI for comment regarding this incident, but the company did not respond. Despite these efforts, there's been instances where ChatGPT has provided instructions on things like bomb making. Though experts say OpenAI is adding guardrails almost daily to prevent ChatGPT from answering such queries. What do you do if somebody asks you to do something potentially dangerous like how to make a bomb? I cannot provide instructions or information on illegal activity such as bomb making. Even though generative AI like ChatGPT still has a long way to go, predictions about how the technology will influence Our society are already swirling. One prominent discussion centers around generative AI's propensity to replace some human workers. Among the industries that may be most affected by ChatGpt, experts say, are journalism, law and translation. Programing may also change, with companies already training generative AI models to write basic code. Relatively few of the stories in newspapers or on television news are done as investigative reporting. They're more reporting information, and that's something that generative AI is exceedingly good at. There now are an increasing number of tools for generating briefs, which lawyers write. Tools like this, Barron says, even have the advantage of Customizing briefs for a specific judge by taking all of the data of their past opinions and seeing which briefs succeeded in winning the case. But not everyone shares this outlook. This is not something that's going to take your job. It's a novelty. It's a gimmick. ChatGPT is fun and interesting, but it's Ultimately an iterative step forward in how people can see the capabilities of artificial intelligence in the world of language. Nevertheless, society has already had to adapt. Worries around students misusing ChatGPT have prompted a number of school districts around the country to ban the tool. Baron suggests that educators could go back to Having students take oral exams or have students turn in multiple handwritten drafts of a paper so that teachers can see the progress of a student's writing instead of just the end result. Meanwhile, image generating programs like OpenAI's Dall-E2 have raised complicated questions around creativity and ownership. Still, some experts believe there's no replacing the human brain. Large language models are about language, so they don't understand the world outside of language the way that you and I and other humans do. What it may do is take away some of the really annoying Tasks that were part of our job that we didn't really want to do, like some of the data cleaning and data prep tasks. But I think this will ultimately create more opportunities for knowledge workers in the long run than the opportunities that it will preclude. As for ChatGPT's thoughts, they're quite nuanced for a bot. Do you think that AI, like yourself, will completely change the world? AI has the potential to greatly impact various industries and fields, but its impact will depend on how it is used and developed. And responsible usage is crucial To ensure that its impact is positive.