“Even inside the company, the chatbot’s popularity has come
as something of a shock.
One day in mid-November, workers at OpenAI got an unexpected
assignment: Release a chatbot, fast.
The chatbot, an executive announced, would be known as “Chat
with GPT-3.5,” and it would be made available free to the public. In two weeks.
The announcement confused some OpenAI employees. All year,
the San Francisco artificial intelligence company had been working toward the
release of GPT-4, a new A.I. model that was stunningly good at writing essays,
solving complex coding problems and more. After months of testing and
fine-tuning, GPT-4 was nearly ready. The plan was to release the model in early
2023, along with a few chatbots that would allow users to try it for
themselves, according to three people with knowledge of the inner workings of
OpenAI.
But OpenAI’s top executives had changed their minds. Some
were worried that rival companies might upstage them by releasing their own
A.I. chatbots before GPT-4, according to the people with knowledge of OpenAI.
And putting something out quickly using an old model, they reasoned, could help
them collect feedback to improve the new one.
So they decided to dust off and update an unreleased chatbot
that used a souped-up version of GPT-3, the company’s previous language model,
which came out in 2020.
Thirteen days later, ChatGPT was born.
In the months since its debut, ChatGPT (the name was, mercifully,
shortened) has become a global phenomenon. Millions of people have used it to
write poetry, build apps and conduct makeshift therapy sessions. It has been
embraced (with mixed results) by news publishers, marketing firms and business
leaders. And it has set off a feeding frenzy of investors trying to get in on
the next wave of the A.I. boom.
It has also caused controversy. Users have complained that
ChatGPT is prone to giving biased or incorrect answers. Some A.I. researchers
have accused OpenAI of recklessness. And school districts around the country,
including New York City’s, have banned ChatGPT to try to prevent a flood of
A.I.-generated homework.
Yet little has been said about ChatGPT’s origins, or the
strategy behind it. Inside the company, ChatGPT has been an earthshaking
surprise — an overnight sensation whose success has created both opportunities
and headaches, according to several current and former OpenAI employees, who
requested anonymity because they were not authorized to speak publicly.
An OpenAI spokesman, Niko Felix, declined to comment for
this column, and the company also declined to make any employees available for
interviews.
Before ChatGPT’s launch, some OpenAI employees were
skeptical that the project would succeed. An A.I. chatbot that Meta had
released months earlier, BlenderBot, had flopped, and another Meta A.I.
project, Galactica, was pulled down after just three days. Some employees,
desensitized by daily exposure to state-of-the-art A.I. systems, thought that a
chatbot built on a two-year-old A.I. model might seem boring.
But two months after its debut, ChatGPT has more than 30
million users and gets roughly five million visits a day, two people with
knowledge of the figures said. That makes it one of the fastest-growing
software products in memory. (Instagram, by contrast, took nearly a year to get
its first 10 million users.)
The growth has brought challenges. ChatGPT has had frequent
outages as it runs out of processing power, and users have found ways around
some of the bot’s safety features. The hype surrounding ChatGPT has also
annoyed some rivals at bigger tech firms, who have pointed out that its
underlying technology isn’t, strictly speaking, all that new.
ChatGPT is also, for now, a money pit. There are no ads, and
the average conversation costs the company “single-digit cents” in processing
power, according to a post on Twitter by Sam Altman, OpenAI’s chief executive,
likely amounting to millions of dollars a week. To offset the costs, the
company announced this week that it would begin selling a $20 monthly
subscription, known as ChatGPT Plus.
Despite its limitations, ChatGPT’s success has vaulted
OpenAI into the ranks of Silicon Valley power players. The company recently
reached a $10 billion deal with Microsoft, which plans to incorporate the
start-up’s technology into its Bing search engine and other products. Google
declared a “code red” in response to ChatGPT, fast-tracking many of its own
A.I. products in an attempt to catch up.
Mr. Altman has said his goal at OpenAI is to create what is
known as “artificial general intelligence,” or A.G.I., an artificial
intelligence that matches human intellect. He has been an outspoken champion of
A.I., saying in a recent interview that its benefits for humankind could be “so
unbelievably good that it’s hard for me to even imagine.” (He has also said
that in a worst-case scenario, A.I. could kill us all.)
As ChatGPT has captured the world’s imagination, Mr. Altman
has been put in the rare position of trying to downplay a hit product. He is
worried that too much hype for ChatGPT could provoke a regulatory backlash or
create inflated expectations for future releases, two people familiar with his
views said. On Twitter, he has tried to tamp down excitement, calling ChatGPT
“incredibly limited” and warning users that “it’s a mistake to be relying on it
for anything important right now.”
He has also discouraged employees from boasting about
ChatGPT’s success. In December, days after the company announced that more than
a million people had signed up for the service, Greg Brockman, OpenAI’s
president, tweeted that it had reached two million users. Mr. Altman asked him
to delete the tweet, telling him that advertising such rapid growth was unwise,
two people who saw the exchange said.
OpenAI is an unusual company, by Silicon Valley standards.
Started in 2015 as a nonprofit research lab by a group of tech leaders
including Mr. Altman, Peter Thiel, Reid Hoffman and Elon Musk, it created a
for-profit subsidiary in 2019 and struck a $1 billion deal with Microsoft. It
has since grown to around 375 employees, according to Mr. Altman — not counting
the contractors it pays to train and test its A.I. models in regions like
Eastern Europe and Latin America.
From the start, OpenAI has billed itself as a mission-driven
organization that wants to ensure that advanced A.I. will be safe and aligned
with human values. But in recent years, the company has embraced a more
competitive spirit — one that some critics say has come at the expense of its
original aims.
Those concerns grew last summer when OpenAI released its
DALL-E 2 image-generating software, which turns text prompts into works of digital
art. The app was a hit with consumers, but it raised thorny questions about how
such powerful tools could be used to cause harm. If creating hyper-realistic
images was as simple as typing in a few words, critics asked, wouldn’t
pornographers and propagandists have a field day with the technology?
To allay these fears, OpenAI outfitted DALL-E 2 with
numerous safeguards and blocked certain words and phrases, such as those
related to graphic violence or nudity. It also taught the bot to neutralize
certain biases in its training data — such as making sure that when a user
asked for a photo of a C.E.O., the results included images of women.
These interventions prevented trouble, but they struck some
OpenAI executives as heavy-handed and paternalistic, according to three people
with knowledge of their positions. One of them was Mr. Altman, who has said he
believes that A.I. chatbots should be personalized to the tastes of the people
using them — one user could opt for a stricter, more family-friendly model,
while another could choose a looser, edgier version.
OpenAI has taken a less restrictive approach with ChatGPT,
giving the bot more license to weigh in on sensitive subjects like politics,
sex and religion. Even so, some right-wing conservatives have accused the
company of overstepping. “ChatGPT Goes Woke,” read the headline of a National
Review article last month, which argued that ChatGPT gave left-wing responses
to questions about topics such as drag queens and the 2020 election. (Democrats
have also complained about ChatGPT — mainly because they think A.I. should be
regulated more heavily.)
As regulators swirl, Mr. Altman is trying to keep ChatGPT
above the fray. He flew to Washington last week to meet with lawmakers,
explaining the tool’s strengths and weaknesses and clearing up misconceptions
about how it works.
Back in Silicon Valley, he is navigating a frenzy of new
attention. In addition to the $10 billion Microsoft deal, Mr. Altman has met
with top executives at Apple and Google in recent weeks, two people with
knowledge of the meetings said. OpenAI also inked a deal with BuzzFeed to use
its technology to create A.I.-generated lists and quizzes. (The announcement
more than doubled BuzzFeed’s stock price.)
The race is heating up. Baidu, the Chinese tech giant, is
preparing to introduce a chatbot similar to ChatGPT in March, according to
Reuters. Anthropic, an A.I. company started by former OpenAI employees, is
reportedly in talks to raise $300 million in new funding. And Google is racing
ahead with more than a dozen A.I. tools.
Then there’s GPT-4, which is still scheduled to come out
this year. When it does, its abilities may make ChatGPT look quaint. Or maybe,
now that we’re adjusting to a powerful new A.I. tool in our midst, the next one
won’t seem so shocking.”
Most importantly, ChatGPT gives shorter and clearer answers than what a poor teacher or a Google search can provide. Therefore, it can be used for faster and better learning. Especially useful in medical and legal sciences and writing apps.
Komentarų nėra:
Rašyti komentarą