Connect with us

Business

OpenAI ChatGPT and Anthropic Claude chatbot usage studies may signal job losses ahead

Published

on



Hello and welcome to Eye on AI…In this edition: OpenAI and Anthropic detail chatbot usage trends…AI companies promise big investments in the U.K….and the FTC probes chatbots’ impact on kids.

Yesterday saw the release of dueling studies from OpenAI and Anthropic about the usage of their respective AI chatbots, ChatGPT and Claude. The studies provide a good snapshot of who is using AI chatbots and what they are using them for. But the two reports were also a study in contrasts, with OpenAI clearly emerging as primarily a consumer product, while Claude’s use cases were more professionally oriented.

The ChatGPT study confirmed the huge reach OpenAI has, with 700 million active weekly users, or almost 10% of the global population, exchanging some 18 billion messages with the chatbot every week. And the majority of those messages—70%—were classified by the study’s authors as “non-work” queries. Of these, about 80% of the messages fell into three big categories: practical guidance, writing help, and seeking information. Within practical guidance, teaching or tutoring queries accounted for more than a third of messages. How many of these were students using ChatGPT to “help” with homework or class assignments was unclear—but ChatGPT has a young user base, with nearly half of all messages coming from those under the age of 26.

Educated professionals more likely to be using ChatGPT for work

When ChatGPT was used for work, it was most likely to be used by highly educated users working in high-paid professions. While this is perhaps not surprising, it is a bit depressing.

There is a vision of our AI future, one which I outline in my book, Mastering AI, in which the technology becomes a leveling force. With the help of AI copilots and decision-support systems, people with fewer qualifications or experience could take on some of the work currently performed by more skilled and experienced professionals. They might not earn as much as those more qualified individuals, but they could still earn a good middle-class income. To some extent, this already happens in law, with paralegals, and in medicine, with nurse practitioners. But this model could be extended to other professions, for instance accounting and finance—democratizing access to professional advice and helping shore up the middle class.

There’s another vision of our AI future, however, where the technology only makes economic inequality worse, with the most educated and credentialed using AI to become even more productive, while everyone else falls farther behind. I fear that, as this ChatGPT data suggests, that’s the way things may be heading.

While there’s been a lot of discussion lately of the benefits and dangers of using chatbots for companionship, or even romance, OpenAI’s research showed messages classified as being about relationships constituted just 2.4% of messages, personal reflection 1.9%, and role-playing and games 0.4%.

Interestingly, given how fiercely all the leading AI companies—including OpenAI—compete with one another on coding benchmarks and tout the coding performance of their models, coding was a relatively small use case for ChatGPT, constituting just 4.2% of the messages the researchers analyzed. (One big caveat here is that the research only looked at the consumer versions of ChatGPT—its free, premium, and pro tiers—but not usage of the OpenAI API or enterprise ChatGPT subscriptions, which is how many business users may access ChatGPT for professional use cases.)

Meanwhile, coding constituted 39% of Claude.ai’s usage. Software development tasks also dominated the use of Anthropic’s API.

Automation rather than augmentation dominates work usage

Read together, both studies also hinted at an intriguing contrast in how people were using chatbots in work contexts, compared to more personal ones.

ChatGPT messages classified as non-work related were more about what the researchers called “asking”—which involved seeking information or advice—as opposed to “doing” prompts, where the chatbot was asked to complete a task for the user. But in work-related messages, “doing” prompts were more common, constituting 56% of message traffic.

For Anthropic, where work-related messages seemed more dominant to begin with, there was a clear trend for users to ask the chatbot to complete tasks for them, and in fact the majority of Anthropic’s API usage (some 77%) was classified as automation requests. Anthropic’s research also indicated that many of the tasks that were most popular with business users of Claude also were those that were most expensive to run, indicating that companies are probably finding—despite some other survey and anecdotal evidence to the contrary—that the value of automating tasks with AI is indeed worth the money.

The studies also indicate that in business contexts people increasingly want AI models to automate tasks for them, not necessarily offer decision support or expert advice. This could have significant implications for economies as a whole: If companies mostly use the technology to automate tasks, the negative effect of AI on jobs is likely to be far greater.

There were lots of other interesting tidbits in the two studies. For instance, whereas previous usage data had shown a significant gender gap, with men far more likely than women to be using ChatGPT, the new study shows that gap has now disappeared. Anthropic’s research shows interesting geographic divergence in Claude usage too—usage is concentrated on the coasts, which is to be expected, but there are also hotspots in Utah and Nevada.

With that, here’s more AI news.

Jeremy Kahn
jeremy.kahn@fortune.com
@jeremyakahn

FORTUNE ON AI

China says Nvidia violated antitrust laws as it ratchets up pressure ahead of U.S. trade talks—by Jeremy Kahn

AI chatbots are harming young people. Regulators are scrambling to keep up.—by Beatrice Nolan

OpenAI’s deal with Microsoft could pave the way for a potential IPO—by Beatrice Nolan

EYE ON AI NEWS

Alphabet announces $6.8 billion investment in U.K.-based AI initiatives, other tech companies also announce U.K. investments alongside Trump’s state visit. Google’s parent company announced a £5 billion ($6.8 billion) investment in the U.K. over the next two years, funding AI infrastructure, a new $1 billion AI data center that is set to open this week, and more funding for research at Google DeepMind, its advanced AI lab that continues to be headquartered in London. The BBC reports that the investments were unveiled ahead of President Trump’s state visit to Britain. Many other big U.S. tech companies are expected to make similar investments over the next few days. For instance, Nvidia, OpenAI and U.K. data center provider Nscale also announced a multi-billion-dollar data center project this week. More on that here from Bloomberg. Meanwhile, Salesforce said it was increasing a previously announced package of investments in the U.K., much of it around AI, from $4 billion to $6 billion.

FTC launches inquiry into AI chatbot effects on children amid safety concerns. The U.S. Federal Trade Commission has started an inquiry into how AI chatbots affect children, sending detailed questionnaires to six major companies including OpenAI, Alphabet, Meta, Snap, xAI, and Character.AI. Regulators are seeking information on issues such as sexually themed responses, safeguards for minors, monetization practices, and how companies disclose risks to parents. The move follows rising concerns over children’s exposure to inappropriate or harmful content from chatbots, lawsuits and congressional scrutiny, and comes as firms like OpenAI have pledged new parental controls. Read more here from the New York Times.

Salesforce backtracks, reinstates team that helped customers adopt AI agents. The team, called Well-Architected, had displeased Salesforce CEO Marc Benioff by suggesting to customers that deploying AI agents successfully would take extensive planning and significant work, a position that contradicted Benioff’s own pitch to customers that, with Salesforce, deploying AI agents was a cinch. Now, according to a story in The Information, the software company has had to reconstitute the team, which provided advisory and consulting help to companies implementing Agentforce. The company is finding Agentforce adoption is lagging its expectations—with fewer than 5% of its 150,000 clients currently paying for the AI agent product, the publication reported—amid complaints that the product is too expensive, too difficult to implement, and too prone to accuracy issues and errors. Having invested heavily in the pivot to Agentforce, Benioff is now under pressure from investors to deliver.

Humanoid robotics startup Figure AI valued at $39 billion in new funding deal. Figure AI, a startup developing humanoid robots, has raised over $1 billion in a new funding round that values the company at $39 billion, making it one of the world’s most valuable startups, Bloomberg reports. The round was led by Parkway Venture Capital with participation from major backers including Nvidia, Salesforce, Brookfield, Intel, and Qualcomm, alongside earlier supporters like Microsoft, OpenAI, and Jeff Bezos. Founded in 2022, Figure aims to build general-purpose humanoid robots, though Fortune’s Jason del Rey questioned whether the company was exaggerating the extent to which its robots were being deployed with BMW.

EYE ON AI RESEARCH

Can AI replace my job? Journalists are certainly worried about what AI is doing to the profession. Mostly, though, after some initial concerns that AI would directly replace journalists, the concern has largely shifted to fears that AI will further undermine the business models that fund good journalism (see Brain Food below). But recently a group of AI researchers in Japan and Taiwan created a benchmark called NEWSAGENT to see how well LLMs can do at actually taking source material and composing accurate news stories. It turned out that the models could, in many cases, do an ok job.

But the most interesting thing about the research is how the scientists, none of whom were journalists, characterized the results. They found that Alibaba’s open weight model, Qwen-3 32B, did best stylistically, but that GPT 4-o did better on metrics like objectivity and factual accuracy. And they write that human-written stories did not consistently outperform those drafted by the AI models in overall win rates, but that the human-written stories “emphasize factual accuracy.” The human-written stories were also often judged to be more objective than the AI-written ones.

The problem here is that in the real world, factual accuracy is the bedrock of journalism, and objectivity would be a close second. If the models fall down on accuracy, they should lose in every case to the human-written stories, even if evaluators preferred the AI-written ones stylistically.

This is why computer scientists should not be left to create benchmarks for real world professional tasks without deferring to expert advice from people working in those professions.  Otherwise you get distorted views of what AI models can and can’t do. You can read the NEWSAGENT research here on arxiv.org.

AI CALENDAR

Oct. 6-10: World AI Week, Amsterdam

Oct. 21-22: TedAI San Francisco.

Nov. 10-13: Web Summit, Lisbon. 

Nov. 26-27: World AI Congress, London.

Dec. 2-7: NeurIPS, San Diego

Dec. 8-9: Fortune Brainstorm AI San Francisco. Apply to attend here.

BRAIN FOOD

Is Google the most malevolent AI actor? A lot of publishing execs are starting to say so. At Fortune Brainstorm Tech in Deer Valley, Utah, last week, Neil Vogel, the CEO of magazine publisher People Inc. said that Google was “the worst” when it came to using publishers’ content without permission to train AI models. The problem, Vogel said, is that Google used the same web crawlers to index sites for Google Search as it did to scrape content to feed its Gemini AI models. While other AI vendors have increasingly been cutting multi-million dollar annual licensing deals to pay for publishers’ content, Google has refused to do so. And publishers’ can’t block Google’s bots without losing search traffic on which they currently depend for revenue.
You can read more on Vogel’s comments here



Source link

Continue Reading

Business

Why Jerome Powell’s latest rate cut still won’t help you get a lower mortgage rate

Published

on



For the third meeting in a row, the Federal Reserve cut interest rates—a “hawkish” move in an effort to help a softening labor market. The 0.25% cut brought the interest rate range to 3.5% to 3.75%—but economists and housing experts warn that’s not going to affect mortgage rates in the way potential homebuyers were hoping for. 

Chen Zhao, head of economics research at Redfin, wrote in a Wednesday post that the Fed’s December interest rate cut won’t move mortgage rates “because markets have already priced it in.” 

The Federal Reserve controls the Federal funds rate, which is a rate that banks charge each other and is more closely tied to credit cards, personal loans, and home-equity lines. A standard 30-year mortgage, on the other hand, is a long-term loan, and the pricing of those loans are tied more closely to yields on longer-term bonds like the 10-year Treasury and mortgage-backed securities. 

“Since this rate cut was no surprise, the markets have taken it in stride,” 43-year mortgage industry veteran Melissa Cohn, regional vice president of William Raveis Mortgage, told Fortune. She said more dropping shoes in terms of economic data will be the real turning point: “The future of bond yields and mortgage rates will be determined as new data on jobs and inflation get released.”

The current mortgage rate is 6.3%, according to Mortgage News Daily, which is of course much higher than the sub-3% rate that homebuyers from the pandemic era remember, although it’s also a far cry from the 8% peak in October 2023

“The committee’s projections and Chair Jerome Powell’s remarks indicate that this will be the last interest cut for a while,” Zhao wrote. “Given the underlying economic fundamentals of 3% inflation coupled with a weakening—but not recessionary—labor market, the Fed is likely to hold steady in the near future.

“Mortgage rates are unlikely to fall or rise by much,” she continued.

How mortgage rates affect housing affordability

Mortgage rates are just one piece of the housing affordability puzzle. While it may feel as if it’s the major roadblock in the ability to buy a home—especially having a recent memory of the pandemic housing boom—mortgage rates are only one factor. 

To put it in perspective, Zillow reported earlier this year not even a 0% mortgage rate would make buying a house affordable in several major U.S. cities. 

Let that sink in. 

Even without any interest accrued on a loan, homebuying is still out of reach for the typical American. Much of the affordability crisis has to do with home prices, which are more than 50% higher than in 2020. This has locked out new homebuyers from entering the market and current homeowners from selling. 

The mortgage rate drop required to make an average home affordable (to about 4.43%) for the typical buyer is “unrealistic,” according to Zillow economic analyst Anushna Prakash.  

“It’s unlikely rates will drop to the mid-[4% range] anytime soon,” Arlington, Va.–based real estate agent Philippa Main told Fortune. “And even if they did, housing prices are still at historic highs.” With 11 years of experience, Main is also a licensed mortgage loan officer.

To be sure, some economists see some light at the end of the tunnel for homebuyers plagued by high mortgage rates and home prices.

“For prospective buyers who have been waiting on the sidelines, the housing market is finally starting to listen,” wrote First American chief economist Mark Fleming in an Aug. 29 blog post. First American’s analysis takes into account inflation, and Fleming said: “The price of a house today is not directly comparable to the price of that same house 30 years ago.”



Source link

Continue Reading

Business

OpenAI debuts GPT-5.2 in effort to silence concerns it is falling behind its rivals

Published

on



OpenAI, under increasing competitive pressure from Google and Anthropic, has debuted a new AI model, GPT-5.2, that it says beats all existing models by a substantial margin across a wide range of tasks.

The new model, which is being released less than a month after OpenAI debuted its predecessor, GPT-5.1, performed particularly well on a benchmark of complicated professional tasks across a range of “knowledge work”—from law to accounting to finance—as well as on evaluations involving coding and mathematical reasoning, according to data OpenAI released.

Fidji Simo, the former InstaCart CEO who now serves as OpenAI’s CEO of applications, told reporters that the model should not been seen as a direct response to Google’s Gemini 3 Pro AI model, which was released last month. That release prompted OpenAI CEO Sam Altman to issue a “code red,” delaying the rollout of several initiatives in order to focus more staff and computing resources on improving its core product, ChatGPT.

“I would say that [the Code Red] helps with the release of this model, but that’s not the reason it is coming out this week in particular, it has been in the works for a while,” she said.

She said the company had been building GPT-5.2 “for many months.” “We don’t turn around these models in just a week. It’s the result of a lot of work,” she said. The model had been known internally by the code name “Garlic,” according to a story in The Information. The day before the model’s release Altman teased its imminent rollout by posting to social media a video clip of him cooking a dish with a large amount of garlic.

OpenAI executives said that the model had been in the hands of “Alpha customers” who help test its performance for “several weeks”—a time period that would mean the model was completed prior to Altman’s “code red” declaration.

These testers included legal AI startup Harvey, note-taking app Notion, and file-management software company Box, as well as Shopify and Zoom.

OpenAI said these customers found GPT-5.2 demonstrated a “state of the art” ability to use other software tools to complete tasks, as well as excelling at writing and debugging code.

Coding has become one of the most competitive use cases for AI model deployment within companies. Although OpenAI had an early lead in the space, Anthropic’s Claude model has proved especially popular among enterprises, exceeding OpenAI’s marketshare according to some figures. OpenAI is no doubt hoping to convince customers to turn back to its models for coding with GPT-5.2.

Simo said the “Code Red” was helping OpenAI focus on improving ChatGPT. “Code Red is really a signal to the company that we want to marshal resources in one particular area, and that’s a way to really define priorities and define things that can be deprioritized,” she said. “So we have had an increase in resources focused on ChatGPT in general.”

The company also said its new model is better than the company’s earlier ones at providing “safe completions”—which it defines as providing users with helpful answers while not saying things that might contribute to or worsen mental health crises.

“On the safety side, as you saw through the benchmarks, we are improving on pretty much every dimension of safety, whether that’s self harm, whether that’s different types of mental health, whether that’s emotional reliance,” Simo said. “We’re very proud of the work that we’re doing here. It is a top priority for us, and we only release models when we’re confident that the safety protocols have been followed, and we feel proud of our work.”

The release of the new model came on the same day a new lawsuit was filed against the company alleging that ChatGPT’s interactions with a psychologically troubled user had contributed to a murder-suicide in Connecticut. The company also faces several other lawsuits alleging ChatGPT contributed to people’s suicides. The company called the Connecticut murder-suicide “incredibly heartbreaking” and said it is continuing to improve “ChatGPT’s training to recognize and respond to signs of mental or emotional distress, de-escalate conversations and guide people toward real-world support.” 

GPT-5.2 showed a large jump in performance across several benchmark tests of interest to enterprise customers. It met or exceeded human expert performance on a wide range of difficult professional tasks, as measured by OpenAI’s GDPval benchmark, 70.9% of the time. That compares to just 38.8% of the time for GPT-5, a model that OpenAI released in August; 59.6% for Anthropic’s Claude Opus 4.5; and 53.3% for Google’s Gemini 3 Pro.

On the software development benchmark, SWE-Bench Pro, GPT-5.2 scored 55.6%, which was almost 5 percentage points better than its predecessor, GPT-5.1, and more than 12% better than Gemini 3 Pro.

OpenAI’s Aidan Clark, vice president of research (training), declined to answer questions about exactly what training methods had been used to upgrade GPT-5.2’s performance, although he said that the company had made improvements across the board, including in pretraining, the initial step in creating an AI model.

When Google released its Gemini 3 Pro model last month, its researchers also said the company had made improvements in pretraining as well as post-training. This surprised some in the field who believed that AI companies had largely exhausted the ability to wring substantial improvements out of the pretraining stage of model building, and it was speculated that OpenAI may have been caught off guard by Google’s progress in this area.



Source link

Continue Reading

Business

OpenAI and Disney just ended the ‘war’ between AI and Hollywood with their $1 billion Sora deal

Published

on



Disney’s $1 billion investment in OpenAI, announced Thursday morning—and its decision to let more than 200 Disney, Pixar, Marvel, and Star Wars characters appear inside the Sora video generator—is more than a licensing deal. According to copyright and AI law expert Matthew Sag, who teaches at Emory University’s law school, the deal marks a strategic realignment that could reshape how Hollywood protects its IP in the face of AI-generated content that threatens to leech on their legally-protected magic. 

“AI companies are either in a position where they need to aggressively filter user prompts and model outputs to make sure that they don’t accidentally show Darth Vader, or strike deals with the rights holders to get permission to make videos and images of Darth Vader,” Sag told Fortune. “The licensing strategy is much more of a win-win.” 

The three-year agreement gives OpenAI the right to ingest hundreds of Disney-owned characters into Sora and ChatGPT Image. Disney will also receive equity warrants and become a major OpenAI customer, while deploying ChatGPT internally.

Sag said the deal itself will be a kind of “revenue-sharing.”

“OpenAI hasn’t figured out the revenue model,” Sag said. “So I think making this just an investment deal, in some ways, simplifies it. For Disney … [OpenAI] will figure out a way to make this profitable at some point, and [Disney will] get a cut of that.”

Why this deal matters: the ‘Snoopy problem’

For more than a year, the biggest legal threat to large-scale generative AI has centered on what Sag calls the “Snoopy problem”: It is extremely difficult to train powerful generative models without some degree of memorization, and copyrightable characters are uniquely vulnerable because copyright protects them in the abstract.

Sag was careful to outline a key distinction. AI companies aren’t licensing the right to train on copyrighted works; they’re licensing the right to create outputs that would otherwise be infringing.

That’s because the case for AI companies training their models on unlicensed content is “very strong,” Sag said. Two recent court rulings involving Anthropic and Meta have strengthened those arguments.  

The real stumbling block, Sag said, has always been outputs, not training. If a model can accidentally produce a frame that looks too much like Darth Vader, Homer Simpson, Snoopy, or Elsa, the fair use defense begins to fray.

“If you do get too much memorization, if that memorization finds its way into outputs, then your fair-use case begins to just crumble,” Sag said.

While it’s impossible to license enough text to train an LLM (“that would take a billion” deals, Sag said), it is possible to build image or video models entirely from licensed data if you have the right partners. This is why deals like Disney’s are crucial: They turn previously illegal outputs into legal ones, irrespective of whether the training process itself qualifies as fair use.

“The limiting principle is going to be essentially about whether—in their everyday operation—these models reproduce substantial portions of works from their training data,” Sag said.

The deal, Sag says, is also a hedge against Hollywood’s lawsuits. This announcement is “very bad” for Midjourney, who Disney is suing for copyright infringement, because it upholds OpenAI’s licensing deal as the “responsible” benchmark for AI firms. 

This is also a signal about the future of AI data

Beyond copyright risk, the deal exposes another trend: the drying up of high-quality, unlicensed data on the public internet.

In a blog post, Sag wrote:

“The low-hanging fruit of the public internet has been picked,” he wrote. “To get better, companies like OpenAI are going to need access to data that no one else has. Google has YouTube; OpenAI now has the Magic Kingdom.”

This is the core of what he calls the “data scarcity thesis.” OpenAI’s next leap in model quality may require exclusive content partnerships, as opposed to more scraping. 

“By entangling itself with the world’s premier IP holder, OpenAI makes itself indispensable to the very industry that threatened to sue it out of existence,” Sag wrote. 

AI and Hollywood have spent three years locked in a cold war over training data, likeness rights and infringement. With Disney’s $1 billion investment, that era appears to be ending.

“This is the template for the future,” Sag wrote. “We are moving away from total war between AI and content, toward a negotiated partition of the world.”



Source link

Continue Reading

Trending

Copyright © Miami Select.