Connect with us

Business

OpenAI’s new AI safety tools could give a false sense of security

Published

on



OpenAI last week unveiled two new free-to-download tools that are supposed to make it easier for businesses to construct guardrails around the prompts users feed AI models and the outputs those systems generate.

The new guardrails are designed so a company can, for instance, more easily set up contorls to prevent a customer service chatbot responding with a rude tone or revealing internal policies about how it should make decisions around offering refunds, for example.

But while these tools are designed to make AI models safer for business customers, some security experts caution that the way OpenAI has released them could create new vulnerabilities and give companies a false sense of security. And, while OpenAI says it has released these security tools for the good of everyone, some question whether OpenAI’s motives aren’t driven in part by a desire to blunt one advantage that its AI rival Anthropic, which has been gaining traction among business users in part because of a perception that its Claude models have more robust guardrails than other competitors.

The OpenAI security tools—which are called gpt-oss-safeguard-120b and gpt-oss-safeguard-20b—are themselves a type of AI model known as a classifier, which is designed to assess whether the prompt a user submits to a larger, more general-purpose AI model as well as that larger AI model produces meet a set of rules. Companies that purchase and deploy AI models could, in the past, train these classifiers themselves, but the process was time-consuming and potentially expensive, since the developers would have to collect examples of content that violates the policy in order to train the classifier. And then, if the company wanted to adjust the policies used for the guardrails, they would have to collect new examples of violations and retrain the classifier.

OpenAI is hoping the new tools can make that process faster and more flexible. Rather than being trained to follow one fixed rulebook, these new security classifiers can simply read a written policy and apply it to new content.

OpenAI says this method, which it calls “reasoning-based classification,” allows companies to adjust their safety policies as easily as editing the text in a document instead of rebuilding an entire classification model. The company is positioning the release as a tool for enterprises that want more control over how their AI systems handle sensitive information, such as medical records or personnel records.

However, while the tools are supposed to be safer for enterprise customers, some safety experts say that they instead may give users a false sense of security. That’s because OpenAI has open-sourced the AI classifiers. That means they have made all the code for the classifiers available for free, including the weights, or the internal settings of the AI models.

Classifiers act like extra security gates for an AI system, designed to stop unsafe or malicious prompts before they reach the main model. But by open-sourcing them, OpenAI risks sharing the blueprints to those gates. That transparency could help researchers strengthen safety mechanisms, but it might also make it easier for bad actors to find the weak spots and risks, creating a kind of false comfort.

“Making these models open source can help attackers as well as defenders,” David Krueger, an AI safety professor at Mila, told Fortune. It will make it easier to develop approaches to bypassing the classifiers and other similar safeguards.”

For instance, when attackers have access to the classifier’s weights, they can more easily develop what are known as “prompt injection” attacks, where they develop prompts that trick the classifier into disregarding the policy it is supposed to be enforcing. Security researchers have found that in some cases even a string of characters that look nonsensical to a person can, for reasons researchers don’t entirely understand, convince an AI model to disregard its guardrails and do something it is not supposed to, such as offer advice for making a bomb or spew racist abuse.

Representatives for OpenAI directed Fortune to the company’s blog post announcement and technical report for the models.

Short-term pain for long-term gains

Open-source can be a double-edged sword when it comes to safety. It allows researchers and developers to test, improve, and adapt AI safeguards more quickly, increasing transparency and trust. For instance, there may be ways in which security researchers could adjust the model’s weights to make it more robust to prompt injection without degrading the model’s performance.

But it can also make it easier for attackers to study and bypass those very protections—for instance, by using other machine learning software to run through hundreds of thousands of possible prompts until it finds ones that will cause the model to jump its guardrails. What’s more, security researchers have found that these kinds of automatically-generated prompt injection attacks developed on open source AI models will also sometimes work against proprietary AI models, where the attackers don’t have access to the underlying code and model weights. Researchers have speculated this is because there may be something inherent in the way all large language models encode language that similar prompt injections will have success against any AI model.

In this way, open sourcing the classifiers may not just give users a false sense of security that their own system is well-guarded, it may actually make every AI model less secure. But experts said that this risk was probably worth taking because open-sourcing the classifiers should also make it easier for all of the world’s security experts to find ways to make the classifiers more resistant to these kinds of attacks.

“In the long term, it’s beneficial to kind of share the way your defenses work— it may result in some kind of short-term pain. But in the long term, it results in robust defenses that are actually pretty hard to circumvent,” Vasilios Mavroudis, principal research scientist at the Alan Turing Institute, said.

Mavroudis said that while open-sourcing the classifiers could, in theory, make it easier for someone to try to bypass the safety systems on OpenAI’s main models, the company likely believes this risk is low. He said that OpenAI has other safeguards in place, including having teams of human security experts continually trying to test their models’ guardrails in order to find vulnerabilities and hopefully improve them.

“Open-sourcing a classifier model gives those who want to bypass classifiers an opportunity to learn about how to do that. But determined jailbreakers are likely to be successful anyway,” Robert Trager, co-director of the Oxford Martin AI Governance Initiative, said.

“We recently came across a method that bypassed all safeguards of the major developers around 95% of the time — and we weren’t looking for such a method. Given that determined jailbreakers will be successful anyway, it’s useful to open-source systems that developers can use for the less determined folks,” he added.

The enterprise AI race

The release also has competitive implications, especially as OpenAI looks to challenge rival AI company Anthropic’s growing foothold among enterprise customers. Anthropic’s Claude family of AI models have become popular with enterprise customers partly because of their reputation for stronger safety controls compared to other AI models. Among the safety tools Anthropic uses are “constitutional classifiers” that work similarly to the ones OpenAI just open-sourced.

Anthropic has been carving out a market niche with enterprise customers, especially when it comes to coding. According to a July report from Menlo Ventures, Anthropic holds 32% of the enterprise large language model market share by usage compared to OpenAI’s 25%. In coding‑specific use cases, Anthropic reportedly holds 42%, while OpenAI has 21%. By offering enterprise-focused tools, OpenAI may be attempting to win over some of these business customers, while also positioning itself as a leader in AI safety.

Anthropic’s “constitutional classifiers,” consist of small language models that check a larger model’s outputs against a written set of values or policies. By open-sourcing a similar capability, OpenAI is effectively giving developers the same kind of customizable guardrails that helped make Anthropic’s models so appealing.

“From what I’ve seen from the community, it seems to be well received,” Mavroudis said. “They see the model as potentially a way to have auto-moderation. It also comes with some good connotation, as in, ‘we’re giving to the community.’ It’s probably also a useful tool for small enterprises where they wouldn’t be able to train such a model on their own.”

Some experts also worry that open-sourcing these safety classifiers could centralize what counts as “safe” AI.

“Safety is not a well-defined concept. Any implementation of safety standards will reflect the values and priorities of the organization that creates it, as well as the limits and deficiencies of its models,” John Thickstun, an assistant professor of computer science at Cornell University, told VentureBeat. “If industry as a whole adopts standards developed by OpenAI, we risk institutionalizing one particular perspective on safety and short-circuiting broader investigations into the safety needs for AI deployments across many sectors of society.”



Source link

Continue Reading

Business

Why the timing was right for Salesforce’s $8 billion acquisition of Informatica — and for the opportunities ahead

Published

on



The must-haves for building a market-leading business include vision, talent, culture, product innovation and customer focus. But what’s the secret to success with a merger or acquisition? 

I was asked about this in the wake of Salesforce’s recently completed $8 billion acquisition of Informatica. In part, I believe that people are paying attention because deal-making is up in 2025. M&A volume reached $2.2 trillion in the first half of the year, a 27% increase compared to a year ago, according to JP Morgan. Notably, 72% of that volume involved deals greater than $1 billion. 

There will be thousands of mergers and acquisitions in the United States this year across industries and involving companies of all sizes. It’s not unusual for startups to position themselves to be snapped up. But Informatica, founded in 1993, didn’t fit that mold. We have been building, delivering, supporting and partnering for many years. Much of the value we bring to Salesforce and its customers is our long-earned experience and expertise in enterprise data management. 

Although, in other respects, a “legacy” software company like ours — founded well before cloud computing was mainstream — and early-stage startups aren’t so different. We all must move fast and differentiate. And established vendors and growth-oriented startups have a few things in common when it comes to M&A, as well. 

First and foremost is a need to ensure that the strategies of the two companies involved are in alignment. That seems obvious, but it’s easier said than done. Are their tech stacks based on open protocols and standards? Are they cloud-native by design? And, now more than ever, are they both AI-powered and AI-enabling? All of these came together in the case of Salesforce and Informatica, including our shared belief in agentic AI as the next major breakthrough in business technology.

Don’t take your foot off the gas

In the days after the acquisition was completed, I was asked during a media interview if good luck was a factor in bringing together these two tech industry stalwarts. Replace good luck with good timing, and the answer is a resounding, “Yes!”

As more businesses pursue the productivity and other benefits of agentic AI, they require high-quality data to be successful. These are two areas where Salesforce and Informatica excel, respectively. And the agentic AI opportunity — estimated to grow to $155 billion by 2030 — is here and now. So the timing of the acquisition was perfect. 

Tremendous effort goes into keeping an organization on track, leading up to an acquisition and then seeing it through to a smooth and successful completion. In the few months between the announcement of Salesforce’s intent to acquire Informatica and the close, we announced new partnerships and customer engagements and a fall product release that included autonomous AI agents, MCP servers and more. 

In other words, there’s no easing into the new future. We must maintain the pace of business because the competitive environment and our customers require it. That’s true whether you’re a small, venture-funded organization or, like us, an established firm with thousands of employees and customers. Going forward we plan to keep doing what we do best: help organizations connect, manage and unify their AI data. 

Out with the old, in with the new

It’s wrong to think of an acquisition as an end game. It’s a new chapter. 

Business leaders and employees in many organizations have demonstrated time and again that they are quite good at adapting to an ever-changing competitive landscape. A few years ago, we undertook a company-wide shift from on-premises software to cloud-first. There was short-term disruption but long-term advantage. It’s important to develop an organizational mindset that thrives on change and transformation, so when the time comes, you’re ready for these big steps. 

So, even as we take pride in all that we accomplished to get to this point, we now begin to take on a fresh identity as part of a larger whole. It’s an opportunity to engage new colleagues and flourish professionally. And importantly, customers will be the beneficiaries of these new collaborations and synergies. On the day Informatica was welcomed into the Salesforce family and ecosystem, I shared my feeling that “the best is yet to come.” That’s my North Star and one I recommend to every business leader forging ahead into an M&A evolution — because the truest measure of success ultimately will be what we accomplish next.

The opinions expressed in Fortune.com commentary pieces are solely the views of their authors and do not necessarily reflect the opinions and beliefs of Fortune.



Source link

Continue Reading

Business

The ‘Great Housing Reset’ is coming: Income growth will outpace home-price growth in 2026

Published

on



Homebuyers may experience a reprieve in 2026 as price normalization and an increase in home sales over the next year will take some pressure off the market—but don’t expect homebuying to be affordable in the short run for Gen Z and young families.

The “Great Housing Reset” will start next year, with income growth outpacing home-price growth for a prolonged period for the first time since the Great Recession era, according to a Redfin report released this week. 

The residential real estate brokerage sees mortgage rates in the low-6% range, down from down from the 2025 average of 6.6%; a median home sales price increase of just 1%, down from 2% this year; and monthly housing payments growth that will lag behind wage growth, which will remain steady at 4%.

These trends toward increased affordability will likely bring back some house hunters to the market, but many Gen Zers and young families will opt for nontraditional living situations, according to the report. 

More adult children will be living with their parents, as households continue to shift further away from a nuclear family structure, Redfin predicted.

“Picture a garage that’s converted into a second primary suite for adult children moving back in with their parents,” the report’s authors wrote. “Redfin agents in places like Los Angeles and Nashville say more homeowners are planning to tailor their homes to share with extended family.”

Gen Z and millennial homeownership rates plateaued last year, with no improvement expected. Just over one-quarter of Gen Zers owned their home in 2024, while the rate for millennial owners was 54.9% in the same year.

Meanwhile, about 6% of Americans who struggled to afford housing as of mid-2025 moved back in with their parents, while another 6% moved in with roommates. Both trends are expected to increase in 2026, according to the report.

Obstacles to home affordability 

Despite factors that could increase affordability for prospective homebuyers, C. Scott Schwefel, a real estate attorney at Shipman, Shaiken & Schwefel, LLC, told Fortune that income growth and home-price growth are just a few keys to sustainable homeownership. 

An improved income-to-price ratio is welcome, but unless tax bills stabilize, many households may not experience a net relief, Schwefel said.

“Prospective buyers need to recognize that affordability is not just price versus income…it’s price, mortgage rate and the annual bill for living in a place—and that bill includes property taxes,” he added.

In November, voters—especially young ones—showed lowering housing costs is their priority, the report said. But they also face high sale prices and mortgage rates, inflated insurance premiums, and potential utility costs hikes due to a data center construction boom that’s driving up energy bills. The report’s authors expect there to be a bipartisan push to help remedy the housing affordability crisis.

Still, an affordable housing market for first-time home buyers and young families still may be far away.

“The U.S. housing market should be considered moving from frozen to thawing,” Sergio Altomare, CEO of Hearthfire Holdings, a real estate private equity and development company, told Fortune

“Prices aren’t surging, but they’re no longer falling,” he added. “We are beginning to unlock some activity that’s been trapped for a couple of years.”



Source link

Continue Reading

Business

Nvidia’s CEO says AI adoption will be gradual, but we still may all end up making robot clothing

Published

on



Nvidia CEO Jensen Huang doesn’t foresee a sudden spike of AI-related layoffs, but that doesn’t mean the technology won’t drastically change the job market—or even create new roles like robot tailors.

The jobs that will be the most resistant to AI’s creeping effect will be those that consist of more than just routine tasks, Huang said during an interview with podcast host Joe Rogan this week. 

“If your job is just to chop vegetables, Cuisinart’s gonna replace you,” Huang said.

On the other hand, some jobs, such as radiologists, may be safe because their role isn’t just about taking scans, but rather interpreting those images to diagnose people.

“The image studying is simply a task in service of diagnosing the disease,” he said.

Huang allowed that some jobs will indeed go away, although he stopped short of using the drastic language from others like Geoffrey Hinton a.k.a. “the Godfather of AI” and Anthropic CEO Dario Amodei, both of whom have previously predicted massive unemployment thanks to the improvement of AI tools.

Yet, the potential, AI-dominated job market Huang imagines may also add some new jobs, he theorized. This includes the possibility that there will be a newfound demand for technicians to help build and maintain future AI assistants, Huang said, but also other industries that are harder to imagine.

“You’re gonna have robot apparel, so a whole industry of—isn’t that right? Because I want my robot to look different than your robot,” Huang said. “So you’re gonna have a whole apparel industry for robots.”

The idea of AI-powered robots dominating jobs once held by humans may sound like science fiction, and yet some of the world’s most important tech companies are already trying to make it a reality. 

Tesla CEO Elon Musk has made the company’s Optimus robot a central tenet of its future business strategy. Just last month, Musk predicted money will no longer exist in the future and work will be optional within the next 10 to 20 years thanks to a fully fledged robotic workforce. 

AI is also advancing so rapidly that it already has the potential to replace millions of jobs. AI can adequately complete work equating to about 12% of U.S. jobs, according to a Massachusetts Institute of Technology (MIT) report from last month. This represents about 151 million workers representing more than $1 trillion in pay, which is on the hook thanks to potential AI disruption, according to the study.

Even Huang’s potentially new job of AI robot clothesmaker may not last. When asked by Rogan whether robots could eventually make apparel for other robots, Huang replied: “Eventually. And then there’ll be something else.”



Source link

Continue Reading

Trending

Copyright © Miami Select.