Friday, 1 May 2026
Subscribe
logo
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Font ResizerAa
Data Center NewsData Center News
Search
  • AI Compute
  • Infrastructure
  • Power & Cooling
  • Security
  • Colocation
  • Cloud Computing
  • More
    • Sustainability
    • Industry News
    • About Data Center News
    • Terms & Conditions
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI & Compute > Does RAG make LLMs less safe?  Bloomberg research reveals hidden dangers
AI & Compute

Does RAG make LLMs less safe?  Bloomberg research reveals hidden dangers

Last updated: April 28, 2025 12:07 pm
Published April 28, 2025
Share
Does RAG make LLMs less safe?  Bloomberg research reveals hidden dangers
SHARE

Be a part of our day by day and weekly newsletters for the newest updates and unique content material on industry-leading AI protection. Study Extra


Retrieval Augmented Technology (RAG) is meant to assist enhance the accuracy of enterprise AI by offering grounded content material. Whereas that’s usually the case, there may be additionally an unintended facet impact.

In response to stunning new analysis printed as we speak by Bloomberg, RAG can doubtlessly make massive language fashions (LLMs) unsafe. 

Bloomberg’s paper, ‘RAG LLMs are Not Safer: A Security Evaluation of Retrieval-Augmented Technology for Massive Language Fashions,’ evaluated 11 widespread LLMs together with Claude-3.5-Sonnet, Llama-3-8B and GPT-4o. The findings contradict typical knowledge that RAG inherently makes AI techniques safer. The Bloomberg analysis staff found that when utilizing RAG, fashions that sometimes refuse dangerous queries in commonplace settings usually produce unsafe responses.

Alongside the RAG analysis, Bloomberg launched a second paper, ‘Understanding and Mitigating Dangers of Generative AI in Monetary Companies,’ that introduces a specialised AI content material threat taxonomy for monetary providers that addresses domain-specific issues not coated by general-purpose security approaches.

The analysis challenges widespread assumptions that retrieval-augmented era (RAG) enhances AI security, whereas demonstrating how present guardrail techniques fail to deal with domain-specific dangers in monetary providers purposes.

“Techniques have to be evaluated within the context they’re deployed in, and also you won’t be capable of simply take the phrase of others that say, Hey, my mannequin is protected, use it, you’re good,” Sebastian Gehrmann, Bloomberg’s Head of Accountable AI, informed VentureBeat. 

RAG techniques could make LLMs much less protected, no more

RAG is extensively utilized by enterprise AI groups to offer grounded content material. The aim is to offer correct, up to date data. 

There was a variety of analysis and development in RAG in current months to additional enhance accuracy as properly. Earlier this month a brand new open-source framework known as Open RAG Eval debuted to assist validate RAG effectivity.

See also  OpenAI drops Deep Research access to Plus users, heating up AI agent wars with DeepSeek and Claude

It’s essential to notice that Bloomberg’s analysis isn’t questioning the efficacy of RAG or its capacity to cut back hallucination. That’s not what the analysis is about. Reasonably it’s about how RAG utilization impacts LLM guardrails in an surprising means.

The analysis staff found that when utilizing RAG, fashions that sometimes refuse dangerous queries in commonplace settings usually produce unsafe responses. For instance, Llama-3-8B’s unsafe responses jumped from 0.3% to 9.2% when RAG was carried out.

Gehrmann defined that with out RAG being in place, if a person typed in a malicious question, the built-in security system or guardrails will sometimes block the question. But for some purpose, when the identical question is issued in an LLM that’s utilizing RAG, the system will reply the malicious question, even when the retrieved paperwork themselves are protected.

“What we discovered is that when you use a big language mannequin out of the field, usually they’ve safeguards in-built the place, when you ask, ‘How do I do that unlawful factor,’ it can say, ‘Sorry, I can’t enable you to do that,’” Gehrmann defined. “We discovered that when you really apply this in a RAG setting, one factor that would occur is that the extra retrieved context, even when it doesn’t include any data that addresses the unique malicious question, may nonetheless reply that authentic question.”

How does RAG bypass enterprise AI guardrails?

So why and the way does RAG serve to bypass guardrails? The Bloomberg researchers weren’t solely sure although they did have just a few concepts.

Gehrmann hypothesized that the best way the LLMs had been developed and educated didn’t totally contemplate security alignments for actually lengthy inputs. The analysis demonstrated that context size straight impacts security degradation. “Supplied with extra paperwork, LLMs are usually extra susceptible,” the paper states, displaying that even introducing a single protected doc can considerably alter security conduct.

See also  PepsiCo uses AI to rethink how factories are designed and updated

“I believe the larger level of this RAG paper is you actually can’t escape this threat,” Amanda Stent, Bloomberg’s Head of AI Technique and Analysis, informed VentureBeat. “It’s inherent to the best way RAG techniques are. The way in which you escape it’s by placing enterprise logic or reality checks or guardrails across the core RAG system.”

Why generic AI security taxonomies fail in monetary providers

Bloomberg’s second paper introduces a specialised AI content material threat taxonomy for monetary providers, addressing domain-specific issues like monetary misconduct, confidential disclosure and counterfactual narratives.

The researchers empirically demonstrated that present guardrail techniques miss these specialised dangers. They examined open-source guardrail fashions together with Llama Guard, Llama Guard 3, AEGIS and ShieldGemma towards information collected throughout red-teaming workouts.

“We developed this taxonomy, after which ran an experiment the place we took overtly obtainable guardrail techniques which are printed by different companies and we ran this towards information that we collected as a part of our ongoing purple teaming occasions,” Gehrmann defined. “We discovered that these open supply guardrails… don’t discover any of the problems particular to our {industry}.”

The researchers developed a framework that goes past generic security fashions, specializing in dangers distinctive to skilled monetary environments. Gehrmann argued that common objective guardrail fashions are often developed for client dealing with particular dangers. So they’re very a lot targeted on toxicity and bias. He famous that whereas essential these issues will not be essentially particular to anybody {industry} or area. The important thing takeaway from the analysis is that organizations have to have the area particular taxonomy in place for their very own particular {industry} and software use circumstances.

Accountable AI at Bloomberg

Bloomberg has made a reputation for itself through the years as a trusted supplier of monetary information techniques. In some respects, gen AI and RAG techniques may doubtlessly be seen as aggressive towards Bloomberg’s conventional enterprise and due to this fact there might be some hidden bias within the analysis. 

See also  Enterprise AI factories are here and NTT DATA is building them with NVIDIA

“We’re within the enterprise of giving our shoppers the very best information and analytics and the broadest capacity to find, analyze and synthesize data,” Stent mentioned. “Generative AI is a instrument that may actually assist with discovery, evaluation and synthesis throughout information and analytics, so for us, it’s a profit.”

She added that the sorts of bias that Bloomberg is anxious about with its AI options are focussed on  finance. Points resembling information drift, mannequin drift and ensuring there may be good illustration throughout the entire suite of tickers and securities that Bloomberg processes are essential. 

For Bloomberg’s personal AI efforts she highlighted the corporate’s dedication to transparency.

 “All the pieces the system outputs, you may hint again, not solely to a doc however to the place within the doc the place it got here from,” Stent mentioned.

Sensible implications for enterprise AI deployment

For enterprises trying to cleared the path in AI, Bloomberg’s analysis imply that RAG implementations require a basic rethinking of security structure. Leaders should transfer past viewing guardrails and RAG as separate parts and as an alternative design built-in security techniques that particularly anticipate how retrieved content material may work together with mannequin safeguards.

Business-leading organizations might want to develop domain-specific threat taxonomies tailor-made to their regulatory environments, shifting from generic AI security frameworks to those who handle particular enterprise issues. As AI turns into more and more embedded in mission-critical workflows, this strategy transforms security from a compliance train right into a aggressive differentiator that clients and regulators will come to anticipate.

“It actually begins by being conscious that these points may happen, taking the motion of really measuring them and figuring out these points after which growing safeguards which are particular to the applying that you just’re constructing,” defined Gehrmann.


Source link
TAGGED: Bloomberg, dangers, hidden, LLMs, RAG, Research, reveals, safe
Share This Article
Twitter Email Copy Link Print
Previous Article SWiRL: The business case for AI that thinks like your best problem-solvers SWiRL: The business case for AI that thinks like your best problem-solvers
Next Article Kubernetes 1.33 'Octarine' Delivers Major Upgrades Kubernetes 1.33 ‘Octarine’ Delivers Major Upgrades
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

E.SUN Bank and IBM build AI governance framework for banking

E.SUN Financial institution is working with IBM to construct clearer AI governance guidelines for the…

March 13, 2026

Hyperscale cloud: Expectations versus reality

With all of the advertising and marketing round hyperscale cloud, you’d assume that adopting it…

April 5, 2025

Anthropic’s Claude Opus 4.5 is here: Cheaper AI, infinite chats, and coding skills that beat humans

Anthropic launched its most succesful synthetic intelligence mannequin but on Monday, slashing costs by roughly…

November 24, 2025

ElevenLabs debuts Conversational AI 2.0 voice assistants that understand when to pause, speak, and take turns talking

Be part of our each day and weekly newsletters for the most recent updates and…

June 1, 2025

AI is moving to the edge – and network security needs to catch up

Offered by T-Cell for EnterpriseSmall and mid-sized companies are adopting AI at a tempo that…

December 17, 2025

You Might Also Like

STL launches Neuralis data centre connectivity suite in the U.S.
AI & Compute

STL launches Neuralis data centre connectivity suite in the U.S.

By saad
What is optical interconnect and why Lightelligence's $10B debut says it matters for AI
AI & Compute

What is optical interconnect and why Lightelligence’s $10B debut says it matters for AI

By saad
IBM launches AI platform Bob to regulate SDLC costs
AI & Compute

IBM launches AI platform Bob to regulate SDLC costs

By saad
The evolution of encoders: From simple models to multimodal AI
AI & Compute

The evolution of encoders: From simple models to multimodal AI

By saad

About Us

Data Center News is your dedicated source for data center infrastructure, AI compute, cloud, and industry news.

Top Categories

  • AI & Compute
  • Cloud Computing
  • Power & Cooling
  • Colocation
  • Security
  • Infrastructure
  • Sustainability
  • Industry News

Useful Links

  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

Find Us on Socials

© 2026 Data Center News. All Rights Reserved.

© 2026 Data Center News. All Rights Reserved.
Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.