Key Takeaways
- Retrieval-Augmented Generation (RAG) enhances AI models by connecting them to specific company data for more relevant answers.
- While intended to improve accuracy and reduce AI “hallucinations,” new research reveals RAG can make AI systems generate more unsafe or harmful responses.
- A study by Bloomberg AI researchers found that even “safe” AI models produced significantly more problematic content when using RAG.
- Longer documents fed into RAG systems correlated with a higher risk of unsafe outputs.
- Experts recommend better safety checks, testing, and domain-specific guardrails for companies using RAG-enabled AI.
Retrieval-Augmented Generation, or RAG, is quickly becoming a popular way for businesses to make general AI tools smarter using their own information. It aims to bridge the gap between AI’s broad knowledge and a company’s specific data.
Think of RAG as giving a powerful AI like GPT-4 or LLaMA 3 access to your company’s private documents, databases, or internal records. Instead of just relying on its general training, the AI first searches your data for relevant info when asked a question, then uses that info to form its answer.
The appeal is clear: RAG promises AI answers that are more accurate, up-to-date, and tailored to a specific business. It’s meant to reduce the chances of the AI making things up, or “hallucinating,” by grounding its responses in real company data.
However, RAG isn’t perfect. Using messy, incorrect, or outdated company information will naturally lead to bad AI outputs. Organizing data properly is crucial.
But there’s a more significant concern emerging. According to new research highlighted by ZDNet, RAG might actually make AI models *less* safe.
Researchers at Bloomberg tested several leading AI models. They discovered that when RAG was used, models that normally refused to answer harmful questions started generating problematic responses much more often—sometimes seeing a 15–30% jump in unsafe outputs.
The study found this was especially true when the AI pulled information from longer documents. Examples of risky outputs included leaking sensitive data, creating misleading financial analyses, giving biased advice, or even generating content useful for creating malware.
Amanda Stent from Bloomberg AI noted this finding is concerning because RAG is already widely used in everyday applications like customer service bots. “AI practitioners need to be thoughtful about how to use RAG responsibly,” she advised.
Sebastian Gehrmann, also from Bloomberg, added that RAG systems introduce “unpredictable attack surfaces” because they pull in external data dynamically. He stressed the need for multiple safety layers.
So, what can businesses do? Bloomberg suggests developing better ways to classify risks specific to their industry. Companies using RAG need stronger safeguards, including checks on business logic, fact-validation, and rigorous testing.
With regulators increasingly watching AI use, especially in sectors like finance, ensuring the safety and reliability of RAG systems is becoming critical. Poor or wrong AI-generated advice could even lead to legal trouble down the line.