包阅导读总结
1. 关键词:
– 索赔处理
– 向量搜索
– 大型语言模型
– 数据整合
– 保险应用
2. 总结:
本文探讨将 MongoDB Atlas 向量搜索和大型语言模型用于保险索赔处理,介绍了面临的挑战及解决方案,包括数据组织、架构和流程,强调其带来的速度、灵活性、自然交互和数据可访问性优势,并指出该模式还适用于保险领域的其他场景。
3. 主要内容:
– 索赔处理的挑战
– 信息分散在不同系统和数据格式中,难以整合和索引。
– 保险公司积累大量非结构化数据但未充分利用。
– 解决方案
– 利用 Atlas 向量搜索和大型语言模型的集成(RAG)。
– 数据准备:确保数据可访问且安全,组织在 MongoDB 集合中。
– 架构和流程:用户输入自然语言提示,经向量化后由 Atlas 向量搜索检索相似文档,与提示一起作为上下文传递给 LLM,LLM 生成回答。
– 优势
– 提高速度和准确性。
– 增加灵活性。
– 实现自然交互。
– 改善数据可访问性。
– 应用扩展
– 适用于客户服务、客户自助服务、承保等场景。
– 相关资源
– 提供快速入门指南等资源。
思维导图:
文章地址:https://www.mongodb.com/blog/post/rag-claim-processing-combing-mongodb-atlas-vector-search-llms
文章来源:mongodb.com
作者:Jeff Needham
发布时间:2024/8/8 16:19
语言:英文
总字数:1057字
预计阅读时间:5分钟
评分:86分
标签:MongoDB,Atlas 向量搜索,大语言模型,理赔处理,检索增强生成
以下为原文内容
本内容来源于用户推荐转载,旨在分享知识与观点,如有侵权请联系删除 联系邮箱 media@ilingban.com
Following up on our previous blog, AI, Vectors, and the Future of Claims Processing: Why Insurance Needs to Understand The Power of Vector Databases, we’ll pick up the conversation right where we left it. We discussed extensively how Atlas Vector Search can benefit the claim process in insurance and briefly covered Retrieval Augmented Generation (RAG) and Large Language Models (LLMs).
Check out our AI resource page to learn more about building AI-powered apps with MongoDB.
One of the biggest challenges for claim adjusters is pulling and aggregating information from disparate systems and diverse data formats. PDFs of policy guidelines might be stored in a content-sharing platform, customer information locked in a legacy CRM, and claim-related pictures and voice reports in yet another tool. All of this data is not just fragmented across siloed sources and hard to find but also in formats that have been historically nearly impossible to index with traditional methods. Over the years, insurance companies have accumulated terabytes of unstructured data in their data stores but have failed to capitalize on the possibility of accessing and leveraging it to uncover business insights, deliver better customer experiences, and streamline operations. Some of our customers even admit they’re not fully aware of all the data in their archives. There’s a tremendous opportunity to leverage this unstructured data to benefit the insurer and its customers.
Our image search post covered part of the solution to these challenges, opening the door to working more easily with unstructured data. RAG takes it a step further, integrating Atlas Vector Search and LLMs, thus allowing insurers to go beyond the limitations of baseline foundational models, making them context-aware by feeding them proprietary data. Figure 1 shows how the interaction works in practice: through a chat prompt, we can ask questions to the system, and the LLM returns answers to the user and shows what references it used to retrieve the information contained in the response. Great! We’ve got a nice UI, but how can we build an RAG application? Let’s open the hood and see what’s in it!
Architecture and flow
Before we start building our application, we need to ensure that our data is easily accessible and in one secure place. Operational Data Layers (ODLs) are the recommended pattern for wrangling data to create single views. This post walks the reader through the process of modernizing insurance data models with Relational Migrator, helping insurers migrate off legacy systems to create ODLs.
Once the data is organized in our MongoDB collections and ready to be consumed, we can start architecting our solution. Building upon the schema developed in the image search post, we augment our documents by adding a few fields that will allow adjusters to ask more complex questions about the data and solve harder business challenges, such as resolving a claim in a fraction of the time with increased accuracy. Figure 2 shows the resulting document with two highlighted fields, “claimDescription”
and its vector representation, “claimDescriptionEmbedding”
. We can now create a Vector Search index on this array, a key step to facilitate retrieving the information fed to the LLM.
Having prepared our data, building the RAG interaction is straightforward; refer to this GitHub repository for the implementation details. Here, we’ll just discuss the high-level architecture and the data flow, as shown in Figure 3 below:
-
The user enters the prompt, a question in natural language.
-
The prompt is vectorized and sent to Atlas Vector Search; similar documents are retrieved.
-
The prompt and the retrieved documents are passed to the LLM as context.
-
The LLM produces an answer to the user (in natural language), considering the context and the prompt.
It is important to note how the semantics of the question are preserved throughout the different steps. The reference to “adverse weather” related accidents in the prompt is captured and passed to Atlas Vector Search, which surfaces claim documents whose claim description relates to similar concepts (e.g., rain) without needing to mention them explicitly. Finally, the LLM consumes the relevant documents to produce a context-aware question referencing rain, hail, and fire, as we’d expect based on the user’s initial question.
So what?
To sum it all up, what’s the benefit of combining Atlas Vector Search and LLMs in a Claim Processing RAG application?
-
Speed and accuracy: Having the data centrally organized and ready to be consumed by LLMs, adjusters can find all the necessary information in a fraction of the time.
-
Flexibility: LLMs can answer a wide spectrum of questions, meaning applications require less upfront system design. There is no need to build custom APIs for each piece of information you’re trying to retrieve; just ask the LLM to do it for you.
-
Natural interaction: Applications can be interrogated in plain English without programming skills or system training.
-
Data accessibility: Insurers can finally leverage and explore unstructured data that was previously hard to access.
Not just claim processing
The same data model and architecture can serve additional personas and use cases within the organization:
-
Customer Service: Operators can quickly pull customer data and answer complex questions without navigating different systems. For example, “Summarize this customer’s past interactions,” “What coverages does this customer have?” or “What coverages can I recommend to this customer?”
-
Customer self-service: Simplify your members’ experience by enabling them to ask questions themselves. For example, “My apartment is flooded. Am I covered?” or “How long do windshield repairs take on average?”
-
Underwriting: Underwriters can quickly aggregate and summarize information, providing quotes in a fraction of the time. For example, “Summarize this customer claim history.” “I Am renewing a customer policy. What are the customer’s current coverages? Pull everything related to the policy entity/customer. I need to get baseline info. Find relevant underwriting guidelines.”
VIDEO
If you would like to discover more about Converged AI and Application Data Stores with MongoDB, take a look at the following resources:
Head over to our quick-start guide to get started with Atlas Vector Search today.