0:02 want to dive in.
0:06 >> Let's go ahead and get started here.
0:08 >> All right. Um, today we are going to be
0:11 discussing how to build enterprise ready
0:14 rag systems. Um,
0:16 yeah. Uh, before we dive into the actual
0:18 content, some really quick introductions
0:20 on our side.
0:22 Oh, my name is Kevin. I'm on the go to
0:24 market team here at Unstructured. I'm
0:26 the Ryan Serest of our webinar series.
0:28 I'm joined by my better half, Daniel
0:31 Scoffield. Um he is a principal
0:35 solutions architect. Um awesome.
0:38 The agenda for today's conversation, a
0:39 little bit of background on who
0:41 Unstructured is, what are the problems
0:46 that we solve? Um also what is Rag and
0:48 what are the problems that we set up to
0:50 solve for and then how do we implement
0:52 RAG systems successfully across the enterprise.
0:54 enterprise.
0:56 [snorts] Um some quick housekeeping.
0:57 Make sure that chat goes in the chat
0:59 box. If you have specific questions you
1:00 want us to get to at the end of the
1:02 webinar, there is a separate tab for
1:05 that. We will send a recording of the
1:07 webinar after the conversation with
1:08 resources to get started on
1:11 unstructured. And if you do want to talk
1:13 with somebody like Daniel about your use
1:15 cases, go to our website, schedule a
1:22 All right, a little bit of background on
1:25 unstructured. We started three years ago
1:27 as an open- source product looking to
1:28 answer the question, do companies
1:31 struggle to take unstructured data and
1:32 convert it into a structured JSON
1:34 output? Oh, Daniel, do you mind going to
1:37 the next slide?
1:40 Um, you flash forward to today and it's
1:42 been a pretty incredible ride. We have
1:44 amazing customers. We have over 50
1:46 million product downloads of our open
1:48 source product and we have some pretty
1:50 incredible backers who are helping us
1:52 solve this problem at scale. Um, and
1:56 without further ado, Daniel, I will turn
1:58 it over to you.
2:00 >> Thank you for that fantastic
2:01 introduction, Kevin. Um, as Kevin said,
2:03 my name is Daniel Scoffield. I'm a
2:04 principal solutions architect at
2:07 Unstructured. I've been with an uh been
2:09 involved with countless enterprise rag
2:10 deployments. So, I'm excited to be able
2:12 to share some of my insights about what
2:14 it takes to actually build an enterprise
2:16 ready rag system. I'd like to start
2:18 today's talk with a quick overview of
2:19 the current state of RAG here in late
2:22 2025 and then we'll get into the meat of
2:24 today's talk. Uh to kick us off though,
2:27 I'd like to address uh a common idea you
2:29 probably encountered periodically uh
2:31 especially the LinkedIn AI space over
2:33 the past couple years. And that is this
2:36 periodic yet persistent notion that rag
2:39 is dead. And this idea comes in various
2:42 forms and variations. Um such as that
2:45 rag is now somehow obsolete or outdated.
2:47 Uh you might hear something like rag is
2:49 so 2024.
2:52 Um, and it's also uh common when uh you
2:55 know you see such things as like a
2:59 sizable context window uh increase where
3:00 um you know for example when Google
3:02 released the first Gemini model with a 1
3:05 million token context window uh it
3:06 nearly broke the internet with all the
3:08 rag is dead chanting from AI
3:10 influencers. And it's also of course
3:12 whenever a new flavor of rag gets
3:15 introduced it's the same thing rag is
3:18 dead long live corrective rag. Um, so
3:19 you can quote me on this. Rag is not
3:21 going away as long as large language
3:24 models are still the dominant technology
3:26 powering the AI landscape. So the next
3:28 time you hear some clickbaity
3:30 announcement stating that rag is dead, I
3:32 want you to remember back to the many
3:34 '90s era rappers who are way ahead of
3:37 their time that said it best when they
3:40 said rag till we die. Uh, you even have
3:42 this nice graphic to remember that by.
3:44 >> Is that what they said?
3:46 >> That's exactly what they said, Kevin. Okay.
3:47 Okay.
3:49 >> Uh, of course, this is just a fun idea,
3:51 but I honestly would not be surprised if
3:53 RAG indeed does actually outlive us all.
3:55 Um, looking at the projections for the
3:58 back half of this decade, uh, the growth
4:00 of the fundamental rag pattern actually
4:02 looks stronger than ever. Of course,
4:03 these are just projections, but
4:05 according to the wellrespected Grand
4:07 View Research, uh the REGG market is
4:10 expected to grow at an impressive uh
4:12 near 40% compound compounded annual
4:15 growth rate uh going into 2030, reaching
4:19 over a 10 billion uh market size. So,
4:22 rest assured, RAG is not only not dead,
4:23 but it would seem that it's actually
4:26 only just getting started. Um and why is
4:29 that? So what problems does rag actually
4:33 solve uh for the enterprise? Um and
4:35 really the answer starts with a simple
4:38 problem. Um and that problem is that LM
4:41 are stateless. The only realistic way to
4:43 make an LM smart about your business
4:45 data. Um aside from fine-tuning, which
4:48 is impractical for a lot of reasons, um
4:50 the answer is to inject that information
4:52 into its context window at runtime.
4:54 Couple that with the fact that there is
4:57 an explosion of LM adoption where every
4:58 company is in a race to take advantage
5:00 of AI's capabilities. There's an urgent
5:02 demand for making a company's
5:04 proprietary context available to these
5:07 models. So if you have this demand, how
5:10 do you actually solve that securely? Um
5:11 this will appeal to the architects out
5:14 there. Um if you just fine-tune a model
5:16 on all your sensitive HR and finance
5:18 data, you've effectively just collapsed
5:20 all your access controls and
5:22 traceability. So, REGGG actually
5:24 provides this auditability and a
5:26 separation of concerns around the
5:28 security of data. And because of the
5:30 security advantage, REGGG has become the
5:32 dominant pattern for enterprise
5:35 grounding. And on top of that, we're at
5:37 the dawn of the agentic era where AIs
5:38 don't just answer questions, but they
5:40 also take actions. And so, even in this
5:42 new world order, rag is still the
5:44 dominant pattern for grounding these now
5:47 agents with your enterprise context.
5:49 And within aentic systems, reggg is
5:51 actually not only being used to access
5:53 enterprise knowledges, knowledge bases,
5:55 but it's also actually being used to
5:57 endow the agent itself with memory so it
5:59 can recall its past actions and conversations.
6:01 conversations.
6:03 Um, but what does a mature production
6:05 grade version of this architecture
6:07 actually look like here in 2025?
6:09 Something like this. I assume that this
6:11 audience is already familiar with rag,
6:12 so I won't spin long here. But just to
6:14 get us all on the same page, here's a
6:16 quick 30 second tour of a somewhat
6:19 standard rag production system in 2025.
6:20 So it has two parts. First, you have the
6:22 offline ingestion uh in the top right
6:25 here uh where we take our data, we chunk
6:27 it, embed it into vectors, and then load
6:29 it in a vector database. That becomes
6:31 the enterprise knowledge base. And then
6:34 secondly, we have an online query flow.
6:35 Um this can actually be divided into two
6:38 steps. the retrieval step on the top
6:40 here and then the generation step mostly
6:42 on the bottom there. And the query step,
6:45 we analyze the query's intent, its
6:47 keywords. Um, we then search the vector
6:49 database to get the top most relevant
6:50 chunks. And then finally, in the
6:52 generation flow, we don't just jump
6:54 those chunks in. We rerank them. We
6:56 maybe run them against an anti-h
6:58 hallucination algorithm. and then only
7:00 the best context makes its way to the
7:03 prompt where it then uh is used to
7:05 retrieve uh and synthesize a final
7:07 answer. So this complete flow creates a
7:10 sound trustworthy system. Um so with
7:12 this blueprint in mind uh let's actually
7:15 talk about the larger rag ecosystem. So
7:17 the most important thing to understand
7:18 about rag today is that it's no longer
7:21 just referring to one thing. It's an
7:22 entire field of research that has
7:24 rapidly evolved over the last 5 years.
7:26 So this is roughly the journey. It all
7:28 started back in 2020 with Meta's famous
7:30 paper laying out the approach. Um that's
7:34 in hindsight been na named naive rag. Um
7:35 and naive rag is basically just the
7:37 previous slide if you remove a lot of
7:39 the intermediary steps. That's quickly
7:42 moved into advanced rag and modular rag
7:44 um to address a lot of the quality and
7:46 scalability shortcomings of naive rag.
7:48 And this is where most production
7:50 systems actually start today. Um, and
7:51 this is where we introduce critical
7:53 components like the reranking, the query
7:56 transformations and hybrid search. And
7:57 of course, you notice what happens.
7:58 You'll see the accuracy jumps, but so
8:01 does the cost and complexity as well.
8:02 And finally, we're now at
8:04 self-reflective and agentic rag, where
8:06 the system doesn't just retrieve, but it
8:08 analyzes its own results, decides if
8:09 they're good enough, and then even
8:12 re-queries if needed. Um, and this
8:13 five-stage evolution isn't just actually
8:15 a simple line. Um, it's really more of
8:17 an entire universe of techniques and
8:20 architectures. So this is what the
8:22 actual universe looks like today. Uh you
8:24 don't have to read all the 25 of these,
8:26 but uh this grid is here just to reveal
8:27 all the different flavors of rag that
8:30 you as an architect um basically you
8:32 have at your disposal when you're
8:34 choosing an implementation pattern for a
8:36 given use case. So do you need uh
8:38 conversational rag if you're building a
8:41 chatbot or if you're synthesizing across
8:42 multiple complex documents you might
8:46 want to look at multihop rag and so on.
8:48 But even once you've chosen the right
8:50 flavor of rag for a given use case,
8:52 you're still going to need a strong ROI
8:55 uh focused business case as well as uh
8:58 you know robust security governance uh
9:00 evaluation frameworks and so on in order
9:03 to move that pilot into production. So
9:05 here is our road map for the next 10 to
9:07 12 minutes. We're going to cover the
9:09 remaining sort of four pillars of a
9:11 production system um the top four here
9:14 and then also quickly touch on uh rag
9:16 for a gentic systems uh which is going
9:17 to be light because we're going to
9:19 actually discuss this more in depth in a
9:20 future webinar.
9:24 Um so with that um we're going to start
9:26 with the most important questions for of
9:29 any enterprise rag system and that is
9:32 ROI. What are we building and why? what
9:35 is the expected ROI and what meaning
9:37 business uh meaningful business outcomes
9:38 are we aiming to achieve in building
9:40 this system.
9:42 So this is actually the step that has a
9:43 tendency not to receive the most
9:45 rigorous scrutiny and disciplined
9:47 approach that it deserves and it's
9:49 reportedly one of the biggest reasons
9:50 projects fail to move past the pilot
9:53 stage. So either the cost of the system
9:54 doesn't end up justifying its delivered
9:56 value or else it had a weak value
9:59 proposition from the start. So, how do
10:01 you find a use case that's worth
10:03 pursuing? The core idea is simple. You
10:06 want to target um high value friction in
10:08 the organization. You're not just uh
10:11 trying to go around automating tasks.
10:12 You want to find a process where your
10:15 most expensive employees are wasting the
10:17 most time looking for the most important
10:19 information. So, it's really that
10:21 intersection of high value and high
10:24 friction. That's your sweet spot. Um and
10:25 of course, this is going to look
10:27 different based on your business domain.
10:29 Um but let's just look at a few example
10:31 use cases here. Um so take the customer
10:33 support use case. Uh you have support
10:36 agents toggling five screens looking at
10:38 five different sources of information.
10:40 That's friction where a rag system can
10:42 help. Or take a sales enablement use
10:45 case where you have sales reps hunting
10:47 down scattered internal repositories
10:50 searching for uh answers to security
10:52 questionnaire in order to answer in
10:54 order to close a high stakes deal. Kevin
10:57 can probably uh attest to this that he's
10:58 been there and done that. That's
11:00 friction. If you find the
11:01 [clears throat] friction, apply rag, you
11:03 get ROI.
11:06 Um and so on. Um but once you've found
11:08 your high friction use case, how do you
11:11 then prove that it's actually working?
11:12 That's where success metrics come into
11:15 play. This is all about measuring what
11:16 actually matters. And the biggest
11:18 mistake a team makes and we see this
11:20 actually again and again is only
11:22 focusing on the technical metrics like
11:26 accuracy or uh engagement with the tool.
11:28 Nobody on the business side of the org
11:30 actually cares about a blue score
11:32 though. Success for them is measured by
11:34 business outcomes. So you really need to
11:37 target the business outcome metrics. Um
11:39 and so these are going to be the KPIs of
11:42 your VP or director um that they already
11:44 have on a dashboard somewhere. You're
11:45 not creating new metrics here. You're
11:47 actually moving the needle on their
11:50 metrics. So in support, it's things like
11:52 decreasing average handle time. In
11:54 sales, you're decreasing the sales cycle
11:56 length. Um this is actually the language
11:58 of ROI. And if you aren't moving the
12:00 business outcome, the next two don't
12:02 actually matter. So of course, if you
12:04 are creating that ROI, then technical
12:06 metrics also become of great importance.
12:07 So we'll cover two of those. The first
12:10 is the enduser experience. Um how do you
12:12 actually quantify that though? This is
12:14 all about whether users like and trust
12:16 the tool and if it actually creates a
12:18 positive impact in their workflow. Um
12:20 the single most important metric in this
12:21 area is going to be the simple thumbs up
12:23 thumbs down rating um that you embed
12:26 within the application itself. And this
12:27 is your real world sort of continuous
12:30 feedback loop. Finally, we have our
12:32 system performance metrics. These are
12:34 the ones that we as engineers love. Um
12:36 context relevance, did we find the right
12:38 stuff and answer faithfulness, did the
12:41 LM stick to the script. So these are
12:42 essential but of course you have to
12:43 remember their only purpose is to serve
12:46 the two categories above.
12:48 Okay. So now that we have a use case in
12:50 mind with a strong ROI potential we need
12:53 to shift our mindset uh to security. A
12:55 pilot is only going to give uh get off
12:57 the ground if it doesn't represent a
12:59 massive security risk uh to the organization.
13:01 organization.
13:03 Um so how do you actually secure a rag
13:06 system? Now the most critical part, you
13:08 cannot treat an LM based application
13:10 that connects to your enterprise data
13:12 like a standard application. [snorts]
13:13 These systems really require a
13:16 three-stage security posture. So first
13:17 you have to have your pre- retrieval
13:19 guard rails. This is going to be what
13:21 authenticates the user ensuring that
13:23 they are who they say they are as well
13:25 as authorization where you're checking
13:26 whether they should have access to the
13:29 system at all in the first place. And
13:30 this one's actually pretty
13:31 straightforward. You'll see this across
13:33 many applications. Um but the next one
13:35 is actually very rags specific and
13:37 that's the retrieval time guardrails.
13:38 This is all around securing the data
13:41 itself. So this one comes with some uh
13:44 challenges. Um but security in this step
13:45 basically means that you're ensuring
13:48 that both the user as well as the LLM
13:50 only see documents or pieces of
13:52 documents that they're allowed to see.
13:54 And finally there's post retrieval which
13:56 is more around securing the answer. And
13:58 this is u basically filtering out bad
14:00 results um you know flagging incorrect
14:03 information possibly toxic content
14:05 before it reaches the end user. So now
14:06 of all these the most important to
14:08 actually get right and also
14:09 unfortunately the one that comes with
14:11 most unique security challenges is the
14:14 second one securing the data. So let's
14:17 zoom into that one uh for just a minute.
14:19 Uh so within this one um before you
14:21 write any code you really want to
14:22 understand what the security model needs
14:25 to be for the system. So, this is going
14:28 to often be use case specific. Um, and
14:29 within that use case, you need to ask
14:31 yourself who is the audience of this
14:33 tool and what is the sensitivity of the
14:35 data. If this is a public chatbot or a
14:37 generic helper, you might be able to get
14:39 away with just setting up a new simple
14:41 permissions system from scratch, such as
14:44 a basic arbback system where you're uh
14:46 implementing by tagging documents with
14:48 certain metadata during ingestion and
14:50 then filtering on those tags during
14:52 retrieval. That's actually the easy
14:54 path. Um, but for a lot of internal uh
14:56 enterprise use cases, you often hit
14:58 what's called the access control
15:00 mirroring problem. Um, or at least
15:02 that's what we call it. Um, and
15:04 basically what is that? Um, so it's
15:05 essentially stems from the fact that
15:07 with a rag system, you're actually
15:09 making a vectorzed copy of your data. So
15:12 if your rag system contains data from
15:14 SharePoint, from Salesforce, from Slack,
15:17 uh, you can't just invent new rules. You
15:19 should mirror the existing ones. So uh
15:21 you know for example if Jack can't see
15:24 the HR folder in SharePoint he probably
15:26 should not be allowed to query uh it in
15:29 your rag system. Um so how do you
15:33 actually achieve that uh in in the wild?
15:34 Um you effectively have three
15:36 architectural choices. So there's pre-
15:38 retrieval filtering. Um this is where
15:40 you actually bake the permissions um
15:43 from the source uh systems record into
15:45 the vector metadata. Um now this
15:46 approach actually allows uh for the
15:48 fastest query time. But if you go this
15:50 route, you often end up with um syncing
15:53 issues and you know re-indexing the data
15:56 which can of course have uh negative um
15:59 impact on cost and also latency in some
16:01 ways. Um the next approach is actually
16:03 to filter the retrieve content um once
16:05 it's been actually fetched. So in this
16:07 case you would uh fetch everything and
16:09 then check permissions using some sort
16:12 of external authorization service. So
16:14 this is very secure, easy to keep up to
16:16 date. Uh but this can actually add
16:17 latency and it can actually be very
16:22 challenging at scale. Um so uh typically
16:24 uh in full-blown production systems you
16:27 often see a hybrid approach. Um where
16:28 you basically do some sort of course
16:30 filter during the pre- retrieval. Um
16:32 this can be filtering on a metadata tag
16:34 such as like the department. Um and then
16:36 you might couple that with a fine grain
16:38 check via the post retrieval approach to
16:41 check for the user's actual permissions.
16:44 Um, so baking the permissions into the
16:46 vector database probably doesn't need
16:48 much explanation, but for the post
16:49 filter and retrieval option, you might
16:51 still be less left with a lot of
16:54 questions there. Um, oh, I actually
16:55 forgot I skipped that slide. We'll go
16:58 ahead and uh skip that for now. Um, and
17:00 just move on to evaluation. But um, I
17:01 had a slide in here about uh, basically
17:03 the different rag security models and
17:05 you know tool calling versus uh,
17:07 permissions graph ACL table and so
17:10 forth. Um but let's go ahead and jump on
17:12 to evaluations.
17:15 Um so uh this is a critical two-part
17:18 challenge. Um you can't just feel if
17:20 your rag system is accurate. Uh you need
17:22 a practical two-pronged approach to
17:25 prove it's working. So first uh is your
17:28 offline uh first is your offline
17:30 pre-eployment testing. So this is going
17:32 to be what's baked into your CI/CD and
17:34 it operates as your rag system safety
17:37 net. So in this set you in this step you
17:38 basically build what's called a golden
17:40 set. So this is maybe a 100 to a few
17:43 hundred Q&A pairs of your most important
17:44 questions. Uh these are the questions
17:46 that you can't get wrong. And then
17:48 before you deploy any change whether
17:50 that's a new prompt a new chunking
17:53 strategy etc you want to run your new
17:55 build against this golden set
17:59 programmatically scoring if those uh if
18:00 using those technical success metrics we
18:02 saw earlier. So this is your context
18:05 relevance and your answer faithfulness.
18:06 This is what lets you quantify
18:08 improvement. So now you can say things
18:11 like our new chunking strategy is 8%
18:13 more effective than the old one. Uh but
18:14 you can't think of everything. So that's
18:16 why you also have to couple that with an
18:20 online inproduction feedback um loop. So
18:21 this is your thumbs up thumbs down
18:24 button. And that is the start of your
18:25 most important workflow which is the
18:28 continuous improvement workflow.
18:30 Um and so this is what that looks like.
18:32 Um, and you can basically think of this
18:34 as like the thumbs down workflow. So
18:36 whenever your system produces a bad
18:38 response and a user flags as such with a
18:40 thumbs down, that's going to kick off a
18:41 task that goes into a review queue where
18:44 it gets triaged. So the goal of triage
18:47 is to is to determine if the process
18:49 failed at either the retrieval step or
18:51 the generation step. So from there it
18:53 can be addressed accordingly. If it was
18:54 a failed retrieval then you want to look
18:57 at basically the retrieval components uh
18:59 the ingestion, the chunking, the ranking
19:01 and so forth. Um, and if it failed at
19:02 generation, you want to actually look at
19:04 the the prompt. So, you might need to
19:06 fine-tune the prompt, adjust the logic,
19:08 and so forth. Um, and then, uh, this is
19:10 a critical step. Once the fix has been
19:12 made, that question and its
19:14 corresponding answer can now be added to
19:16 the golden set. So, you're effectively
19:18 strengthening your test coverage of the
19:21 system over time. So, finally, uh, that
19:23 takes us to the last pillar in our
19:26 discussion today around organizational
19:29 alignment. Um, and this is perhaps the
19:31 most painful lessons that we've seen a
19:33 lot of companies have to learn. Um, and
19:34 we've observed it time again in the
19:37 field. So, we've talked about ROI, we
19:39 talked about security and evaluation,
19:40 but how do teams actually try to build
19:43 these systems uh for their enterprises?
19:45 From what we've seen, this is uh what
19:47 usually happens. We call it the DIY's
19:49 rat's nest. Because rag is such an
19:51 emerging field, there aren't many good
19:54 all-in-one platforms out there that can
19:55 meet the stringent security requirements
19:58 of enterprises. So most organizations
20:00 have opted to build it their own little
20:02 pipelines in house. The problem is that
20:05 for enterprises when every one of their
20:07 individual teams is doing this the
20:10 result is organizational IT chaos. The
20:12 legal team writes a Python script for
20:14 PDFs. The sales team finds a hacky way
20:16 to scrape Salesforce. You end up with
20:18 custom code for chunking, custom scripts
20:22 for embedding, and zero standardization.
20:24 So, we've talked to organizations with
20:26 literally thousands of PC rag pilots
20:29 globally, all using different stacks.
20:31 So, we can't align on security because
20:33 every pipeline is different. You can't
20:34 align on quality because everyone is
20:37 using different chunking logic. Um, and
20:38 I can't tell you how many times we've
20:40 been called in just to entangle this
20:42 exact mess. So, how do you actually
20:45 prevent this outcome?
20:47 Well, the way that uh enterprises
20:49 typically invariably solve this is by
20:51 arriving at an organizational alignment
20:53 around an internal genai stack that can
20:56 be leveraged across all teams. So
20:58 instead of 50 custom DIY solutions
21:00 across 50 different departments, you
21:03 align the organization around one secure
21:06 stable ingestion uh and rag ETL layer.
21:08 Um and then that simplifies the
21:09 architecture ensuring that the stack is
21:11 kept secure and up to date and is
21:13 flexible enough to serve a wide range of
21:15 use cases. And of course uh that's where
21:17 we like to make a plug for ourselves. Uh
21:20 we offer just such a platform um on the
21:23 uh ingestion and rag ETL layer. So uh
21:25 what that gives you is instead of a
21:27 tangled nightmare of mess uh you end up
21:30 with an effortless unstructured data ETL
21:32 feeding as many highv value rack use
21:35 cases as your organization decides to
21:38 bite off. U so the benefit of here of
21:39 course choosing a platform like
21:40 unstructured is that there's no vendor
21:43 lockin for data sources um also for data
21:46 destinations the the available large uh
21:48 language models that you want to use in
21:49 your system the embedding models and so
21:51 on. Plus, it also features
21:53 enterprisegrade security and controls
21:55 and offers a number of flexible
21:57 deployment models. So, what this means
21:59 is that uh your developers can stop
22:00 writing maintenance scripts for their
22:03 DIY rag systems and start building
22:05 actual AI products. It also turns a
22:07 maintenance nightmare into a predictable
22:10 manageable utility. And in short, that
22:13 is how you scale rag from a single demo
22:15 to an enterprisewide capability that is
22:19 ready for the AI future of tomorrow.
22:21 Um, and so that's basically it for like
22:23 the main part of the sort of enterprise
22:26 uh ready uh overview. Um, but I do want
22:29 to touch also lightly as I said earlier
22:32 on rag for identic systems. Um, I'm only
22:33 going to lightly touch on it because
22:34 we're going to cover this topic in much
22:37 more detail in a future webinar. Um, but
22:40 the basic idea here um is that uh you
22:42 know you're introducing an intelligent
22:45 agent between the user and the rag
22:47 system. And so uh what this looks like
22:48 is that you're typically going to wrap
22:51 the rag system with an MCP server so
22:53 that can be called by an agent as one of
22:56 its tools. So the advantage here and I
22:57 sort of also touched on that earlier
22:59 with like the reflection and agentic rag
23:03 uh slide. Uh but the agent can actually
23:05 um review its response and then make a
23:07 subsequent follow-up queries to the rag
23:09 system. Uh and all that is basically to
23:11 ensure that it has all the information
23:13 it needs to effectively either answer
23:15 the question posed by the user or else
23:17 if it was solving a task it can use that
23:20 information and retrieve data uh to
23:23 effectively solve that task. Again we'll
23:25 cover this more in depth in a future uh
23:27 webinar but since agentic rag is
23:28 exploding in popularity I did just want
23:31 to touch on it here.
23:34 And with that um that concludes uh the
23:39 talk for today. Um if you enjoyed uh the
23:41 discussion uh make sure you check out uh
23:43 one of our past webinars. Um this is the
23:46 most recent um making your data work for
23:49 you rag strategies that scale as well as
23:51 um AJ and Kevin are back uh after the
23:54 break to discuss rag over evolving
23:56 enterprise knowledge. >> Awesome.
23:56 >> Awesome.
23:58 >> And with that we can take some questions.
23:59 questions.
24:02 >> Awesome. Um Daniel that was fantastic.
24:04 Thank you so so much. We're going to
24:06 have to make this a rapidfire round of
24:07 questions. So, we're going to go through
24:09 these pretty quickly. Um, thank you to
24:12 everyone who submitted questions. Um,
24:13 the first one is, do you have a
24:15 recommended end-to-end blueprint for
24:17 integrating unstructured with common
24:19 enterprise stacks, Snowflake, data
24:21 bricks or data lakes so that rag stays
24:24 in sync with upstream data? Um, so we
24:26 are the blueprint or that's like what we
24:28 are intending to be. So, we have staple
24:30 connectors to all of the sources that
24:32 you mentioned. Um, and we can run
24:33 incremental syncing. so that you're only
24:35 syncing updates and deletions to avoid
24:37 blowing up your ingress and egress
24:40 costs. Quickly on to you, Daniel. How do
24:42 you benchmark unstructured pipelines
24:44 impact on rag quality versus a naive
24:48 ingestion pipeline? Um, eg basic text
24:51 extraction and fixedsized chunking.
24:54 >> Yeah, I I sort of touched on this with
24:57 the eval section. Um, and also a little
25:01 bit on the ROI section. Um so
25:05 basically with like when you're doing uh
25:08 an evaluation of the ingestion portion
25:11 um in the context of reggg you can
25:12 actually miss a lot of the pictures
25:15 because like if you're your um golden
25:17 set is too small you know it could
25:19 actually be that it's able to answer all
25:21 the questions correctly even with a
25:24 basic extraction. Um but over time as
25:26 that long tail if you have that
25:28 continuous improvement and your golden
25:30 set begins to grow it'll start to reveal
25:33 cracks in the ingestion. Um and so uh
25:35 that will basically like identify those.
25:38 Um we also have if you want to just like
25:40 individually focus on the ing injection
25:42 metrics which we recommend. Uh we just
25:45 released um a new technique for doing so
25:47 called score. Um and definitely check it
25:49 out. It's all over our social um you
25:50 can't miss it. Um, but it's a very
25:53 powerful system for just evaluating uh
25:55 na naive ingestion.
25:58 >> Awesome. Um, we have some really
26:00 additional thoughtful slides. I don't
26:01 know if we're going to be able to get to
26:03 them because we're coming up on time,
26:05 but I just want to say thank you so much
26:07 for everyone who joined. Thank you for
26:10 the questions. We will respond in email
26:11 to the individuals who submitted
26:14 questions ahead of time. Um, and we will