You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: image_processing/README.md
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,10 +1,12 @@
1
-
# AI Search Indexing with Azure Document Intelligence
1
+
# Image Processing for RAG - AI Search Indexing with Azure Document Intelligence
2
2
3
3
This portion of the repo contains code for linking Azure Document Intelligence with AI Search to process complex documents with charts and figures, and uses multi-modal models (gpt-4o-mini) to interpret and understand these.
4
4
5
5
The implementation in Python, although it can easily be adapted for C# or another language. The code is designed to run in an Azure Function App inside the tenant.
6
6
7
-
**This approach makes use of Azure Document Intelligence v4.0 which is still in preview.**
7
+
> [!NOTE]
8
+
>
9
+
> See `GETTING_STARTED.md` for a step by step guide of how to use the accelerator.
8
10
9
11
## High Level Workflow
10
12
@@ -37,6 +39,7 @@ Once the Markdown is obtained, several steps are carried out:
37
39
3.**Cleaning of Markdown**. The final markdown content is cleaned of any characters or unsupported Markdown elements that we do not want in the chunk e.g. non-relevant figures.
38
40
39
41
> [!NOTE]
42
+
>
40
43
> For scalability, the above steps are performed across 5 differnet function app endpoints that are orchestrated by AI search.
41
44
42
45
## Sample Output
@@ -52,9 +55,6 @@ Using the [Phi-3 Technical Report: A Highly Capable Language Model Locally on Yo
52
55
53
56
The Figure 4 content has been interpreted and added into the extracted chunk to enhance the context for a RAG application. This is particularly powerful for applications where the documents are heavily imaged or chart based.
54
57
55
-
> [!NOTE]
56
-
> See `GETTING_STARTED.md` for a step by step guide of how to use the accelerator.
57
-
58
58
## Provided Notebooks \& Utilities
59
59
60
60
-`./function_app` provides a pre-built Python function app that communicates with Azure Document Intelligence, Azure OpenAI etc to perform the Markdown conversion, extraction of figures, figure understanding and corresponding cleaning of Markdown.
Copy file name to clipboardExpand all lines: text_2_sql/README.md
+5-4Lines changed: 5 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,7 +6,7 @@ The sample provided works with Azure SQL Server, although it has been easily ada
6
6
7
7
> [!NOTE]
8
8
>
9
-
> - Previous versions of this approach have now been moved to `previous_iterations/semantic_kernel`. These will not be updated.
9
+
> See `GETTING_STARTED.md` for a step by step guide of how to use the accelerator.
10
10
11
11
## Why Text2SQL instead of indexing the database contents?
12
12
@@ -31,6 +31,10 @@ To solve these issues, a Multi-Shot approach is developed. Below is the iteratio
31
31
32
32

33
33
34
+
> [!NOTE]
35
+
>
36
+
> - Previous versions of this approach have now been moved to `previous_iterations/semantic_kernel`. These will not be updated or maintained.
37
+
34
38
Our approach has evolved as the system has matured into an multi-agent approach that brings improved reasoning, speed and instruction following capabilities. With separation into agents, different agents can focus on one task only, and provide a better overall flow and response quality.
35
39
36
40
Using Auto-Function calling capabilities, the LLM is able to retrieve from the plugin the full schema information for the views / tables that it considers useful for answering the question. Once retrieved, the full SQL query can then be generated. The schemas for multiple views / tables can be retrieved to allow the LLM to perform joins and other complex queries.
@@ -39,9 +43,6 @@ To improve the scalability and accuracy in SQL Query generation, the entity rela
39
43
40
44
For the query cache enabled approach, AI Search is used as a vector based cache, but any other cache that supports vector queries could be used, such as Redis.
41
45
42
-
> [!NOTE]
43
-
> See `GETTING_STARTED.md` for a step by step guide of how to use the accelerator.
44
-
45
46
### Full Logical Flow for Agentic Vector Based Approach
46
47
47
48
The following diagram shows the logical flow within mutlti agent system. In an ideal scenario, the questions will follow the _Pre-Fetched Cache Results Path** which leads to the quickest answer generation. In cases where the question is not known, the group chat selector will fall back to the other agents accordingly and generate the SQL query using the LLMs. The cache is then updated with the newly generated query and schemas.
0 commit comments