generated from cotes2020/chirpy-starter
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Santiago Olivar
committed
Feb 16, 2024
1 parent
5d8f157
commit fccd89c
Showing
2 changed files
with
47 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,47 @@ | ||
--- | ||
title: Intro to RAGs and LlamaIndex | ||
date: 2024-02-15 12:00:00 -0 | ||
categories: [LlamaIndex, Learning] | ||
tags: [RAGs, LlamaIndex] | ||
--- | ||
|
||
## What is LlamaIndex? | ||
|
||
LlamaIndex provides abstractions to make the ingestion, structuring, and access of data more straightforward, enabling seamless integration into Large Language Models (LLMs) for more accurate text generation. | ||
|
||
While LLMs are trained on extensive datasets, they lack training on private or domain-specific data. Fine-tuning a LLM with your data is costly, making it challenging to keep the model up-to-date with the latest information. Additionally, it often lacks observability.[^footnote] | ||
|
||
## What is RAG? | ||
|
||
Retrieval-Augmented Generation (RAG) provides an alternative1. This technique focuses on context augmentation to achieve more precise text generation tailored to your specific data. | ||
|
||
Key steps in RAG implementation: | ||
|
||
- **Get Information**: Nodes serve as the atomic unit of data, and connectors assist in ingesting these nodes. | ||
|
||
- **Index and Add to Query as Context**: This involves generating vector embeddings and incorporating them into your query as context. | ||
|
||
- **Pass Improved Prompt to LLM**: Further details will be covered in upcoming posts, addressing querying, retrievers, routers, postprocessors, and synthesizers. | ||
|
||
The beauty of RAG lies in its absence of training, ensuring it's always up-to-date by adding your data to the existing data accessible to LLMs. Another vital aspect is the ability to store your indexes, mainly accomplished through specialized databases known as vector stores. Finally a RAG framework must effectively perform proper evaluation on the accuracy of the responses.[^fn-nth-2] | ||
|
||
![RAG](/images/RAG/basic_rag.png)_Source: LLamaIndex Documentation. RAG Diagram_ | ||
|
||
## LlamaIndex Tools | ||
|
||
- **Data Connectors**: Facilitate data ingestion. | ||
|
||
- **Data Indexes**: Provide representations of the data usable in LLMs. | ||
|
||
- **Engines**: Serve as end-to-end pipelines for interaction with LLMs. | ||
|
||
- **Data Agents**: LLM-based decision-makers offering flexibility to tackle complex problems. | ||
|
||
- **Application Integrations**: Enable seamless integration of LlamaIndex into other frameworks. | ||
|
||
|
||
## References | ||
|
||
[^footnote]: https://docs.llamaindex.ai/en/stable/getting_started/concepts.html# | ||
|
||
[^fn-nth-2]: https://docs.llamaindex.ai/en/stable/getting_started/concepts.html |
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.