From 9102d58e6a348a723ba9e28adf8f791aa2f69f55 Mon Sep 17 00:00:00 2001 From: kevleininger Date: Mon, 4 Dec 2023 01:01:46 -0500 Subject: [PATCH] Missing a key word in the initial description of RAG article. (#891) --- examples/evaluation/Evaluate_RAG_with_LlamaIndex.ipynb | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/evaluation/Evaluate_RAG_with_LlamaIndex.ipynb b/examples/evaluation/Evaluate_RAG_with_LlamaIndex.ipynb index 8c86c98..5d77753 100644 --- a/examples/evaluation/Evaluate_RAG_with_LlamaIndex.ipynb +++ b/examples/evaluation/Evaluate_RAG_with_LlamaIndex.ipynb @@ -31,7 +31,7 @@ "source": [ "**Retrieval Augmented Generation (RAG)**\n", "\n", - "LLMs are trained on vast datasets, but these will include your specific data. Retrieval-Augmented Generation (RAG) addresses this by dynamically incorporating your data during the generation process. This is done not by altering the training data of LLMs, but by allowing the model to access and utilize your data in real-time to provide more tailored and contextually relevant responses.\n", + "LLMs are trained on vast datasets, but these will not include your specific data. Retrieval-Augmented Generation (RAG) addresses this by dynamically incorporating your data during the generation process. This is done not by altering the training data of LLMs, but by allowing the model to access and utilize your data in real-time to provide more tailored and contextually relevant responses.\n", "\n", "In RAG, your data is loaded and and prepared for queries or “indexed”. User queries act on the index, which filters your data down to the most relevant context. This context and your query then go to the LLM along with a prompt, and the LLM provides a response.\n", "\n",