This article discusses methods to measure and improve the accuracy of Large Language Model (LLM) applications, focusing on building an SQL Agent where precision is crucial. It covers setting up the environment, creating a prototype, evaluating accuracy, and using techniques like self-reflection and retrieval-augmented generation (RAG) to enhance performance.
Notebook detailing the use of RAG (Retrieval-Augmented Generation) via function calling.
This article discusses the development of multimodal Retrieval Augmented Generation (RAG) systems which allow for the processing of various file types using AI. The article provides a beginner-friendly guide with example Python code and explains the three levels of multimodal RAG systems.
Turn your Pandas data frame into a knowledge graph using LLMs. Learn how to build your own LLM graph-builder, implement LLMGraphTransformer by LangChain, and perform QA on your knowledge graph.
This article discusses how traditional machine learning methods, particularly outlier detection, can be used to improve the precision and efficiency of Retrieval-Augmented Generation (RAG) systems by filtering out irrelevant queries before document retrieval.
This article explains how to use Large Language Models (LLMs) to perform document chunking, dividing a document into blocks of text that each express a unified concept or 'idea', to create a knowledge base with independent elements.
The article explores how smaller language models like the Meta 1 Billion model can be used for efficient summarization and indexing of large documents, improving the performance and scalability of Retrieval-Augmented Generation (RAG) systems.
The article discusses the misconception that integrating complex graph databases (DBs), query languages (QLs), and analytics tools are necessary for Graph RAG. It emphasizes the distinction between traditional graph use cases and generative AI applications, and the need for a simpler tech stack.
An open-source project offering a functional RAG UI for document QA, suitable for both end-users and developers. It supports various LLM providers, is customizable, and offers multi-modal QA, citations, and complex reasoning methods.
Discussion in r/LocalLLaMA about finding a self-hosted, local RAG (Retrieval Augmented Generation) solution for large language models, allowing users to experiment with different prompts, models, and retrieval rankings. Various tools and resources are suggested, such as Open-WebUI, kotaemon, and tldw.