This paper addresses the misalignment between traditional IR evaluation metrics and the requirements of modern Retrieval-Augmented Generation (RAG) systems. It proposes a novel annotation schema and the UDCG metric to better evaluate retrieval quality for LLM consumers.
Plural is bringing AI into the DevOps lifecycle with a new release that leverages a unified GitOps platform as a RAG engine. This provides AI-powered troubleshooting, natural language infrastructure querying, autonomous upgrade assistance, and agentic workflows for infrastructure modification, all with enterprise-grade guardrails.
A curated collection of Awesome LLM apps built with RAG, AI Agents, Multi-agent Teams, MCP, Voice Agents, and more. This repository features LLM apps that use models from OpenAI, Anthropic, Google, xAI and open-source models like Qwen or Llama.
The article explores whether combining a command-line agent (like Claude Code or Gemini CLI) with Unix-like file system tools and SemTools is sufficient for complex tasks, particularly document search. It details a benchmark testing the limits of coding agents with and without SemTools, focusing on search, cross-referencing, and temporal analysis. The conclusion is that CLI access is powerful and SemTools enhances agent capabilities for document search and RAG.
Nvidia’s NeMo Retriever models and RAG pipeline make quick work of ingesting PDFs and generating reports based on them. Chalk one up for the plan-reflect-refine architecture.
Sparse Priming Representations (SPR) is a research project focused on developing and sharing techniques for efficiently representing complex ideas, memories, or concepts using a minimal set of keywords, phrases, or statements, enabling language models or subject matter experts to quickly reconstruct the original idea with minimal context.
Scaling a simple RAG pipeline from simple notes to full books. This post elaborates on how to utilize larger files with your RAG pipeline by adding an extra step to the process — chunking.
An end-to-end raw text-to-graph pipelines. This blog explores the limitations of LangChain extraction when using smaller quantized models, and how BAML can improve extraction success rates.
This article details 10 open-source AI tools for developers, covering their benefits, features, and use cases. It emphasizes transparency, offline capabilities, and community support as key advantages of open-source AI.
| **Tool Name** | **Description** | **Key Features** | **What I Like About It** |
|---|---|---|---|
| **Talkd.ai** | Prototyping AI Agents | No-code, JSON/YAML config, API integration | Fast prototyping, no backend needed |
| **Marimo** | Python Notebooks for Apps | Reactive cells, version control, UI widgets | Stable, shareable, version-controlled apps |
| **Unsloth AI** | LLM Fine-Tuning | Memory-optimized training, supports Llama 3 | Accessible fine-tuning on modest hardware |
| **HackingBuddyGPT** | AI for Ethical Hacking | Offline operation, recon tools, payload generation | Offline security, privacy |
| **Giskard** | AI Testing & Debugging | Test case creation, continuous monitoring | Engineering discipline for AI quality |
| **OpenWebUI** | Self-Hosted ChatGPT UI | Local LLMs, plugin support, persistent memory | Privacy, local control |
| **Axolotl** | LLM Fine-Tuning | YAML config, supports QLORA/PEFT/LORA | Simplified fine-tuning, reproducibility |
| **FastRAG** | RAG Pipeline | Local operation, fast query times | Quick, lightweight RAG setup |
| **Nav2** | Robot Navigation Framework | Real-time obstacle detection, multi-robot coordination | Flexible, modern ROS 2 integration |
| **MindsDB** | Machine Learning in Database | SQL-based training/inference, supports various DBs | Easy integration with existing SQL workflows |
MarkItDown is an open-source Python utility that simplifies converting diverse file formats into Markdown, designed to prepare data for LLMs and RAG systems. It handles various file types, preserves document structure, and integrates with LLMs for tasks like image description.