Tags: llm* + rag*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. This article explains the differences between Model Context Protocol (MCP), Retrieval-Augmented Generation (RAG), and AI Agents, highlighting that they solve different problems at different layers of the AI stack. It also covers how ChatGPT routes prompts and handles modes, agent skills, architectural concepts for developers, and service deployment strategies.
  2. A collection of prompts designed to be used with AI coding assistants to build various use cases, ranging from personal CRM and knowledge bases to content pipelines and social media research.
  3. Here’s the simplest version — key sentence extraction:


    <pre>
    ```
    def extract_relevant_sentences(document, query, top_k=5):
    sentences = document.split('.')
    query_embedding = embed(query)
    scored = »
    for sentence in sentences:
    similarity = cosine_sim(query_embedding, embed(sentence))
    scored.append((sentence, similarity))
    scored.sort(key=lambda x: x 1 » , reverse=True)
    return '. '.join( s[0 » for s in scored :top_k » ])
    ```
    </pre>

    For each sentence, compute similarity to the query. Keep the top 5. Discard the rest
  4. The article discusses the evolution from RAG (Retrieval-Augmented Generation) to 'context engineering' in the field of AI, particularly with the rise of agents. It explores how companies like Contextual AI are building platforms to manage context for AI agents and highlights the shift from prompt engineering to managing the entire context state.
  5. Logs, metrics, and traces aren't enough. AI apps require visibility into prompts and completions to track everything from security risks to hallucinations.
  6. An AI-powered document search agent that explores files like a human would — scanning, reasoning, and following cross-references. Unlike traditional RAG systems that rely on pre-computed embeddings, this agent dynamically navigates documents to find answers.
  7. FailSafe is an open-source, modular framework designed to automate the verification of textual claims. It employs a multi-stage pipeline that integrates Large Language Models (LLMs) with retrieval-augmented generation (RAG) techniques.
  8. This article details how to build a 100% local MCP (Model Context Protocol) client using LlamaIndex, Ollama, and LightningAI. It provides a code walkthrough and explanation of the process, including setting up an SQLite MCP server and a locally served LLM.
  9. An extensible Model Context Protocol (MCP) server that provides intelligent semantic code search for AI assistants. Built with local AI models using Matryoshka Representation Learning (MRL) for flexible embedding dimensions.
  10. A curated repository of AI-powered applications and agentic systems showcasing practical use cases of Large Language Models (LLMs) from providers like Google, Anthropic, OpenAI, and self-hosted open-source models.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "llm+rag"

About - Propulsed by SemanticScuttle