A post-retrieval temporal layer designed to improve RAG systems by addressing time-blindness in vector searches. This library implements validity filtering, document kind classification, and exponential decay scoring to ensure retrieved information is fresh and accurate. It functions downstream of existing vector search systems without requiring re-indexing or new infrastructure.
gitcrawl is a local-first GitHub triage tool and a drop-in caching shim for the gh CLI. It mirrors repository issues and pull requests into a local SQLite database, enabling semantic clustering and full-text search while preventing API rate limit exhaustion. This setup allows maintainers and AI agents to perform heavy read operations against a local cache rather than live GitHub servers.
Main features:
Local SQLite storage for all issue, PR, and commit metadata.
A gh-compatible shim that handles most read-only calls locally.
Semantic clustering using OpenAI embeddings to group related reports.
An interactive terminal UI for cluster browsing.
JSON support for easy automation with AI agents.
OpenKB is an open-source command-line system designed to transform raw documents into a structured, interlinked wiki-style knowledge base using Large Language Models. Unlike traditional RAG systems that rediscover information with every query, OpenKB compiles knowledge once into a persistent format where summaries, concept pages, and cross-references are automatically maintained and updated.
Key features and capabilities include:
- Vectorless long document retrieval powered by PageIndex tree indexing.
- Native multi-modality for understanding figures, tables, and images.
- Broad format support including PDF, Word, Markdown, PowerPoint, HTML, and Excel.
- Automated wiki compilation that creates summaries and synthesizes concepts across documents.
- Interactive chat sessions with persisted history and Obsidian compatibility via wikilinks.
- Health check tools (linting) to identify contradictions, gaps, or stale content within the knowledge base.
Personal website of Alex L. Zhang, a PhD student at MIT CSAIL focusing on the efficiency and utilization of language models. His research spans ML systems, language model benchmarks, and specialized model development.
Key areas of work include:
- Recursive Language Models (RLMs) and Project Popcorn
- GPU programming competitions via KernelBot and GPU MODE
- Benchmarking capabilities through VideoGameBench and KernelBench
- Development of models like Neo-1 and KernelLLM-8B
This article explores how a team built an AI-powered emoji list generator during a Rubber Duck Thursday live stream. The tool runs in the terminal, takes a list of bullet points, and uses AI to intelligently replace them with relevant emojis before copying the result to the clipboard.
Key highlights include:
- Use of GitHub Copilot CLI for rapid development via plan and autopilot modes.
- Integration of @opentui/core for the terminal user interface.
- Leveraging the GitHub Copilot SDK to provide intelligent emoji selection.
- Implementation of a multi-model workflow using different LLMs for planning and execution.
A Python package designed to provide production-ready templates for Generative AI agents on Google Cloud. It allows developers to focus on agent logic by automating the surrounding infrastructure, including CI/CD pipelines, observability, security, and deployment via Cloud Run or Agent Engine.
Key features and offerings include:
- Pre-built agent templates such as ReAct, RAG (Retrieval-Augmented Generation), multi-agent systems, and real-time multimodal agents using Gemini.
- Automated CI/CD integration with Google Cloud Build and GitHub Actions.
- Data pipelines for RAG using Terraform, supporting Vertex AI Search and Vector Search.
- Support for various frameworks including Google's Agent Development Kit (ADK) and LangGraph.
- Integration with the Gemini CLI for architectural guidance directly in the terminal.
A single CLAUDE.md file to improve Claude Code behavior, derived from Andrej Karpathy's observations on LLM coding pitfalls.
graphify is an AI coding assistant skill that transforms codebases, documents, and images into a structured, queryable knowledge graph. By utilizing deterministic AST parsing via tree-sitter for code and multimodal LLM capabilities for unstructured data like PDFs and screenshots, it creates a comprehensive map of concepts and relationships. This allows developers to understand complex architectures faster and find the "why" behind design decisions. A key advantage is its massive reduction in token usage per query compared to reading raw files, making it highly efficient for large-scale projects. The tool supports 19 programming languages and integrates seamlessly with platforms like Claude Code and Codex, providing an interactive, persistent, and highly organized way to navigate any codebase or research corpus.
This article introduces Codebase Navigator, a tool designed to simplify the process of understanding large, unfamiliar GitHub repositories. By pasting a repository URL, users can interact with an AI assistant that provides a live dependency graph built from actual import statements, a code viewer, and a full file tree. Unlike standard AI assistants that often hallucinate file paths, this tool uses real data to visualize connections between files in real time. Built with a modern tech stack including Next.js, CopilotKit, and React Flow, the project can be run entirely for free using local LLMs via Ollama. The author provides a deep dive into the architecture, the technical implementation of the dependency resolution, and how the tool maintains state across multiple UI panels.
This project, `autoresearch-opencode`, is an autonomous experiment loop designed for use with OpenCode. It's a port of `pi-autoresearch`, but implemented as a pure skill, eliminating the need for an MCP server and relying solely on instructions the agent follows using its built-in tools. The skill allows users to automate optimization tasks, as demonstrated by the example of optimizing the BogoSort algorithm which achieved a 7,802x speedup by leveraging Python's `bisect` module for sorted-state detection.
The system maintains state using a JSONL file, enabling resume/pause functionality and detailed experiment tracking. It provides a dashboard for monitoring progress and ensures data integrity through atomic writes and validation checks.