A list of Discord alternatives, categorized by self-hosted and centralized options, with notes on encryption and usability. Includes options for users seeking more privacy and control over their data.
HookCats is a self-hosted webhook routing server that acts as the central hub between your infrastructure and your team chat. It receives webhooks from any supported source, formats the messages nicely, and delivers them to your preferred chat platform.
An LLM-powered personal diary. Write your thoughts, and Reminor analyzes them, remembers them, and chats with you about them.
This article explores the architecture enabling AI chatbots to perform web searches, covering retrieval-augmented generation (RAG), vector databases, and the challenges of integrating search with LLMs.
An open-source, multi-model AI chat playground built with Next.js App Router. It allows users to switch between providers and models, compare outputs, and use web search and image attachments. It supports Gemini, OpenRouter, and Docker.
This repository contains the source code for the summarize-and-chat project. This project provides a unified document summarization and chat framework with LLMs, aiming to address the challenges of building a scalable solution for document summarization while facilitating natural language interactions through chat interfaces.
A no-install needed web-GUI for Ollama. It provides a web-based interface for interacting with Ollama, offering features like markdown rendering, keyboard shortcuts, a model manager, offline/PWA support, and an optional API for accessing more powerful models.
Pure C++ implementation of several models for real-time chatting on your computer (CPU), based on ggml.
This pull request adds StreamingLLM support for llamacpp and llamacpp_HF models, aiming to improve performance and reliability. The changes allow indefinite chatting with the model without re-evaluating the prompt.
This PR implements the StreamingLLM technique for model loaders, focusing on handling context length and optimizing chat generation speed.