The Lucid Vision Extension integrates advanced vision models into textgen-webui, enabling contextualized conversations about images and direct communication with vision models.
This pull request adds StreamingLLM support for llamacpp and llamacpp_HF models, aiming to improve performance and reliability. The changes allow indefinite chatting with the model without re-evaluating the prompt.
This PR implements the StreamingLLM technique for model loaders, focusing on handling context length and optimizing chat generation speed.
This project provides Dockerised deployment of oobabooga's text-generation-webui with pre-built images for Nvidia GPU, AMD GPU, Intel Arc, and CPU-only inference. It supports various extensions and offers easy deployment and updates.
A benchmark of large language models, sorted by size (on disk) for each score. Highlighted entries are on the Pareto frontier.
A web search extension for Oobabooga's text-generation-webui (now with nougat) that allows for web search integration with the AI.
Steer LLM outputs towards a certain topic/subject and enhance response capabilities using activation engineering by adding steering vectors, now in oobabooga text generation webui!
An extension for Oobabooga's Text-Generation Web UI that retrieves and adds web content to the context of prompts for more informative AI responses.
An extension for oobabooga/text-generation-webui that enables the LLM to search the web using DuckDuckGo
An extension that automatically unloads and reloads your model, freeing up VRAM for other programs.