klotz: oobabooga*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. This pull request adds StreamingLLM support for llamacpp and llamacpp_HF models, aiming to improve performance and reliability. The changes allow indefinite chatting with the model without re-evaluating the prompt.
    2024-11-26 Tags: , , , , by klotz
  2. This PR implements the StreamingLLM technique for model loaders, focusing on handling context length and optimizing chat generation speed.
    2024-11-26 Tags: , , , , , by klotz
  3. This project provides Dockerised deployment of oobabooga's text-generation-webui with pre-built images for Nvidia GPU, AMD GPU, Intel Arc, and CPU-only inference. It supports various extensions and offers easy deployment and updates.
  4. A benchmark of large language models, sorted by size (on disk) for each score. Highlighted entries are on the Pareto frontier.
    2024-09-03 Tags: , , by klotz
  5. A web search extension for Oobabooga's text-generation-webui (now with nougat) that allows for web search integration with the AI.
  6. Steer LLM outputs towards a certain topic/subject and enhance response capabilities using activation engineering by adding steering vectors, now in oobabooga text generation webui!
  7. An extension for Oobabooga's Text-Generation Web UI that retrieves and adds web content to the context of prompts for more informative AI responses.
  8. An extension for oobabooga/text-generation-webui that enables the LLM to search the web using DuckDuckGo
  9. An extension that automatically unloads and reloads your model, freeing up VRAM for other programs.
  10. Training PRO extension for oobabooga WebUI - recent dev version. Key features and changes from the main Training in WebUI include:
    - Chunking: precise raw text slicer (PRTS) uses sentence splitting and making sure things are clean on all ends
    - Overlapping chunking: this special overlapping will make additional overlap block based on logical rules
    - Custom scheduler: FP_low_epoch_annealing keeps the LR constant for the first epoch and uses cosine for the rest
    - Target selector: Normal LORA is q, v, and it should be used with (q k v o) or (q k v)
    - DEMENTOR LEARNING (experimental) is an experimental chunking to train long-form text in low numbers of epochs
    2024-06-29 Tags: , , , , by klotz

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: oobabooga

About - Propulsed by SemanticScuttle