klotz: streamingllm*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. This pull request adds StreamingLLM support for llamacpp and llamacpp_HF models, aiming to improve performance and reliability. The changes allow indefinite chatting with the model without re-evaluating the prompt.
    2024-11-26 Tags: , , , , by klotz
  2. This PR implements the StreamingLLM technique for model loaders, focusing on handling context length and optimizing chat generation speed.
    2024-11-26 Tags: , , , , , by klotz

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: streamingllm

About - Propulsed by SemanticScuttle