klotz: llama.cpp*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. Large Model Proxy is designed to make it easy to run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources.
    2024-07-22 Tags: , , , , by klotz
  2. This page provides information about LLooM, a tool that uses raw LLM logits to weave threads in a probabilistic way. It includes instructions on how to use LLooM with various environments, such as vLLM, llama.cpp, and OpenAI. The README also explains the parameters and configurations for LLooM.
  3. An explanation of the quant names used in the llama.cpp implementation, as well as information on the different types of quant schemes available.
    2024-06-23 Tags: , , by klotz
  4. Retrochat is chat application that supports Llama.cpp, Kobold.cpp, and Ollama. It highlights new features, commands for configuration, chat management, and models, and provides a download link for the release.
    2024-06-14 Tags: , , , , , , , by klotz
  5. Utilities for Llama.cpp, OpenAI, Anthropic, Mistral-rs. A collection of tools for interacting with various large language models. The code is written in Rust and includes functions for loading models, tokenization, prompting, text generation, and more.
  6. llm-tool provides a command-line utility for running large language models locally. It includes scripts for pulling models from the internet, starting them, and managing them using various commands such as 'run', 'ps', 'kill', 'rm', and 'pull'. Additionally, it offers a Python script named 'querylocal.py' for querying these models. The repository also come
  7. - create a custom base image for a Cloud Workstation environment using a Dockerfile
    . Uses:

    Quantized models from
  8. The "LLM" toolkit offers a versatile command-line utility and Python library that allows users to work efficiently with large language models. Users can execute prompts directly from their terminals, store the outcomes in SQLite databases, generate embeddings, and perform various other tasks. In this extensive tutorial, topics covered include setup, usage, OpenAI models, alternative models, embeddings, plugins, model aliases, Python APIs, prompt templates, logging, related tools, CLI references, contributing, and change logs.
    2024-02-08 Tags: , , , by klotz

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: llama.cpp

About - Propulsed by SemanticScuttle