Tags: gguf*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. Mistral.rs is a fast LLM inference platform supporting inference on a variety of devices, quantization, and easy-to-use application with an Open-AI API compatible HTTP server and Python bindings. It supports the latest Llama and Phi models, as well as X-LoRA and LoRA support. The project aims to provide the fastest LLM inference platform possible.
    2024-04-29 Tags: , , , , by klotz
  2. - create a custom base image for a Cloud Workstation environment using a Dockerfile
    . Uses:

    Quantized models from
  3. A deep dive into model quantization with GGUF and llama.cpp and model evaluation with LlamaIndex
  4. Exploring Pre-Quantized Large Language Models
    2023-11-15 Tags: , , by klotz

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "gguf"

About - Propulsed by SemanticScuttle