Tags: llama.cpp* + llm*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. Unify your existing devices into one powerful GPU: iPhone, iPad, Android, Mac, NVIDIA, Raspberry Pi, pretty much any device!

  2. A USB stick equipped with a Raspberry Pi Zero W runs a large language model using llama.cpp. The project involves porting the model to an ARMv6 architecture and setting up the device as a composite that presents a filesystem to the host, allowing users to interact with the LLM by creating text files that are automatically filled with generated content.

  3. This PR implements the StreamingLLM technique for model loaders, focusing on handling context length and optimizing chat generation speed.

    2024-11-26 Tags: , , , , , by klotz
  4. A collection of lightweight AI-powered tools built with LLaMA.cpp and small language models.

  5. A guide on how to download, convert, quantize, and use Llama 3.1 8B model with llama.cpp on a Mac.

    2024-09-28 Tags: , , , by klotz
  6. A step-by-step guide on building llamafiles from Llama 3.2 GGUFs, including scripting and Dockerization.

  7. Tutorial on enforcing JSON output with Llama.cpp or the Gemini’s API for structured data generation from LLMs.

    2024-08-25 Tags: , , , , , by klotz
  8. Large Model Proxy is designed to make it easy to run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources.

    2024-07-22 Tags: , , , , by klotz
  9. This page provides information about LLooM, a tool that uses raw LLM logits to weave threads in a probabilistic way. It includes instructions on how to use LLooM with various environments, such as vLLM, llama.cpp, and OpenAI. The README also explains the parameters and configurations for LLooM.

  10. An explanation of the quant names used in the llama.cpp implementation, as well as information on the different types of quant schemes available.

    2024-06-23 Tags: , , by klotz

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "llama.cpp+llm"

About - Propulsed by SemanticScuttle