klotz: self-hosted* + llm*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. Msty offers a simple and powerful interface to work with local and online AI models without the hassle of setup or configuration, ensuring privacy and reliability with offline capabilities.
    2025-01-08 Tags: , , by klotz
  2. Discover how to run AI models locally with ease using tools like Msty, which simplifies the process of setting up, running, and managing local AI models on various operating systems.
    2025-01-08 Tags: , , , , by klotz
  3. Persys is a locally-run device designed to function as a second brain. The repository includes the backend server (Linux only) and the Electron-based desktop application for accessing the server.
  4. A comparison of frameworks, models, and costs for deploying Llama models locally and privately.

    - Four tools were analyzed: HuggingFace, vLLM, Ollama, and llama.cpp.
    - HuggingFace has a wide range of models but struggles with quantized models.
    - vLLM is experimental and lacks full support for quantized models.
    - Ollama is user-friendly but has some customization limitations.
    - llama.cpp is preferred for its performance and customization options.
    - The analysis focused on llama.cpp and Ollama, comparing speed and power consumption across different quantizations.
    2024-11-03 Tags: , , , , , by klotz
  5. The article discusses the challenges and strategies for load testing and infrastructure decisions when self-hosting Large Language Models (LLMs).
  6. Discussion in r/LocalLLaMA about finding a self-hosted, local RAG (Retrieval Augmented Generation) solution for large language models, allowing users to experiment with different prompts, models, and retrieval rankings. Various tools and resources are suggested, such as Open-WebUI, kotaemon, and tldw.
    2024-10-13 Tags: , , , , by klotz
  7. Tabby is an open-source, self-hosted AI coding assistant that is easy to configure and deploy with a simple TOML config. It is powered by Rust for speed and safety.
  8. A step-by-step guide to run Llama3 locally with Python. Discusses the benefits of running local LLMs, including data privacy, cost-effectiveness, customization, offline functionality, and unrestricted use.
    2024-07-12 Tags: , , , by klotz
  9. llm-tool provides a command-line utility for running large language models locally. It includes scripts for pulling models from the internet, starting them, and managing them using various commands such as 'run', 'ps', 'kill', 'rm', and 'pull'. Additionally, it offers a Python script named 'querylocal.py' for querying these models. The repository also come

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: self-hosted + llm

About - Propulsed by SemanticScuttle