klotz: llama-2* + llm*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. This tutorial demonstrates how to fine-tune the Llama-2 7B Chat model for Python code generation using QLoRA, gradient checkpointing, and SFTTrainer with the Alpaca-14k dataset.

  2. A comparison of frameworks, models, and costs for deploying Llama models locally and privately.

    • Four tools were analyzed: HuggingFace, vLLM, Ollama, and llama.cpp.
    • HuggingFace has a wide range of models but struggles with quantized models.
    • vLLM is experimental and lacks full support for quantized models.
    • Ollama is user-friendly but has some customization limitations.
    • llama.cpp is preferred for its performance and customization options.
    • The analysis focused on llama.cpp and Ollama, comparing speed and power consumption across different quantizations.
    2024-11-03 Tags: , , , , , by klotz
  3. "This is one of the best 13B models I've tested. (for programming, math, logic, etc) speechless-llama2-hermes-orca-platypus-wizardlm-13b"

    2023-10-02 Tags: , , , , , , by klotz
  4. 2023-09-01 Tags: , , , by klotz
  5. 2023-08-28 Tags: , , , by klotz
  6. 2023-08-19 Tags: , , , , by klotz
  7. 2023-08-03 Tags: , , by klotz
  8. 2023-07-23 Tags: , , , , , by klotz
  9. 2023-07-22 Tags: , , , , , , by klotz
  10. 2023-07-19 Tags: , , , , by klotz

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: llama-2 + llm

About - Propulsed by SemanticScuttle