Tags: llama-cpp*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. An exploration of the new Qwen3.6-27B open weight model, which claims flagship-level agentic coding performance that surpasses previous larger MoE models while being significantly smaller in size. The author tests a quantized version using llama-server and demonstrates its impressive ability to generate complex SVG graphics locally.
    Key points:
    - Qwen3.6-27B outperforms the older Qwen3.5-397B-A17B on coding benchmarks.
    - Dramatic reduction in model size from 807GB to approximately 55.6GB for the base version.
    - Successful local execution using a 16.8GB quantized GGUF version via llama.cpp.
    - High-quality SVG generation capabilities for complex prompts like a pelican riding a bicycle.
  2. LlamaBarn is a macOS menu bar app for running local LLMs. It provides a simple way to install and run models locally, connecting to apps via an OpenAI-compatible API.
  3. The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). It provides a simple yet robust interface using llama-cpp-python, allowing users to chat with LLM models, execute structured function calls and get structured output.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "llama-cpp"

About - Propulsed by SemanticScuttle