Tags: hardware* + reddit*

0 bookmark(s) - Sort by: Date ↓ / Title /

  1. A user shares their experience running the GPT-OSS 120b model on Ollama with an i7 6700, 64GB DDR4 RAM, RTX 3090, and a 1TB SSD. They note slow initial token generation but acceptable performance overall, highlighting it's possible on a relatively modest setup. The discussion includes comparisons to other hardware configurations, optimization techniques (llama.cpp), and the model's quality.

    >I have a 3090 with 64gb ddr4 3200 RAM and am getting around 50 t/s prompt processing speed and 15 t/s generation speed using the following:
    >
    >`llama-server -m <path to gpt-oss-120b> --ctx-size 32768 --temp 1.0 --top-p 1.0 --jinja -ub 2048 -b 2048 -ngl 99 -fa 'on' --n-cpu-moe 24`
    > This about fills up my VRAM and RAM almost entirely. For more wiggle room for other applications use `--n-cpu-moe 26`.
  2. A user is asking about the compatibility of different modules (GPS, GPIO, I2C, 'Port C' UART) with the Cardputer, specifically whether modules using 'Port C' can be used.
  3. A user review of the Nerd Miner, a small device with minimal mining capabilities but allows users to participate in the BTC network.
  4. Easy Forth
  5. Works with GxEPD library using GxGDEW0213Z16 display type.
    Not GxEPD2 :-(

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: tagged with "hardware+reddit"

About - Propulsed by SemanticScuttle