A user shares their experience running the GPT-OSS 120b model on Ollama with an i7 6700, 64GB DDR4 RAM, RTX 3090, and a 1TB SSD. They note slow initial token generation but acceptable performance overall, highlighting it's possible on a relatively modest setup. The discussion includes comparisons to other hardware configurations, optimization techniques (llama.cpp), and the model's quality.
>I have a 3090 with 64gb ddr4 3200 RAM and am getting around 50 t/s prompt processing speed and 15 t/s generation speed using the following:
>
>`llama-server -m <path to gpt-oss-120b> --ctx-size 32768 --temp 1.0 --top-p 1.0 --jinja -ub 2048 -b 2048 -ngl 99 -fa 'on' --n-cpu-moe 24`
> This about fills up my VRAM and RAM almost entirely. For more wiggle room for other applications use `--n-cpu-moe 26`.
   
    
 
 
  
   
   A user is asking about the compatibility of different modules (GPS, GPIO, I2C, 'Port C' UART) with the Cardputer, specifically whether modules using 'Port C' can be used.
   
    
 
 
  
   
   A user review of the Nerd Miner, a small device with minimal mining capabilities but allows users to participate in the BTC network.
   
    
 
 
  
   
   Works with GxEPD library using GxGDEW0213Z16 display type.
Not GxEPD2 :-(