klotz: llms*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. An article discussing the rise in interest among enterprises to build their own large language models (LLMs) using publicly available models as a starting point. The article discusses the challenges and benefits of this approach, as well as the need for enterprises to prepare for the integration of AI into their businesses.
    2024-06-12 Tags: , by klotz
  2. A discussion post on Reddit's LocalLLaMA subreddit about logging the output of running models and monitoring performance, specifically for debugging errors, warnings, and performance analysis. The post also mentions the need for flags to output logs as flat files, GPU metrics (GPU utilization, RAM usage, TensorCore usage, etc.) for troubleshooting and analytics.
  3. Learn how to repurpose an old PC to generate AI text and images, with a focus on using Ollama with Stable Diffusion. The guide covers installation, configuration, and setting up a web UI for a more organized user interface.
    2024-06-10 Tags: , , , , by klotz
  4. This article discusses the latest open LLM (large language model) releases, including Mixtral 8x22B, Meta AI's Llama 3, and Microsoft's Phi-3, and compares their performance on the MMLU benchmark. It also talks about Apple's OpenELM and its efficient language model family with an open-source training and inference framework. The article also explores the use of PPO and DPO algorithms for instruction finetuning and alignment in LLMs.
  5. efficient method for fine-tuning LLM using LoRA and QLoRA, making it possible to train them even on consumer hardware

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: llms

About - Propulsed by SemanticScuttle