klotz: llm* + openai*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. LLM Council works together to answer your hardest questions. A local web app that uses OpenRouter to send queries to multiple LLMs, have them review/rank each other's work, and finally a Chairman LLM produces the final response.
  2. OpenAI releases GPT-5.1 Instant and GPT-5.1 Thinking, upgrades to the GPT-5 series focusing on improved intelligence, conversational style, and customization options for ChatGPT. Includes new tone presets and the ability to fine-tune characteristics.
  3. OpenAI releases gpt-oss-safeguard, an open-source AI model for content moderation that allows developers to define their own safety policies instead of relying on pre-trained models. It operates by reasoning about content based on custom policies, offering a more flexible and nuanced approach to moderation.
  4. An in-depth look at the architecture of OpenAI's GPT-OSS models, detailing tokenization, embeddings, transformer blocks, Mixture of Experts, attention mechanisms (GQA and RoPE), and quantization techniques.
  5. This paper explains that hallucinations in large language models (LLMs) aren’t due to flawed data, but to the way these models are trained and evaluated. LLMs are incentivized to guess rather than admit uncertainty, leading to errors that are statistically predictable. The authors frame this as a binary classification problem – correctly identifying valid outputs – and demonstrate a link between misclassification rate and hallucination rate. They argue that fixing this requires a shift in evaluation metrics, moving away from rewarding overconfidence and towards accepting uncertainty, to build more trustworthy models.
  6. This article provides a practical guide to JSON prompting for Large Language Models (LLMs), demonstrating how structuring prompts with JSON improves consistency, accuracy, and scalability. It includes Python coding examples comparing free-form and JSON prompts, and provides access to full code notebooks.
    2025-08-27 Tags: , , , , by klotz
  7. This tutorial explores implementing the LLM Arena-as-a-Judge approach to evaluate large language model outputs using head-to-head comparisons. It demonstrates using OpenAI’s GPT-4.1 and Gemini 2.5 Pro, judged by GPT-5, in a customer support scenario.
  8. OpenAI's release of GPT-OSS marks their first major open source LLM since GPT-2, featuring improvements in reasoning, tool usage, and problem-solving capabilities. The article explores its architecture, message formatting, reasoning modes, and tokenizer details.
  9. A user demonstrates how to run a 120B model efficiently on hardware with only 8GB VRAM by offloading MOE layers to CPU and keeping only attention layers on GPU, achieving high performance with minimal VRAM usage.
  10. A 120 billion parameter OpenAI model can now run on consumer hardware thanks to the Mixture of Experts (MoE) technique, which significantly reduces memory requirements and allows processing on CPUs while offloading key parts to modest GPUs.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: llm + openai

About - Propulsed by SemanticScuttle