klotz: gpt-4*

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. This article details the performance of Unsloth Dynamic GGUFs on the Aider Polyglot benchmark, showcasing how it can quantize LLMs like DeepSeek-V3.1 to as low as 1-bit while outperforming models like GPT-4.5 and Claude-4-Opus. It also covers benchmark setup, comparisons to other quantization methods, and chat template bug fixes.
  2. This article provides a practical guide to JSON prompting for Large Language Models (LLMs), demonstrating how structuring prompts with JSON improves consistency, accuracy, and scalability. It includes Python coding examples comparing free-form and JSON prompts, and provides access to full code notebooks.
    2025-08-27 Tags: , , , , by klotz
  3. This tutorial explores implementing the LLM Arena-as-a-Judge approach to evaluate large language model outputs using head-to-head comparisons. It demonstrates using OpenAI’s GPT-4.1 and Gemini 2.5 Pro, judged by GPT-5, in a customer support scenario.
  4. Python tutorial for reproducible labeling of cutting-edge topic models with GPT4-o-mini. The article details training a FASTopic model and labeling its results using GPT-4.0 mini, emphasizing reproducibility and control over the labeling process.
  5. Meta releases Llama 3.1, its largest and best model yet, surpassing GPT-4o on several benchmarks. Zuckerberg believes this marks the 'Linux moment' in AI, opening the door for open-source models to flourish.
  6. This tutorial provides a step-by-step guide on building an LLM router to balance the use of high-quality closed LLMs like GPT-4 and cost-effective open-source LLMs, achieving high response quality while minimizing costs. The approach includes preparing labeled data, finetuning a causal LLM classifier, and offline evaluation using the RouteLLM framework.
    2024-07-04 Tags: , , , , , by klotz
  7. OpenAI introduces GPT-4, a new large language model that surpasses human performance on various tasks. Although not yet publicly available, the article provides insights into its capabilities and how it sets a new standard for AI.
    2024-05-15 Tags: , , , by klotz
  8. Researchers from NYU Tandon School of Engineering investigated whether modern natural language processing systems could solve the daily Connections puzzles from The New York Times. The results showed that while all the AI systems could solve some of the puzzles, they struggled overall.
  9. This tutorial introduces promptrefiner, a tool created by Amirarsalan Rajabi that uses the GPT-4 model to create perfect system prompts for local LLMs.
  10. - Demonstrates how to improve two pretrained models' proficiency in the Dafny verified programming language.
    - Uses 178 programming problems from the MBPP dataset for prompting GPT-4 and PaLM-2 to generate methods in Dafny.
    - Three types of prompts were used: a direct contextless prompt, one that includes a signature of the method and test cases, and a third one that decomposes the problem into steps and includes dynamically chosen similar examples.
    - GPT-4 was able to generate verified (and human-evaluated) Dafny methods in 58% of the cases with the third prompt.
    - Contributes a collection of 153 MBPP problems implemented and formally verified in Dafny, 50 written by authors and 103 synthesized by GPT-4.

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: gpt-4

About - Propulsed by SemanticScuttle