This article explores how to use LLMLingua, a tool developed by Microsoft, to compress prompts for large language models, reducing costs and improving efficiency without retraining models.
This article details the performance of Unsloth Dynamic GGUFs on the Aider Polyglot benchmark, showcasing how it can quantize LLMs like DeepSeek-V3.1 to as low as 1-bit while outperforming models like GPT-4.5 and Claude-4-Opus. It also covers benchmark setup, comparisons to other quantization methods, and chat template bug fixes.
This article provides a practical guide to JSON prompting for Large Language Models (LLMs), demonstrating how structuring prompts with JSON improves consistency, accuracy, and scalability. It includes Python coding examples comparing free-form and JSON prompts, and provides access to full code notebooks.
This tutorial explores implementing the LLM Arena-as-a-Judge approach to evaluate large language model outputs using head-to-head comparisons. It demonstrates using OpenAI’s GPT-4.1 and Gemini 2.5 Pro, judged by GPT-5, in a customer support scenario.
Python tutorial for reproducible labeling of cutting-edge topic models with GPT4-o-mini. The article details training a FASTopic model and labeling its results using GPT-4.0 mini, emphasizing reproducibility and control over the labeling process.
Meta releases Llama 3.1, its largest and best model yet, surpassing GPT-4o on several benchmarks. Zuckerberg believes this marks the 'Linux moment' in AI, opening the door for open-source models to flourish.
This tutorial provides a step-by-step guide on building an LLM router to balance the use of high-quality closed LLMs like GPT-4 and cost-effective open-source LLMs, achieving high response quality while minimizing costs. The approach includes preparing labeled data, finetuning a causal LLM classifier, and offline evaluation using the RouteLLM framework.
OpenAI introduces GPT-4, a new large language model that surpasses human performance on various tasks. Although not yet publicly available, the article provides insights into its capabilities and how it sets a new standard for AI.
Researchers from NYU Tandon School of Engineering investigated whether modern natural language processing systems could solve the daily Connections puzzles from The New York Times. The results showed that while all the AI systems could solve some of the puzzles, they struggled overall.
This tutorial introduces promptrefiner, a tool created by Amirarsalan Rajabi that uses the GPT-4 model to create perfect system prompts for local LLMs.