This document provides guidelines for maintaining high-quality Python code, specifically for AI coding agents. It covers principles, tools, style, documentation, testing, and security best practices.
A guide to common pitfalls and best practices when starting with Playwright and Python, covering topics like browser context, waiting strategies, and handling different environments.
This article details how to use Playwright MCP and GitHub Copilot to reproduce and debug web app issues. It covers setup, a sample scenario, and the benefits of this debugging approach.
Get LLMs to do things from Emacs with gptel. The project seeks testers to help evolve tool use within the gptel interface for Emacs.
Ensuring the quality and stability of Large Language Models (LLMs) is crucial. This article explores four open-source repositories - DeepEval, OpenAI SimpleEvals, OpenAI Evals, and RAGAs - each providing special tools and frameworks for assessing LLMs and RAG applications.