Ensuring the quality and stability of Large Language Models (LLMs) is crucial. This article explores four open-source repositories - DeepEval, OpenAI SimpleEvals, OpenAI Evals, and RAGAs - each providing special tools and frameworks for assessing LLMs and RAG applications.