klotz: evaluation*

Bookmarks on this page are managed by an admin user.

0 bookmark(s) - Sort by: Date ↓ / Title / - Bookmarks from other users for this tag

  1. Langfuse is an open-source LLM engineering platform that offers tracing, prompt management, evaluation, datasets, metrics, and playground for debugging and improving LLM applications. It is backed by several renowned companies and has won multiple awards. Langfuse is built with security in mind, with SOC 2 Type II and ISO 27001 certifications and GDPR compliance.
  2. Discover how to build custom LLM evaluators for specific real-world needs
    2024-04-20 Tags: , by klotz
  3. Learn about the importance of evaluating classification models and how to use the confusion matrix and ROC curves to assess model performance. This post covers the basics of both methods, their components, calculations, and how to visualize the results using Python.
  4. A ready-to-run tutorial in Python and scikit-learn to evaluate a classification model compared to a baseline model
  5. Why evaluating LLM apps matters and how to get started
    2023-11-10 Tags: , , , by klotz
  6. 2023-10-13 Tags: , , by klotz

Top of the page

First / Previous / Next / Last / Page 1 of 0 SemanticScuttle - klotz.me: Tags: evaluation

About - Propulsed by SemanticScuttle