Sort by

Newest

Oldest

Popular

Open-source LLM tracing, evals and prompt optimization with Evidently
00:15:55
8. Tutorial: Adversarial testing for LLM applications
00:13:24
7. Tutorial: Building and evaluating an AI agent
00:17:35
6.2. Tutorial: Building and evaluating a RAG system
00:16:13
6.1 How to evaluate a RAG system: methods and metrics
00:07:08
5. Tutorial: Evaluating LLMs on content generation tasks. Tracing and experiments.
00:26:01
4. Tutorial: Evaluating LLMs on classification tasks
00:19:01
3. Tutorial: How to create an LLM judge and align with human labels
00:24:24
2.3. Tutorial on LLM evaluation methods: Reference-free evals.
00:14:07
2.2. Tutorial on LLM evaluation methods: Reference-based evals.
00:10:26
2.1. Tutorial on LLM evaluation methods. Overview and Basic API.
00:10:13
1. Introduction to LLM evaluations in 10 key ideas
00:11:43
LLM evaluation for builders - Course announcement
00:01:43
How to run LLM evals with no code | PRACTICE
00:12:27
How to continuously improve LLM products?
00:02:34
A business case for LLM evaluations
00:04:06
AI agent and RAG evaluation
00:06:05
LLM observability in production: tracing and online evals
00:05:09
LLM safety and red-teaming
00:05:11
LLM-as-a-judge: evaluating LLMs with LLMs
00:05:41
LLM evaluation methods and metrics
00:05:10
LLM evaluation datasets: test cases and synthetic data
00:06:06
How to evaluate an LLM application
00:06:00
LLM evaluation benchmarks
00:03:07
What is an LLM-powered product?
00:03:08
Welcome to the LLM evaluation course
00:01:32
Open-source LLM Evaluation with Evidently - Intro
00:02:30
LLM Evaluation Tutorial with Evidently
00:35:45
6.5. Connecting the dots: full-stack ML observability.
00:05:21
6.4. ML monitoring with Evidently and Grafana. [OPTIONAL CODE PRACTICE]
00:25:03