From the course: Python for AI Projects: From Data Exploration to Impact
Unlock the full course today
Join today to access over 24,900 courses taught by industry experts.
Model metrics - Python Tutorial
From the course: Python for AI Projects: From Data Exploration to Impact
Model metrics
- [Tutor] To evaluate any model, whether it's a traditional classifier or a powerful LLM, we need labeled reference data. That means we already have examples with a known good output, usually written or chosen by a human. In traditional NLP, labeled data might be, "The food was terrible and disgusting." Negative sentiment. In LLM evaluation, labeled reference data often includes entire text outputs. These could be summaries, translations, or answers paired with the input they respond to. The input might be a complete customer review talking about dishes, how close the restaurant was to key public transport, and how the waiter was not very attentive. The reference summary could be as simple as, convenient location, but service needs improvement. These references act as the gold standard, and we use them to compare the model's output using specific metrics. When evaluating models, it's important to remember…
Contents
-
-
-
-
(Locked)
Data exploration4m 48s
-
Training data pipeline4m 49s
-
(Locked)
Turning Raw Text into Business Insights with Python and NLP8m 34s
-
(Locked)
Model fitting7m 9s
-
(Locked)
Building Smarter Search: From Keywords to Semantic AI2m 33s
-
(Locked)
Model metrics10m 41s
-
(Locked)
From Search to Answers: Building AI Knowledge Solutions5m 21s
-
(Locked)
-
-
-