Skip to content Skip to sidebar Skip to footer

RAG-LLM Evaluation & Test Automation for Beginners

RAG-LLM Evaluation & Test Automation for Beginners

Understand, Evaluate & Test RAG - LLM's (AI based Systems) from Scratch using RAGAS-Python-Pytest Framework

LLMs are everywhere! Every business is building its own custom AI-based RAG-LLMs to improve customer service. But how are engineers testing them? Unlike traditional software testing, AI-based systems need a special methodology for evaluation.

This course starts from the ground up, explaining the architecture of how AI systems (LLMs) work behind the scenes. Then, it dives deep into LLM evaluation metrics.

This course shows you how to effectively use the RAGAS framework library to evaluate LLM metrics through scripted examples. This allows you to use Pytest assertions to check metric benchmark scores and design a robust LLM Test/evaluation automation framework.

What you'll learn

  • How Custom Large Language Models (LLM) are designed using Retrieval Augmented Generation (RAG) Architecture
  • Common Benchmarks/Metrics used in Evaluating RAG based LLM’s
  • Introduction to RAGAS Evaluation framework for evaluating/test LLM’s
  • Practical Scripts generation to automate and assert the Metrics Score of LLM’s.
  • Automate Scenarios such as Single turn interactions and Multi turn interactions with LLM’s using RAGAS Framework
  • Generate Test Data for evaluating the Metrics of LLM using RAGAS Framework.
  • Create RAGAS Pytest Evaluation Framework to assert the Metrics of RAG- (Custom) LLM’s

What will you learn from the course?
  • High level overview on Large Language Models (LLM)
  • Understand how Custom LLM’s are built using Retrieval Augmented Generation (RAG) Architecture
  • Common Benchmarks/Metrics used in Evaluating RAG based LLM’s
  • Introduction to RAGAS Evaluation framework for evaluating/test LLM’s
  • Practical Scripts generation to automate and assert the Metrics Score of LLM’s.
  • Automate Scenarios such as Single turn interactions and Multi turn interactions with LLM’s using RAGAS Framework
  • Generate Test Data for evaluating the Metrics of LLM using RAGAS Framework.
By end of the course, you will be able to create RAGAS Pytest Evaluation Framework to assert the Metrics of RAG- (Custom) LLM’s

Post a Comment for "RAG-LLM Evaluation & Test Automation for Beginners"