TrueschoTruescho
All Courses
Building and Evaluating Data Agents
Coursera
Guided Project
Unknown

Building and Evaluating Data Agents

DeepLearning.AI

Learn to build and evaluate data agents that connect to multiple sources using multi-agent workflows and LLM-based evaluation.

Unknown1 weeksEnglish

About this Course

Learn how to build and evaluate a data agent in “Building and Evaluating Data Agents,” a course created in collaboration with Snowflake, and taught by Anupam Datta, AI Research Lead, and Josha Reini, Developer Advocate at Snowflake. You’ll design a data agent that connects to data sources (databases, files) and performs web searches to respond to users’ queries. The agent will consist of sub-agents, each specialized in connecting to a particular data source, and other sub-agents that summarize or visualize the results. To answer a particular query, the agent will use a planner that identifies which sub-agents to call and in what order. You’ll add observability to the agent’s workflow and evaluate the quality of its output. Using an LLM-as-a-judge approach, you’ll assess whether the final answer is relevant to the user’s query and grounded in the collected data. You’ll also evaluate the process by determining whether the agent’s goal, plan, and actions (GPA) are all aligned. Finally, you’ll apply inline evaluations to evaluate the agent’s performance during runtime. At every retrieval step, you’ll evaluate if the collected data is relevant to the user’s query. The agent will use this evaluation score to decide if it needs to adjust its plan. What you’ll do, in detail: Understand what data agents are and how they can be trustworthy when their goal, plan, and actions are properly aligned. Build a data agent that plans, performs web searches ,and visualizes or summarizes the results, using a multi-agent workflow implemented in LangGraph. Expand the agent’s capabilities by adding a Cortex sub-agent that retrieves information from structured and unstructured data stored in Snowflake. Add tracing to the agent’s workflow to log the steps it takes to answer a query. Evaluate the context relevance of the retrieved results, the groundedness of the final answer, and its relevance to the user’s query. Measure the alignment of the agent’s goal, plan, and actions (GPA) by computing metrics such as plan quality, plan adherence, logical consistency, and execution efficiency. Improve the agent’s performance by adding inline evaluations and updating the agent’s prompt. By the end, you’ll know how to build, trace, and evaluate a multi-agent workflow that plans tasks, pulls context from structured and unstructured data, performs web search, and summarizes or visualizes the final results

What You'll Learn

  • Design a data agent with multi-agent planning and specialized sub-agents
  • Trace and evaluate the agent’s answer quality and alignment
  • Improve performance through prompt updates and inline evaluations

Prerequisites

  • Basic familiarity with the software or workflow used in the project
  • Ability to follow step-by-step instructions in English

Instructors

J

Josh Reini

Developer Advocate

A

Anupam Datta

AI Research Lead

Topics

Software Development
Computer Science
Algorithms
Data Validation
Model Evaluation
Large Language Modeling
Generative AI Agents
LLM Application
Prompt Engineering
Data Visualization

Course Info

PlatformCoursera
LevelUnknown
PacingUnknown
PriceFree

Skills

تطوير البرمجيات
علوم الحاسوب
الخوارزميات
التحقق من البيانات
تقييم النماذج
نمذجة لغوية كبيرة
أنظمة وكلاء
تطبيقات نماذج لغوية كبيرة
Prompt Engineering
Data Visualization

Start Learning Now