About Evaluating Large Outputs from Language Models: A Practical Guide course
This course covers the evaluation of Large Language Models (LLMs), starting with fundamental evaluation methods, exploring advanced techniques using Vertex AI tools like Automatic Metrics and AutoSxS, and looking ahead to the future of generative AI evaluation.
This course is ideal for AI product managers looking to optimize LLM applications, data scientists interested in advanced methods for evaluating AI models, AI ethicists and policymakers looking to drive responsible AI adoption, and academic researchers studying the impact of generative AI across a variety of domains. A basic understanding of artificial intelligence, machine learning concepts, and familiarity with natural language processing (NLP) are recommended. The course explores practical applications, integrating human insight with automated methods, and preparing for future trends in AI evaluation across a variety of media, including text, images, and audio. This comprehensive approach ensures that you are prepared to effectively evaluate LLMs to drive business strategy and innovation.