Large language models show impressive capabilities, but ensuring their safe and reliable deployment remains challenging. This talk will cover evaluation techniques to assess and improve LLM reliability across key vectors like groundedness and faithfulness. It will also explore detecting vulnerabilities to attacks like prompt injection and PII leaks. Attendees will learn how to build custom evaluations tailored to their use cases.
Speaker: Shiv Sakuja is a former Google engineer, and co-founder of Athina AI, an LLM observability and evaluation platform that helps developers safeguard LLMs in production.
Not a Meetup member? Sign up to attend the next event:
https://voxel51.com/computer-vision-ai-meetups/
Recorded on May 2, 2024 at the AI, Machine Learning and Data Science Meetup.
Top comments (0)