AI Safety & Evaluation
Project One Liner: Coming up with evaluation frameworks and understanding model learning representations w.r.t safety
Status: ongoing
Project Theme: safety
Project Areas:
Team: Gokul S Krishnan, Balaraman Ravindran
Collaborators: Precog Labs, IIIT Hyderabad
Short Description: As Large Language Models (LLMs) have been increasingly used in various sectors and applications, it is important to evaluate and improve safety aspects. Current evaluation methodologies for AI models/systems standardized mechanisms that can be applied across diverse stages of AI lifecycle. Several gaps hinder the capacity of both developers and deployers to assess the performance, fairness, and other safety aspects of AI solutions. For solving these issues, we work on designing evaluation frameworks for AI models and systems. We also focus on understanding model learning representations w.r.t AI safety aspects and we expect that this understanding will help us improve AI models in terms of trustworthiness.