AI Frontiers: Measuring and mitigating harms with Hanna Wallach
Powerful large-scale AI models like GPT-4 are showing dramatic improvements in reasoning, problem-solving, and language capabilities. This marks a phase change for artificial intelligence—and a signal of accelerating progress to come.
In this Microsoft Research Podcast series, AI scientist and engineer Ashley Llorens hosts conversations with his collaborators and colleagues about what these models—and the models that will come next—mean for our approach to creating, understanding, and deploying AI, its applications in areas such as healthcare and education, and its potential to benefit humanity.
This episode features Partner Research Manager Hanna Wallach, whose research into fairness, accountability, transparency, and ethics in AI and machine learning has helped inform the use of AI in Microsoft products and services for years. Wallach describes how she and a team of applied scientists expanded their tools for measuring fairness-related harms in AI systems to address harmful content more broadly during their involvement in the deployment of Bing Chat; her interest in filtering, a technique for mitigating harms that she describes as widely used but not often talked about; and the cross-company collaboration that brings policy, engineering, and research together to evolve and execute the Microsoft approach to developing and deploying AI responsibly.
Learn more:
Published on:
Learn more