{"id":736096,"date":"2021-03-11T13:36:26","date_gmt":"2021-03-11T21:36:26","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-research-item&p=736096"},"modified":"2021-03-25T13:45:28","modified_gmt":"2021-03-25T20:45:28","slug":"fairness-related-harms-in-ai-systems-examples-assessment-and-mitigation","status":"publish","type":"msr-video","link":"https:\/\/www.microsoft.com\/en-us\/research\/video\/fairness-related-harms-in-ai-systems-examples-assessment-and-mitigation\/","title":{"rendered":"Fairness-related harms in AI systems: Examples, assessment, and mitigation"},"content":{"rendered":"
AI has transformed modern life via previously unthinkable feats, from machines that can master the ancient board game Go and self-driving cars to developments we experience more routinely, such as virtual agents and personalized product recommendations. Simultaneously, these new opportunities have raised new challenges\u2014most notably, challenges that have highlighted the potential for AI systems to cause fairness-related harms. Indeed, the fairness of AI systems is one of the key concerns facing society as AI continues to influence our lives in new ways.<\/p>\n
In this webinar, Microsoft researchers Hanna Wallach and Miroslav Dud\u00edk will guide you through how AI systems can lead to a variety of fairness-related harms. They will then dive deeper into assessing and mitigating two specific types: allocation harms and quality-of-service harms. Allocation harms occur when AI systems allocate resources or opportunities in ways that can have significant negative impacts on people\u2019s lives, often in high-stakes domains like education, employment, finance, and healthcare. Quality-of-service harms occur when AI systems, such as speech recognition or face detection systems, fail to provide a similar quality of service to different groups of people.<\/p>\n
Together, you\u2019ll explore:<\/p>\n
Resource list:<\/strong><\/p>\n *This on-demand webinar features a previously recorded Q&A session and open captioning.<\/p>\n\n