{"id":4135,"date":"2018-12-04T13:00:19","date_gmt":"2018-12-04T13:00:19","guid":{"rendered":"https:\/\/www.microsoft.com\/en-gb\/industry\/blog\/?p=4135"},"modified":"2019-03-21T17:50:02","modified_gmt":"2019-03-21T17:50:02","slug":"how-to-implement-an-ethical-framework-in-ai","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-gb\/industry\/blog\/cross-industry\/2018\/12\/04\/how-to-implement-an-ethical-framework-in-ai\/","title":{"rendered":"How to implement an ethical framework in AI"},"content":{"rendered":"
<\/p>\n
AI is already showing its potential for good causes<\/a>. Whether it’s predicting weather impact<\/a>, or optimising transport<\/a> in an intelligent way. Or projecting when maintenance of operational machines will be required<\/a>\u00a0and parts needed in manufacturing, revolutionising healthcare using genomics and microbiome R&D<\/a>. Or supporting\u00a0small businesses with smart and fast access to capital.<\/span><\/a>\u00a0It’s even helping to prevent blindness<\/a>, assisting deaf or hard of hearing students<\/a>, and aiding cancer research.<\/a> Incredibly, we’re also seeing AI being used to help in our mission to save endangered species<\/a> and understand climate change.\u00a0<\/span><\/span><\/p>\n Yet it’s not without challenges. To ensure AI can only do good, we must first understand the risks and a host of ethical issues related to creating thinking machines and relying on them to make important decisions that affect humans and society in general. We must ask ourselves not what AI can<\/em> do, but what it should<\/em> do.<\/p>\n “‘Should’<\/em> companies have been shown to outperform ‘can’<\/em> companies by 9%”<\/p>\n – Maximising the AI opportunity, Microsoft UK.<\/p><\/blockquote>\n <\/p>\n The ability to dissect a conclusion that AI takes, together with the predictability of such an intelligent algorithm and our trust in building and operationalising it, along with a robust legal framework addressed adequately by legislation, is key to future proofing the success of AI.<\/p>\n Satya Nadella<\/span><\/span><\/a> rightly says, “Unfortunately the corpus of human data is full of biases”. At Build 2018 he also mentioned that Microsoft’s internal AI ethics team\u2019s job is to ensure that the company\u2019s foray into cutting-edge techniques, like deep learning, don\u2019t unintentionally perpetuate societal biases in their products, among other tasks. <\/span><\/span><\/p>\n Some of the fundamentals of computer science have not been changed by AI, such as \u201c<\/span><\/span>garbage<\/span><\/span> in, garbage out\u201d. However, machine learning and deep learning – that power many AI systems – learn from large data sets.<\/span><\/span> In most situations the more data, the better the predictions and quality of the results.<\/span><\/span>\u00a0If the input data that\u2019s used in training the model has some bias, it’s likely that the outcome will also be biased.<\/span><\/span><\/p>\n Let\u2019s look at the key elements required for an ethical framework in AI.<\/p>\n You need to ensure AI is built and executed with a fairness lens by considering the following:<\/p>\n Reliability of predictions and the safety of AI models are key for us to have trust in AI. Follow these steps to build these in from the very start:<\/p>\n As AI becomes embedded in almost everything we do, it’s fundamentally important for developers and vendors to consider privacy and security requirements as the number one priority:<\/p>\n The problem with bias in data is that it considers too much one of thing or type as the ground truth, and vice versa. Being inclusive in your design and ensuring your data is inclusive of all attributes that will use or benefit from your AI solution is key to the success of AI. Here are some simple examples you must consider:<\/p>\n Transparency is key in building trust. To install transparency when building or operating AI systems, you should consider the following:<\/p>\n “The big challenge for us, and anyone looking to use AI, is building trust. People who give their data to the kind of infrastructure that we are developing are inherently cautious about how that data is going to be used and who will have access to it” – Richard Tiffin, Chief Scientific Officer, Agrimetrics<\/p><\/blockquote>\n Ultimately humans are building AI systems, and we ourselves should have accountability for what we are building. Here are some examples of what must do:<\/p>\n Quoting Satya Nadella again, \u201c<\/span><\/span>AI isn\u2019t just another piece of technology. It could be one of the world\u2019s most fundamental pieces of technology <\/span><\/span>the human race<\/span><\/span> has ever created<\/span><\/span>\u201d<\/span><\/span>. We are at a singularity, and it is <\/span><\/span>up to<\/span><\/span> us\u00a0<\/span><\/span>to build the AI systems that <\/span><\/span>augment<\/span><\/span> what we do, and in the <\/span><\/span>manner in which<\/span><\/span> we do, <\/span><\/span>in a positive way<\/span><\/span>.<\/span><\/span><\/p>\n To learn more, read <\/span><\/span>Future Computed,<\/span><\/span><\/a> our newly launched report<\/a> to help you maximise the AI opportunity.<\/p>\n The AI opportunity in healthcare<\/a><\/p>\n Join our AI for Good accelerator programme for startups<\/a><\/p>\nEthics is key to the future success of AI<\/h2>\n
So, how do we build an ethical framework for AI?<\/h2>\n
Fairness<\/h3>\n
\n
Reliability & safety<\/h3>\n
\n
Privacy & security<\/h3>\n
\n
Inclusiveness<\/h3>\n
\n
Transparency<\/h3>\n
\n
Accountability<\/h3>\n
\n
The opportunity<\/h2>\n
Find out more<\/h2>\n