A new study from the Capgemini Research Institute reveals that the ethical use of AI is becoming fundamental to gaining the people’s trust as consumers, employees and citizens are watching closely and are ready to reward or punish AI behavior accordingly. Researchers reveal:
Companies using AI in an ethical way will be rewarded
The good news is that companies who demonstrate trustworthiness in AI will be rewarded.
Among consumers surveyed, 62% said they would place higher trust in a company whose AI interactions they perceived as ethical, 61% said they would share positive experiences with friends and family, 59% said that they would have higher loyalty to the company, and 55% said that they would purchase more products and provide high ratings and positive feedback on social media.
By contrast, when consumers’ AI interactions result in ethical issues, it threatens both reputation and the bottom line: 41% said they would complain if an AI interaction resulted in ethical issues, 36% would demand an explanation and 34% would stop interacting with the company.
Ethical issues resulting from AI systems have already been observed
In a rush to automate, executives in nine out of 10 organizations believe that ethical issues have resulted from the use of AI technology. Over the last 2-3 years, respondents cited examples such as collection of personal patient data without consent in healthcare, and over-reliance on machine-led decisions without disclosure in banking and insurance.
Executives suggested the reasons for this include pressure to urgently implement AI, the failure to consider ethics when constructing AI systems, and a lack of resources dedicated to ethical AI systems.
Consumers, employees and citizens are worried about ethics & want some form of AI regulation
Almost half of respondents surveyed (47%) believe they have experienced at least two types of uses of AI that resulted in ethical issues in the last 2-3 years.
75% told researchers they expect more transparency when a service is powered by AI, and to know if AI is treating them fairly (73%). Over three quarters (76%) of consumers think there should be further regulation on how companies use AI.
Organizations are starting to realize the importance of ethics in AI
It seems that executives understand the public concern on ethics and will take this on board.
51% of executives agree that it is important to ensure that AI systems are ethical and transparent.
Organizations are also taking concrete actions when ethical issues are raised, with 41% of senior executives reporting they have already abandoned an AI system altogether when an ethical issue had been raised.
“Many organizations find themselves at a crossroads in their use of AI. Consumers, employees and citizens are increasingly open to interacting with the technology but are mindful of potential ethical implications. This research shows that organizations must create ethical systems and practices for the use of AI if they are to gain people’s trust. This is not just a compliance issue, but one that can create a significant benefit in terms of loyalty, endorsement and engagement,” said Anne-Laure Thieullent, AI and Analytics Group Offer Leader at Capgemini. “To achieve this, organizations need to focus on putting the right governance structures in place, they must not only define a code of conduct based on their own values, but also implement it as an ‘ethics-by-design’ approach, and, above all, focus on informing and empowering people in how they interact with AI solutions.”
Capgemini: organisations must ensure AI is ethical and fit for purpose
To address consumer concerns and unlock the trust dividend of ethical AI use, organizations must take a focused approach to making systems fit for purpose.
Capgemini recommends a three-pronged approach to build a strategy for ethics in AI that embraces all key stakeholders:
- For CxOs, business leaders and those with a remit for trust and ethics: Establish a strong foundation with a strategy and code of conduct for ethical AI; develop policies that define acceptable practices for the workforce and AI applications; create ethics governance structures and ensure accountability for AI systems; and build diverse teams to ensure sensitivity towards the full spectrum of ethical issues
- For the customer and employee-facing teams, such as HR, marketing, communications and customer service: Ensure ethical usage of AI application; educate and inform users to build trust in AI systems; empower users with more control and the ability to seek recourse; and proactively communicate on AI issues internally and externally to build trust
- For AI, data and IT leaders and their teams: Seek to make AI systems as transparent and understandable as possible to gain users’ trust; practice good data management and mitigate potential biases in data; continuously monitor for precision and accuracy; and use technology tools to build ethics in AI.
Luciano Floridi, director of Digital Ethics Labs, Oxford Internet Institute, said,
“A classic way of gaining trust, with AI interactions in particular, can be summarized in three words: transparency, accountability, and empowerment. That means transparency so that people can see what you are doing; accountability because you take responsibility for what you are doing; and empowerment because you put people in charge to tell you if something you did was not right or not good.”
A copy of the report can be downloaded here.