AI bias could harm society, so we need to tackle it now
AI hasn’t quite come of age, but it’s now at a point where most people understand what the benefits are.However, for all of those potential benefits, companies looking to take advantage of AI must still make ethical considerations and the avoidance of bias their top priority. That’s according to a panel session held this week at Salesforce’s Dreamforce event in San Francisco.
“Accuracy levels are so high now that the kind of things you can do in one year weren’t possible years ago with hundreds of people,” said Richard Socher, chief scientist at Salesforce. “Now that this stuff is working, we really need to think about the ethical implications.”
Kathy Baxter, an architect in Salesforce’s Ethical AI Practice concurred, adding: “How do we rebuild software that truly has a positive impact on the people it serves? AI can do so much tremendous good, but it can have the potential to unknowingly harm individuals. We can’t expect AI to magically exclude bias in society – bias is baked in.”
“How do we represent the world that we want and not the world as it is?”, Baxter continued.
Given AI essentially needs to learn, it will take its lead from human beings. It’s our responsibility to act ethically and do the right thing when it comes to developing AI development. Or so the panel, moderated by Salesforce futurist Peter Schartz, believes.
Baxter stressed that in particular, there’s a need to ensure that people are not adversely impacted because of factors they cannot change or control, such as gender or race.
This isn’t the first time that issues around AI ethics have come into play. Elon Musk, for all of his eccentricities, has been championing this stance on AI for years and earlier this month the World Economic Forum’s head of AI echoed similar sentiments.
The Dreamforce panel highlighted that it will be just as important to educate people on the shortcomings of AI and potential bias as it is to promote the benefits of smart systems. Ultimately, like with technology today, the results you get out are only as good as the data that’s put in. The same is true of AI as it stands now.
“AI will have a bigger impact than the internet on humanity,” Socher added. “AI will pick up bias and either amplify it or keep it going. We have to educate people that AI is only as good as the training data.”
When it comes to that so-called training data, Baxter said Salesforce recognised its role in boosting awareness and education levels. Using Trailhead, as well as other AI-focused resources, the cloud firm hopes to help open peoples’ eyes to the potential and the pitfalls so they can make informed decisions.
“The quality of that training data is key. It helps customers see and understand the data so they can identify if there is any bias there if there are any errors, so they can correct it,” Baxter added.
“Ethics is a mindset, not a checklist, and we need to instil it early on.”