Real or artificial? Tech titans declare AI ethics concerns

FILE- In this Jan. 5, 2019, file photo people stand in front of the Google tent during preparations for CES International in Las Vegas. Google employees have had more success than other tech workers at demanding change at the company. Google dropped a contract with the Pentagon after employees pushed back on the ethical implications of using company technology to analyze drone video. (AP Photo/John Locher, File)

the Associated Press

The biggest tech companies want you to know that they’re taking special care to ensure that their use of artificial intelligence to sift through mountains of data, analyze faces or build virtual assistants doesn’t spill over to the dark side.

But their efforts to assuage concerns that their machines may be used for nefarious ends have not been universally embraced. Some skeptics see it as mere window dressing by corporations more interested in profit than what’s in society’s best interests.

“Ethical AI” has become a new corporate buzz phrase, slapped on internal review committees, fancy job titles, research projects and philanthropic initiatives. The moves are meant to address concerns over racial and gender bias emerging in facial recognition and other AI systems, as well as address anxieties about job losses to the technology and its use by law enforcement and the military.

But how much substance lies behind the increasingly public ethics campaigns? And who gets to decide which technological pursuits do no harm?

Google was hit with both questions when it formed a new board of outside advisers in late March to help guide how it uses AI in products. But instead of winning over potential critics, it sparked internal rancor. A little more than a week later, Google bowed to pressure from the backlash and dissolved the council.

The outside board fell apart in stages. One of the board’s eight inaugural members quit within days and another quickly became the target of protests from Google employees who said her conservative views don’t align with the company’s professed values.

As thousands of employees called for the removal of Heritage Foundation President Kay Coles James, Google disbanded the board last week.

“It’s become clear that in the current environment, (the council) can’t function as we wanted,” the company said in a statement.

That environment is one of increasing concern that the corporate AI ethics campaigns lack teeth.

“I think (Google’s decision) reflects a broader public understanding that ethics involves more than just creating an ethics board without an institutional framework to provide for accountability,” AI researcher Ben Wagner said.

Google’s original initiative fell into a tech industry trend that Wagner calls “ethics-washing,” which he describes as a superficial effort that’s mostly a show for the public or lawmakers.

“It’s basically an attempt to pretend like you’re doing ethical things and using ethics as a tool to reach an end, like avoiding regulation,” said Wagner, an assistant professor at the Vienna University of Economics and Business. “It’s a new form of self-regulation without calling it that by name.”

Big companies have made an increasingly visible effort to discuss their AI efforts in recent years.

Microsoft, which often tries to position itself as an industry leader on ethics and privacy issues, published its principles around developing AI, released a short book that discussed the societal implications of the technology and has called for some government regulation of AI technologies.

The company’s president even met with Pope Francis earlier this year to discuss industry ethics. Amazon recently announced it is helping fund federal research into “algorithmic fairness,” and Salesforce employs an “architect” for ethical AI practice, as well as a “chief ethical and human use” officer. It’s hard to find a brand-name tech firm without similar initiatives.

It’s a good thing that companies are studying the issue and seeking perspectives on industry ethics, said Oren Etzioni, CEO of the Allen Institute for Artificial Intelligence, a research organization. But ultimately, he said, a company’s CEO is tasked with deciding what suggestions on AI ethics to incorporate in business decisions.

“I think overall it’s a positive step rather than a fig leaf,” he said. “That said, the proof is in successful implementation. I think the jury is still out on that.”

The impact artificial intelligence can have on society has never been more clear, Etzioni said, and companies are reacting to studies about the power of recommendation algorithms and gender bias in AI.

But as Google’s attempt shows, discussing the issues in the public eye also invites public scrutiny.

Google employees have had more success than other tech workers at demanding change at their company. The internet search behemoth dropped a contract with the Pentagon after employees pushed back on the ethical implications of using the company’s AI technology to analyze drone video.

And after more than 2,400 Google employees signed a petition calling for James to be taken off the board, Google scrapped the board altogether.