(Gizmodo)— It’s been nearly three months since many Google employees—and the public—learned about the company’s decision to provide artificial intelligence to a controversial military pilot program known as Project Maven, which aims to speed up analysis of drone footage by automatically classifying images of objects and people. Now, about a dozen Google employees are resigning in protest over the company’s continued involvement in Maven.
The resigning employees’ frustrations range from particular ethical concerns over the use of artificial intelligence in drone warfare to broader worries about Google’s political decisions—and the erosion of user trust that could result from these actions. Many of them have written accounts of their decisions to leave the company, and their stories have been gathered and shared in an internal document, the contents of which multiple sources have described to Gizmodo.
The employees who are resigning in protest, several of whom discussed their decision to leave with Gizmodo, say that executives have become less transparent with their workforce about controversial business decisions and seem less interested in listening to workers’ objections than they once did.
In the case of Maven, Google is helping the Defense Department implement machine learning to classify images gathered by drones. But some employees believe humans, not algorithms, should be responsible for this sensitive and potentially lethal work—and that Google shouldn’t be involved in military work at all.
Historically, Google has promoted an open culture that encourages employees to challenge and debate product decisions. But some employees feel that their leadership no longer as attentive to their concerns, leaving them to face the fallout.
“Over the last couple of months, I’ve been less and less impressed with the response and the way people’s concerns are being treated and listened to,” one employee who resigned said.
“I realized if I can’t recommend people join here, then why am I still here?”
There’s precedent for employee pushback resulting in product changes—in 2015, employees and users successfully challenged Google’s ban on sexually explicit content posted to Blogger. But these are the first known mass resignations at Google in protest against one of the company’s business decisions, and they speak to the strongly felt ethical concerns of the employees who are departing.
In addition to the resignations, nearly 4,000 Google employees have voiced their opposition to Project Maven in an internal petition that asks Google to immediately cancel the contract and institute a policy against taking on future military work.
However, the mounting pressure from employees seems to have done little to sway Google’s decision—the company has defended its work on Maven and is thought to be one of the lead contenders for another major Pentagon cloud computing contract, the Joint Enterprise Defense Infrastructure, better known as JEDI, that is currently up for bids.
Employees’ demands that Google end its Pentagon contract are also complicated by the fact that Google claims it is only providing open-source software to Project Maven, which means the military would be able to still use the technology, even if Google didn’t accept payment or offer technical assistance.
Still, the resigning employees believe that Google’s work on Maven is fundamentally at odds with the company’s do-gooder principles. “It’s not like Google is this little machine-learning startup that’s trying to find clients in different industries,” a resigning employee said. “It just seems like it makes sense for Google and Google’s reputation to stay out of that.”
Many Google employees first learned the company was working on Maven when word of the controversial project began to spread internally in late February. At the time, a Google spokesperson told Gizmodo that the company was in the process of drafting “policies and safeguards” around its use of machine learning, but that policy document has yet to materialize, sources said.
One employee explained that Google staffers were promised an update on the ethics policy within a few weeks, but that progress appeared to be locked in a holding pattern. The ethical concerns “should have been addressed before we entered this contract,” the employee said.
Google has emphasized that its AI is not being used to kill, but the use of artificial intelligence in the Pentagon’s drone program still raises complex ethical and moral issues for tech workers and for academics who study the field of machine learning.
In addition to the petition circulating inside Google, the Tech Workers Coalition launched a petition in April demanding that Google abandon its work on Maven and that other major tech companies, including IBM and Amazon, refuse to work with the U.S. Defense Department.
“We can no longer ignore our industry’s and our technologies’ harmful biases, large-scale breaches of trust, and lack of ethical safeguards,” the petition reads. “These are life and death stakes.”
More than 90 academics in artificial intelligence, ethics, and computer science released an open letter today that calls on Google to end its work on Project Maven and to support an international treaty prohibiting autonomous weapons systems. Peter Asaro and Lucy Suchman, two of the authors of the letter, have testified before the United Nations about autonomous weapons; a third author, Lilly Irani, is a professor of science and a former Google employee.
Google’s contributions to Project Maven could accelerate the development of fully autonomous weapons, Suchman told Gizmodo. Although Google is based in the U.S., it has an obligation to protect its global user base that outweighs its alignment with any single nation’s military, she said.
“If ethical action on the part of tech companies requires consideration of who might benefit from a technology and who might be harmed, then we can say with certainty that no topic deserves more sober reflection—no technology has higher stakes—than algorithms meant to target and kill at a distance and without public accountability,” the letter states.
“Google has moved into military work without subjecting itself to public debate or deliberation, either domestically or internationally. While Google regularly decides the future of technology without democratic public engagement, its entry into military technologies casts the problems of private control of information infrastructure into high relief.”
Executives at Google have made efforts to defend Project Maven to employees. At a meeting shortly after the project became public, Google Cloud CEO Diane Greene spoke in support of Project Maven, multiple sources told Gizmodo. More recently, Greene and other employees have hosted several sessions to debate and discuss the project.
These sessions featured speakers who supported and opposed Maven and stressed the difficulty of drafting policy about the ethical use of machine learning, an attendee explained.
“Actions speak louder than words, and that’s a standard I hold myself to as well.”
There are other reputational concerns factoring into employees’ decisions to leave Google. The company’s recent political fumbles, like its sponsorship of the Conservative Political Action Conference and its struggle to address internal diversity concerns, have also played a role.
“At some point, I realized I could not in good faith recommend anyone join Google, knowing what I knew. I realized if I can’t recommend people join here, then why am I still here?” a resigning Google employee said.
“I tried to remind myself right that Google’s decisions are not my decisions. I’m not personally responsible for everything they do. But I do feel responsibility when I see something that I should escalate it,” another added.
“An important part of our culture is having employees who are actively engaged in the work that we do. We know that there are many open questions involved in the use of new technologies, so these conversations—with employees and outside experts—are hugely important and beneficial,” a Google spokesperson said in an April statement.
“The technology is used to flag images for human review and is intended to save lives and save people from having to do highly tedious work. Any military use of machine learning naturally raises valid concerns. We’re actively engaged across the company in a comprehensive discussion of this important topic and also with outside experts, as we continue to develop our policies around the development and use of our machine learning technologies.”
A Google spokesperson did not immediately respond to a request for comment about the resignations. But employees want to see action from the company, in the form of an ethics policy, a canceled contract, or both.
“Actions speak louder than words, and that’s a standard I hold myself to as well,” a resigning employee said. “I wasn’t happy just voicing my concerns internally. The strongest possible statement I could take against this was to leave.”