Fuller wouldn’t comment on specific data privacy controversies dogging Google lately, but says she shares other concerns many have about Big Tech: Cyberbullying. Hate speech amplified online. The impact of artificial intelligence on everything, from jobs to warfare.
“As a consumer myself, as part of the general public, as a mother, it’s very important to understand ... what am I seeing, what are my children seeing,” she said in an interview with The Associated Press in Paris, where she’s announcing new grant winners Tuesday for projects aimed at teaching digital skills to poor, immigrant or elderly users.
The philanthropic arm she runs, Google.org, is like the company’s conscience, spending $100 million a year on non-profit groups that use technology to try to counteract problems the tech world is accused of creating, abetting or exacerbating.
“Across the world we want to make sure we’re a responsible citizen,” she said. But can Google’s do-good arm make up for everything else? At least it’s trying, she argues. “The company is having a lot of conversations around things like access to information and access to data and making sure there’s no algorithmic bias,” she said.
Public outrage has grown over Google’s use of consumer data and domination of the online search market, with governments stepping up scrutiny of the company. Just in the past week, nine groups called for the U.S. government to block Google’s $2.1 billion acquisition of fitness-gadget maker Fitbit, citing privacy and antitrust concerns. Then Google came under fire for a partnership with U.S. health care system Ascension that the Wall Street Journal says gives the search giant access to thousands of patient health records without doctors’ knowledge. Both companies say the deal is compliant with health-privacy law.
Fuller wouldn’t comment specifically on either case, but said, “We take our users’ trust very seriously.” She also insisted that the company has a very “vibrant discussion” internally about sexual misconduct, human rights and other problems that have tarnished Google’s reputation.
Its philanthropic arm is focused lately on limiting damage from breakneck developments in artificial intelligence, notably after employee departures and public pressure over a Pentagon contract pushed the company to pledge it wouldn’t use AI in weapons development.
Among projects Google.org is funding are those that help users create and share digital resumes or map job opportunities, as the company tries to figure out “how can we anticipate some of the impacts of AI in an economy, and understand how can we make sure that everyone has access to jobs that are not only interesting now but jobs that are going to be here in the future,” Fuller said.
Google is also holding a competition this year in Europe for projects on “how we can keep children safe,” she said. Digital literacy is crucial, she said: “All of us need to discern what is truthful of what I see online. How do I ask the questions of who is sponsoring this content.”
Despite its philanthropic efforts, Google’s critics remain legion — even within the tech universe. Former Google design ethicist Tristan Harris argues technology is shortening our attention spans and pushing people toward more extreme views. He couldn’t get Google to tackle these problems when he was there, so he quit and is pushing for change from the outside, through his Center for Human Technologies.
He says companies like Google won’t change voluntarily but that the tech world has undergone a “sea change” in awareness of problems it’s caused, thanks in part to pressure from a frustrated public.