Salesforce empowered an AI ethics team to keep it honest. Dealing with third-party apps is trickier.

Kathy Baxter and her product design ethics team at Salesforce have a lot of sway over when tech is ready to ship and when it’s not. But there are lots of technologies living on the Salesforce platform beyond their control.

Kathy Baxter

Some ethics roles are relegated to a monthly governance or privacy check-in. However, Salesforce's AI ethics team has a surprising amount of input into decisions made throughout the organization.

Photo: Salesforce

Salesforce’s AI software had it way wrong.

When the company’s engineers were incorporating tools that use natural-language processing and machine learning to analyze people’s sentiments toward products or brands in social media or reviews, they discovered a problem. The system automatically labeled content featuring words such as gay, Black or Muslim as negative. It even happened if the statements around them were positive, as in “I’m gay and proud.”

Kathy Baxter and her Ethical AI team, part of the Ethics by Design group inside Salesforce, stepped in to assess the algorithmic tool. But despite the potential for discrimination, some product management decision-makers at the company still wanted to roll out the product in beta anyway to get customer feedback.

In the end, the ethics group won. “We said no — we need to wait on this,” said Baxter, Salesforce’s ethical AI principal architect. “And so the team held back.”

It was a rare example of actual power for tech ethics watchdogs over product decisions that could affect countless people. In the end, Baxter’s ethical design group worked with engineers to conduct bias mitigation on the product’s machine-learning models. They tested and watched how those changes took effect before finally launching the tool as part of its Einstein software in 2019, which is now a component of the company’s Tableau CRM software.

However, there are limits to what her team has the power to change. There are third-party technologies living on the Salesforce platform that they simply cannot inspect. And as the company’s software application universe continues to expand, watching out for problematic AI will only become more challenging.

Since she started her ethics role after joining the company in 2015 as a user experience researcher, Baxter’s team has been folded into a larger Office of Ethical and Humane Use of Technology that sits inside its product division and includes people with backgrounds from philosophy and international studies to more traditional tech disciplines like computer and data science.

“In 2016, when Marc Benioff, our CEO, announced that we were going to become an AI-first company, and we started acquiring a bunch of AI companies, my concern was, how do we ensure that we are going to use this technology in a way that's responsible?” Baxter said.

Problem solving

Some of the companies that are subject to the most intense scrutiny of their AI practices say they have embedded ethics into product development. Facebook parent Meta has an interdisciplinary responsible AI group that works with its product teams to address fairness in its AI-enabled products. Microsoft’s responsible AI office seeds governance processes across the company, while another group there is dedicated to incorporating the company’s responsible AI rules into engineering work.

Google famously shuttered its AI ethics committee following criticism over its choices for committee members. The company also angered the AI ethics community after firing one of the industry's most respected AI ethics researchers. But today it still has a Responsible AI and Human-Centered Technology group that works with product and engineering teams.

In general, however, for many companies putting AI principles into practice is not as common, and many are looking to government to create regulations to guide that process.

Some ethics roles sit inside a legal compliance team or are relegated to a monthly governance or privacy check-in. However, Baxter’s team has a surprising amount of day-to-day interaction and input into decisions made throughout the organization.

When Salesforce was developing templates in its Einstein Discovery software for predictive analytics, also now part of Tableau CRM, her team convinced engineers to remove data fields showing zip codes. That neighborhood-level data can be a proxy for race, and therefore risked enabling racial bias in how the system made predictions.

“Although from a legal perspective, it probably would have been okay, from an ethical perspective, we said no. And so the team had to remove those fields,” she said.

Baxter even assists in sales interactions; for instance, if a customer wants to do something with the company’s software that doesn’t seem quite kosher, she said. “We will actually sit in on calls with our sales engineers,” said Baxter, who said she tries to find alternatives to misguided approaches.

“It’s important to me to never come out of the gate and say no – that's against our acceptable-use policy,” she said. “Instead, tell me what the problem is you're trying to solve, because sometimes what they're trying to solve, their idea for the solution is just not the best one.” So, she said, she tries to work with the customer to find “a better solution.”

In general, Salesforce has what Baxter calls “red lines” for how its AI products can be used. For instance, its acceptable-use policy prohibits customers from using its computer vision, natural-language processing or predictive-analytics tools to estimate someone’s race or ethnic origin, religious or political beliefs, gender, sexual orientation, health or financial status or to predict likelihood to commit a crime.

Beyond their control

But Salesforce isn’t just what is built by Salesforce.

Like many cloud platforms, Salesforce is a dense forest of software applications and tools. The company’s AppExchange hosts applications built by other companies that do anything from giving automated financial planning advice to managing patients with addictions.

Salesforce gently nudges tech partners toward more ethical practices. For instance, Baxter authored a paper published last year that presented steps for organizations to take to address ethical questions during the AI development and implementation phases. The company also offers a course for Responsible Creation of Artificial Intelligence for Salesforce users. But that educational module includes just four short classes lasting 45 minutes in total.

Indeed, Baxter’s team can only go so far to ensure that everything happening in the vast Salesforce universe is fair, non-discriminatory and privacy-preserving.

“There are real legal restrictions,” said Baxter, pointing to the company’s status as a data processor rather than a data controller under European data privacy regulations. “From a contractual standpoint as well, we can't look at our customers’ data or their models, or futz with them or tell them, no, you're not allowed to do that.”

Baxter said the pandemic put those limitations in stark relief. “We recognized that at the very beginning of the pandemic, we were probably going to see a whole lot of apps popping up on AppExchange. And there could be real concerns as far as: What data would they collect? How would they handle that data? Would it be inclusive?” said Baxter.

Salesforce put up some special barriers and created guidelines for application partners in a Vaccine Cloud it launched a year ago to provide technologies to help public health agencies and businesses manage COVID-19 vaccine programs, and handle employee health and in-person events. Baxter said the company required app providers in that health and vaccine category to have conversations with Salesforce about how they handle data and other aspects of how their technologies are built and how they work.

Today, the Vaccine Cloud lets customers access all sorts of partner technologies used for things like digital vaccine credentialing for international travel or employee vaccination and testing management. “In that particular case, we recognize that it is very important to ensure that those apps are inclusive and as helpful and privacy-preserving as possible,” said Baxter.

For now, along with Salesforce’s internal product development policies and stopgaps, Baxter said education is the company’s primary tool for fostering more ethical practices among partners in its sprawling app marketplace. She also said in 2022 the company’s account reps are expected to offer customers additional AI ethics education and services.

In coming years, Baxter said, more software providers will have internal ethics staff, especially as regulators in Europe and elsewhere demand more transparent and accountable AI. For now, “It's a nice-to-have, it's a pretty rare thing,” to have someone on staff dedicated to ethics in development of products or services, Baxter said. “But in a few years, it will be madness not to have at least one of these in your company.”

This story was updated to clarify the organizational structure of Salesforce's ethics team and to correct the year the product flagged by the team was released.


Netflix’s layoffs reveal a larger diversity challenge in tech

Netflix just laid off 150 full-time employees and a number of agency contractors. Many of them were the company’s most marginalized employees.

It quickly became clear that many of the laid-off contractors possessed marginalized identities.

Illustration: Christopher T. Fong/Protocol

After Netflix’s first round of layoffs, there was a brief period of relief for the contractors who ran Netflix’s audience-oriented social media channels, like Strong Black Lead, Most and Con Todo. But the calm didn’t last.

Last week, Netflix laid off 150 full-time employees and a number of agency contractors. The customary #opentowork posts flooded LinkedIn, many coming from impacted members of Netflix’s talent and recruiting teams. A number of laid-off social media contractors also took to Twitter to share the news. It quickly became clear that similar to the layoffs at Tudum, Netflix’s entertainment site, many of the affected contractors possessed marginalized identities. The channels they ran focused on Black, LGBTQ+, Latinx and Asian audiences, among others.

Keep Reading Show less
Lizzy Lawrence

Lizzy Lawrence ( @LizzyLaw_) is a reporter at Protocol, covering tools and productivity in the workplace. She's a recent graduate of the University of Michigan, where she studied sociology and international studies. She served as editor in chief of The Michigan Daily, her school's independent newspaper. She's based in D.C., and can be reached at llawrence@protocol.com.

Sponsored Content

Why the digital transformation of industries is creating a more sustainable future

Qualcomm’s chief sustainability officer Angela Baker on how companies can view going “digital” as a way not only toward growth, as laid out in a recent report, but also toward establishing and meeting environmental, social and governance goals.

Three letters dominate business practice at present: ESG, or environmental, social and governance goals. The number of mentions of the environment in financial earnings has doubled in the last five years, according to GlobalData: 600,000 companies mentioned the term in their annual or quarterly results last year.

But meeting those ESG goals can be a challenge — one that businesses can’t and shouldn’t take lightly. Ahead of an exclusive fireside chat at Davos, Angela Baker, chief sustainability officer at Qualcomm, sat down with Protocol to speak about how best to achieve those targets and how Qualcomm thinks about its own sustainability strategy, net zero commitment, other ESG targets and more.

Keep Reading Show less
Chris Stokel-Walker

Chris Stokel-Walker is a freelance technology and culture journalist and author of "YouTubers: How YouTube Shook Up TV and Created a New Generation of Stars." His work has been published in The New York Times, The Guardian and Wired.


Crypto doesn’t have to be red or blue

Sens. Cynthia Lummis and Kirsten Gillibrand are backing bipartisan legislation that establishes regulatory clarity for cryptocurrencies. This is the right way to approach a foundational technology.

"Crypto doesn’t neatly fall along party lines because, as a foundational technology, it is — or should be — inherently nonpartisan," says Diogo Mónica, co-founder and president of Anchorage Digital.

Photo: Anchorage Digital

Diogo Mónica is president and co-founder of Anchorage Digital.

When I moved from Portugal to the United States to work at Square, it was hard to wrap my head around the two-party system that dominates American politics. As I saw at home, democracies, by their very nature, can be messy. But as an outsider looking in, I can’t help but worry that the ever-widening gap between America’s two major parties looms over crypto’s future.

Keep Reading Show less
Diogo Mónica
Diogo Mónica is the co-founder and president of Anchorage Digital, the premier digital asset platform for institutions. He holds a Ph.D. in computer science from the Technical University of Lisbon, and has worked in software security for over 15 years. As an early employee at Square, he helped build security architecture that now moves $100 billion annually. At Docker, he helped secure core infrastructure used in global banks, governments and the three largest cloud providers.

What downturn? A16z raises $4.5 billion for latest crypto fund

The new fund is more than double the $2.2 billion fund the VC firm raised just last June.

A16z general partner Arianna Simpson said that despite the precipitous drop in crypto prices in recent months, the firm is looking to stay active in the market and isn’t worried about short-term price changes.

Photo: Andreessen Horowitz

Andreessen Horowitz has raised $4.5 billion for two crypto venture funds. They’re the industry’s largest ever and represent an outsized bet on the future of Web3 startups, even with the industry in the midst of a steep market downturn.

The pool of money is technically two separate funds: a $1.5 billion fund for seed deals and a $3 billion fund for broader venture deals. That’s more than other megafunds recently raised by competitors such as Paradigm and Haun Ventures.

Keep Reading Show less
Tomio Geron

Tomio Geron ( @tomiogeron) is a San Francisco-based reporter covering fintech. He was previously a reporter and editor at The Wall Street Journal, covering venture capital and startups. Before that, he worked as a staff writer at Forbes, covering social media and venture capital, and also edited the Midas List of top tech investors. He has also worked at newspapers covering crime, courts, health and other topics. He can be reached at tgeron@protocol.com or tgeron@protonmail.com.


How Amazon built its kid-focused Glow video calling projector

Robots, laser pointers, talking stuffies: Amazon’s devices team went through many iterations while developing its very first interactive projection device.

The Amazon Glow is the first interactive projection device sold by Amazon, and it could be a stepping stone for the company to use the technology in other areas.

Illustration: Christopher T. Fong/Protocol

Cats love chasing laser pointers. So why not have kids do the same?

When a small team within Amazon’s devices group began exploring the idea of a kid-focused video calling device nearly five years ago, they toyed with a lot of far-out ideas, a laser pointer controlled by an adult calling from afar being one of them. The suggestion was quickly dismissed over eye safety concerns, but it did lead the team down a path exploring projection technologies.

Keep Reading Show less
Janko Roettgers

Janko Roettgers (@jank0) is a senior reporter at Protocol, reporting on the shifting power dynamics between tech, media, and entertainment, including the impact of new technologies. Previously, Janko was Variety's first-ever technology writer in San Francisco, where he covered big tech and emerging technologies. He has reported for Gigaom, Frankfurter Rundschau, Berliner Zeitung, and ORF, among others. He has written three books on consumer cord-cutting and online music and co-edited an anthology on internet subcultures. He lives with his family in Oakland.

Latest Stories