How Axon's plans for Taser drones blindsided its AI ethics board

“It scares the living daylights out of a lot of us.”

Axon Taser drone

Axon’s decision to ignore its own handpicked panel of experts’ explicit guidance is an object lesson on the shortcomings of tech companies’ own ethical AI efforts.

Image: Axon

Late Tuesday night, NYU law professor Barry Friedman called an emergency Zoom meeting with members of the AI ethics board for Taser-maker Axon.

Just a few weeks before, the board — which includes academics, civil liberties advocates and two former chiefs of police — had voted against a proposal by Axon to develop Taser-equipped drones and run a limited pilot program with law enforcement. The board had been mulling the possibility of such a pilot for about a year, according to Friedman; ultimately, a majority of the board decided the risks outweighed the benefits.

But on Tuesday, an email landed in Friedman’s inbox from an Axon employee, alerting him that the company was forging ahead with the plan anyway. Not only was Axon going to develop Taser drones, it planned to pitch them as an answer to school shootings, in the wake of the Uvalde tragedy.

The board had about 48 hours to respond. “This came out of the blue,” Friedman said.

“We were told and given two days to react to something very different than something that we had reacted to. And we already said no to it,” said Danielle Citron, a law professor at University of Virginia who is also a member of the board. “It scares the living daylights out of a lot of us.”

Friedman scrambled to organize the board’s response and “repeatedly” pushed the company not to go forward, he said.

On Thursday morning, Axon CEO Rick Smith announced his company’s intention to develop Taser drones anyway, sharing a press release in which the board’s earlier opposition was buried about 1,200 words deep. Hours later, after receiving a wave of negative responses, Axon tweeted the board’s full response roundly condemning the decision, which the board submitted after Axon’s public announcement.

In a statement to Protocol, Smith said, “I understand and agree with the board’s concerns that there are many questions we will need to answer to ensure these systems are designed for maximum safety and with equity in mind. That’s the exact reason why I decided to go public: to broaden the discussion with many stakeholders.”

Smith encouraged concerned citizens and lawmakers to engage with the company through the development process, including during a Reddit ask me anything session on Friday. “I want to hear from legislators, public safety leaders, school administrators, and parents and members of the concerned public,” Smith wrote.

But whether the board’s members will stay on to have those discussions remains an open question. Some members are now actively considering whether working with the company is still worth their time. “In the past, we were helpful and listened to and [our] feedback was relevant, and maybe not so much anymore,” Citron said. “Maybe this was a period of time, and it’s not meant to be forever.”

“We’re all having conversations about that,” Friedman said.

Axon’s decision to ignore its own handpicked panel of experts’ explicit guidance is an object lesson on the shortcomings of tech companies’ own ethical AI efforts. Short of regulation or laws governing the use of AI and other forms of surveillance, even the most accomplished advisers ultimately only have so much power to push back against companies’ competing priorities.

In a video explaining the company’s decision, Smith attributed it to the horrors of the Uvalde, Texas, shooting, which left 19 children and two teachers dead and 17 others wounded. “When I heard about the latest shooting in an elementary school, I held my kids and my wife, and we cried. That could have been us. It’s so frustrating this just keeps happening,” he said. “So, I’m done waiting for politicians to solve this problem, and we’re going to solve it.”

He described Taser-equipped drones as part of a “three-point plan” to address the scourge of school shootings. That plan also includes sharing footage from cameras in schools with first responders and building VR active shooter training for law enforcement. (In Uvalde, police took part in active shooter training just months before the rampage.)

During his AMA, Smith rejected the idea that Axon is pitching this product as a way to profit from tragedy. “Frankly, there are much easier ways to make money than solving intractable problems like this,” he wrote. “We are engaged out of a passionate belief that we can make technology that is safer, more ethical, and more controlled than today’s solution of adding more people with more guns.”

Citron said she and others had joined Axon’s ethics board because they believed Smith’s ultimate goal is to cut down on shootings with less lethal technology than guns. “I really actually believe him. His end goal is less death by bullets,” Citron said.

And members of the board felt they were making headway in guiding the company’s stance on AI issues, she said, including its commitment not to use facial recognition in body cameras. “That they were interested in hearing our opinions about the kinds of legal imprimatur they should support was gratifying,” Citron said.

It’s been clear to Citron and others, however, that the board’s authorities were limited. That’s by design, Smith said in his AMA. “The purpose of this board is to bring in police-skeptical view points, and our company makes tools for police,” he wrote. “If the board has governing rights over the company, then we would have to make sure the board had a stronger balance of pro-public safety views … which would undermine the very reason for having this advisory board.”

Smith noted that the former police chiefs on the board did support the drone proposal, and he emphasized that the concept is still in the idea phase, not the product phase. “The ethics board will have a say in this decision,” Smith wrote.

But Axon’s dismissal of the majority’s recommendations regarding drones has shaken members’ faith in the board’s overall purpose. Despite their objections, Axon decided to develop this technology not for law enforcement, but for an entirely different and unvetted context: schools. “It’s going to fall on the shoulders of marginalized kids, without question, and couple that with a drone with Taser in a classroom that a kid could hack?” Citron said. “It boggles the mind.”


Judge Zia Faruqui is trying to teach you crypto, one ‘SNL’ reference at a time

His decisions on major cryptocurrency cases have quoted "The Big Lebowski," "SNL," and "Dr. Strangelove." That’s because he wants you — yes, you — to read them.

The ways Zia Faruqui (right) has weighed on cases that have come before him can give lawyers clues as to what legal frameworks will pass muster.

Photo: Carolyn Van Houten/The Washington Post via Getty Images

“Cryptocurrency and related software analytics tools are ‘The wave of the future, Dude. One hundred percent electronic.’”

That’s not a quote from "The Big Lebowski" — at least, not directly. It’s a quote from a Washington, D.C., district court memorandum opinion on the role cryptocurrency analytics tools can play in government investigations. The author is Magistrate Judge Zia Faruqui.

Keep ReadingShow less
Veronica Irwin

Veronica Irwin (@vronirwin) is a San Francisco-based reporter at Protocol covering fintech. Previously she was at the San Francisco Examiner, covering tech from a hyper-local angle. Before that, her byline was featured in SF Weekly, The Nation, Techworker, Ms. Magazine and The Frisc.

The financial technology transformation is driving competition, creating consumer choice, and shaping the future of finance. Hear from seven fintech leaders who are reshaping the future of finance, and join the inaugural Financial Technology Association Fintech Summit to learn more.

Keep ReadingShow less
The Financial Technology Association (FTA) represents industry leaders shaping the future of finance. We champion the power of technology-centered financial services and advocate for the modernization of financial regulation to support inclusion and responsible innovation.

AWS CEO: The cloud isn’t just about technology

As AWS preps for its annual re:Invent conference, Adam Selipsky talks product strategy, support for hybrid environments, and the value of the cloud in uncertain economic times.

Photo: Noah Berger/Getty Images for Amazon Web Services

AWS is gearing up for re:Invent, its annual cloud computing conference where announcements this year are expected to focus on its end-to-end data strategy and delivering new industry-specific services.

It will be the second re:Invent with CEO Adam Selipsky as leader of the industry’s largest cloud provider after his return last year to AWS from data visualization company Tableau Software.

Keep ReadingShow less
Donna Goodison

Donna Goodison (@dgoodison) is Protocol's senior reporter focusing on enterprise infrastructure technology, from the 'Big 3' cloud computing providers to data centers. She previously covered the public cloud at CRN after 15 years as a business reporter for the Boston Herald. Based in Massachusetts, she also has worked as a Boston Globe freelancer, business reporter at the Boston Business Journal and real estate reporter at Banker & Tradesman after toiling at weekly newspapers.

Image: Protocol

We launched Protocol in February 2020 to cover the evolving power center of tech. It is with deep sadness that just under three years later, we are winding down the publication.

As of today, we will not publish any more stories. All of our newsletters, apart from our flagship, Source Code, will no longer be sent. Source Code will be published and sent for the next few weeks, but it will also close down in December.

Keep ReadingShow less
Bennett Richardson

Bennett Richardson ( @bennettrich) is the president of Protocol. Prior to joining Protocol in 2019, Bennett was executive director of global strategic partnerships at POLITICO, where he led strategic growth efforts including POLITICO's European expansion in Brussels and POLITICO's creative agency POLITICO Focus during his six years with the company. Prior to POLITICO, Bennett was co-founder and CMO of Hinge, the mobile dating company recently acquired by Match Group. Bennett began his career in digital and social brand marketing working with major brands across tech, energy, and health care at leading marketing and communications agencies including Edelman and GMMB. Bennett is originally from Portland, Maine, and received his bachelor's degree from Colgate University.


Why large enterprises struggle to find suitable platforms for MLops

As companies expand their use of AI beyond running just a few machine learning models, and as larger enterprises go from deploying hundreds of models to thousands and even millions of models, ML practitioners say that they have yet to find what they need from prepackaged MLops systems.

As companies expand their use of AI beyond running just a few machine learning models, ML practitioners say that they have yet to find what they need from prepackaged MLops systems.

Photo: artpartner-images via Getty Images

On any given day, Lily AI runs hundreds of machine learning models using computer vision and natural language processing that are customized for its retail and ecommerce clients to make website product recommendations, forecast demand, and plan merchandising. But this spring when the company was in the market for a machine learning operations platform to manage its expanding model roster, it wasn’t easy to find a suitable off-the-shelf system that could handle such a large number of models in deployment while also meeting other criteria.

Some MLops platforms are not well-suited for maintaining even more than 10 machine learning models when it comes to keeping track of data, navigating their user interfaces, or reporting capabilities, Matthew Nokleby, machine learning manager for Lily AI’s product intelligence team, told Protocol earlier this year. “The duct tape starts to show,” he said.

Keep ReadingShow less
Kate Kaye

Kate Kaye is an award-winning multimedia reporter digging deep and telling print, digital and audio stories. She covers AI and data for Protocol. Her reporting on AI and tech ethics issues has been published in OneZero, Fast Company, MIT Technology Review, CityLab, Ad Age and Digiday and heard on NPR. Kate is the creator of RedTailMedia.org and is the author of "Campaign '08: A Turning Point for Digital Media," a book about how the 2008 presidential campaigns used digital media and data.

Latest Stories