AI Ethics Weekly – Jan 28: Spotlight on Privacy, Bias in AI, and the Battle for Ethical AI

AI Ethics Weekly – Jan 28: Spotlight on Privacy, Bias in AI, and the Battle for Ethical AI
January 27, 2020 LH3_Admin

Here’s your weekly round-up of the latest developments in AI Ethics!

Sign up for our weekly AI ethics newsletter to get this and more exclusive content delivered to your inbox every Tuesday!

Picture credit:

Opinion | You Are Now Remotely Controlled h/t Liz O’Sullivan

These data flows empty into surveillance capitalists’ computational factories, called “artificial intelligence,” where they are manufactured into behavioral predictions that are about us, but they are not for us. Instead, they are sold to business customers in a new kind of market that trades exclusively in human futures. More on this

‘Anonymised’ data can never be totally anonymous, says study h/t Mia Dand

“Anonymised” data lies at the core of everything from modern medical research to personalised recommendations and modern AI techniques. Unfortunately, according to a paper, successfully anonymising data is practically impossible for any complex dataset.” Learn more.

IBM unveils Policy Lab, advocates ‘precision regulation’ of AI h/t Hessie Jones

“…the IBM Policy Lab calls for what it describes as “precision regulation” of AI, or laws that require companies to develop and operate “trustworthy” systems.  Our approach is grounded in the belief that tech can continue to disrupt and improve civil society while protecting individual privacy,” wrote Hagemann and Leclerc in a joint statement. “As technological innovation races ahead, our mission to raise the bar for a trustworthy digital future could not be more urgent.” Read more.

Ring Doorbell App Packed with Third-Party Trackers h/t Alaric Aloor

“Ring isn’t just a product that allows users to surveil their neighbors. The company also uses it to surveil its customers. An investigation by EFF of the Ring doorbell app for Android found it to be packed with third-party trackers sending out a plethora of customers’ personally identifiable information (PII).” Read more.

Clearview AI Says Its Facial Recognition Software Identified A Terrorism Suspect. The Cops Say That’s Not True  h/t Hessie Jones

“There has to be some personal or professional responsibility here,” said Liz O’Sullivan, an artificial intelligence researcher and technology director at the Surveillance Technology Oversight Project. “The consequences of a false positive is that someone goes to jail.” Read more.

Worried About Privacy at Home? There’s an AI for That h/t @MrFresh

“Moore’s law being Moore’s law, AI hardware and software have improved dramatically in recent years. Now there’s a new breed of neural net that can run entirely on cheap, low-power microprocessors. It can do all the AI tricks we need, yet never send a picture or your voice into the cloud.” Read more.

San Diego’s 7-year experiment with facial recognition technology appears to be a flop  h/t Liz O’Sullivan

“Since 2012, the city’s law enforcement agencies have compiled over 65,000 face scans and tried to match them against a massive mugshot database. But it’s almost completely unclear how effective the initiative was, with one spokesperson saying they’re unaware of a single arrest or prosecution that stemmed from the program.” Read more.

AI Isn’t a Solution to All Our Problems h/t @DorotheaBaur

“Beyond AI farms, online crowdsourcing projects are able to create robust tools because thousands of people come together to curate data. However, people bring biases and subjectivity that can influence AI, intentionally or not.” Read more.

Spotlight – Coded Bias Documentary h/t Meredith Broussard

“Corporations & governments are deploying #AI in many ways, but they have failed to ensure technologies work for All. AJL United’s work to expose the resulting harms is featured in the Shalini Kantayya’s film Coded Bias at Sundance.” Get screening details.

The battle for ethical AI at the world’s biggest machine-learning conference  h/t Abeba Birhane

“Bias and the prospect of societal harm increasingly plague artificial-intelligence research — but it’s not clear who should be on the lookout for these problems.” Read more.

Sign up for our weekly AI ethics newsletter to get more exclusive content delivered to your inbox every Tuesday!