A Xoogler May Question the Google about Responsible and Ethical Smart Software

Write a research paper. Get colleagues to provide input. Well, ask colleagues do that work and what do you get. How about “Looks good.” Or “Add more zing to that chart.” Or “I’m snowed under so it will be a while but I will review it…” Then the paper wends its way to publication and a senior manager type reads the paper on a flight from one whiz kid town to another whiz kid town and says, “This is bad. Really bad because the paper points out that we fiddle with the outputs. And what we set up is biased to generate the most money possible from clueless humans under our span of control.” Finally, the paper is blocked from publication and the offending PhD is fired or sent signals that your future lies elsewhere.


Will this be a classic arm wrestling match? The winner may control quite a bit of conceptual territory along with knobs and dials to shape information.

Could this happen? Oh, yeah.

Ex Googler Timnit Gebru Starts Her Own AI Research Center” documents the next step, which may mean that some wizards undergarments will be sprayed with eau de poison oak for months, maybe years. Here’s one of the statements from the Wired article:

“Instead of fighting from the inside, I want to show a model for an independent institution with a different set of incentive structures,” says Gebru, who is founder and executive director of Distributed Artificial Intelligence Research (DAIR). The first part of the name is a reference to her aim to be more inclusive than most AI labs—which skew white, Western, and male—and to recruit people from parts of the world rarely represented in the tech industry. Gebru was ejected from Google after clashing with bosses over a research paper urging caution with new text-processing technology enthusiastically adopted by Google and other tech companies.

The main idea, which Wired and Dr. Gebru delicately sidestep, is that there are allegations of an artificial intelligence or machine learning cabal drifting around some conference hall chatter. On one side is the push for what I call the SAIL approach. The example I use to illustrate how this cost effective, speedy, and clever short cut approach works is illustrated in some of the work of Dr. Christopher Ré, the captain of the objective craft SAIL. Oh, is the acronym unfamiliar to you? SAIL is short version of Stanford Artificial Intelligence Laboratory. SAIL fits on the Snorkel content diving gear I think.

On the other side of the ocean, are Dr. Timnit Gebru’s fellow travelers. The difference is that Dr. Gebru believes that smart software should not reflect the wit, wisdom, biases, and general bro-ness of the high school science club culture. This culture, in my opinion, has contributed to the fraying of the social fabric in the US, caused harm, and erodes behaviors that are supposed to be subordinated to “just what people do to make a social system function smoothly.”

Does the Wired write up identify the alleged cabal? Nope.

Does the write up explain that the Ré / Snorkel methods sacrifice some precision in the rush to generate good enough outputs? (Good enough can be framed in terms of ad revenue, reduced costs, and faster time to market testing in my opinion.) Nope.

Does Dr. Gebru explain how insidious the short cut training of models is and how it will create systems which actively harm those outside the 60 percent threshold of certain statistical yardsticks? Heck, no.

Hopefully some bright researchers will explain what’s happening with a “deep dive”? Oh, right, Deep Dive is the name of a content access company which uses Dr. Ré’s methods. Ho, ho, ho. You didn’t know?

Beyond Search believes that Dr. Gebru has important contributions to make to applied smart software. Just hurry up already.

Stephen E Arnold, December 2, 2021


DarkCyber, March 29, 2022: An Interview with Chris Westphal, DataWalk

Chris Westphal is the Chief Analytics Officer of DataWalk, a firm providing an investigative and analysis tool to commercial and government organizations. The 12-minute interview covers DataWalk’s unique capabilities, its data and information resources, and the firm’s workflow functionality. The video can be viewed on YouTube at this location.

Stephen E Arnold, March 29, 2022

Latest News

Pixel and Emergency Number Dialing: Is Google Leaving Money on the Table?

I read “Very Scary Issue Dialing 911 on Google Pixel 6 Cell Phones.” The write up may not be representative because it relates data from an undefined sample.... Read more »

November 25, 2022 | Comment

Is There a Horse Named Intel PR?

I noted the information in “Intel Introduces Real-Time Deepfake Detector.” I like the real time angle. The subtitle caught my attention: Intel’s deepfake detector... Read more »

November 25, 2022 | Comment

Is Cyber Security Lagging a Grade Behind Other Technology?

The average computer user is unaware of how invasive and harmful cyber attacks are. Forbes details how little individuals and companies know about cyber crime in,... Read more »

November 25, 2022 | Comment

Cyber Security? That Is a Good Question

This is not ideal. We learn from Yahoo Finance, “Russian Software Disguised as American Finds Its Way into U.S. Army, CDC Apps.” Reuters journalists... Read more »

November 25, 2022 | Comment

The Zuck Play: Why Not Fire Thousands with Twitter As Cover?

Here’s the answer: Mark Zuckerberg’s pet project, Reality Labs, may be his company’s downfall. TechSpot reports, “Meta Value Down $520 Billion... Read more »

November 24, 2022 | Comment

Are Governments Behaving Like Sheep?

North Korea, China, and possibly Russia are incarnates of Orwell’s Big Brother from the dystopian 1984 novel. The US government is compared to Big Brother (and... Read more »

November 24, 2022 | Comment

Will Decision Intelligence Lead to Better Decision Making?

After years of hype, it turns out big data is not paying off as promised. Not yet. Marc Warner, CEO of AI firm Faculty, asserts, “Data-Driven Decision Making... Read more »

November 24, 2022 | Comment

The iPhone Is Magic

I believe everything I read about the Apple iPhone. My knowledge junk bun includes such items as: Apple has a secret $275 billion deal with China. China is, of... Read more »

November 23, 2022 | Comment

Snorkel: Now Humans Are a Benefit?

Snorkel emerged from Stanford University’s AI lab. Some at the Google are ga-ga over Snorkel’s approach to reducing the cost of creating training sets for machine... Read more »

November 23, 2022 | Comment

AI: Black Boxes ‘R Us

Humans design and make AI. Because humans design and make AI, we should know how they work. For some reason, humans do not know how AI works. Motherboard on Vice... Read more »

November 23, 2022 | Comment

  • Archives

  • Recent Posts

  • Meta