A Xoogler May Question the Google about Responsible and Ethical Smart Software

Write a research paper. Get colleagues to provide input. Well, ask colleagues do that work and what do you get. How about “Looks good.” Or “Add more zing to that chart.” Or “I’m snowed under so it will be a while but I will review it…” Then the paper wends its way to publication and a senior manager type reads the paper on a flight from one whiz kid town to another whiz kid town and says, “This is bad. Really bad because the paper points out that we fiddle with the outputs. And what we set up is biased to generate the most money possible from clueless humans under our span of control.” Finally, the paper is blocked from publication and the offending PhD is fired or sent signals that your future lies elsewhere.


Will this be a classic arm wrestling match? The winner may control quite a bit of conceptual territory along with knobs and dials to shape information.

Could this happen? Oh, yeah.

Ex Googler Timnit Gebru Starts Her Own AI Research Center” documents the next step, which may mean that some wizards undergarments will be sprayed with eau de poison oak for months, maybe years. Here’s one of the statements from the Wired article:

“Instead of fighting from the inside, I want to show a model for an independent institution with a different set of incentive structures,” says Gebru, who is founder and executive director of Distributed Artificial Intelligence Research (DAIR). The first part of the name is a reference to her aim to be more inclusive than most AI labs—which skew white, Western, and male—and to recruit people from parts of the world rarely represented in the tech industry. Gebru was ejected from Google after clashing with bosses over a research paper urging caution with new text-processing technology enthusiastically adopted by Google and other tech companies.

The main idea, which Wired and Dr. Gebru delicately sidestep, is that there are allegations of an artificial intelligence or machine learning cabal drifting around some conference hall chatter. On one side is the push for what I call the SAIL approach. The example I use to illustrate how this cost effective, speedy, and clever short cut approach works is illustrated in some of the work of Dr. Christopher Ré, the captain of the objective craft SAIL. Oh, is the acronym unfamiliar to you? SAIL is short version of Stanford Artificial Intelligence Laboratory. SAIL fits on the Snorkel content diving gear I think.

On the other side of the ocean, are Dr. Timnit Gebru’s fellow travelers. The difference is that Dr. Gebru believes that smart software should not reflect the wit, wisdom, biases, and general bro-ness of the high school science club culture. This culture, in my opinion, has contributed to the fraying of the social fabric in the US, caused harm, and erodes behaviors that are supposed to be subordinated to “just what people do to make a social system function smoothly.”

Does the Wired write up identify the alleged cabal? Nope.

Does the write up explain that the Ré / Snorkel methods sacrifice some precision in the rush to generate good enough outputs? (Good enough can be framed in terms of ad revenue, reduced costs, and faster time to market testing in my opinion.) Nope.

Does Dr. Gebru explain how insidious the short cut training of models is and how it will create systems which actively harm those outside the 60 percent threshold of certain statistical yardsticks? Heck, no.

Hopefully some bright researchers will explain what’s happening with a “deep dive”? Oh, right, Deep Dive is the name of a content access company which uses Dr. Ré’s methods. Ho, ho, ho. You didn’t know?

Beyond Search believes that Dr. Gebru has important contributions to make to applied smart software. Just hurry up already.

Stephen E Arnold, December 2, 2021


DarkCyber, March 29, 2022: An Interview with Chris Westphal, DataWalk

Chris Westphal is the Chief Analytics Officer of DataWalk, a firm providing an investigative and analysis tool to commercial and government organizations. The 12-minute interview covers DataWalk’s unique capabilities, its data and information resources, and the firm’s workflow functionality. The video can be viewed on YouTube at this location.

Stephen E Arnold, March 29, 2022

Latest News

Elephants Recognize One Another and When They Stomp Around, Grass Gets Trampled

I find the coverage of the Twitter, Apple, and Facebook hoe down a good example of self serving and possibly dysfunctional behavior. What caught my attention in... Read more »

December 1, 2022 | Comment

WikiLeaks: Oh, Oh, Some Folks Are Not Happy

I read “WikiLeaks Website Is Struggling to Stay Online—As Millions of Documents Disappear.” If the write up is on the money, one lesson from this alleged cancel... Read more »

December 1, 2022 | Comment

Sesamy for Content in Small Bites

Here is good news for anyone who would like to purchase a piece of content without a long-term relationship with its host platform. The Next Web reports, “Swedish... Read more »

December 1, 2022 | Comment

AI: Opaqueness ‘R Us Unless You Are Special

Humans design and make AI. Because humans design and make AI, we should know how they work. For some reason, humans do not know how AI works. Motherboard on Vice... Read more »

December 1, 2022 | Comment

Blue Chip Consulting: An Interesting Question with a Painfully Obvious Answer

I read “Why Is Booz Allen Renting Us Back Our Own National Parks?” The author is asking a BIG question with what may be a tiny answer. The essay states: Today... Read more »

November 30, 2022 | Comment

France and US Businesses: Semi Permanent Immiscibility?

Unlike a pendulum, the French government and two US high-technology poster kids don’t see eye to eye. However, governments, particularly those in France, are not... Read more »

November 30, 2022 | Comment

Collusion? What Do You Mean Collusion?

Ah what wise and ethical firms we have at the top of the tech food chain. MacRumors reports, “Amazon and Apple ‘Colluded’ to Make iPhone and iPad... Read more »

November 30, 2022 | Comment

This Is Not About Message Bubbles. We Want Cash, Suggests Apple

Telegram is an encrypted message service that has avoided paying Apple fees, but according to TechRadar that has come to an end: “Telegram Forced To Crack Down... Read more »

November 30, 2022 | Comment

Smart Software: Can Humans Keep Pace with Emergent Behavior ?

For the last six months, I have been poking around the idea that certain behaviors are emergent; that is, give humans a capability or a dataspace, and those humans... Read more »

November 29, 2022 | Comment

Harvard Expert Opines about Harvard Drop Out, the Zuckster

I read a weird news release or self promotional write up from an outfit called Benzinga. The write up is titled “Harvard Expert Says Zuckerberg Is Detailing Facebook:... Read more »

November 29, 2022 | Comment

  • Archives

  • Recent Posts

  • Meta