Facebook: High School Science Club Management Faces Modest Challenge
November 4, 2020
Imagine running an elite high school for really smart people. Now think about having almost half of the people in school think you are a dork. Sound cool?
According to “49% Facebook Employees Don’t Believe It Had Positive Impact On World” seems to suggest that the senior management of ever lovable Facebook has this hurdle to surmount, crawl around, tunnel under, or leap over. The write up states:
Facebook released the results of its internal half-yearly “Pulse survey.” One of the key findings reported by Buzzfeed is that only 51% of employees believe that Facebook is having a positive impact on the world. The survey was taken by 49,000 Facebook employees in a period of two weeks in October.
How does one manage this “half don’t get with the program” issue? The write up does not consider this management question. But we learn:
Recently, the company has been working on various issues with the platform. The company first got rid of anti-vax content, then aimed to remove misinformation about the holocaust. The social media giant also took two good initiatives for flu shots and making U.S. citizens aware of voting. While those are all good things, the issues run deeper.
The challenge, however, is not limited to employees. If “Most Americans Think Social Media Has a Negative Effect on the US” presents accurate data, Facebook faces a larger management task.
Even more intriguing is Facebook’s growth runway. “Facebook Makes More Money per User Than Rivals, But It’s Running Out of Growth Options” asserts that Snap and Pinterest seems to be in “double digit year-over-year growth in users, revenue, and average revenue per user.” The reason? Advertising strategies.
“Move fast and break things” may apply to running a company in an manner that keeps half of the employees in a happy place. Maybe if management buys lunch, more people will feel good about the company’s innovative approach to making the world closer together. Maybe?
Stephen E Arnold, November 4, 2020
Facebook: Be a Good Neighbor
October 27, 2020
Engadget published “Facebook Is Testing a Nextdoor-Like Neighborhoods Feature in Canada.” The write up reports:
Facebook is testing a feature called Neighborhoods that would allow users to join community-based groups, much as you can with Nextdoor.
The article says:
Facebook’s Neighborhoods feature could allow the social network to display hyper-local ads and gather more data on users, judging by the screenshots seen so far.
At the end of the write up, the author reminds me that Facebook is under investigation.
What’s interesting is that the Facebook “innovation,” if the article is accurate, is a good example of me-too; that is, a smaller company figures out a service, attracts users, and, on the surface at least, is successful.
Could the business tactics playbook for an outfit like Facebook contain this tactic: Copy and crush?
Worth watching as the country of Facebook confronts those next door. Now the music for “Won’t You Be My Neighbor” are echoing:
It’s a beautiful day in this neighborhood,
A beautiful day for a neighbor.
Would you be mine?
Could you be mine?…
Stephen E Arnold, October 27, 2020
Algorithm Tuning: Zeros and Ones Plus Human Judgment
October 23, 2020
This is the Korg OT-120 Orchestral Tuner. You can buy it on Amazon for $53. It is a chromatic tuner with an eight octave detection range that supports band and orchestra instruments. Physics tune pianos, organs, and other instruments. Science!
This is the traditional piano tuner’s kit.
You will need ears, judgment, and patience. Richard Feynman wrote a letter to a piano tuner. The interesting point in Dr. Feynman’s note was information about the non-zero stiffness of piano strings affects tuning. The implication? A piano tuner may have to factor in the harmonics of the human ear.
The Korg does hertz; the piano tuner does squishy human, wetware, and subjective things.
I thought about the boundary between algorithms and judgment in terms of piano tuning as I read “Facebook Manipulated the News You See to Appease Republicans, Insiders Say”, published by Mother Jones, an information service not happy with the notes generated by the Facebook really big organ. The main idea is that human judgment adjusted zeros, ones, and numerical recipes to obtain desirable results.
The write up reports:
In late 2017, Zuckerberg told his engineers and data scientists to design algorithmic “ranking changes” that would dial down the temperature.
Piano tuners fool around to deliver the “sound” judged “right” for the venue, the score, and the musician. Facebook seems to be grabbing the old-fashioned tuner’s kit, not the nifty zeros and ones gizmos.
The article adds:
The code was tweaked, and executives were given a new presentation showing less impact on these conservative sites and more harm to progressive-leaning publishers
What happened?
We learn:
for more than two years, the news diets of Facebook audiences have been spiked with hyper conservative content—content that would have reached far fewer people had the company not deliberately tweaked the dials to keep it coming, even as it throttled independent journalism. For the former employee, the episode was emblematic of the false equivalencies and anti-democratic impulses that have characterized Facebook’s actions in the age of Trump, and it became “one of the many reasons I left Facebook.”
The specific impact on Mother Jones was, according to the article:
Average traffic from Facebook to our content decreased 37 percent between the six months prior to the change and the six months after.
Human judgment about tool use reveal that information issues once sorted slowly by numerous gatekeepers can be done more efficiently. The ones and zeros, however, resolve to what a human decides. With a big information lever like Facebook, the effort for change may be slight, but the impact significant. The problem is not ones and zeros; the problem is human judgment, intent, and understanding of context. Get it wrong and people’s teeth are set on edge. Unpleasant. Some maestros throw tantrums and seek another tuner.
Stephen E Arnold, October 23, 2020
Amusing, That Facebook: Born to Curate
October 22, 2020
Facebook loves it when users share news, photos, and opinions, unless they speak ill of the social media platform. Vice explains how Facebook limits free speech in: “Facebook Just Forced Its Most Powerful Critics Offline.”
Facebook does not like the Real Facebook Oversight Board, a group founded in September 2020 when the social media company failed to run its own oversight board in time for the US presidential election. Because Facebook did not like the Real Facebook Oversight Board, they used their legal clout to force the group offline. Facebook wrote the group’s ISP to remove its Web site and succeeded.
What is the Real Facebook Oversight Board?
“The group is made up of dozens of prominent academics, activists, lawyers, and journalists whose goal is to hold Facebook accountable in the run-up to the election next month. Facebook’s own Oversight Board, which was announced 13 months ago, will not meet for the first time until later this month, and won’t consider any issues related to the election.”
Facebook complained the Real Facebook Oversight Board was involved in phishing scams. Usually when a request to remove a Web site reaches an ISP, there is a despite resolution process that takes months and ultimately a court order must be obtained to terminate the site. Facebook had another Web site owned by the Real Facebook Oversight Board removed in the past.
Facebook denied responsibility stating the Real Facebook Oversight Board’s Web site was taken offline because it contained the word “facebook” and violated copyright. Email documentation from Facebook proves otherwise. The company is shaping reality in order to protect its public image and troll its critics. Is Facebook’s editorial process veering away from bright, white lines?
Whitney Grace, October 22, 2020
Journalists Do More Than Report: The Covid Determination
October 17, 2020
One of the DarkCyber research team alerted me to “Facebook Greatest Source of Covid-19 Disinformation, Journalists Say.” That’s the factoid, according to the “real” journalists at a British newspaper.
The main point of the write up may be an interesting way to send this message, “Hey, we are not to blame for erroneous Rona info.” I hear the message.
The write up states:
The majority of journalists covering the pandemic say Facebook is the biggest spreader of disinformation, outstripping elected officials who are also a top source, according to an international survey of journalism and Covid-19.
The survey prompted another Guardian article in August 2020.
Let’s assume Facebook and the other social media high pressure data hoses are responsible for bad, weaponized, or just incorrect Rona info. Furthermore, let’s accept these assertions:
Journalism is one of the worst affected industries during the pandemic as hundreds of jobs have been lost and outlets closed in Australia alone. Ninety per cent of journalists surveyed said their media company had implemented austerity measures including job losses, salary cuts and outlet closures.
The impression the write up creates in the malleable Play-doh of my mind is that journalists are no longer reporting the news. “Real” journalists are making the news, and it is about time!
The sample probably reflects the respondents reaction to the questions on the survey, which remain unknown to me. The survey itself may have been structured as a dark pattern. What better way to explain that bad things are happening to “real” journalists.
What’s interesting is that “real” journalists know that Facebook and other social media systems are bad.
One question, “How long has it taken “real” journalists to figure out the harsh realities of digital streams of users unfettered by internal or external constraints.
Maybe the news is: “It is too late.” Maybe the working hypothesis is that “better late than never”?
Stephen E Arnold, October 17, 2020
Facebook: Interesting Data If Accurate
October 16, 2020
DarkCyber spotted a factoid of interest to law enforcement professionals in “Facebook Responsible for 94% of 69 Million Child Sex Abuse Images Reported by US Tech Firms.”
Facebook has previously announced plans to fully encrypt communications in its Messenger app, as well as its Instagram Direct service – on top of WhatsApp, which is already encrypted – meaning no one apart from the sender and recipient can read or modify messages.
Now about Facebook’s content curation procedures? End-to-end encryption of ad supported private messaging services appears to benefit bad actors.
Stephen E Arnold, October 16, 2020
Avaaz Facebook Report: Another Road Map for Bad Actors?
October 14, 2020
DarkCyber is intrigued by research reports which try to alert the public to an issue. Often the reports provide a road map for bad actors who are looking for a new method or angle to practice their dark arts. “Facebook’s Algorithm: A Major Threat to Public Health” may be a recent example of doing right going wrong.
Avaaz is, according to the organization’s Web site:
a global web movement to bring people-powered politics to decision-making everywhere.
A 33-page study, published in late August 2020, is available without charge at this link. The publication covers health misinformation through the lens of Facebook’s apparently flawed content curation mechanisms.
For bad actors (real or would be), the document explains:
- The relationship between Web pages with “disinformation” and Facebook sites
- The amplification function of content generators
- Utility of a message output in multiple languages
- The role of backlinks
- A list of “gaps in Facebook’s” content curation method.
Interesting report and one which may help some individuals operate more effectively. Facebook’s content curation has some flaws. The company flagged a photograph of onions as salacious. No, really.
Stephen E Arnold, October 14, 2020
Facebook: Merging Apps Before the Call to Break It Up
October 7, 2020
DarkCyber noted “Facebook Is Merging Messenger into Instagram.” The write up explains:
…Facebook is starting to unify its messaging platforms across apps. They will start including more of Messenger’s features into Instagram’s direct messaging chat platform. It will also add the ability to send messages across the two apps.
DarkCyber believes that unified messaging may have some downstream consequences. On one hand, certain government requests for data may be more helpful if Facebook provides the requested information. On the other hand, breaking up the company could become more difficult.
Stephen E Arnold, October 7, 2020
Facebook Is Nothing If Not Charming
October 5, 2020
Facebook spies on its users by collecting their personal information from hobbies, birthdays, relationships, and vacation spots. Facebook users voluntarily share this information publicly and/or privately. As a result, the company sells that information to advertisers. Facebook also spies on its competitors, but it does so in a more sophisticated way says the BBC article “Facebook Security App Used To ‘Spy’ On Competitors.”
Facebook apparently used its cross-party Onavo VPN to collect information on its competitors knowingly and in violation of anti-piracy laws. The Commons Committee discussed the incident in a report that is more than one hundred pages. Here is the gist of the report:
“The Digital, Culture, Media and Sport Committee wrote that through the use of Onavo, which was billed as a way to give users an extra layer of security, Facebook could ‘collect app usage data from its customers to assess not only how many people had downloaded apps, but how often they used them”.
The report added:
‘This knowledge helped them to decide which companies were performing well and therefore gave them invaluable data on possible competitors. They could then acquire those companies, or shut down those they judged to be a threat.”
Even more alarming are the details about ways Facebook could shut down services it provides to its competition. Twitter’s video sharing app Vine is an example of how Facebook destroyed a competitor. Twitter wanted Vine users to find friends via their Facebook accounts, but Zuckerberg nixed that idea. Vine shuttered in 2016.
Facebook does something equally nefarious with a white list of approved apps that are allowed to use Facebook user data. Among the 5,000 approved apps are Netflix, Airbnb, and Lyft. These app companies supposedly spend $250,000 on Facebook advertising to keep their coveted position.
Zuckerburg wrote in an email:
“I think we leak info to developers, but I just can’t think of any instances where that data has leaked from developer to developer and caused a real issue for us.”
There was the Cambridge Analytica scandal where voter information was collected through a personality quiz. The data of users and their friends was stolen and it profiled 82 million Americans, then that information was sold to the Cambridge Analytica company. The United Kingdom fined Facebook 500,000 pounds and the company apologized.
It will not be the first time Facebook steals and sells user information. We wonder how their competition spies on users and sells their data.
Whitney Grace, October 5, 2020
Facebook WhatsApp: Disappearing Media. Really? Gone for Good?
September 28, 2020
Facebook is endlessly entertaining. On one tentacle, the octopus company seeks to lessen the competitive threats from next generation social media like TikTok-type videos. On another tentacle, Facebook suggests that those in the European Union can do without Facebook. DarkCyber thinks of this as “the take my ball and go home” tactic. Ten year olds with minimal physical coordination but a stock of high end athletic equipment have been known to trot out this argumentative chestnut. Another tentacle semi cooperates with government officials. Another tentacle balances on a knife edge in order to keep employees happy with the wonderful social environment within the helpful company’s digital walled garden. There are other tentacles too, but I want to focus your attention on “WhatsApp Expiring Media Feature Details Tipped via New Beta Version.” Keep in mind that “beta” does not mean something a thumbtyper will be able to use.
The write up explains:
WhatsApp 2.20.201.6 beta for Android has been released with further references for a new feature called “Expiring Media.” The feature, as its name suggests, would make media files such as images, videos, and GIFs sent to the recipient’s phone over WhatsApp disappear once they are viewed.
Interesting. Just one question:
If media are disappeared for users, are those data deleted from the Facebook servers?
One hopes not; otherwise, some investigations will be slowed or halted.
Stephen E Arnold, September 28, 2020