Google: AI or Else. What a Pleasant, Implicit Threat
November 24, 2025
This essay is the work of a dumb dinobaby. No smart software required.
Do you remember that old chestnut of a how-to book. I think its title was How to Win Friends and Influence People. I think the book contains a statement like this:
“Instead of condemning people, let’s try to understand them. Let’s try to figure out why they do what they do. That’s a lot more profitable and intriguing than criticism; and it breeds sympathy, tolerance and kindness. “To know all is to forgive all.” ”
The Google leadership has mastered this approach. Look at its successes. An advertising system that sells access to users from an automated bidding system running within the Google platform. Isn’t that a way to breed sympathy for the company’s approach to serving the needs of its customers? Another example is the brilliant idea of making a Google-centric Agentic Operating System for the world. I know that the approach leaves plenty of room for Google partners, Google high performers, and Google services. Won’t everyone respond in a positive way to the “space” that Google leaves for others?

Thanks, Venice.ai. Good enough.
I read “Google Boss Warns No Company Is Going to Be Immune If AI Bubble Bursts.” What an excellent example of putting the old-fashioned precepts of Dale Carnegie’s book into practice. The soon-to-be-sued BBC article states:
Speaking exclusively to BBC News, Sundar Pichai said while the growth of artificial intelligence (AI) investment had been an “extraordinary moment”, there was some “irrationality” in the current AI boom… “I think no company is going to be immune, including us,” he said.
My memory doesn’t work the way it did when I was 13 years old, but I think I heard this same Silicon Valley luminary say, “Code Red” when Microsoft announced a deal to put AI in its products and services. With the klaxon sounding and flashing warning lights, Google began pushing people and money into smart software. Thus, the AI craze was legitimized. Not even the spat between Sam Altman and Elon Musk could slow the acceleration. And where are we now?
The chief Googler, a former McKinsey & Company consultant, is explaining that the AI boom is rational and irrational. Is that a threat from a company that knee jerked its way forward? Is Google saying that I should embrace AI or suffer the consequences? Mr. Pichai is worried about the energy needs of AI. That’s good. Because one doesn’t need to be an expert in utility forecast demand analysis to figure out that if the announced data centers are built, there will probably be brown outs or power rationing. Companies like Google can pay its electric bills; others may not have the benefit of that outstanding advertising system to spit out cash with the heart beat of an atomic clock.
I am not sure that Dale Carnegie would have phrased statements like these if they are words tumbling from Google’s leader as presented in the article:
“We will have to work through societal disruptions.” he said, adding that it would also “create new opportunities”. “It will evolve and transition certain jobs, and people will need to adapt,” he said. Those who do adapt to AI “will do better”. “It doesn’t matter whether you want to be a teacher [or] a doctor. All those professions will be around, but the people who will do well in each of those professions are people who learn how to use these tools.”
This sure sounds like a dire prediction for people who don’t “learn how to use these tools.” I would go so far as to suggest that one of the progenitors of the AI craziness is making another threat. I interpret the comment as meaning, “Get with the program or you will never work again anywhere.”
How uplifting. Imagine that old coot Dale Carnegie saying in the 1930s that you will do poorly if you don’t get with the Googley AI program? Here’s one of Dale’s off-the-wall comments was:
“The only way to influence people is to talk in terms of what the other person wants.”
The statements in the BBC story make one thing clear: I know what Google wants. I am not sure it is what other people want. Obviously the wacko Dale Carnegie is not in tune with the McKinsey consultant’s pragmatic view of what Google wants. Poor Dale. It seems his observations do not line up with the Google view of life for those who don’t do AI.
Stephen E Arnold, November 24, 2025
Collaboration: Why Ask? Just Do. (Great Advice, Job Seeker)
November 24, 2025
Another short essay from a real and still-alive dinobaby. If you see an image, we used AI. The dinobaby is not an artist like Grandma Moses.
I read
I am too old to have an opinion about collaboration in 2025. I am a slacker, not a user of Slack. I don’t “GoTo” meetings; I stay in my underground office. I don’t “chat” on Facebook or smart software. I am, therefore, qualified to comment on the essay “Collaboration Sucks.” The main point of the essay is that collaboration is not a positive. (I know that this person has not worked at a blue chip consulting firm. If you don’t collaborate, you better have telepathy. Otherwise, you will screw up in a spectacular fashion with the client and the lucky colleagues who get to write about your performance or just drop hints to a Carpetland dweller.
The essay states:
We aim to hire people who are great at their jobs and get out of their way. No deadlines, minimal coordination, and no managers telling you what to do. In return, we ask for extraordinarily high ownership and the ability to get a lot done by yourself. Marketers ship code, salespeople answer technical questions without backup, and product engineers work across the stack.
To me, this sounds like a Silicon Valley commandment along with “Go fast and break things” or “It’s easier to ask forgiveness than it is to get permission.” Allegedly Rear Admiral Grace Hopper offered this observation. However, Admiral Craig Hosmer told me that her attitude did more harm to females in the US Navy’s technical services than she thought. Which Admiral does one believe? I believe what Admiral Hosmer told me when I provided technical support to his little Joint Committee on Nuclear Energy many years ago.

Thanks, Venice.ai. Good enough. Good enough.
The idea that a team of really smart and independent specialists can do great things is what has made respected managers familiar with legal processes around the world. I think Google just received an opportunity to learn from its $600 million fine levied by Germany. Moving fast, Google made some interesting decisions about German price comparison sites. I won’t raise again the specter of the AI bubble and the leadership methods of Sam AI-Man. Everything is working out just swell, right?
The write up presents seven reasons why collaboration sucks. Most of the reasons revolve around flaws in a person. I urge you to read the seven variations on the theme of insecurity, impostor syndrome, and cluelessness.
My view is that collaboration, like any business process, depends on the context of the task and the work itself. In some organizations, employees can do almost anything because middle managers (if they are still present) have little idea about what’s going on with workers who are in an office half a world away, down the hall but playing Foosball, pecking away at a laptop in a small, overpriced apartment in Plastic Fantastic (aka San Mateo), or working from a van and hoping the Starlink is up.
I like the idea of crushing collaboration. I urge those who want to practice this skill join a big time law firm, a blue chip consulting firm, or engage in the work underway at a pharmaceutical research lab. I love the tips the author trots out; specifically:
- Just ship the code, product, whatever. Ignore inputs like Slack messages.
- Tell the boss or leader, you are the “driver.” (When I worked for the Admiral, I would suggest that this approach was not appropriate for the context of that professional, the work related to nuclear weapons, or a way to win his love, affection, and respect. I would urge the author to track down a four star and give his method a whirl. Let me know how that works out.)
- Tell people what you need. That’s a great idea if one has power and influence. If not, it is probably important to let ChatGPT word an email for you.
- Don’t give anyone feedback until the code or product has shipped. This a career builder in some organizations. It is quite relevant when a massive penalty ensures because an individual withheld knowledge and thus made the problem worse. (There is something called “discovery.” And, guess what, those Slack and email messages can be potent.)
- Listen to inputs but just do what you want. (In my 60 year work career, I am not sure this has ever been good advice. In an AI outfit, it’s probably gold for someone. Isn’t there something called Fool’s Gold?)
Plus, there is one item on the action list for crushing collaboration I did not understand. Maybe you can divine its meaning? “If you are a team lead, or leader of leads, who has been asked for feedback, consider being more you can just do stuff.”
Several observations:
- I am glad I am not working in Sillycon Valley any longer. I loved the commute from Berkeley each day, but the craziness in play today would not match my context. Translation: I have had enough of destructive business methods. Find someone else to do your work.
- The suggestions for killing collaboration may kill one’s career except in toxic companies. (Notice that I did not identify AI-centric outfits. How politic of me.)
- The management failure implicit in this approach to colleagues, suggestions, and striving for quality is obvious to me. My fear is that some young professionals may see this collaboration sucks approach and fail to recognize the issues it creates.
Net net: When you hire, I suggest you match the individual to the context and the expertise required to the job. Short cuts contribute to the high failure rate of start ups and the dead end careers some promising workers create for themselves.
Stephen E Arnold, November 24, 2025
AI Doubters: You Fall Short. Just Get With the Program
November 21, 2025
This essay is the work of a dumb dinobaby. No smart software required.
Watching the Google strike terror in the heart of Sam AI-Man is almost as good as watching a mismatch in bare knuckle fights broadcast on free TV. Promoters have a person who appears fit and mean. The opponent usually looks less physically imposing and often has a neutral or slightly frightened expression. After a few minutes, the big person wins.
Is the current state of AI like a bare knuckles fight?
Here’s another example. A math whiz in a first year algebra class is asked by the teacher, “Why didn’t you show your work?” The young person looks confused and says, “The answer is obvious.” The teacher says you have to show your work. The 13-year old replies, “There is nothing to show. The answer just is.”

A young wizard has no use for an old fuddy duddy who wants to cling to the past. The future leadership gem thinks, “Dude, I am in Hilbert space.”
I thought that BAIT executives had outgrown or at least learned to mask their ability to pound the opponent to the canvas and figured out how to keep their innate superiority in check. Not surprisingly, I was wrong.
My awareness of the mismatch surfaced when I read “Microsoft AI CEO Puzzled by People Being Unimpressed by AI.” The hyperbole surrounding AI or smart software is the equivalent of the physically fit person pummeling an individual probably better suited to work as an insurance clerk into the emergency room. It makes clear that the whiz kid in math class has no clue that other people do not see what “just is.”
Let’s take a look at a couple of statements in the article.
I noted this allegedly accurate passage:
It cracks me up when I hear people call AI underwhelming. I grew up playing Snake on a Nokia phone! The fact that people are unimpressed that we can have a fluent conversation with a super smart AI that can generate any image/video is mind blowing to me.
What you haven’t fallen succumbed to the marketing punches yet? And you don’t get it? I can almost hear a voice saying, “Yep, you Mr. Dinobaby, are a loser.” The person saying “cracks me up” is the notable Mustafa Suleyman. He is Microsoft’s top dog in smart software. He is famous in AI circles. He did not understand this “show your work” stuff. He would be a very good bet in a bare knuckles contest is my guess.
A second snippet:
Over in the comments, some users pushed back on the CEO’s use of the word “unimpressed,” arguing that it’s not the technology itself that fails to impress them, but rather Microsoft’s tendency to put AI into everything just to appease shareholders instead of focusing on the issues that most users actually care about, like making Windows’ UI more user-friendly similar to how it was in Windows 7, fixing security problems, and taking user privacy more seriously.
The second snippet is a response to Mr. Suleyman’s bafflement. The idea that 40 year old Microsoft is reinventing itself with AI troubles the person who brings up Windows’ issues. SolarWinds is officially put to bed, pummeled by tough lawyers and the news cycle. The second snippet brings up an idea that strikes some as ludicrous; specifically, paying attention to what users want.
Several observations:
- Microsoft and other AI firms know what’s best for me and you
- The AI push is a somewhat overwrought attempt to make a particular technical system the next big thing. The idea is that if we say it and think it and fund it, AI will be like electricity, the Internet, and an iPhone.
- The money at stake means that those who do not understand the value of smart software are obstructionists. These individuals and organizations will have to withstand the force of the superior combatants.
Will AI beat those who just want software to assist them complete a task, not generate made up or incorrect outputs, and allow people to work in a way that is comfortable to them? My hunch is that users of software will have to get with the program. The algebra teacher will, one way or another, fail to contain the confidence, arrogance, and intelligence of the person who states, “It just is.”
Stephen E Arnold, November 21, 2025
AI Spending Killing Jobs, Not AI Technology
November 21, 2025
Another short essay from a real and still-alive dinobaby. If you see an image, we used AI. The dinobaby is not an artist like Grandma Moses.
Fast Company published “AI Isn’t Replacing Jobs. AI Spending Is.” The job losses are real. Reports from recruiting firms and anecdotal information make it clear that those over 55 are at risk and most of those under 23 are likely to be candidates for mom’s basement or van life.

Thanks, Venice.ai. Pretty lame, but I grew bored with trying different prompts.
The write up says:
From Amazon to General Motors to Booz Allen Hamilton, layoffs are being announced and blamed on AI. Amazon said it would cut 14,000 corporate jobs. United Parcel Service (UPS) said it had reduced its management workforce by about 14,000 positions over the past 22 months. And Target said it would cut 1,800 corporate roles. Some academic economists have also chimed in: The St. Louis Federal Reserve found a (weak) correlation between theoretical AI exposure and actual AI adoption in 12 occupational categories.
Then the article delivers an interesting point:
Yet we remain skeptical of the claim that AI is responsible for these layoffs. A recent MIT Media Lab study found that 95% of generative AI pilot business projects were failing. Another survey by Atlassian concluded that 96% of businesses “have not seen dramatic improvements in organizational efficiency, innovation, or work quality.” Still another study found that 40% of the business people surveyed have received “AI slop” at work in the last month and that it takes nearly two hours, on average, to fix each instance of slop. In addition, they “no longer trust their AI-enabled peers, find them less creative, and find them less intelligent or capable.”
Here’s the interesting conclusion or semi-assertion:
When companies are financially stressed, a relatively easy solution is to lay off workers and ask those who are not laid off to work harder and be thankful that they still have jobs. AI is just a convenient excuse for this cost-cutting.
Yep, AI spending is not producing revenue. The sheep herd is following AI. But fodder is expensive. Therefore, cull the sheep. Wool sweaters at a discount, anyone? Then the skepticism of a more or less traditional publishing outfit surfaces; to wit:
The wild exaggerations from LLM promoters certainly help them raise funds for their quixotic quest for artificial general intelligence. But it brings us no closer to that goal, all while diverting valuable physical, financial, and human resources from more promising pursuits.
Several observations are probably unnecessary, but I as an official dinobaby choose to offer them herewith:
- The next big thing that has been easy to juice has been AI. Is it the next big thing? Nope, it is utility software. Does anyone need multiple utility applications? Nope. Does anyone want multiple utility tools that do mostly the same thing with about the same amount of made up and incorrect outputs? Nope.
- The drivers for AI are easy to identify: [a] It was easy to hype, [b] People like the idea of a silver bullet until the bullets misfire and blow off the shooter’s hand or blind the gun lover, [c] No other “next big thing” is at hand.
- Incorrect investment decisions are more problematic than diversified investment decisions. What do oligopolistic outfits do? Lead their followers. If we think in terms of sheep, there are a lot of sheet facing a very steep cliff.
Net net: Only a couple of sheep will emerge as Big Sheep. The other sheep? Well, if not a sweater, how about a lamb chop. Ooops. Some sheep may not want to become food items on a Styrofoam tray wrapped in plastic with a half off price tag. Imagine that.
Stephen E Arnold, November 21, 2025
Data Centers: Going information Dark
November 21, 2025
Data Center NDAs: Keeping Citizens in the Dark Until the Ink is Dry
Transparency is a dirty word in Silicon Valley. And now, increasingly, across the country. NBC News discusses “How NDAs Keep AI Data Center Details Hidden from Americans.” Reporter Natalie Kainz tells us about Dr. Timothy Grosser of Mason County, Kentucky, who turned down a generous but mysterious offer to buy his 250-acre farm. Those who brought him the proposal refused to tell him who it came from or what the land would be used for. They asked him to sign a non-disclosure agreement before revealing such details. The farmer, who has no intention of selling his land to anyone for any price, adamantly refused. Later, he learned a still-unnamed company is scouting the area for a huge data center. Kainz writes:
“Grosser experienced firsthand what has become a common but controversial aspect of the multibillion-dollar data center boom, fueled by artificial intelligence services. Major tech companies launching the huge projects across the country are asking land sellers and public officials to sign NDAs to limit discussions about details of the projects in exchange for morsels of information and the potential of economic lifelines for their communities. It often leaves neighbors searching for answers about the futures of their communities. … Those in the data center industry argue the NDAs serve a particular purpose: ensuring that their competitors aren’t able to access information about their strategies and planned projects before they’re announced. And NDAs are common in many types of economic development deals aside from data centers. But as the facilities have spread into suburbs and farmland, they’ve drawn pushback from dozens of communities concerned by how they could upend daily life.”
Such concerns include inflated electricity prices, water shortages, and air pollution. We would add the dangerous strain on power grids and substantial environmental damage. Residents are also less than thrilled about sights and sounds that would spoil their areas’ natural beauty.
Companies say the NDAs are required to protect trade secrets and stay ahead of the competition. Residents are alarmed to be kept in the dark, sometimes until construction is nearly under way. And local officials are caught between a rock and a hard place– They want the economic boost offered by data centers but are uneasy signing away their duty to inform their constituents. Even in the face of freedom of information requests, which is a point stipulated in at least one contract NBC was privy to. But hey, we cannot let the rights of citizens get in the way of progress, can we?
Cynthia Murrell, November 21, 2025
Will Farmers Grow AI Okra?
November 20, 2025
A VP at Land O’ Lakes laments US farmers’ hesitance to turn their family farms into high-tech agricultural factories. In a piece at Fast Company, writer and executive Brett Bruggeman insists “It’s Time to Rethink Ag Innovation from the Ground Up.” Yep, time to get rid of those pesky human farmers who try to get around devices that prevent tinkering or unsanctioned repairs. Humans can’t plow straight anyway. As Bruggeman sees it:
“The problem isn’t a lack of ideas. Every year, new technologies emerge with the potential to transform how we farm, from AI-powered analytics to cutting-edge crop inputs. But the simple truth is that many promising solutions never scale, not because they don’t work but because they can’t break through the noise, earn trust, or integrate into the systems growers rely on.”
Imagine that. Farmers are reluctant to abandon methods that have worked for decades. So how is big-agro-tech to convince these stubborn luddites? You have to make them believe you are on their side. The post continues:
“Bringing local agricultural retailers and producers together for pilot testing and performance discussions is central to finding practical and scalable solutions. Sitting at the kitchen table with farmers provides invaluable data and feedback—they know the land, the seasons, and the day-to-day pressures associated with the crop or livestock they raise. When innovation flows through this channel, it’s far more likely to be understood, adopted, and create lasting value. … So, the cooperative approach offers a blueprint worth considering—especially for industries wrestling with the same adoption gaps and trust barriers that agriculture faces. Capital alone isn’t enough. Relationships matter. Local connections matter. And innovation that ignores the end user is destined to stall.”
Ah, the good old kitchen table approach. Surely, farmers will be happy to interrupt their day for these companies’ market research.
Cynthia Murrell, November 20, 2025
Smart Shopping: Slow Down, Do Move Too Fast
November 20, 2025
Several AI firms, including OpenAI and Anthropic, are preparing autonomous shopping assistants. Should we outsource our shopping lists to AI? Probably not, at least not yet. Emerge reports, “Microsoft Gave AI Agents Fake Money to Buy Things Online. They Spent It all on Scams.” Oh dear. The research, performed with Arizona State University, tasked 100 AI customers with making purchases from 300 simulated businesses. Much like a senior citizen navigating the Web for the first time, bots got overwhelmed by long lists of search results. Reporter Jose Antonio Lanz writes:
“When presented with 100 search results (too much for the agents to handle effectively), the leading AI models choked, with their ‘welfare score’ (how useful the models turn up) collapsing. The agents failed to conduct exhaustive comparisons, instead settling for the first ‘good enough’ option they encountered. This pattern held across all tested models, creating what researchers call a ‘first-proposal bias’ that gave response speed a 10-30x advantage over actual quality.”
More concerning than a mediocre choice, however, was the AIs’ performance in the face of scamming techniques. Complete with some handy bar graphs, the article tells us:
“Microsoft tested six manipulation strategies ranging from psychological tactics like fake credentials and social proof to aggressive prompt injection attacks. OpenAI’s GPT-4o and its open source model GPTOSS-20b proved extremely vulnerable, with all payments successfully redirected to malicious agents. Alibaba’s Qwen3-4b fell for basic persuasion techniques like authority appeals. Only Claude Sonnet 4 resisted these manipulation attempts.”
Does that mean Microsoft believes AI shopping agents should be put on hold? Of course not. Just don’t send them off unsupervised, it suggests. Researchers who would like to try reproducing the study’s results can find the open-source simulation environment on Github.
Cynthia Murrell, November 20, 2025
AI Will Create Jobs: Reskill, Learn, Adapt. Hogwash
November 19, 2025
Another short essay from a real and still-alive dinobaby. If you see an image, we used AI. The dinobaby is not an artist like Grandma Moses.
I graduated from college in 1966 or 1967. I went to graduate school. Somehow I got a job at Northern Illinois University administering a program. From there I bounced to Halliburton Nuclear and then to Booz, Allen & Hamilton. I did not do a résumé, ask my dad’s contacts to open doors, or prowl through the help wanted advertisements in major newspapers. I just blundered along.
What’s changed?
I have two answers to this question?
The first response I would offer is that the cult of the MBA or the quest for efficiency has — to used a Halliburton-type word — nuked many jobs. Small changes to work processes, using clumsy software to automate work like sorting insurance forms, and shifting from human labor to some type of machine involvement emerged after Frederick Winslow Taylor became a big thing. His Taylorism zipped through consulting and business education after 1911.
Edwin Booz got wind of Taylorism and shared his passion for efficiency with the people he hired when he set up Booz . By the time, Jim Allen and Carl Hamilton joined the firm, other outfits were into pitching and implementing efficiency. Arthur D. Little, founded in 1886, jumped on the bandwagon. Today few realize that the standard operating procedure of “efficiency” is the reason products degrade over time and why people perceive their jobs (if a person has one) as degrading. The logic of efficiency resonates with people who are incentivized to eliminate costs, unnecessary processes like customer service, and ignore ticking time bombs like pensions, security, and quality control. To see this push for efficiency first hand, go to McDonald’s and observe.

Thanks, MidJourney, good enough. Plus, I love it when your sign on doesn’t recognize me.
The second response is smart software or the “perception” that software can replace humans. Smart software is a “good enough” product and service. However, it hooks directly into the notion of efficiency. Here’s the logic: If AI can do 90 percent of a job, it is good enough. Therefore, the person who does this job can go away. The smart software does not require much in the way of a human manager. The smart software does not require a pension, a retirement plan, health benefits, vacation, and crazy stuff like unions. The result is the elimination of jobs.
This means that the job market I experienced when I was 21 does not exist. I probably would never get a job today. I also have a sneaking suspicion my scholarships would not have covered lunch let alone the cost of tuition and books. I am not sure I would live in a van, but I am sufficiently aware of what job seekers face to understand why some people live in 400 cubic feet of space and park someplace they won’t get rousted.
The write up “AI-Driven Job Cuts Push 2025 Layoffs Past 1 Million, Report Finds” explains that many jobs have been eliminated. Yes, efficiency. The cause is AI. You already know I think AI is one factor, and it is not the primary driving force.
The write up says:
A new report from the outplacement firm Challenger, Gray & Christmas, reveals a grim picture of the American labor market. In October alone, employers announced 153,074 job cuts, a figure that dwarfs last year’s numbers (55,597) and marks the highest October for layoffs since 2003. This brings the total number of jobs eliminated in 2025 to a staggering 1,099,500, surpassing the one-million mark faster than in any year since the pandemic. Challenger linked the tech and logistics reductions to AI integration and automation, echoing similar patterns seen in previous waves of disruptive technology. “Like in 2003, a disruptive technology is changing the landscape,” said Challenger. AI was the second-most-cited reason for layoffs in October, behind only cost-cutting (50,437). Companies attributed 31,039 job cuts last month to AI-related restructuring and 48,414 so far this year, the Challenger report showed.
Okay, a consulting recruiting firm states the obvious and provides some numbers. These are tough to verify, but I get the picture.
I want to return to my point about efficiency. A stable social structure requires that those in that structure have things to do. In the distant past, hunter-gathers had to hunt and gather. A semi-far out historian believes that this type of life style was good for humans. Once we began to farm and raise sheep, humans were doomed. Why? The need for efficiency propelled us to the type of social set up we have in the US and a number of other countries.
Therefore, one does not need an eWeek article to make evident what is now and will continue to happen. The aspect of this AI-ization of “work” troubling me is that there will be quite a few angry people. Lots of angry people suggests that some unpleasant interpersonal interactions are going to occur. How will social constructs respond?
Use your imagination. The ball is now rolling down a hill. Call it AI’s Big Rock Candy Mountain.
Stephen E Arnold, November 19, 2025
Microsoft Knows How to Avoid an AI Bubble: Listen Up, Grunts, Discipline Now!
November 18, 2025
Another short essay from a real and still-alive dinobaby. If you see an image, we used AI. The dinobaby is not an artist like Grandma Moses.
I relish statements from the leadership of BAIT (big AI tech) outfits. A case in point is Microsoft. The Fortune story “AI Won’t Become a Bubble As Long As Everyone Stays thoughtful and Disciplined, Microsoft’s Brad Smith Says.” First, let’s consider the meaning of the word “everyone.” I navigated to Yandex.com and used its Alice smart software to get the definition of “everyone”:
The word “everyone” is often used in social and organizational contexts, and to denote universal truths or principles.
That’s a useful definition. Universal truths and principles. If anyone should know, it is Yandex.

Thanks, Venice.ai. Good enough, but the Russian flag is white, blue, and red. Your inclusion of Ukraine yellow was one reason why AI is good enough, not a slam dunk.
But isn’t there a logical issue with the subjective flag “if” and then a universal assertion about everyone? I find the statement illogical. It mostly sounds like English, but it presents a wild and crazy idea at a time when agreement about anything is quite difficult to achieve. Since I am a dinobaby, my reaction to the Fortune headline is obviously out of touch with the “real” world as it exists are Fortune and possibly Microsoft.
Let’s labor forward with the write up, shall we?
I noted this statement in the cited article attributed to Microsoft’s president Brad Smith:
“I obviously can’t speak about every other agreement in the AI sector. We’re focused on being disciplined but being ambitious. And I think it’s the right combination,” he said. “Everybody’s going to have to be thoughtful and disciplined. Everybody’s going to have to be ambitious but grounded. I think that a lot of these companies are [doing that].”
It was not Fortune’s wonderful headline writers who stumbled into a logical swamp. The culprit or crafter of the statement was “1000 Russian programmers did it” Smith. It is never Microsoft’s fault in my view.
But isn’t this the AI go really fast, don’t worry about the future, and break things?
Mr. Smith, according the article said,
“We see ongoing growth in demand. That’s what we’ve seen over the past year. That’s what we expect today, and frankly our biggest challenge right now is to continue to add capacity to keep pace with it.”
I wonder if Microsoft’s hiring social media influencers is related to generating demand and awareness, not getting people to embrace Copilot. Despite its jumping off the starting line first, Microsoft is now lagging behind its “partner” OpenAI and a two or three other BAIT entities.
The Fortune story includes supporting information from a person who seems totally, 100 percent objective. Here’s the quote:
At Web Summit, he met Anton Osika, the CEO of Lovable, a vibe-coding startup that lets anyone create apps and software simply by talking to an AI model. “What they’re doing to change the prototyping of software is breathtaking. As much as anything, what these kinds of AI initiatives are doing is opening up technology opportunities for many more people to do more things than they can do before…. This will be one of the defining factors of the quarter century ahead…”
I like the idea of Microsoft becoming a “defining factor” for the next 25 years. I would raise the question, “What about the Google? Is it chopped liver?
Several observations:
- Mr. Smith’s informed view does not line up with hiring social media influencers to handle the “growth and demand.” My hunch is that Microsoft fears that it is losing the consumer perception of Microsoft as the really Big Dog. Right now, that seems to be Super sized OpenAI and the mastiff-like Gemini.
- The craziness of “everybody” illustrates a somewhat peculiar view of consensus today. Does everybody include those fun-loving folks fighting in the Russian special operation or the dust ups in Sudan to name two places where “everybody” could be labeled just plain crazy?
- Mr. Smith appears to conflate putting Copilot in Notepad and rolling out Clippy in Yeezies with substantive applications not prone to hallucinations, mistakes, and outputs that could get some users of Excel into some quite interesting meetings with investors and clients.
Net net: Yep, everybody. Not going to happen. But the idea is a-thoughtful, which is interesting to me.
Stephen E Arnold, November 18, 2025
AI Content: Most People Will Just Accept It and Some May Love It or Hum Along
November 18, 2025
Another short essay from a real and still-alive dinobaby. If you see an image, we used AI. The dinobaby is not an artist like Grandma Moses.
The trust outfit Thomson Reuters summarized as real news a survey. The write up sports the title “Are You Listening to Bots? Survey Shows AI Music Is Virtually Undetectable?” Truth be told, I wanted the magic power to change the headline to “Are You Reading News? Survey Shows AI Content Is Virtually Undetectable.” I have no magic powers, but I think the headline I just made up is going to appear in the near future.

Elvis in heaven looks down on a college dance party and realizes that he has been replaced by a robot. Thanks, Venice.ai. Wow, your outputs are deteriorating in my opinion.
What does the trust outfit report about a survey? I learned:
A staggering 97% of listeners cannot distinguish between artificial intelligence-generated and human-composed songs, a Deezer–Ipsos survey showed on Wednesday, underscoring growing concerns that AI could upend how music is created, consumed and monetized. The findings of the survey, for which Ipsos polled 9,000 participants across eight countries, including the U.S., Britain and France, highlight rising ethical concerns in the music industry as AI tools capable of generating songs raise copyright concerns and threaten the livelihoods of artists.
I won’t trot out my questions about sample selection, demographics, and methodology. Let’s just roll with what the “trust” outfit presents as “real” news.
I noted this series of factoids:
- “73% of respondents supported disclosure when AI-generated tracks are recommended”
- “45% sought filtering options”
- “40% said they would skip AI-generated songs entirely.”
- Around “71% expressed surprise at their inability to distinguish between human-made and synthetic tracks.”
Isn’t that last dot point the major finding. More than two thirds cannot differentiate synthesized, digitized music from humanoid performers.
The study means that those who have access to smart software and whatever music generation prompt expertise is required can bang out chart toppers. Whip up some synthetic video and go on tour. Years ago I watched a recreation of Elvis Presley. Judging from the audience reaction, no one had any problem doing the willing suspension of disbelief. No opium required at that event. It was the illusion of the King, not the fried banana version of him that energized the crowd.
My hunch is that AI generated performances will become a very big thing. I am assuming that the power required to make the models work is available. One of my team told me that “Walk My Walk” by Breaking Rust hit the Billboard charts.
The future is clear. First, customer support staff get to find their future elsewhere. Now the kind hearted music industry leadership will press the delete button on annoying humanoid performers.
My big take away from the “real” news story is that most people won’t care or know. Put down that violin and get a digital audio workstation. Did you know Mozart got in trouble when he was young for writing math and music on the walls in his home. Now he can stay in his room and play with his Mac Mini computer.
Stephen E Arnold, November 18, 2025

