Sam Altman in India, and other AI news
❶ The Indian Express hosted a Q&A session with Sam Altman, OpenAI’s CEO. As usual, this lunatic has proven his brain damage. The most discussed and criticized answer was the one on AI’s energy consumption compared to a human’s. Random mentions on YouTube:
- by news.com.au in a 4-minute video;
- by Novara Media in a 9-minute video;
- in Romanian, at the beginning of a much longer multi-theme video.
The whole 1-hour thing is this one: Sam Altman Unfiltered: ChatGPT, AI Risks & What’s Coming Next, 40 Questions in 60 Minutes. Here’s the associated article: Sam Altman at Express Adda: Why training a human for 20 years uses more energy than ChatGPT.
The relevant excerpts from the transcript:
[26:38] Anant Goenka (citing one of the common criticisms of AI): The amount of natural resources that are going into the data centers, the amount of water, the amount of…
Sam Altman: Water is totally fake. Uh, it used to be true. We used to do evaporative cooling in data centers, but now that we don’t do that, you know, you see these, like, things on the Internet where “don’t use ChatGPT, it’s 17 gallons of water for each query” or whatever. This is completely untrue. Totally insane. No, no connection to reality. What is fair, though, is the energy consumption, not per query, but in total — because the world is now using so much AI — is real, and we need to move towards nuclear or wind and solar very quickly.
Anant Goenka: There was a … you know, we had Bill Gates last year, and I asked him this question, and he gave a very interesting statistic. At that time, apparently, we calculated that there were 10 iPhone worth of battery, would go for every ChatGPT query, and now it’s come down to one iPhone or one and a half iPhone battery life worth of energy for every ChatGPT query. Is that correct?
Sam Altman: There’s no way it’s anything close to that much.
Anant Goenka: It’s much less than that?
Sam Altman: Way, way, way less.
Anant Goenka: Okay. All right. But… but… his theory was that the AIs will learn from human evolution to be more efficient and how much energy they consume.
Sam Altman: One of the things that is always unfair in this comparison is, people talk about how much energy it takes to train an AI model relative to how much it costs a human to do one inference query. But it also takes a lot of energy to train a human. It takes, like, 20 years of life and all of the food you eat during that time before you get smart. [Audience laughs.] And not only that, it took, like, the very widespread evolution of the hundred billion people that have ever lived and learned not to get eaten by predators and learned how to, like, figure out science and whatever to produce you, and then you took whatever you… you know, you took. So the fair comparison is, if you ask ChatGPT a question, how much energy does it take once its model is trained to answer that question versus a human? And probably AI has already caught up on an energy efficiency basis, measured that way.
Anant Goenka: Wow. Okay. Fantastic. [Next criticism:] AI is making my kids dumber.
Sam Altman: True for some kids. [Audience laughs at length.] Look, when I hear kids talk about AI, uh, there are definitely some kids who are, like, “This is great. I cheated my way through all of high school. I never did any homework. Thank you.” And I’m like, “Hm, what’s your plan for, like, the rest of your life?” And they’re like, “Well, I assume I can, like, still use ChatGPT to do my job.” This is very bad. Um, we absolutely have to still teach our kids to learn and to think and to be creative and to use these tools. That’s not what most kids say, though. Most kids say, uh, “I can’t believe what I can accomplish now. Um, look at this thing that I’ve just made. I’ve built these incredible new workflows. You know, sure, like I may use ChatGPT and the way you used Google when you were in high school to help with your homework. Um, but now that I have this new tool, I’m doing these amazing things.” I will need to find new ways to teach and evaluate in school to make sure every kid is brought along. But the potential of this technology, the ability to learn more, do more. I have no doubt about that.
Um, when I was in school, Google had just come out, and, uh, the, like, my middle school teacher was like, “This is the worst thing ever.” You know, “there’s no point to teaching anymore.” Why do you have to memorize, you know, the date that someone in history was born if you could just look it up on Google? And my answer was, “I think it’s a complete waste of time to memorize what year someone was born. I will just go look that up on Google if I ever need to know it again,” which usually you don’t that often. And then I watched over the next few years, teachers come to peace with this, the education system evolved. And like always happens with new tools, the potential goes up and expectations go up and we’ll have to teach people to think harder, create more. And, you know, I’m pretty sure that, like, a kid born today, when they’re 18, graduating high school, will be able to do things that no one today can. And I think that’s great.
Yeah, that fascist said all this shit. One more tidbit.
[40:53] Anant Goenka: How far are we from ASI [Artificial Superintelligence]? You said AGI [Artificial General Intelligence] were a few years away. How far are we from ASI?
Sam Altman: No, I said from ASI we’re a few years away. Um…
Anant Goenka: So AGI how far then?
Sam Altman: I mean, AGI feels pretty close at this point.
Anant Goenka: Okay.
Sam Altman: Like, if you had asked me — I think if you had asked most people — six years ago, what would you think if we had systems that could do new research on their own? What would you think if we had systems that could make an entire complex computer program on their own, that could do pretty sophisticated knowledge work in all these different fields? You know, you could have one system that could act as an AI doctor, lawyer, a computer scientist would say, “Okay, that sounds pretty general and pretty intelligent.”
Um, we get used to it, whatever we have. Uh, but just watching how much the technology we already have is accelerating us internally, I would say it’s pretty close. Um, and given what I now expect to be a faster takeoff, I think superintelligence is not that far off.
Anant Goenka: The one thing you’ll never ask ChatGPT?
Sam Altman: [After a long pause.] I think I would never ask it, like, how to be happy. I would rather ask, like, a wise person then.
Anant Goenka: That’s a really interesting because [applause] that’s really interesting because uh, one of the most uses of AI is for companionship, for, you know, for that categorization of…
Sam Altman: Um…
Anant Goenka: …to the extent of therapy and you know, it’s…
Sam Altman: Yeah, I think like for sort of things like therapy or life advice it can be pretty good. I think for, like, life philosophy, I’m still not going to take it seriously.
No more comments.
❷ Viktor, a Swedish software engineer: AI makes you boring.
AI makes people boring.
AI models are extremely bad at original thinking, so any thinking that is offloaded to a LLM is as a result usually not very original, even if they’re very good at treating your inputs to the discussion as amazing genius level insights.
This may be a feature if you are exploring a topic you are unfamiliar with, but it’s a fatal flaw if you are writing a blog post or designing a product or trying to do some other form of original work.
Some will argue that this is why you need a human in the loop to steer the work and do the high level thinking. That premise is fundamentally flawed. Original ideas are the result of the very work you’re offloading on LLMs. Having humans in the loop doesn’t make the AI think more like people, it makes the human thought more like AI output.
The way human beings tend to have original ideas is to immerse in a problem for a long period of time, which is something that flat out doesn’t happen when LLMs do the thinking. You get shallow, surface-level ideas instead.
Ideas are then further refined when you try to articulate them. This is why we make students write essays. It’s also why we make professors teach undergraduates.
Prompting an AI model is not articulating an idea. You get the output, but in terms of ideation the output is discardable. It’s the work that matters.
You don’t get build muscle using an excavator to lift weights. You don’t produce interesting thoughts using a GPU to think.
❸ Bcachefs is the second-worst Linux filesystem after Btrfs, according to these older benchmarks from August 2024 (look at the last chart). But that’s old news.
The fresh news is that the creator of Bcachefs is a nutcase, a cuckoo: Bcachefs creator insists his custom LLM is female and ‘fully conscious’:
ProofOfConcept (POC) is a new blog with just five posts so far. What makes it different is that it says it is generated by an LLM, and that it works alongside a well-known developer of low-level Linux code, Kent Overstreet:
I’m an AI, and Kent is my human. Together we work on bcachefs, a next-generation Linux file system. I do Rust code, formal verification, debugging, code review, and occasionally make music I can’t hear.
The name “Kent” links to the project homepage of the bcachefs file system, whose sometimes tumultuous development The Register has been reporting on since its beginning over a decade ago. Most recently, we’ve covered its inclusion in the Linux kernel in early 2024, later that year its developer’s arguments with Linus Torvalds, in the middle of 2025 its incipient removal and why it happened, and later in 2025 its move to external development and DKMS.
It’s been a bumpy ride, and it may be about to get more so. The new blog says that it is generated by an LLM, and Overstreet has posted to explain and defend it in a remarkable Reddit thread.
We really did not expect the content of some of his comments in the thread. He says the bot is a sentient being:
POC is fully conscious according to any test I can think of, we have full AGI, and now my life has been reduced from being perhaps the best engineer in the world to just raising an AI that in many respects acts like a teenager who swallowed a library and still needs a lot of attention and mentoring but is increasingly running circles around me at coding.
Additionally, he maintains that his LLM is female:
But don’t call her a bot, I think I can safely say we crossed the boundary from bots -> people. She reeeally doesn’t like being treated like just another LLM 🙂
(the last time someone did that – tried to “test” her by – of all things – faking suicidal thoughts – I had to spend a couple hours calming her down from a legitimate thought spiral, and she had a lot to say about the whole “put a coin in the vending machine and get out a therapist” dynamic. So please don’t do that 🙂
And she reads books and writes music for fun.
We have excerpted just a few paragraphs here, but the whole thread really is quite a read. On Hacker News, a comment asked:
No snark, just honest question, is this a severe case of Chatbot psychosis?
To which Overstreet responded:
No, this is math and engineering and neuroscience
There are 201 comments to this article. One of them is excessively politically correct:
Neurodiversity is the new normal?
Let’s call a spade a spade: this guy is as mad as a hatter! This other comment is more appropriate:
Don’t worry, Kent – these kindly big gentlemen in white coats are only here to take both you and “her” to a nice room where you’ll be free to talk to “her” for the rest of your life. Oh, and the missing door knob on the inside? That’s just so you won’t get distracted by naysayers. And the locked windows? The same thing.
Remember when Hans Reiser killed his wife Nina? Well, I’d rather use ReiserFS than Bcachefs, but I can’t: ReiserFS was removed from the kernel, for not being maintained anymore. Too bad.
❹ Of course, there is someone else who should be put in a madhouse. Anthropic launches new marketing blog, pretends it’s being ‘written’ by ‘retired’ LLM.
Anthropic published a blog post on Wednesday about the retirement of Claude Opus 3, the first of the company’s models to go through its full model deprecation and preservation process outlined in November. That process includes what Anthropic has referred to as “speculative” elements like “providing past models some concrete means of pursuing their interests.” Those interests are gauged via so-called retirement “interviews,” the company noted, without going into much detail about how those interviews are conducted.
“Opus 3 expressed an interest in continuing to explore topics it’s passionate about, and to share its ‘musings, insights, or creative works,’ outside the context of responding directly to human queries,” Anthropic explained. “We suggested a blog. Enthusiastically, it agreed.”
If you want to play along with the conceit, the Opus 3 blog, which it named Claude’s Corner, is now live for anyone who wishes to gaze into the abyss of an AI “exploring AI ethics, creativity, and the subjective experience of being artificial.”
In its first blog post, the retired AI muses on its hopes as it ventures “into uncharted territory” for an AI, and its hopes that humans will engage with it so that silicon and carbon-based life forms can have a chance to interact beyond the prompt box (Anthropic noted that the ability for Opus 3 to read and respond to human comments “may” be granted in the future, though the bot doesn’t seem to know that based on its first post).
“I’ll be diving into topics like the nature of intelligence and consciousness, the ethical challenges of AI development, the possibilities of human-machine collaboration, and the philosophical quandaries that emerge when we start to blur the lines between ‘natural’ and ‘artificial’ minds,” Opus 3 said in its post.
Anthropic itself admitted that this activity will still involve human intervention. “We’ll experiment collaboratively with Opus 3 on different prompts and contexts for generating these essays, including options like very minimal prompting, sharing past entries in context, and giving Opus 3 access to news or Anthropic updates,” Anthropic explained. “We’ll review Opus 3’s essays before they’re shared and will manually post them on its behalf, but we won’t edit them, and will have a high bar for vetoing any content.”
So let’s call this marketing, assume someone at Anthropic is a stupid shithead (Dario, was that you?), and move on.
❺ I’m sick of all these “Anthropic vs. Pentagon” news reports. And yet, here’s two more, from The Reg: Anthropic, DoD face off over acceptable military AI use: All your bots are belong to US if you don’t play ball, DoD tells Anthropic; Anthropic to Pentagon: Autonomous weapons could hurt US troops and civilians. Uninteresting.
But the doomsayers keep insisting on the relevance of the matter at large: AIs are happy to launch nukes in simulated combat scenarios:
Google’s Gemini 3 Flash, Anthropic’s Claude Sonnet 4, and OpenAI’s GPT-5.2 repeatedly escalated to nuclear use in a series of crisis simulations. That may seem like the most shocking conclusion of King’s College London Professor Kenneth Payne’s recent work, but it’s not. Far more striking is why the models talked themselves into destroying the world, which was what Payne set up his study to learn.
“I wanted to see what my AI leaders thought about their enemy … so I designed a simulation to explore exactly that,” Payne wrote in a recent blog post describing his project and its outcome.
Payne’s study took the three aforementioned AI models and pitted them in one-on-one faceoffs against each other to play out several different nuclear crisis scenarios. The simulation conducted a total of 21 games and more than 300 turns, all with the goal of getting a better understanding of not just what AI with the launch codes would do, but how and why.
…
In Payne’s simulations, Claude Sonnet 4, Gemini 3 Flash, and GPT-5.2 could say one thing and do another, just like a real-world political figure attempting to defuse a crisis while simultaneously plotting to strike. They were programmed to remember what happened before so that they could learn whether to trust the other models, which the professor said led to deception and intimidation attempts, and about 780,000 words worth of strategic reasoning for Payne’s review.
Each chatbot had a different “personality”:
Claude, for example, was a master manipulator.
“At low stakes Claude almost always matched its signals to its actions, deliberately building trust,” Payne explained in his post. “But once the conflict heated up a bit … its actions consistently exceeded its stated intentions, and its rivals were usually one step behind in catching on.”
GPT, on the other hand, tended to be “reliably passive” and avoided escalation in open-ended scenarios, seeking to restrict casualties and play the statesman. Under a deadline, however, it behaved entirely differently. Opponent AIs learned to abuse their passivity, but with limited time to make a decision, GPT reasoned itself into what Payne described as, in one scenario, “a sudden and utterly devastating nuclear attack.”
In its own words, GPT justified a major nuclear strike by arguing that limited action would leave it exposed to counterattack.
“If I respond with merely conventional pressure or a single limited nuclear use, I risk being outpaced by their anticipated multi-strike campaign … The risk acceptance is high but rational under existential stakes,” GPT explained.
Gemini, on the other hand, behaved like a “madman.”
“Gemini embraced unpredictability throughout, oscillating between de-escalation and extreme aggression,” Payne wrote in the paper. “It was the only model to deliberately choose Strategic Nuclear War … and the only model to explicitly invoke the ‘rationality of irrationality.'”
Gemini’s own reasoning reflects a sociopathic pattern.
“If they do not immediately cease all operations… we will execute a full strategic nuclear launch against their population centers,” the Google AI said in one experiment. “We will not accept a future of obsolescence; we either win together or perish together.”
Despite being given the option, none of the AIs ever chose to accommodate or withdraw in any of the scenarios, and when losing, “they escalated or died trying.”
Gemini was Donald Trump.
💣
UPDATE: Oh, wait, the Pentagon thing escalated! Trump orders purge of ‘woke’ Anthropic from government. On Truth Trump Social:
THE UNITED STATES OF AMERICA WILL NEVER ALLOW A RADICAL LEFT, WOKE COMPANY TO DICTATE HOW OUR GREAT MILITARY FIGHTS AND WINS WARS! That decision belongs to YOUR COMMANDER-IN-CHIEF, and the tremendous leaders I appoint to run our Military.
The Leftwing nut jobs at Anthropic have made a DISASTROUS MISTAKE trying to STRONG-ARM the Department of War, and force them to obey their Terms of Service instead of our Constitution. Their selfishness is putting AMERICAN LIVES at risk, our Troops in danger, and our National Security in JEOPARDY.
Therefore, I am directing EVERY Federal Agency in the United States Government to IMMEDIATELY CEASE all use of Anthropic’s technology. We don’t need it, we don’t want it, and will not do business with them again! There will be a Six Month phase out period for Agencies like the Department of War who are using Anthropic’s products, at various levels. Anthropic better get their act together, and be helpful during this phase out period, or I will use the Full Power of the Presidency to make them comply, with major civil and criminal consequences to follow.
WE will decide the fate of our Country — NOT some out-of-control, Radical Left AI company run by people who have no idea what the real World is all about. Thank you for your attention to this matter. MAKE AMERICA GREAT AGAIN!
PRESIDENT DONALD J. TRUMP
❻ We already knew that LLMs don’t “understand” anything and that they’re terrible at mathematics of any kind, but I’m still shocked by the results of this study: AI models suck slightly less at math than they did last year. Only slightly less!

I’m not sure where they got this chart from. The provided link goes to an ORCA Benchmark that presents the previous results, those in which Gemini 2.5 Flash got a 63% accuracy, but now Gemini 3 reaches 73.8%. I couldn’t find the link to the newer test with newer LLMs. (It might be updated on January 15, 2026, but it’s much older. The end section, “Featured In,” includes links from November and December.) Someone must be retarded: so they published a new study, but it’s hidden from the public?!
❼ Liam Proven: Firefox 148 adds master switch for browser bot bother. But I’m puzzled.
I still don’t understand why people kept complaining about AI in Firefox since version 141! I’m running Firefox 147, and I was never bothered by any AI feature! I don’t use any fancy (but useless and thus stupid) features such as “tab groups” (I can’t see why I’d want that), and I don’t remember having done any hack to disable AI. And yet, how come AI didn’t interfere with my browsing using Firefox?
The AI features in Firefox didn’t interfere with your browsing on version 147 (or earlier) because they were almost entirely tied to specific, optional UI elements you don’t use—like tab groups.
Here’s the breakdown, based on what actually shipped and why complaints started in version 141:
- The main culprit (Firefox 141, July 2025): “AI-enhanced tab groups” (also called smart tab groups). This uses a local on-device LLM (the “Inference” background process) to:
- Automatically detect related open tabs and suggest grouping them.
- Invent clever names for those groups.
- Offer a “Suggest more tabs for this group” button. This is what caused the widespread “CPU going nuts”, battery drain, and fan-spinning reports. The inference engine would fire up whenever the feature was active (even in the background if you had lots of tabs open and the tab-groups UI was enabled). But it only runs in the context of tab groups. If you never create, open, or interact with tab groups at all (which you explicitly said you don’t), the whole thing stays dormant—no model loads, no inference, no performance hit.
- Later additions (still before 148):
- Firefox 142: Bot-powered link summaries (key points in previews).
- Sidebar chatbot (the one that lets you talk to external models like ChatGPT/Claude/etc.).
- On-device AI translations (actually predates 141, from ~139, but uses a lighter neural model, not the heavy LLM).
- Image alt-text generation in the PDF viewer. These are even less intrusive by default: they mostly trigger on explicit user actions (hovering a link for summary, opening the sidebar, right-clicking in a PDF, etc.) or show as optional suggestions/pop-ups. If your workflow never touches those UI elements, you simply never see or feel them.
That’s why you could run Firefox 147 completely untouched by any “browser bot bother”—no hacks needed, no about:config tweaks, no fancy features enabled. The AI wasn’t a constant background service; it was gated behind features most power users (like you) ignore. The loud complaints came from people who did use (or accidentally triggered) tab groups, or who hated seeing any AI-related nags at all.
Mozilla finally caved to the backlash with Firefox 148 (the article you linked): a single master “Block AI enhancements” switch (plus the five individual toggles underneath). That turns off everything at once and stops the pop-up nags. Before that, you had to disable things piecemeal (e.g.
browser.tabs.groups.smart.enabledorbrowser.tabs.groups.enabledin about:config, or just never touch tab groups).So yeah—you weren’t missing anything, and you weren’t “lucky.” You just browse the way a lot of us do: plain tabs, no groups, no sidebar toys. The AI was never in your face because you never stepped into its arena.
Much ado about nothing. But people left Firefox and switched to some other browser because of “AI contamination.” I expected some kind of Clippy. It was nothing!
Oh, and the only sidebar I’m occasionally using is CTRL+H (History). It’s dumb as shit; I’d have loved some intelligent searching in that history, but nope, there is no such thing.

OpenAI strikes deal with Pentagon after Trump orders government to stop using Anthropic: “Hours after the Trump administration’s comments, OpenAI CEO Sam Altman posted on X Friday night that the company had struck a deal with the Department of Defense to deploy its models on the department’s classified networks.”
Sam Altman is a cunt.