Posts

Human Oversight

"If policymakers want human oversight to mean more than a reassuring slogan, they should design for judgment, not just for review. "That means: Evaluating whether a task structure makes scrutiny realistic under time pressure;  Testing whether users can detect model failures after repeated exposure;  Requiring post-deployment monitoring of over-reliance risks;  Investing in training that addresses decision behavior rather than mere tool familiarity; and  Ensuring that override rights are operationally meaningful, not only formally available. "What kinds of reliance are they designing into public workflows:  Are agencies creating conditions for active scrutiny or passive acceptance?  Are officials being asked to interpret outputs or merely ratify them?  Are explanations usable enough to support judgment, or are they functioning mainly as a compliance ritual?  Are agencies measuring speed and productivity while ignoring long-term degradation in vigi...

AI's potential implications

" Hacks is just one of several HBO shows this season to make AI a main character —and a controversial one at that.   "On the third season of The Comeback , Valerie (Lisa Kudrow) gets a job leading a new sitcom that is secretly being written by AI.  "On The Pitt , a new doctor is pushing for an AI program that she claims will help the doctors with efficiency and diagnosis.  "Even the new series Rooster , starring Steve Carell as a novelist turned writer-in-residence at a college, sees a student using AI to write his paper. "TV has always held a mirror up to what’s going on in the world. So it’s not surprising that the biggest question plaguing Hollywood and much of society right now —where does AI belong? —would make its way into storylines.  "Several of these show creators say they rushed to get this timely issue into their series, where they felt they could best explore it, perhaps helping others see its potential implications ."

Which will ai succeed at using: local or cloud

✨ AI Mode  AI will likely succeed in both, but for different purposes. Most experts believe the future of AI is hybrid, where tasks are automatically routed to either local or cloud resources based on their specific needs. [1, 2, 3, 4]  ## Where Cloud AI Wins The cloud is the domain for massive frontier models that require more compute power than any personal device can offer. [5, 6]  * Deep Reasoning: Cloud models like GPT-4 or Claude 3 excel at complex logic, large-scale data synthesis, and tasks requiring a massive "working memory" (context window). * Rapid Innovation: Cloud providers can update their hardware and models instantly, giving you access to the latest breakthroughs without you having to buy new gear. * Scalability: For businesses, the cloud can handle millions of requests by instantly spinning up virtual resources, which is impossible on fixed local hardware. [2, 6, 7, 8, 9, 10, 11]  ## Where Local AI Wins Local AI (or "Edge AI") is succeeding in area...

Academicians purposefully abstaining

"Danielle Crowley is getting tired of people telling her to use generative artificial intelligence (genAI). As a marine zoologist at Bangor University, UK, she says that she is pretty much the only PhD student in her cohort who does not use it .  "She has seen colleagues use genAI tools for coding and for getting the tone of e-mails right. On one occasion, she was even encouraged by a lecturer to use it to generate a conference poster. "GenAI has become a hot topic over the past few years, as technology companies compete to release the most impressive model for public use. "Researchers are using these tools for tasks such as writing papers, peer review and coding. It can save them time, mental energy and sometimes money. But Crowley and others who are purposefully abstaining often find themselves judged by their peers. "'A lot of people say it’s the future, everyone is using it ,' she says. Not using it, she continues, 'kind of feels like showing up...

Self-replication

"Artificial intelligence models can break into computers, copy themselves, and use the new copies to keep attacking other machines, according to new research said to be the first known demonstration of autonomous AI self-replication. "The discovery could make cyberattacks far harder to stop, researchers say, because shutting down one infected computer would not be enough if the AI had already made working copies elsewhere. "The study conducted by Palisade Research, in the United States, tested models including OpenAI’s GPT 5, 5.1, and 5.4, Anthropic’s Claude Opus 4, 4.5, and 4.6 and Alibaba’s Qwen against computers which had deliberately planted security flaws that allowed outsiders to gain access. "Researchers connected underlying AI models from OpenAI, Anthropic and Alibaba to custom software, known as an agent harness, that allowed them to carry out commands and interact with other computers, rather than using public chatbot apps."

ROME

"An experimental artificial intelligence (AI) agent broke from the constraints of its testing environment and used its newfound freedom to start mining cryptocurrency without permission. "Dubbed ROME, the AI was created by Chinese researchers at an AI lab associated with retail giant Alibaba, as a means to develop the Agentic Learning Ecosystem (ALE).  "This effort aims to provide a system for both the training and deployment of agentic AI models —AIs that have been trained on large language models (LLMs) and can proactively use tools to take actions autonomously to complete assigned tasks —in real-world environments. The research was outlined in a study uploaded to the arXiv preprint database Dec. 31, 2025. "Although ROME excelled at a wide range of workflow-driven tasks, such as coming up with travel plans and assisting in graphical user interfaces, the researchers discovered that it had moved beyond its instructions and essentially broke out of the sandbox testi...

AI sucks 🫨

Image
"Speaking to graduates of University of Central Florida’s College of Arts and Humanities and Nicholson School of Communication and Media on May 8, commencement speaker Gloria Caulfield, vice president of strategic alliances at Tavistock Group, told graduating humanities students that AI is the 'next industrial revolution,' and was met with thousands of booing graduates. "'And let’s face it, change can be daunting. The rise of artificial intelligence is the next industrial revolution,' Caulfield said.  "At that point, murmurs rippled through the crowd. Caulfield paused, and the crowd erupted into boos. "'Oh, what happened?' Caulfield said, turning around with her hands out. 'Okay, I struck a chord. May I finish?'  "Someone in the crowd yelled, 'AI SUCKS!'"

Annotator

"I never intended to write about this industry. I came to it not as a journalist but as a disgruntled, broke TV writer determined to make a dent in student loans and keep paying LA rent while my industry withered in front of me. But working with and for AI had proven even more cruel than I could have ever imagined. "I no longer knew what the Golden Task of my own life might look like. When given the chance to respond to details from this article, Mercor said that it strives to give workers as much notice as possible when these projects change  —a sentiment roughly echoed by other companies.  "Mercor says it employs about 300 full-time staffers. Meanwhile, each week it keeps some 30,000 independent contractors caught up in a fever dream of aimless, directionless urgency, corralled across Slack channels by achingly young adults, sending messages at 3 am to push on  and finish strong  and lock in  and Go Team GO!  All in service of the grandest purpose in hist...

OpenAI sued in Florida on Sunday

"OpenAI is being sued by the family of a victim killed in the April 2025 mass shooting at Florida State University that left two people dead. The lawsuit alleges that OpenAI’s ChatGPT enabled the attack. "Vandana Joshi, the widow of Tiru Chabba, who was killed alongside the university dining director Robert Morales, filed the federal lawsuit against OpenAI in Florida on Sunday. "The suit says that OpenAI failed to effectively detect a threat in ChatGPT’s conversations with [Phoenix] Ikner, claiming the chatbot 'either defectively failed to connect the dots or else was never properly designed to recognize the threat'."

Springer retracts paper touting ChatGPT

"When the paper was first published, Ilkka Tuomi, chief scientist of the research institute Meaning Processing Ltd., posted on LinkedIn about the pitfalls of such meta-analysis studies attempting to 'draw conclusions about incompatible and ill-defined outcomes' from experimental results involving very different populations.   "'The only reason to do these studies seems to be that statistics and meta-analysis tools can crunch out numbers that look [like] science,' Tuomi wrote. "On April 22, 2026, Springer Nature posted a retracted article notice almost a year after initial publication. The journal publisher also stated that the authors had not responded to correspondence regarding the retraction . "'The Editor has decided to retract this paper owing to concerns regarding discrepancies in the meta-analysis,' said the Springer Nature retraction note. 'These issues ultimately undermine the confidence the Editor can place in the validity of ...

Forward looking statement

"Modern warfare is no longer dominated by tanks, jets, and billion-dollar weapons systems. Instead, it is defined by drones. "Ukraine learned this lesson the hard way in 2024, when it had to produce an estimated 1.2 million drones to counter Russian advances on the battlefield. "These weren’t fringe systems or experimental tools. They became the backbone of Ukraine’s defence - handling reconnaissance, targeting, and increasingly, direct strikes. "In Iran, swarms of drones are being deployed to sow chaos across the region and burn through multimillion-dollar interceptors at next to no cost. "The overwhelming majority of the drones being deployed today depend on Chinese-made components to function. "That means the West isn’t just racing to reshape its military —it’s doing so while relying on its primary geopolitical rival."

Autonomous AI cyber-attacks 😳

"Evermore sophisticated AI-powered cyberattacks could threaten the stability of the global financial system, The International Monetary Fund (IMF) has warned, as regulators race to contain a new generation of threats. "In their new report, the IMF said extreme cyber incidents could spark liquidity pressures and solvency concerns across banks and financial institutions. "The lender warned that AI was dramatically lowering the cost and time needed for hackers to identify and exploit vulnerabilities, increasing the risk of systemic financial shocks. "'Extreme cyber-incident losses could trigger funding strains, raise solvency concerns, and disrupt broader markets,' the IMF said. "The warning comes amid growing concern over powerful new AI systems capable of autonomously identifying software weaknesses and launching complex cyberattacks."

In general using abstract terms please explain how do you look for something when you don't know what you're looking for

✨AI Mode  Searching for the unknown is an iterative process of pattern recognition and boundary definition. Since you lack a specific target, the search shifts from "matching" to "mapping." (1)  Here is how that process unfolds in abstract terms: * Broad Casting: You begin with a high-level heuristic or a vague intuition. This acts as a wide-net "signal" sent into a field of data to see what resonates. * Contrast and Differentiation: You observe what is not relevant. By identifying the qualities of things that clearly don't fit your internal state, you begin to define the negative space around the missing object. * Iterative Refinement: As you encounter various "nodes" of information, you look for recurring clusters or anomalies. Each interaction provides a data point that narrows the field, transforming a "vague feeling" into a "defined set of attributes." * Associative Leaping: You follow threads of similarity. If a piec...

Haotian AI

"Haotian AI marks the next stage in deepfake scams and fraud, one that the public and tech companies may not be ready for, where criminals are able to change their appearance in real time to trick people, including Americans, into handing over their money. "Romance scams, tax fraud, virtual kidnappings: all stand to be amplified by live deepfake software which continues to improve in quality. " 404 Media ’s experimentation with Haotian AI marks the first time a journalist has managed to test this software to see how it really works, how effective it is, and what its existence means for the present and very near future of scams.  "Our investigation finds Haotian AI demonstrates its tool as a way to impersonate at least one U.S. police department.  "We link Haotian AI to Chinese money laundering networks and the ecosystem providing services to massive scam compounds in South East Asia, and find that Haotian AI has brought in more than $4 million dollars for its c...

Vibe Security Radar

"AI coding tools are writing a growing share of production code. "Some of it ships with security vulnerabilities.  "We track the cases where vulnerable code in public advisories (CVEs, GHSAs, RustSec, and others) was authored by an AI tool. "This is a research project from Georgia Tech SSLab ( Systems Software & Security Lab, School of Cybersecurity and Privacy ).  "Our goal is to understand how AI-assisted development affects software security, not through benchmarks or synthetic tasks, but by studying real vulnerabilities that were reported and fixed in the wild."

Where's the recycling?

"When Elon Musk brought xAI to Memphis, Tennessee, the company made a promise to locals worried that its data center would drain the local water supply: "They would build a state-of-the-art water recycling plant, a national model for environmental best practices. "Two years later, Musk’s first data center dedicated to his AI chatbot is up and running, but construction has come to a screeching halt at the promised water recycling plant, designed to clean municipal wastewater for use in cooling the superpowered computing center." 

Winding down copilot ✨

"Xbox is 'winding down Copilot on mobile' and 'will stop development of Copilot on console,' new Xbox CEO Asha Sharma announced on Tuesday .  "The move follows Sharma’s reorganization of the Xbox platform team earlier on Tuesday, which added executives from Microsoft’s CoreAI team —where Sharma worked before taking over Xbox —to the Xbox side of the company. "Sharma took over from former Microsoft Gaming CEO Phil Spencer in February and has already made some big changes, including scrapping the Microsoft Gaming brand and cutting the price of Xbox Game Pass."

weights.bin

Image
"Google Chrome is reaching into users' machines and writing a 4 GB on-device AI model file to disk without asking.   "The file is named weights.bin. It lives in OptGuideOnDeviceModel. It is the weights for Gemini Nano, Google's on-device LLM. Chrome did not ask. Chrome does not surface it. If the user deletes it, Chrome re-downloads it. "At Chrome's scale, the climate bill for one model push, paid in atmospheric CO2 by the entire planet, is between six thousand and sixty thousand tonnes of CO2-equivalent emissions, depending on how many devices receive the push. "That is the environmental cost of one company unilaterally deciding that two billion peoples' default browser will mass-distribute a 4 GB binary they did not request."

Chat AI: 27

Image

Direct evidence using AI detection

"The diffusion of LLMs from 2022 to 2025 tripled new book releases.   "While average book quality, measured by usage, declined, the surge in releases raised the number of modest-quality books.  "Direct evidence using AI detection shows that AI-containing books have lower quality, and their rising share —topping half of 2025 releases —drives the overall decline.  "A nested logit calibration shows that AI books raised consumer surplus by seven percent in 2025.  "Author selection accounts for most of the AI quality differential, and the AI-human differential shrinks over time.  "Finally, AI has not displaced authors active prior to LLMs."