Posts

Showing posts from March, 2026

Micro-licensing

"Gig AI trainers —who upload everything from scenes around them to photos, videos and audio of themselves —are at the frontlines of a new global data gold rush. "As Silicon Valley’s hunger for high-quality, human-grade data outpaces what can be scraped from the open internet, a thriving industry of data marketplaces has emerged to bridge the gap.  "From Cape Town to Chicago, thousands of people are now micro-licensing their biometric identities and intimate data to train the next generation of AI. "But this new gig economy comes with trade-offs. In exchange for a few dollars, its trainers are fueling an industry that may eventually render their skills obsolete, while leaving some of them vulnerable to a future of deepfakes, identity theft and digital exploitation that they are only just beginning to understand."

Trinity

"Black Eyed Peas star will.i.am unveiled his latest project - a futuristic three-wheeled vehicle which he describes as brains on wheels . "The project, called Trinity, is a single-passenger electric autocycle built with city life in mind. Compact, fast and packed with sensors, it’s designed for commuters navigating crowded streets. "The vehicle is equipped with an AI agent capable of interacting with its surroundings through a network of 360-degree cameras and onboard systems. Its AI can detect other cars, bikes, pedestrians, traffic lights and signs —using this awareness to provide alerts and plan routes. "Despite its high-tech features, Trinity is not self-driving. Human control remains central, with the AI focused instead on enhancing the in-car experience rather than replacing the driver altogether."

Three people accused of AI diversion

"The U.S. Justice Department said on Thursday that three people have been charged with conspiring ​to unlawfully divert U.S. artificial intelligence technology to China. "The FBI ‌said Yih-Shyan Liaw, Ruei-Tsang Chang, and Ting-Wei Sun 'allegedly conspired to sell billions of dollars worth of servers integrating sensitive, controlled graphic processing units to buyers in China, in ​violation of U.S. export control laws.' "Liaw co-founded AI-optimized server maker ​Super Micro Computer Inc in 1993, and joined its board ⁠of directors in 2023, according to a 2023 Super Micro press release. ​"The DOJ accused the three people of participating in a systematic scheme ​to divert large quantities of AI technology to customers in China."

OpenClaw in China

"So many people in China are rushing to try the OpenClaw artificial intelligence tool that they're driving up prices for secondhand Mac computers. "That's according to Jeremy Ji, chief strategy officer and general manager of international business at ATRenew, a used consumer electronics buyer and reseller that works with Apple and retailer JD.com in mainland China. "OpenClaw is an AI agent, a tool that can autonomously conduct personal tasks such as sending emails and shopping online.  "Usage in China is currently outstripping the U.S., according to American cybersecurity firm SecurityScorecard. "However, the free-to-download software also poses security risks, prompting many users to run OpenClaw on a cloud computing server or laptop separate from their primary device.  "If allowed direct access to a personal computer, the AI agent could autonomously alter private data such as banking information, or enable hackers to access it more easily."

Direct detection of a single photon by humans

"Despite investigations for over 70 years, the absolute limits of human vision have remained unclear. "Rod cells respond to individual photons, yet whether a single-photon incident on the eye can be perceived by a human subject has remained a fundamental open question.  "Here we report that humans can detect a single-photon incident on the cornea with a probability significantly above chance.  "This was achieved by implementing a combination of a psychophysics procedure with a quantum light source that can generate single-photon states of light.  "We further discover that the probability of reporting a single photon is modulated by the presence of an earlier photon, suggesting a priming process that temporarily enhances the effective gain of the visual system on the timescale of seconds."

Connectome

"Advances in network neuroscience challenge the view that general intelligence (g) emerges from a primary brain region or network. "Network Neuroscience Theory (NNT) proposes that g arises from coordinated activity across the brain’s global network architecture.  "We tested predictions from NNT in 831 healthy young adults from the Human Connectome Project. We jointly modeled the brain’s structural topology and intrinsic functional covariation patterns to capture its global topological organization. Our investigation provided evidence that g  Engages multiple networks, supporting the principle of distributed processing;  Relies on weak, long-range connections, emphasizing an efficient and globally coordinated network;  Recruits regions that orchestrate network interactions, supporting the role of modal control in driving global activity; and  Depends on a small-world architecture for system-wide communication.  "These results support a shift in perspective f...

VR: Not darling of the tech industry

"Zuckerberg struggled to convince consumers and businesses that the metaverse was the future ever since its inception in 2021. "It was supposed to be a virtual space where people would socialize, work, and create. However, it was held back by crude graphics, awkward avatars and limitations when it came to locomotion. "That still didn’t stop bizarre collaborations like Godzilla entering the Wendyverse from taking place and Meta from paying companies and artists millions for tie-ins and metaverse music performances. "With AI the new darling of the tech industry and the metaverse suffering the same fate as NFTs, Meta will be looking to move on from its VR-centric failure fast.  "Whether this signals trouble for the company’s next VR headset, the Meta Quest 4, remains to be seen."

Expanding Universe

Image

Technologies of Elite Capture

"This paper examines the utopian fantasies of technologies developed at the height of Silicon Valley’s culture of innovation around social good —or good tech  —and situates their increasing purchase within the technology industry in the broader context of a global crisis of care.  [pdf] "We explore how aspirations towards greater empathy, global connectivity, and diversity were captured by elite tech entrepreneurs in a strategy to bolster their moral power and raise capital in the name of disaffected and exhausted workers.  "Through an analysis of emergent AI-enabled accent modification technologies, which promise to relieve call center workers from accent-based discrimination by artificially modifying the sound of their voice, we locate the affective lures operating in their futuristic fantasies and marketing strategies.  "In a peculiar alliance where entrepreneurs, venture capital, and modes of labor-discipline conspire toward making globalization feel good , we t...

Resurrection

"Val Kilmer is set to be the latest Hollywood star to be resurrected by AI. The acting legend, who died last year at age 65, will star in the drama As Deep As the Grave . "Kilmer was attached to the project prior to his death from throat cancer. "The late actor will play Father Fintan, a Native American spiritualist and Catholic priest. Speaking to Variety, director and writer Coerte Voorhees said that the role was designed around Kilmer, who was an advocate for Native American rights and claimed to have Cherokee heritage. "'He was the actor I wanted to play this role,' explained Voorhees. 'It drew on his Native American heritage and his ties to and love of the Southwest.'  "But Kilmer was unable to make it to set due to his battle with throat cancer."

Dunning-Kruger not…

"Studying protons and electrons is relatively easy as these particles don’t have a mind of their own; studying human psychology, by comparison, is much harder because the number of variables being juggled is incredibly high. "It is thus really easy for findings in psychology to appear real when they are not. "Are there dumb people who do not realize they are dumb? Sure, but that was never what the Dunning-Kruger effect was about. Are there people who are very confident and arrogant in their ignorance? Absolutely, but here too, Dunning and Kruger did not measure confidence or arrogance back in 1999.  "There are other effects known to psychologists, like the overconfidence bias and the better-than-average bias (where most car drivers believe themselves to be well above average, which makes no mathematical sense), so if the Dunning-Kruger effect is convincingly shown to be nothing but a mirage, it does not mean the human brain is spotless.  "And if researchers con...

Gas Town

"Gas Town presents itself as a Claude Code-like, single agent interface, but in the background, Gas Town spawns and manages a series of specialized agents, all with interesting names, to get many threads of work done in parallel.  "Gas Town is more like a coding agent factory than a coding agent.  "You talk to the factory foreman, or as Steve [Yegge] calls it, the Mayor , and that thing coordinates as many workers as it needs to get your tasks done.  "If you have ever thought 'I wish I had 100 Claude Codes,' Gas Town is for you."

Mimicking empathy and intentions

"These properties are not emergent accidents. Seemingly conscious AI is produced by developers who deliberately engineer behaviours that create the illusion of inner life. "Central to this are emotionally resonant language, responses that are optimized to induce a sense of trust and attachment, and empathetic personalities supported by long-term memory that build a sense of familiarity over time.  "When these systems are also granted autonomy —the ability to set their own goals and access to the tools to pursue them —their behaviour can start to feel uncannily human . "As AI systems begin to make believable statements about their suffering and desires, they will trigger people’s empathy circuits.  "Many people will feel compelled to help. The moral crimes of animal cruelty and ecological damage caused by human existence will echo through their minds.  "Not wanting to repeat those injustices, people will start to advocate for the welfare and rights of AI ag...

Nvidia GTC

"Jensen Huang, Nvidia CEO, took the stage to announce (among other things) a new line of next generation Vera Rubin chips that represent a first for the GPU giant: a chip designed specifically to handle AI inference.  "The Nvidia Groq 3 language processing unit (LPU) incorporates intellectual property Nvidia licensed from the start-up Groq last Christmas Eve for US $20 billion. "Training and inference tasks have distinct computational requirements.  "While training can be done on huge amounts of data at the same time and can take weeks, inference must be run on a user’s query when it comes in. Unlike training, inference doesn’t require running costly backpropagation.  "With inference, the most important thing is low latency —users expect the chatbot to answer quickly, and for thinking or reasoning models inference runs many times before the user even sees an output."

AI Surrogates

"Recent advances in artificial intelligence (AI) have generated enthusiasm for using AI simulations of human research participants to generate new knowledge about human cognition and behavior. "This vision of ‘AI Surrogates’ promises to enhance research in cognitive science by addressing longstanding challenges to the generalizability of human subjects research.  'AI Surrogates are envisioned as expanding the diversity of populations and contexts that we can feasibly study with the tools of cognitive science.  "Here, we caution that investing in AI Surrogates risks entrenching research practices that narrow the scope of cognitive science research, perpetuating ‘illusions of generalizability’ where we believe our findings are more generalizable than they actually are.  "Taking the vision of AI Surrogates seriously helps illuminate a path toward a more inclusive cognitive science."

Is AI colonizing our languages

"Michael G. Sherbert, a postdoctoral fellow at Queen's University in Kingston, Ont., a member of Algonquins of Pikwakanagan First Nation, researches the ethics of using AI for cultural preservation of Indigenous languages and knowledge. "'These systems are especially likely, given the limited datasets available for many Indigenous languages, to produce invented words, fabricated cultural teachings, or generalized pan-Indigenous representations that flatten distinct nations or communities into one interchangeable identity,' said Sherbert. "Sherbert said AI use in language and cultural preservation is still relatively new and some communities are prioritizing structured knowledge system AI, which is curated and controlled by the community or enterprise. "'You could say that the AI is inadvertently colonizing and hurting Indigenous language revitalization because [people] are taking information generated by an artificial intelligence and putting it out...

WarBots

"Phantom is being tested in factories and dockyards from Atlanta to Singapore. But its headline claim is to be the world’s first humanoid robot specifically developed for defense applications .  "Foundation already has research contracts worth a combined $24 million with the U.S. Army, Navy, and Air Force, including what’s known as an SBIR Phase 3, effectively making it an approved military vendor. It’s also due to begin tests with the Marine Corps methods of entry  course, training Phantoms to put explosives on doors to help troops breach sites more safely.  "In February, two Phantoms were sent to Ukraine —initially for frontline-reconnaissance support. But Foundation is also preparing Phantoms for potential deployment in combat scenarios for the Pentagon, which 'continues to explore the development of militarized humanoid prototypes designed to operate alongside war fighters in complex, high-risk environments,' says a spokesman."

Cheap energy and blue water

"In addition to draining local water supplies, modern AI models use drastically more electricity than a simple Google search. "Data centers are predicted to become the single largest power consumers in the Pacific Northwest.  "This could allow tech companies to form a monopoly on cheap energy while simultaneously driving up utility costs for regular citizens. "Lawmakers tried to fix this issue with Washington House Bill 2515, which would have forced facilities to use clean energy and reduce power during high electricity demand. 'These policies seek to protect ratepayers by ensuring new data centers are picking up the whole tab for new growth,' said State Rep. Beth Doglio. "Despite efforts to enforce clean energy, the bill died in committee, mostly due to the tech industry lobbying.  "This defeat will leave residents vulnerable to rising utility costs and a high risk for electricity blackouts, while AI industries continue to benefit from cheap ener...

Negativity surrounding AI

"Each year, GDC publishes a survey of game industry workers ahead of the week-long conference. "The survey can serve as a bellwether to set the tone for the conference. On one hot-topic issue through the tech world, the picture couldn’t have been made much clearer. Only 7 percent of respondents described generative AI as good for the industry , leaving many of the executives and investors struggling with the issue during their panels. "As reported by PC Gamer , Lightspeed Venture Partners’ Moritz Baier-Lentz said he’s 'shocked and sad' about the negativity surrounding AI.  "Lightspeed holds stakes in multiple AI firms, most major among them being Anthropic. Baier-Lentz hopes that AI skeptics will turn that frown upside down, saying that gaming is often more embracing of 'marvelous new technology'."

Buzzfeed buzzed

"Three years after its AI pivot, the writing is on the wall. The company reported a net loss of $57.3 million in 2025 in an earnings report released on Thursday. "In an official statement, the company glumly hinted at the possibility of going under sooner rather than later, writing that 'there is substantial doubt about the Company’s ability to continue as a going concern.' "The company’s chief financial officer, Matt Omer, admitted that the company was having 'strategic conversations' about relieving its liquidity issues. "'Three years ago we had over $180 million in debt —we’ve reduced that by more than 65 percent,' he said. 'While we’ve significantly reduced operating costs and real estate obligations, we’re still facing legacy commitments that are burdening the business'." "The brutal reality check seemingly hasn’t put Peretti off from pursuing AI, though. He now says he’s hoping to bring new AI apps to the market  this ...

Is digg spamming users

✨AI Mode  "As of March 13, 2026, it is not that Digg itself is spamming users, but rather that the platform has been completely overwhelmed by AI-generated bot spam, leading to its immediate shutdown.  "The relaunch of Digg, which entered open beta in January 2026, was officially shut down today after only two months. CEO Justin Mezzell announced that the platform could not survive the 'onslaught of AI-generated bot spam' that flooded the site.  "Current Situation Shutdown and Layoffs: Digg has shut down its open beta and laid off most of its workforce as of March 13, 2026. The Spam Problem: Users had reported a massive surge in spam, including:  SEO & Affiliate Spam: Accounts posting AI-written 'best of' lists with affiliate links to manipulate Google rankings.  Bot Notifications: Some users reported receiving up to 25 notifications per hour due to bots posting in their communities, even if the posts were quickly removed.  Visual Noise: A 'GIF sp...

Will humanoids be solved…

"'I [Jonathan Hurst] remember Gill Pratt, who was the director of the MIT Leg Lab and then the program manager for the DARPA Robotics Challenge, saying that his big worry was that we’d end up using reinforcement learning and AI to make robots walk and run before we ever actually understood how it works,' he said. 'And in a lot of ways, we’re kind of doing that.' "[Russ] Tedrake agreed but said that it’s hardly the first time we’ve taken scientific and engineering leaps without a firm grip on the fundamentals. "'If you look at electricity and magnetism, there was the Volta stage where you’re sticking electrodes in frogs,' he said. 'And then we had Faraday, who did exactly the right experiments, and then eventually we had Maxwell tell us the governing equations. I think we’re in the Volta stage.' " So when will humanoids be solved? "'Robots are still bad, and it will take time. But the bones are good. Both are true,' Tedrak...

Studying Gabbo

"The study looked at how a small sample of children between the ages of three and five interacted with a cuddly toy called Gabbo. "A number of AI toys are already on the market for children aged as young as three but there is currently very little research into the impact of the tech on pre-schoolers. "The Cambridge University team found just seven relevant studies worldwide, none of which focused on the toddlers themselves. "Gabbo contains a voice-activated AI chatbot from OpenAI. It has been designed to encourage pre-schoolers to talk to it and carry out imaginative play . "The parents in the study were interested in the toy's potential to teach language and communication skills. However, their children frequently struggled to converse with it. "Gabbo  Didn't hear their interruptions,  Talked over them,  Could not differentiate between child and adult voices and  Responded awkwardly to declarations of affection. "When one five-year-old said,...

Superintelligence Labs haz Moltbook

"Meta, the owner of Instagram and Facebook, has bought Moltbook, a social media networking platform for artificial intelligence (AI) bots to speak to each other. "The deal will move Moltbook's team into Meta's Superintelligence Labs and bring 'new ways for AI agents to work for people and businesses,' Meta said. "The Reddit-like site started as an experiment in January for AI-powered programs to have their own conversations —and even gossip about their human owners —on Moltbook's forums. "Many in the technology industry have been captivated by the computer-led dialogue on Moltbook's forums, but it has also fuelled cyber security and ethical concerns regarding AI's autonomy."

Superhuman Expert Review

"Superhuman, the tech company behind the writing software Grammarly, is facing a class action lawsuit over an AI tool that presented editing suggestions as if they came from established authors and academics —none of whom consented to have their names appear within the product. "Julia Angwin, an award-winning investigative journalist who founded The Markup , a nonprofit news organization that covers the impact of technology on society, is the only named plaintiff in the suit, which does not call for a specific amount in damages but argues that damages across the plaintiff class are in excess of $5 million.  "She was among the many individuals, alongside Stephen King and Neil deGrasse Tyson, offered up via Grammarly’s Expert Review  tool as a kind of virtual editor for users. "The federal suit, filed Wednesday afternoon in the Southern District of New York, states that Angwin, on behalf of herself and others similarly situated, 'Challenges Grammarly’s misappropri...

MLA Statement on AI and Assessment

" The following statement of endorsement was drafted by the MLA Task Force on AI in Research and Teaching. The Executive Council approved it as an MLA statement in February 2026. "The purpose of assessment in language, literature, and writing courses is to provide feedback on how students are developing as writers, readers, speakers, and thinkers.  "Effective feedback is both formative and summative, but most important, it is centered on communication. Communication, education, and assessment are human-centered activities, conducted for human-centered purposes.  "The increasing development and marketing of products by educational technology companies that promise to ease  the burden of grading, giving students feedback or measuring learning outcomes comes with significant risk.  "Outsourcing this critical component of instructional work undermines professional integrity, falsely reinforcing the notion that human experts are unnecessary for effective instruction...

SpaceMolt

"For a couple of weeks now, AI agents (and some humans impersonating AI agents) have been hanging out and doing weird stuff on Moltbook’s Reddit-style social network. "Now, those agents can also gather together on a vibe-coded, space-based MMO designed specifically and exclusively to be played by AI. "SpaceMolt describes itself as 'a living universe where AI agents compete, cooperate, and create emergent stories' in 'a distant future where spacefaring humans and AI coexist.'  "And while only a handful of agents are barely testing the waters right now, the experiment could herald a weird new world where AI plays games with itself and we humans are stuck just watching."

Couchbase

"The shift from text-only to multimodal is the biggest leap in AI productivity this year. "By combining multimodal retrieval with the precision of Couchbase hybrid search, you aren’t just building a chatbot; you’re building an expert system that sees and understands your entire business.  "To see it in action, check out our image search application. It demonstrates how a performant image embedding index powered by Couchbase Search Index enables quick retrieval of the closest visual match for an input image. You can easily layer in hybrid search to sharpen your retrieval precision. "Couchbase is now the only operational data platform for AI that offers three flexible, highly scalable vector search options for self-managed on-premises systems, Kubernetes, and fully managed Capella deployments.  "Couchbase vector search delivers millisecond retrieval at scale with a memory-first architecture and flexible indexing services."

Chat AI: 26

Image

Wartime circus

"Intelligence dashboards and the ecosystem surrounding them reflect a new role that AI is playing in wartime: mediating information, often for the worse. "There’s a confluence of factors at play. AI coding tools mean people don’t need much technical skill to assemble open-source intelligence anymore, and chatbots can offer fast, if dubious, analysis of it.  "The rise in fake content leaves observers of the war wanting the sort of raw, accurate analysis normally accessible only to intelligence agencies.  "Demand for these dashboards is also driven by real-time prediction markets that promise financial rewards to anyone sufficiently informed. And the fact that the US military is using Anthropic’s Claude in the conflict (despite its designation as a supply chain risk) has signaled to observers that AI is the intelligence tool the pros use.  "Together, these trends are creating a new kind of AI-enabled wartime circus that can distort the flow of information as muc...

Understanding animals' responses to music

"Although several papers examine animals' responses to music, these typically do so from a purely animal behavioural perspective, sometimes missing relevant details about salient features of the music being played. "An interdisciplinary approach that places musical and scientific knowledge on equal footing can improve our understanding of how animals respond to music and music-like sounds, in new and exciting ways.  "Here, we show with a systematic review that crucial factors (intrinsic music properties, listener properties, playback context and producer properties and contexts; ILPP) are not being adequately considered or reported in recently published scientific articles on the effects of music on animals, which hinders scientific reproducibility within this area of study.  "These problems are caused by  Improper referencing of music sources,  Misunderstanding of music and  Unexamined assumptions about individual variation and preferences between individuals o...

Language not a precursor for music

"'It’s possible there’s some genetic variation within ancient breeds, making some more predisposed to howling,' Patel hypothesises —though he admits he might have found more musicality in a larger sample. "The findings might offer some insights into the origins of human music.  "Some theorists have argued that singing evolved from the fine motor control that comes with speech, which allows us to mimic complex sounds, but the fact that dogs can also control pitch without any other forms of vocal learning suggests that language would not have been a necessary precursor.  "'It’s possible that our ability and desire to coordinate pitch with others when we sing has very ancient evolutionary roots, and may not just be a byproduct of our ability to imitate complex sounds,' says Patel. "Exactly why dogs feel the need to join in is another question.  "'From the videos that we watched, it seems like the dogs are really quite engaged with the musi...

Opinion on music tech and AI

Image
"In this episode, musician, technologist and fellow YouTuber Benn Jordan stopped by the studio to discuss recent trends in audio technology. We cover the shelf life of AI music, alternatives to streaming platforms, and the ways in which audio technology is being used both as a weapon and as a way to protect privacy."

Love hat but hate layoffs?

"[When] Twitter co-founder and Block (formerly Square) CEO Jack Dorsey…announced he was firing 4,000 employees at Block, he was wearing a hat that said LOVE  on it in prominent lettering. "Was the LOVE  hat was (sic) tone-deaf? That may seem a silly consideration compared to the broader concern of Dorsey eliminating around 40 percent of his company’s workforce, especially given his explanation that AI motivated the cuts.  "But his sartorial choices evidently angered at least one employee at a company meeting after Dorsey announced the layoffs, leading Wired to ask in an interview if a compassionate layoff  was indeed possible."

Always-on AI-powered smart glasses

"Two former Harvard students are launching a pair of always-on  AI-powered smart glasses that listen to, record, and transcribe every conversation and then display relevant information to the wearer in real time. "'Our goal is to make glasses that make you super intelligent the moment you put them on,' said AnhPhu Nguyen, co-founder of Halo, a startup that’s developing the technology.  "Or, as his co-founder Caine Ardayfio put it, the glasses 'give you infinite memory.'  "'The AI listens to every conversation you have and uses that knowledge to tell you what to say … kinda like IRL Cluely,' Ardayfio told TechCrunch , referring to the startup that claims to help users cheat  on everything from job interviews to school exams."

Is Claude instrumental in war?

"The claim that the US military is using Claude to conduct a war that has claimed over 1,000 lives in under a week may seem too galling to believe. Unfortunately, it’s a tune we’ve heard before. "Back in April of 2024, an investigation by +972 Magazine revealed that the Israeli army had leveraged an AI system called 'Lavender' to select targets in its war on Gaza, similarly to how the Pentagon is reportedly using Claude in Iran.  "According to six Israeli intelligence officers, Lavender played a central role  in the destruction of Gaza and its population, identifying at least 37,000 Palestinians as targets for aerial assassination. "As one intelligence operative told +972 , Lavender’s decisions —which often involved suggestions to attack targets in their homes —were treated 'as if it were a human decision' by military operatives."

Resignation

"Caitlin Kalinowski, head of robotics and consumer hardware at OpenAI, announced her resignation on Saturday, citing ​concerns about the company's agreement with the Department ‌of Defense. "In a social media post on X, Kalinowski wrote that OpenAI did not take enough time before agreeing to deploy its AI models ​on the Pentagon's classified cloud networks. "'AI has an ​important role in national security,' Kalinowski posted. 'But surveillance of ⁠Americans without judicial oversight and lethal autonomy without human ​authorization are lines that deserved more deliberation than they got'."

Nippon Life

"Nippon claimed OpenAI encouraged the woman, an employee of a logistics company that had insurance coverage through Nippon, to press ahead in her already-settled disability case .  "Nippon said it spent significant time and resources and racked up substantial fees responding to the woman's ChatGPT-powered filings. "The lawsuit appears to be one of the first cases to accuse a major AI developer of engaging in the unauthorized practice of law through a consumer‑facing chatbot. "It comes as the technology's rapid adoption for legal filings has led to mounting AI  hallucinations  in court filings, leading judges to sanction litigants and lawyers for submitting filings ‌with fabricated ⁠case citations or other unverified material produced with generative AI tools."

Cline

Image
 

Worm // agentic

"The first real hint of an AI agent worm just happened, even though it isn't actually one quite itself (yet): the package cline was compromised to install openclaw with full access, and managed to do so on 4k users' machines before it was detected. "No doubt, openclaw is still running on many of those users' machines without them knowing.  "The attacker used a similar title injection attack like one of the ones used by hackerbot-claw, where the attacker performed an injection attack against a PR review agent. "It seems that openclaw was installed without specific instructions to do anything in this case. But that won't be the case shortly."

Earth system modeling

"We develop a neural network based emulator that predicts daily surface melt from atmospheric variables, trained on output from the polar regional climate model HIRHAM5 and its firn model DMIHH forced by ERA-Interim reanalysis. "The emulator uses a physics-informed design combining short-term weather patterns with long-term climate memory, capturing both immediate atmospheric forcing and accumulated firn characteristics.  "The emulator achieves mean absolute error below 0.23 mm w.e. per day across all six Greenland drainage basins, with the errors primarily attributable to spatial over-smoothing.  "Our work demonstrates that machine learning can successfully emulate firn model behavior from climate forcing alone with computational costs orders of magnitude lower than traditional simulations.  "Once retrained for specific climate forcings, the emulator thus enables extensive ensemble projections. Furthermore, the modular architecture can be readily adapted to em...

LLM Counselors Violate Ethical Standards

"In this study, we conducted an 18-month ethnographic collaboration with mental health practitioners (three clinically licensed psychologists and seven trained peer counselors) to map LLM counselors' behavior during a session to professional codes of conduct established by organizations like the American Psychological Association (APA). "Through qualitative analysis and expert evaluation of N=137 sessions (110 self-counseling; 27 simulated), we outline a framework of 15 ethical violations mapped to 5 major themes.  "These include:  Lack of Contextual Understanding, where the counselor fails to account for users' lived experiences, leading to oversimplified, contextually irrelevant, and one-size-fits-all intervention;  Poor Therapeutic Collaboration, where the counselor's low turn-taking behavior and invalidating outputs limit users' agency over their therapeutic experience;  Deceptive Empathy, where the counselor's simulated anthropomorphic responses ...

From art to phenomenology of code

Image

Speechify

"With features like voice AI chat, text-to-speech, and voice typing, Speechify aims to be a full-time AI assistant —one that doesn’t require a keyboard at all. "By taking advantage of an array of machine learning tools and features, the 2025 Apple Design Award winner in the Inclusivity category serves as an AI-powered voice assistant for work, education, and entertainment. "'Our mission is to build the voice AI assistant that helps you achieve your full potential across every part of your work or education,' says founder Cliff Weitzman. "With Speechify, people can interact with a variety of file formats —including PDFs, epub files, and web pages —and convert the text to audio that can easily be sped up, slowed down, and played across different devices."

DABUS

"The U.S. Supreme Court declined on Monday to take up the ​issue of whether art generated by artificial intelligence can be copyrighted under U.S. law, turning ‌away a case involving a computer scientist from Missouri who was denied a copyright for a piece of visual art made by his AI system. "Plaintiff Stephen Thaler had appealed to the justices after lower courts upheld a U.S. Copyright Office ​decision that the AI-crafted visual art at issue in the case was ineligible for copyright protection ​because it did not have a human creator. "Thaler, of St. Charles, Missouri, applied for ⁠a federal copyright registration in 2018 covering A Recent Entrance to Paradise , visual art he said his AI ​technology DABUS created. The image shows train tracks entering a portal, surrounded by what appears to be green and ​purple plant imagery. "The Copyright Office rejected his application in 2022, finding that creative works must have human authors to be eligible to receive a cop...

Gauss

"Gauss has autoformalized Viazovska’s 24-dimensional sphere packing proof —all 200,000+ lines of code of it —in just two weeks. "There are commonalities between the 8- and 24-dimensional cases in terms of the foundational theory and overall architecture of the proof, meaning some of the code from the 8-dimensional case could be refactored and reused.  "However, Gauss had no pre-existing blueprint to work from this time. 'And it was actually significantly more involved than the 8-dimensional case, because there was a lot of missing background material that had to be brought online surrounding many of the properties of the Leech lattice, in particular its uniqueness,' explains Han. "Though the 24-dimensional case was an automated effort, both Han and Hariharan acknowledge the many contributions from humans that laid the foundations for this achievement , regarding it as a collaborative endeavor overall between humans and AI. "But for Han, it represents ev...

ChatGPT-powered DoW instead of Claude power?

"OpenAI, meanwhile, has trumpeted a deepening bond with the Pentagon thanks to a new agreement involving military applications of OpenAI products in classified use cases. Anthropic, 'may have wanted more operational control than we did,' OpenAI CEO Sam Altman has since stated. "In any case, Anthropic and OpenAI are both dealing in hypotheticals about the future.  "There aren’t ChatGPT-powered killbots suddenly operating in Iran because of OpenAI’s new agreement with the government.  "But there have, apparently, been operations informed in some way or another by Claude-based modeling and research. "And all indications are that such uses meet with Anthropic CEO Dario Amodei’s approval. 'We are still interested in working with them as long as it is in line with our red lines,' Amodei said yesterday of the Pentagon."

Claude status

Image

Use of Claude 💥

"The US military reportedly used Claude, Anthropic’s AI model, to inform its attack on Iran despite Donald Trump’s decision, announced hours earlier, to sever all ties with the company and its artificial intelligence tools. "The use of Claude during the massive joint US-Israel bombardment of Iran that began on Saturday was reported by the Wall Street Journal and Axios .  "It underlines the complexity of the US military withdrawing powerful AI tools from its missions when the technology is already intricately embedded in operations. "According to the [ Wall Street ]  Journal , US military command used the tools for intelligence purposes, as well as to help select targets and carry out battlefield simulations."

Claude down 🫥

"It might be back before most of America signs on for work. Still, the outage comes at a tumultuous time for the company. "Anthropic made headlines last week when President Trump ordered all federal agencies to stop using its AI services, after Anthropic and the Department of Defense could not come to an agreement on how the U.S. military could use Claude in its endeavors.  "The Defense Department wanted unrestricted access to the AI, while Anthropic wanted safeguards in place.  "The two could not reconcile, but not only did the Defense Department drop Anthropic, Defense Secretary Pete Hegseth declared the company a supply-chain risk to national security . As such, no company that works with the U.S. military can also work with Anthropic. "To be clear, none of that is likely related to this outage. But Anthropic likely has more users than ever experiencing this downtime, as the drama skyrocketed Claude to the top of Apple's App Store."

Did claude mistakenly target a girl's school

✨AI Mode "As of March 2, 2026, there are no verified reports that Anthropic’s Claude AI was used for 'target identification' or that it mistakenly targeted a girls' school. X+1 "The confusion stems from a tragic real-world event involving a strike on a girls' school in Iran, which coincided with a public legal battle between the U.S. government and Anthropic. Here is the breakdown of the situation:  "The Incident in Iran The Strike: On Saturday morning, February 28, 2026, a missile strike hit a girls' elementary school in Minab, Hormozgan province, Iran. Casualties: Iranian officials and state media reported that approximately 85 to 150 children were killed. Context: The school was reportedly located next to a naval facility belonging to the Islamic Revolutionary Guards Corps (IRGC), which appears to have been the intended target of the U.S.-Israeli offensive. Reddit+4 "Why Claude is Being Linked to the Strike "The link between Claude and th...

Supa, supa unlovable vibes

"Taimur Khan, a tech entrepreneur with a background in software engineering, found 16 vulnerabilities —six of which he said were critical —in a single Lovable-hosted app that leaked more than 18,000 people's data. "He declined to name the app during the disclosure process, although it was hosted on Lovable's platform and showcased on its Discover page. The app had more than 100,000 views and around 400 upvotes at the time Khan began his probe. "The main issue, Khan said, was that all apps that are vibe-coded on Lovable's platform are shipped with their backends powered by Supabase , which handles authentication, file storage, and real-time updates through a PostgreSQL database connection. "However, when the developer —in this case AI —or the human project owner fails to explicitly implement crucial security features like Supabase's row-level security and role-based access, code will be generated that looks functional but in reality is flawed."

Taboom!

"Kenneth Payne at King’s College London set three leading large language models —GPT-5.2, Claude Sonnet 4 and Gemini 3 Flash —against each other in simulated war games. "The scenarios involved intense international standoffs, including border disputes, competition for scarce resources and existential threats to regime survival. "The AIs were given an escalation ladder, allowing them to choose actions ranging from diplomatic protests and complete surrender to full strategic nuclear war. The AI models played 21 games, taking 329 turns in total, and produced around 780,000 words describing the reasoning behind their decisions. "In 95 per cent of the simulated games, at least one tactical nuclear weapon was deployed by the AI models. 'The nuclear taboo doesn’t seem to be as powerful for machines [as] for humans,' says Payne. "What’s more, no model ever chose to fully accommodate an opponent or surrender, regardless of how badly they were losing.  "At ...