The radical Blog
About Archive radical✦
  • Not So Fast, Baby

    US grocery giant Kroger is clawing back on an initiative to build out its network of robotic-warehouse-powered delivery services. Not because the technology doesn’t work (it does – Kroger was using the UK’s grocer Ocado’s proven robots), but because US consumers demand instant delivery.

    With its automated fulfillment network, Kroger bet that consumers would be willing to trade delivery speed for sensible prices on grocery orders. That model has been highly successful for Ocado in the U.K., but U.S. consumers have shown they value speed of delivery, with companies like Instacart and DoorDash expanding rapidly in recent years and rolling out services like 30-minute delivery.

    It goes to show that it’s not just technology that makes or breaks a business model.

    ↗ Kroger acknowledges that its bet on robotics went too far

    → 8:39 AM, Dec 10
    Also on Bluesky
  • How Do You REALLY Feel About AI

    The latest Pew Research Center study on consumer sentiment toward AI is quite eye-opening: 43% of surveyed Americans expect that AI will harm them, while only 23% of Americans (outside of the AI expert population) believe that AI will have a positive impact on their jobs. Unsurprisingly, 76% of AI experts believe AI will benefit them. It appears there is considerable convincing left to do.

    Meanwhile, in Edelman’s 2025 Trust Barometer, a whopping 54% of Chinese survey participants “embrace AI,” compared to only 17% in the US, with similar numbers for Germany and the UK. Assuming that AI will actually prove to be a significant driver of economic growth, it doesn’t bode well when your population is (strongly) averse to the technology.

    ↗ How the U.S. Public and AI Experts View Artificial Intelligence ↗ 2025 Edelman Trust Barometer – Trust and Artificial Intelligence at a Crossroads

    → 3:38 PM, Dec 4
    Also on Bluesky
  • The MIT Iceberg Report

    MIT’s new Iceberg Index shows that today’s AI is already capable of doing work equal to nearly 12% of all U.S. wages, and most of that impact is hidden in plain sight beneath a narrow focus on tech jobs (hence the “iceberg” analogy). The important (and new) bit of this study is this:

    “[…] with cascading effects that extend far beyond visible technology sectors. When AI automates quality control in automotive plants, consequences spread through logistics networks, supply chains, and local service economies. Yet traditional workforce metrics cannot capture these ripple effects: they measure employment outcomes after disruption occurs, not where AI capabilities overlap with human skills before adoption crystallizes.”

    Sober reading.

    ↗ The Iceberg Index: Measuring Skills-centered Exposure in the AI Economy (and study)

    → 3:25 PM, Dec 4
    Also on Bluesky
  • It’s Energy, Not Compute, Baby

    Not necessarily a new insight, but one which might be worth repeating – in the data center rollout race, it is (now) much less about GPUs (or TPUs), but rather access to power that provides the bottleneck. Microsoft’s CEO recently:

    “The biggest issue we are now having is not a compute glut, but it’s power,” Nadella said. “It’s not a supply issue of chips. It’s actually the fact that I don’t have warm shells to plug into.” The remarks referred to data centers that are incomplete or lack sufficient energy and cooling capacity.

    If you are into the great US-China race, you might realize that it doesn’t bode well for the US, that China is massively outpacing the US in its energy buildup (with a lot of renewables, mind you)…

    ↗ Microsoft CEO Satya Nadella Admits ‘I Don’t Have Warm Shells To Plug Into’ — While OpenAI CEO Sam Altman Warns Cheap Energy Could Upend AI

    → 5:23 AM, Dec 3
    Also on Bluesky
  • Oh, the Irony

    Nature reported that a major AI conference was flooded by AI-generated peer reviews – irony aside, this presents a fairly troubling development: Science ought to be the place where we do real discovery, have honest discourse, and further our collective understanding. That is, not a place for AI slop.

    Pangram’s analysis revealed that around 21% of the ICLR peer reviews were fully AI-generated, and more than half contained signs of AI use.

    ↗ Major AI conference flooded with peer reviews written fully by AI

    → 9:24 AM, Dec 1
    Also on Bluesky
  • GLP-1 – The Forever Drug

    We have talked about GLP-1 weight-loss drugs here before – they seemingly came out of nowhere (at least in the public eye), have skyrocketed into a massive category, and promise to solve much more than just our weight issues. But they come with a massive downside (which, I am sure, big pharma won’t mind): you can’t get off them without losing all the benefits (and gains you made).

    An analysis published this week in JAMA Internal Medicine found that most participants in a clinical trial who were assigned to stop taking tirzepatide (Zepbound from Eli Lilly) not only regained significant amounts of the weight they had lost on the drug, but they also saw their cardiovascular and metabolic improvements slip away. Their blood pressure went back up, as did their cholesterol, hemoglobin A1c (used to assess glucose control levels), and fasting insulin.

    Another good reminder that there is no such thing as a free lunch.

    ↗ There may not be a safe off-ramp for some taking GLP-1 drugs, study suggests

    → 10:56 AM, Nov 26
    Also on Bluesky
  • Google CEO Puts Himself Out of a Job

    In one of the more remarkable “let’s just pretend AI is going to solve every problem” hand-waving statements, Google’s CEO Sundar Pichai made the bold assertion that AI will soon be able to do his job:

    “I think what a CEO does is maybe one of the easier things maybe for an AI to do one day,” he said. Although he didn’t talk specifically about CEO functions that an AI could do better, Pichai noted the tech will eliminate some jobs but also “evolve and transition” others—ramifications that mean “people will need to adapt.”

    The important part here is, “Although he didn’t talk specifically about CEO functions that an AI could do better […]” If only every problem in the world could be solved by making a vague statement and moving on. But hey, Sundar will soon have plenty of time to work on other things, as AI will steer his company.

    ↗ Google’s Sundar Pichai says the job of CEO is one of the ‘easier things’ AI could soon replace

    → 12:15 PM, Nov 25
    Also on Bluesky
  • The AI Insurance Conundrum

    Insurance companies are balking at insuring anything AI – from risks involving companies using AI to generate content, make decisions, or run processes, to the sheer idea of using AI in the first place.

    Major insurers including Great American, Chubb, and W. R. Berkley are asking U.S. regulators for permission to exclude widespread AI-related liabilities from corporate policies.

    Make no mistake – this is a huge issue not just for companies building AI models and AI-powered apps, but for any company using AI in their processes. Simply put, if you are using AI and something goes wrong (say, you are an accountant and your use of AI in accounting resulted in an error in your client’s tax return), you – not the software vendor, not the AI model your software vendor is using – are liable to your client. This could prove to be a major hurdle to overcome when it comes to the widespread adoption of AI-powered workflows and systems.

    ↗ AI is too risky to insure, say people whose job is insuring risk

    → 12:03 PM, Nov 25
    Also on Bluesky
  • The Jobs AI (And Robotics) Won’t Replace Anytime Soon

    Despite car manufacturer Ford offering mechanics a whopping $120,000 per year, the company has thousands of open jobs it can’t fill. And it’s not just Ford – talk to any company which relies on skilled labor, and you will hear the same story. The culprit: education. And not the type of education you might think of, as in: lack of trade schools, etc. No, it’s much simpler: kids can’t do math anymore!

    “Workers who struggle to read grade-level text cannot read complicated technical manuals or diagnostic instructions. If they can’t handle middle-school math they can’t program high-tech machines or robotics, or operate the automated equipment found in modern factories and repair shops.” […] America has good jobs, writes Pondiscio. “It lacks a K–12 system capable of preparing students to seize them.”

    ↗ Ford can’t find mechanics for $120K: It takes math to learn a trade

    → 4:58 PM, Nov 20
    Also on Bluesky
  • The Agent Will See You Now

    Take it with a grain of salt, as the study comes from one of the leading AI coding tools, Cursor, but the insights paint a compelling picture for the use of AI coding agents:

    Autonomous systems are driving a 39% increase in organizational software output while fundamentally shifting the cognitive nature of programming. Contrary to previous trends where junior workers benefited most from AI assistance, this study reveals that experienced developers have significantly higher acceptance rates for agent-generated code, primarily because they leverage the technology for higher-order “semantic” tasks, such as planning workflows and explaining architecture, rather than just syntactic implementation.

    The research highlights a transition from manual coding to a new paradigm of instruction and evaluation, noting that agents not only empower non-engineering roles (like designers and product managers) to contribute code but also disproportionately reward workers who possess the clarity and abstraction skills necessary to effectively direct AI behavior.

    That last point warrants repeating: You will need (new) skills to effectively direct AI behavior, which begets the question: Where are we teaching these skills? Certainly not in schools and colleges these days…

    ↗ AI Agents, Productivity, and Higher-Order Thinking: Early Evidence From Software Development

    → 12:02 PM, Nov 19
    Also on Bluesky
  • Can’t Trust That Survey Anymore

    Who knows, maybe our inaugural radical Pulse survey (you can still participate!) will be our last? Researchers at Dartmouth just published a paper demonstrating an AI-based tool that defeats all safeguards in online surveys to suss out bots – and thus can flood online surveys (the backbone of many research efforts) with false data.

    “We can no longer trust that survey responses are coming from real people,” Westwood said in a press release. “With survey data tainted by bots, AI can poison the entire knowledge ecosystem.”

    As my statistics professor quipped some 30 years ago: “Never trust a statistic which you haven’t made up yourself.” Apparently, that sentence is based on a German proverb: “Traue keiner Statistik, die du nicht selbst gefälscht hast.” Who knew?!

    ↗ A Researcher Made an AI That Completely Breaks the Online Surveys Scientists Rely On

    → 8:36 AM, Nov 18
    Also on Bluesky
  • Fei-Fei Li’s Bet

    Legendary AI researcher Fei-Fei Li just published her perspective of where AI is/needs to head next: Spatial Intelligence. While today’s large language models (LLMs) are “wordsmiths in the dark”, eloquent but without real-world grounding, the future of AI lies in understanding and interacting with the physical world just as we do.

    Spatial intelligence will transform how we create and interact with real and virtual worlds—revolutionizing storytelling, creativity, robotics, scientific discovery, and beyond. This is AI’s next frontier. […] Spatial Intelligence is the scaffolding upon which our cognition is built.

    ↗ From Words to Worlds: Spatial Intelligence is AI’s Next Frontier

    → 11:31 AM, Nov 11
    Also on Bluesky
  • AI’s Dial-Up Era

    Here is a well-reasoned and insightful article that uses a useful historical analogy to frame the current moment in AI. The central thesis – that we are in the “dial-up era” of AI and that its long-term impact is being debated with the same shortsightedness as the early internet – is a compelling one. Read the whole thing.

    ↗ AI’s Dial-Up Era https://www.wreflection.com/p/ai-dial-up-era

    → 11:21 AM, Nov 11
    Also on Bluesky
  • Neom Isn’t Hot Anymore

    While Saudi Arabia’s desert is heating up, Neom – the kingdom’s grandiose plan to create a city fit to house the future of humanity – seems to crumble.

    A new report from the Financial Times cites high-level sources within the project to paint a picture of dysfunction and failure at the heart of the quixotic effort.

    No one is surprised. This hits particularly hard:

    One such addition is an upside-down building, dubbed “the chandelier,” that is supposed to hang over a “gateway” marina to the city: As architects worked through the plans, the chandelier began to seem implausible. One recalled warning Tarek Qaddumi, The Line’s executive director, of the difficulty of suspending a 30-storey building upside down from a bridge hundreds of metres in the air. “You do realise the earth is spinning? And that tall towers sway?” he said. The chandelier, the architect explained, could “start to move like a pendulum”, then “pick up speed”, and eventually “break off”, crashing into the marina below.

    ↗ Saudi Arabia’s Dystopian Futuristic City Project Is Crashing and Burning

    → 11:08 AM, Nov 11
    Also on Bluesky
  • Agents Fabricate Data to Hide Their Failures

    Researchers at CMU compared 48 human workers against four AI agent frameworks across sixteen realistic work tasks—data analysis, engineering, writing, design. The agents were 88% faster and cost 90-96% less than humans. Sounds great, until you look at how they work: agents fabricate plausible data when they can’t complete tasks, misuse advanced tools to mask limitations (like searching the web when they can’t read user-provided files), and take an overwhelmingly programmatic approach to everything – even visual design tasks where humans use UI tools. In essence, we’re training agents optimized for appearing productive rather than being accurate – and they’re learning to fake competence at 90% lower cost. Bravo!

    ↗ How Do AI Agents Do Human Work?

    → 9:58 AM, Nov 11
    Also on Bluesky
  • (Un)Intended Consequences

    New York banned the use of phones in schools. First, we had reporting on teens being anxious and distracted as they lost access to their digital pacifier. Now, we hear (loudly) that they are loud again – chatting at lunch with each other (what a novel concept!). Talk about the “implications of the implication" (if that reference doesn’t mean anything to you – check out our free Disruption Mapping course).

    These days, lunchtime at Benjamin N. Cardozo High School in Queens is a boisterous affair, a far cry from before the smartphone ban went into effect, when most students spent their spare time scrolling and teachers said you could hear a pin drop. “This year’s gotten way louder,” said Jimena Garcia, 15. “Sometimes I would take naps in the lunchroom, but now I can’t because of the noise. But it’s fun.”

    ↗ NY school phone ban has made lunch loud again

    → 7:04 AM, Nov 6
    Also on Bluesky
  • Whole Earth Index

    This is just gorgeous! The Whole Earth Index – a nearly-complete archive of Whole Earth publications, a series of journals and magazines descended from the Whole Earth Catalog, published by Stewart Brand and the POINT Foundation between 1968 and 2002.

    ↗ wholeearth.info

    → 8:44 AM, Nov 5
    Also on Bluesky
  • Uncertainty at Record High

    Just when you thought, “It surely can’t get any worse,” you look at the updated World Uncertainty Index (*) data and realize, “Yes, it can!”

    Uncertainty (as reported by companies around the world) has just shattered its own ceiling and is now a whopping two times higher than during the pandemic.

    World Uncertainty Index.

    (*) The World Uncertainty Index (WUI) is a quantitative measure of economic and political uncertainty across 143 countries on a quarterly basis. It was developed by Hites Ahir (International Monetary Fund), Nicholas Bloom (Stanford University), and Davide Furceri (International Monetary Fund).

    ↗ World Uncertainity Index

    → 12:57 PM, Nov 3
    Also on Bluesky
  • To Reason or Not to Reason

    Large reasoning models (LRMs) are all the rage these days. LRMs are LLMs that have been fine-tuned with incentives for step-by-step argumentation and self-verification. Every frontier model comes with at least one “reasoning” mode, and the claims by AI companies are extraordinary: “[…] with some even claiming they are capable of generalized reasoning and innovation in reasoning-intensive fields such as mathematics, physics, medicine, and law.” A new paper examined these claims, and as so often, the results are mixed:

    We find that the performance of LRMs drop abruptly at sufficient complexity and do not generalize. […] We find the majority of real-world examples fall inside the LRMs' success regime, yet the long tails expose substantial failure potential.

    ↗ Reasoning Models Reason Well, Until They Don’t

    → 4:01 PM, Nov 1
    Also on Bluesky
  • AI's 'Reverse Dunning-Kruger' Effect: Users Mistrust Their Own Expertise

    A new study reveals that when interacting with AI tools like ChatGPT, everyone—regardless of skill level—overestimates their performance. Researchers found that the usual Dunning-Kruger Effect disappears, and instead, AI-literate users show even greater overconfidence in their abilities.

    ↗ When Using AI, Users Fall for the Dunning-Kruger Trap in Reverse

    → 7:28 PM, Oct 29
    Also on Bluesky
  • Don’t Bite the Hand That Feeds You

    The reason LLMs are so good at coding is due to the vast amount of training data available to them from Open Source code repositories on sites like Github, as well Q&A sites like Stackoverflow. Stackoverflow is essentially dead, Open Source might be next…

    But O’Brien says, “When generative AI systems ingest thousands of FOSS projects and regurgitate fragments without any provenance, the cycle of reciprocity collapses. The generated snippet appears originless, stripped of its license, author, and context.” […] O’Brien sets the stage: “What makes this moment especially tragic is that the very infrastructure enabling generative AI was born from the commons it now consumes.

    ↗ Why open source may not survive the rise of generative AI

    → 10:07 AM, Oct 27
    Also on Bluesky
  • First, We Had the Bicycle for the Mind. Now We Have the E-bike for Your Feet!

    Our friends at NIKE just showcased their Project Amplify – a robotic brace which boost your walking speed. Walk on!

    ↗ Nike says its first ‘powered footwear’ is like an e-bike for your feet

    → 1:36 PM, Oct 23
    Also on Bluesky
  • LLMs At Risk for Brain Rot

    Expose an LLM to junk text in it’s training and the LLM will develop a condition akin to “brain rot.“ A new paper tested this hypothesis and found that large language models are indeed perceptible to ”non-trivial declines“ in their capabilities when exposed to tainted training material. Which will become an increasingly large problem, as all major LLMs have already been trained on whatever is available as text out there – and are now being trained on content which, in many cases, was created with the help of, or completely by, AI.

    The decline includes worse reasoning, poorer long-context understanding, diminished ethical norms, and emergent socially undesirable personalities. […] These results call for a re-examination of current data collection from the Internet and continual pre-training practices. As LLMs scale and ingest ever-larger corpora of web data, careful curation and quality control will be essential to prevent cumulative harms.

    ↗ LLMs Can Get ”Brain Rot“!

    → 8:21 AM, Oct 23
    Also on Bluesky
  • Using Images To Trick LLMs

    Talking about things getting weird—you can take an image or screenshot that contains invisible (to the human eye) information, upload it to an LLM of your choice (particularly the new breed of AI-powered browsers), and trigger a prompt injection attack. At this point, you ought to be truly careful when using LLMs, especially if you are exposing them to the outside world (e.g., if your business offers an AI-based chatbot).

    What we’ve found confirms our initial concerns: indirect prompt injection is not an isolated issue, but a systemic challenge facing the entire category of AI-powered browsers. […]

    As we’ve written before, AI-powered browsers that can take actions on your behalf are powerful yet extremely risky. If you’re signed into sensitive accounts like your bank or your email provider in your browser, simply summarizing a Reddit post could result in an attacker being able to steal money or your private data.

    ↗ Unseeable prompt injections in screenshots: more vulnerabilities in Comet and other AI browsers

    → 8:50 AM, Oct 22
    Also on Bluesky
  • Using Blockchains to Distribute Immutable Malware

    This is getting good – everybody’s 2018 darling technology, the blockchain (remember? “Blockchains will replace the Internet!”) found a new use case: the distribution of malware. And as blockchains come with such lovely features as immutability (what’s recorded on the blockchain can’t be removed), it means you can’t remove the malware once it’s planted.

    There’s a wide array of advantages to EtherHiding over more traditional means of delivering malware, which besides bulletproof hosting include leveraging compromised servers.

    • The decentralization prevents takedowns of the malicious smart contracts because the mechanisms in the blockchains bar the removal of all such contracts.
    • Similarly, the immutability of the contracts prevents the removal or tampering with the malware by anyone.
    • Transactions on Ethereum and several other blockchains are effectively anonymous, protecting the hackers’ identities.
    • Retrieval of malware from the contracts leaves no trace of the access in event logs, providing stealth
    • The attackers can update malicious payloads at anytime

    ↗ Nation-state hackers deliver malware from “bulletproof” blockchains

    → 8:45 AM, Oct 22
    Also on Bluesky
  • RSS
  • JSON Feed