The radical Blog
About Archive radical✦
  • AI Productivity Gains Are Coming

    It took a while, but AI is coming for the enterprise for real now…

    June23 Chart 768x432.

    "The Census conducts a biweekly survey of 1.2 million firms, and one question is whether a business has used AI tools such as machine learning, natural language processing, virtual agents, or voice recognition to help produce goods or services in the past two weeks, see chart below. Nine percent of firms reported using AI, and the rising trend in AI adoption increases the likelihood of a rise in productivity over the coming quarters.”

    Source: Apollo Global Management 

    → 3:46 PM, Jun 23
    Also on Bluesky
  • You Sound Like ChatGPT

    Using LLMs such as ChatGPT not only creates an abundance of em-dashes (“—”) in our writing, but it also starts to affect the way we speak, as a new study shows:

    “Words like ‘prowess’ and ‘tapestry,’ which are favored by ChatGPT, are creeping into our vocabulary, while words like ‘bolster,’ ‘unearth,’ and ‘nuance,’ which are less favored by ChatGPT, have declined in use. […] In the 18 months after ChatGPT was released, speakers used words like ‘meticulous,’ ‘delve,’ ‘realm,’ and ‘adept’ up to 51 percent more frequently than in the three years prior.”

     

    “’We internalize this virtual vocabulary into daily communication,’ says Hiromu Yakura, the study’s lead author and a postdoctoral researcher at the Max Planck Institute of Human Development. […] It’s not just that we’re adopting AI language — it’s about how we’re starting to sound... researchers suspect that AI influence is starting to show up in tone, too — in the form of longer, more structured speech and muted emotional expression.”

    In essence: We become (even more) homogeneous. <sigh>

    Link to article and study.

    → 1:10 AM, Jun 22
    Also on Bluesky
  • Toy-Maker Mattel Teams Up With OpenAI

    What could possibly go wrong? Toy maker Mattel (Barbie, anyone?!) teams up with OpenAI to create AI-powered toys…

    By tapping into OpenAI's AI capabilities, Mattel aims to reimagine how fans can experience and interact with its cherished brands, with careful consideration to ensure positive, enriching experiences.

    Sounds like a potentially nightmarish experience…

    Children do not have the cognitive capacity to distinguish fully between reality and play. […] "Endowing toys with human-seeming voices that are able to engage in human-like conversations risks inflicting real damage on children. It may undermine social development, interfere with children's ability to form peer relationships, pull children away from playtime with peers, and possibly inflict long-term harm." - Robert Weissman, Public Citizen co-President

    Link to article.

    → 11:29 AM, Jun 19
    Also on Bluesky
  • Breaking Down the Infinite Workday

    Microsoft’s WorkLab team is out with a new report on what our workplace looks like mid-2025 – and the results are ugly:

    Nearly half of employees (48%) – and more than half of leaders (52%) – say their work feels chaotic and fragmented. […]

    Half (50%) of all meetings take place between 9–11 am and 1–3 pm—precisely when, as research shows, many people have a natural productivity spike in their day. […]

    On average, employees using Microsoft 365 are interrupted every 2 minutes by a meeting, email, or notification.

    And is goes on and on… Particularly hilarious (and sad is this):

    In the final 10 minutes before a meeting, PowerPoint edits spike 122% – the digital equivalent of cramming before an exam.

    The report offers some positive outlook though:

    The future of work won't be defined by how much drudgery we automate, but by what we choose to fundamentally reimagine. […] The most effective organizations know this—and act on it. Frontier Firms are putting the Pareto Principle into practice, focusing on the 20% of work that delivers 80% of the outcomes.

    Link to report.

    → 8:59 AM, Jun 19
    Also on Bluesky
  • Microsoft CEO Admits That AI Is Generating Basically No Value

    Nadella nails it:

    "So, the first thing that we all have to do is, when we say this is like the Industrial Revolution, let's have that Industrial Revolution type of growth," he said. "The real benchmark is: the world growing at 10 percent," he added. "Suddenly productivity goes up and the economy is growing at a faster rate. When that happens, we'll be fine as an industry.” Needless to say, we haven't seen anything like that yet.

    Good reminder to look beyond the hype train. But also do not forget that AI is already changing many aspects of our daily work life.

    Link to article.

    → 7:07 PM, Jun 17
    Also on Bluesky
  • Waymo rides cost more than Uber or Lyft — and people are paying anyway

    Obi, an app that aggregates real-time pricing and pick-up times across multiple ride-hailing services, has just published what it’s calling the “first in-depth examination of Waymo’s pricing strategy.” The company found Waymo’s self-driving car rides to be consistently more expensive than comparative offerings from Uber and Lyft — and it doesn’t seem to matter.

    I get why people ride Waymo’s (novelty), but this is telling:

    “Colloquially, there is an idea that autonomous vehicles are something that will erode driver jobs and put drivers at risk. And I think the irony of what we’ve seen is that it’s actually quite expensive to run an AV, and that that’s not going to be happening, at least in the near term,” she said. 

    Link to article and study.

    → 4:03 PM, Jun 16
    Also on Bluesky
  • The AI future is already here

    Is AI killing jobs or not? The (seemingly) eternal question, which gets answered with yes, no, and “it depends” in equal measures, just received its latest update. In a recent study by Revelio Labs, a workforce data company, the data points to a world where AI already has a direct impact on hiring:

    In short, jobs that AI can perform are disappearing from job boards faster than those that AI can't handle. […] Since OpenAI released ChatGPT in 2022... the hiring downturn has been steeper for high-exposure roles (31%) than for low-exposure roles (25%).

    The important bit here is the distinction between high- and low-exposure roles – or the ability of an AI to do your job.

    Link to study.

    → 10:54 PM, Jun 11
    Also on Bluesky
  • Medicine's Rapid Adoption of AI Has Researchers Concerned

    Looks like your doctor likes his AI not only in his LinkedIn feed but also in all sorts of medical devices and platforms. According to the FDA, more than 1,000 medical AI products have been cleared for use – with interesting (and potentially troubling) consequences.

    "Unlike most other FDA-regulated products, AI tools continue to evolve after approval as they are updated or retrained on new data.”

    It gets worse:

    "...medical algorithms often perform poorly when applied to populations that differ from the ones they were trained on.”

    and:

    "...many hospitals are buying AI tools 'off the shelf' and using them without local validation. That is a recipe for disaster.”

    Brave new world. Maybe ask your doc next time if his diagnosis was aided by AI…

    Link to article in Nature.

    → 11:24 PM, Jun 9
    Also on Bluesky
  • Mary Meeker is Back

    Legendary tech analyst Mary Meeker hasn’t released one of her colossal trend reports since 2019. This week she did – a whopping 340 slides on all things AI.

    https://www.bondcap.com/reports/tai

    → 11:31 AM, Jun 5
    Also on Bluesky
  • Large Language Models Are More Persuasive Than Incentivized Human Persuaders

    A new paper showed that LLMs are now more persuasive than humans when trying to influence others:

    LLM persuaders achieved significantly higher compliance with their directional persuasion attempts than incentivized human persuaders, demonstrating superior persuasive capabilities in both truthful (toward correct answers) and deceptive (toward incorrect answers) contexts.

    Panicking aside, this is both good and bad news. The bad news is pretty obvious (we already live in a world of semi-constant disinformation; it’s not just hard to distinguish truth from lies, but now AI has the upper hand in persuading us), but there are also some interesting upsides: AI could make us more compliant to take our medications (a huge problem in the healthcare industry), save for retirement (another massive problem in the financial services industry), or work out more regularly…

    But there is a very real danger:

    Human persuasion is naturally constrained by effort and opportunity, but AI-generated persuasion can operate continuously and at scale, influencing vast audiences simultaneously.

    Link to study.

    → 7:38 PM, Jun 3
    Also on Bluesky
  • CEOs Know AI Will Shrink Their Teams — They’re Just Too Afraid to Say It, Say 2 Software Investors

    Take this with a grain of salt – as it comes from the very people who would greatly benefit from this statement being true (venture capital investors) – but this sentence in the article stood out:

    Tech companies, in particular, will see "significantly reduced hiring," he added.

    Ironically, the very people building all this stuff might be the ones most affected by it (on a per-sector/industry level; maybe not on the individual level… AI engineers will be fine for a while).

    Link to article.

    → 7:18 PM, Jun 2
    Also on Bluesky
  • German's Use of AI in Workplace Is Lackluster – and Yet Still Crazy High

    New study about the use of AI in German offices shows that Germans don’t use AI in their workplaces all that much and that they are increasingly wary and mistrusting of AI compared to their international peers. That being said, even with those caveats, a solid 42% of white-collar workers are already using AI in their jobs. Given that ChatGPT is just a little over two years old, this is an impressive penetration number.

    Link to article and study.

    → 6:31 PM, Jun 1
    Also on Bluesky
  • Work From Home Is Doing Just Fine, Thank You

    Wondering what the world of “work from home” truly looks like? Despite the rhetoric that it's over, the average (!) American spends 1.6 days per week working from home…

    052725 Chart v2.

    Source: Apollo research​

    → 5:56 PM, May 27
    Also on Bluesky
  • The Dark Side of Artificial Intelligence Adoption

    On the heels of my last post about Google co-founder Sergey Brin’s rather questionable approach to using AI to replace managers (or, at least, replace a lot of what they do), here is new research on the impact AI is having on the workplace.

    From the study “The Dark Side of Artificial Intelligence Adoption: Linking Artificial Intelligence Adoption to Employee Depression via Psychological Safety and Ethical Leadership” published in Nature:

    “While the adoption of AI brings numerous benefits, it also introduces negative aspects that may adversely affect employee well-being, including psychological distress and depression.”

    The study goes on to say that

    “With AI adoption, there is a real risk that employees will perceive threats to crucial resources like job stability, independence in their work, and professional competencies.”

    In summary (and very much contrary to Mr. Brin’s approach):

    “With the continuous advancement and integration of AI technology in our workplaces, it is imperative for organizations and leaders to prioritize the welfare of their employees, foster a supportive and inclusive working environment, and embrace an ethical approach that prioritizes people when incorporating AI.”

    Link to study.

    → 2:42 PM, May 26
    Also on Bluesky
  • "Management is like the easiest thing to do with the AI"

    Google cofounder Sergey Brin recently rambled his way through a conversation on management in the age of AI:

    "Management is like the easiest thing to do with AI," Brin said.

    Apparently, management, for Brin, consists of summarizing meetings and assigning to-dos:

    "It could suck down a whole chat space and then answer pretty complicated questions," he said. "I was like: 'OK, summarize this for me. OK, now assign something for everyone to work on.’”

    So far, so bad. Where it gets really fun is when he lets AI make promotion decisions:

    "It actually picked out this young woman engineer who I didn't even notice; she wasn't very vocal," he said. "I talked to the manager, actually, and he was like, 'Yeah, you know what? You're right. Like she's been working really hard, did all these things.’”

    And as he clearly has outsourced his management to an AI, he doesn’t even really know if this all has happened or not:

    "I think that ended up happening, actually," Brin said of the promotion.

    All in all, it’s a pretty bleak vision for the future.

    Link to article.

    → 10:20 PM, May 23
    Also on Bluesky
  • Artificial Analysis State of AI

    A recent (Q1/2025) report from Artificial Analysis delves into the capabilities and emerging trends of frontier AI models. The report identifies six key themes:

    Continued AI progress across all major labs, the widespread adoption of reasoning models that "think" before answering, increased efficiency through Mixture of Experts architectures, the rise of Chinese AI labs rivaling US capabilities, the growth of autonomous AI agents, and advances in multimodal AI across image, video, and speech.

    There is a lot to unpack in the report, but the key chart might be this:

    All frontier models are converging on the same set of capabilities and the quality of those, which means that we will (and are already seeing) a brutal race to maintain their position in the peloton with further increased price pressure. This brings up the question of how these companies might justify their insane valuations…

    Link to Report.

    → 10:47 PM, May 22
    Also on Bluesky
  • The Wild Dichotomy of AI in Research

    Just this last week saw the announcement of new, sophisticated AI research tools from all the frontier labs, claiming exceptional results. Headlines such as “U. researchers unveil AI-powered tool for disease prediction with ‘unprecedented accuracy’” or “Microsoft's new AI platform to revolutionize scientific research” gush about these new tools’ abilities.

    Meanwhile, Nick McGreivy, a physics and machine learning PhD, shared his own experience with the use of LLMs in scientific discovery – and his story reads very differently:

    “I've come to believe that AI has generally been less successful and revolutionary in science than it appears to be.”

    He elaborates:

    “When I compared these AI methods on equal footing to state-of-the-art numerical methods, whatever narrowly defined advantage AI had usually disappeared. […] 60 out of the 76 papers (79 percent) that claimed to outperform a standard numerical method had used a weak baseline. […] Papers with large speedups all compared to weak baselines, suggesting that the more impressive the result, the more likely the paper had made an unfair comparison.”

    And in summary:

    "I expect AI to be much more a normal tool of incremental, uneven scientific progress than a revolutionary one.”

    And the discussion about what is hype and what is reality in AI continues…

    Link to his blog post.

    → 5:03 PM, May 20
    Also on Bluesky
  • Neal Stephenson on AI: Augmentation, Amputation, and the Risk of Eloi

    Science fiction author Neal Stephenson, who popularized the concept and term “metaverse” in his seminal book Snow Crash (1992), recently spoke at a conference in New Zealand on the promise and peril of AI.

    His (brief but razor-sharp) remarks are well worth reading in full, but this quote stood out:

    “Speaking of the effects of technology on individuals and society as a whole, Marshall McLuhan wrote that every augmentation is also an amputation. […] This is the main thing I worry about currently as far as AI is concerned. I follow conversations among professional educators who all report the same phenomenon, which is that their students use ChatGPT for everything, and in consequence learn nothing. We may end up with at least one generation of people who are like the Eloi in H.G. Wells’s The Time Machine, in that they are mental weaklings utterly dependent on technologies that they don’t understand and that they could never rebuild from scratch were they to break down.”

    Link to his remarks.

    → 3:53 PM, May 19
    Also on Bluesky
  • MIT Backs Away From Paper Claiming Scientists Make More Discoveries with AI

    Remember that MIT paper which showed that researchers leveraging AI are significantly more productive (as in: a higher number of discoveries), yet are less satisfied with their work (as they are relegated to drudgery while AI does all the hard, challenging, and exciting work)?

    Well, it turns out that this was another case of “too good to be true.” MIT just recalled the paper, the researcher who published the paper isn’t affiliated with MIT anymore, and MIT states it “has no confidence in the provenance, reliability, or validity of the data and has no confidence in the veracity of the research contained in the paper.”

    Ouch.

    Link to report.

    → 2:30 PM, May 18
    Also on Bluesky
  • Learn Prompt Engineering From The Pros

    Want to get better at writing prompts (it’s a worthwhile investment of time – the difference in results from a mediocre prompt to a good prompt can be vast)? Here are a couple of excellent sources – straight from the horse’s mouth:

    OpenAI has a whole site dedicated to their “AI Cookbook” - which includes, for example a ChatGPT 4.1 Prompt Guide.

    Google has an excellent resource on prompt design strategies on their "AI for Developers" site.

    And Anthropic (Claude) has something similar on their own developer documentation site – which even includes a prompt generator tool.

    Happy prompt engineering!

    → 2:03 PM, May 18
    Also on Bluesky
  • The (AI) Canary in the Coal Mine

    Here is an interesting question: If AI is truly so good at creating code, why don’t we see many more code contributions in Open Source repositories made with the help of AI?

    Here is Satya Nadella, CEO of Microsoft:

    “Maybe 20 to 30 percent of the code that is inside our repos today in some of our projects is probably all written by software.”

    That’s a lot of “maybe” and “probably”… And we just can’t know, as Microsoft’s code repository is (of course) private. But Open Source code lives in public places like GitHub and thus is inspectable. And when you look closely, you will find very little evidence that code in those repositories is written by AI.

    Admittedly, a lot of Open Source projects aren’t particularly excited about AI-generated pull requests:

    “It’s true that a lot of open source projects really hate AI code. … the biggest one is that users who don’t understand their own lack of competence spam the projects with time-wasting AI garbage.”

    But that aside, when you look at the data, it’s just not there:

    “TL/DR: a lot of noise, a lot of bad advice, and not enough signal, so we switched it off again.”

    In many ways, AI keeps furthering the skill gap:

    “The general comments … were that experienced developers can use AI for coding with positive results because they know what they’re doing. But AI coding gives awful results when it’s used by an inexperienced developer.”

    Overall, a good reminder to look past the marketing and hype…

    Here is the full article.

    → 3:41 PM, May 15
    Also on Bluesky
  • Should We Trust AI With Our Health When It Can’t Even Draw a Simple Map?

    On one hand, we have OpenAI announcing HealthBench, a physician-designed benchmark that rigorously evaluates AI models on 5,000 realistic health conversations across diverse scenarios, thus preparing to pave the way for Doc ChatGPT to become a reality. On the other hand, you have LLM-skeptic Gary Marcus trying something as trivial as having ChatGPT draw a map to hilarious effect:

    It was very good at giving me bullshit, my dog-ate-my-homework excuses, offering me a bar graph after I asked for a map, falsely claiming that it didn’t know how to make maps. A minute later, as I turned to a different question, I discovered that it turns out ChatGPT does know how to draw maps. Just not very well.

    After quite the saga (it’s worth reading Gary’s article in full), he concludes:

    How are you suppose do data analysis with 'intelligent' software that can’t nail something so basic? Surely this is not what we always meant by AGI.

    → 3:43 PM, May 14
    Also on Bluesky
  • The Reports of the Finance Profession’s Death Are Greatly Exaggerated

    Think AIs will come for your CFO and his team? Not quite, not yet… Vals.ai’s Finance Agent Benchmark clearly shows we are some ways off from the professions caving in to our LLM-powered overlords:

    “The foundation models are currently ill-suited to perform open-ended questions expected of entry-level finance analysts”

    But fret not — not all is lost:

    “Models on average performed best in the simple quantitative (37.57% average accuracy) and qualitative retrieval (30.79% average accuracy) tasks. These tasks are easy but time-intensive for finance analysts.”

    Link to benchmark and analysis.

    → 4:04 PM, May 13
    Also on Bluesky
  • The Metaverse is Truly Dead

    Talk about a nail in the coffin: Minecraft, the proto-3D world-building game played by hundreds of millions, is shutting down its VR and Mixed Reality support.

    If there ever was a question of whether the Metaverse is dead (though calling it ‘dead’ assumes it was once alive, which is debatable), this might be the definitive answer.

    “Microsoft’s latest update to Minecraft‘s Bedrock Edition, version 1.21.80, removes support for both virtual reality and Mixed Reality, Microsoft’s own take on AR that it killed off years ago. Now, Minecraft remains as a game that can be played on multiple consoles and platforms — just not VR.”

    Of course, it is not just Minecraft—after losing oodles of money, Meta (the company which conveniently rebranded itself to indicate it is the king of VR) is shuttering large parts of its VR efforts for quite a while.

    Makes me wonder what happened to all the Chief Metaverse Officers (the “new” CMOs)? And yes, THAT was indeed a thing for a hot minute…

    A good reminder to keep applying Chris Yeh’s brilliant “Does it have utility?” framework: Frequency / Density / Friction: How often do you encounter the problem? How much time/energy do you spend in the problem space? How much pain does it cause you? If those factors are too low, you simply don’t have a problem worth solving. Metaverse - Never/None/Zero. Q.E.D.

    Game over.

    → 5:12 PM, May 12
    Also on Bluesky
  • Driverless Freight Trucks Begin Barreling Through Texas

    And so it (finally, maybe, hopefully) begins: after we all get used to Waymo’s robotaxis on our roads (at least if you live in a place like San Francisco), Aurora is now bringing Level 4 autonomous driving to trucks. And just like Waymo, which is retrofitting existing cars, Aurora’s tech is designed to be integrated into OEM trucks.

    “[…] starting with deliveries between Dallas and Houston... aims to expand to El Paso, Texas, and Phoenix, Arizona by the end of the year.”

    And the OEM aspect:

    “[…] the company has already partnered with Volvo and PACCAR (which makes trucks with Kenworth and Peterbilt badges) to bake its hardware and software into their freight vehicles.”

    Link to article.

    → 8:50 PM, May 3
    Also on Bluesky
  • RSS
  • JSON Feed