Search results for: “technium”

  • The dot LLM era?

    The dot LLM era is one of the chunkiest posts that I have written, so I have put it together in a PDF as well that you can download and share freely amongst colleagues and peers.

    The dot LLM era executive summary

    The “dot LLM era” represents a pivotal moment in technological history, drawing striking parallels to the dot-com bubble of the late 1990s. This period is defined by a massive influx of capital into Large Language Models (LLMs) and artificial intelligence infrastructure, which represents clear analogues to the dot-com era “three bubbles” framework: online businesses, open-source ventures, and telecommunications (which represents closest analogue to the current dot LLM era). 

    The Core Thesis

    The current $1 trillion valuation of the AI sector faces two existential challenges:

    • Amortisation Risk: Unlike the dark fibre of the 1990s, which had a useful life of over a decade, modern GPU and TPU hardware becomes technically obsolete within 3 to 5 years.
    • Self-Defeating Economics: If AI-driven automation successfully provides $1 trillion in cost savings through job cuts, the resulting increase in unemployment and drop in GDP could destroy the very macroeconomic environment required to sustain hyperscaler growth.

    A Tale of Three Bubbles

    The document argues that we are conflating three distinct historical analogues:

    • Online Businesses: Recalling the “burn rates” of the early web, where pure-play LLMs are currently providing tokens for less than their marginal cost.
    • Open-Source: Comparing current model proliferation to the rise of Linux, where the ultimate winners may not be the model creators but those providing enterprise-grade support.
    • Telecommunications: The most instructive analogue, involving massive infrastructure build-outs, vendor financing, and potential “Minsky moments” where optimism outstrips sustainable cash flow.

    Geopolitical and Economic Realities

    Unlike the 1990s “Long Boom” characterized by US pre-eminence and budget surpluses, the dot LLM era exists within a climate of high government debt and inflation. Furthermore, US dominance is challenged by Chinese hyperscalers and open-source models like Alibaba’s Qwen, which offer high performance at significantly lower costs.

    Potential Outcomes

    The document outlines seven possible scenarios for the era’s conclusion, ranging from The Breakthrough (total economic transformation) to The Weird Gizmo (total collapse). Currently, “The Moral Hazard”—where AI is deemed “too big to fail” and receives government backing—is viewed as the most likely path (~95% likelihood).

    How this dot LLM exploration started?

    This dot LLM post came out of a number of ideas and vibes. 

    Everyone[i] from commentators[ii] and podcast hosts to friends are talking about a dot-com-type bubble in LLMs, what I’ve termed as shorthand the dot LLM era. The dot LLM era comparison has become a steady tempo of concern. 

    The term AI bubble took off in interest during September of 2025. 

    Change of search volume by week in 2025 for AI bubble

    The dot LLM era is shorthand to move backwards and forwards in time comparing the current AI boom with the dot-com boom of 1990s – 2001. It’s a very different type of ‘Y2K trend’. 

    Many pure-play LLMs customers are currently getting to use tokens for less than their marginal cost[iii],  and this is part of the reason (alongside the high cost of model training) why the likes of OpenAI, C3.ai, Perplexity, Anysphere and Anthropic are raising new rounds of financing[iv]. They have been losing money[v] and continue to do so. 

    Spending by both pure-play LLMs and their hyperscaler partners is driven by the effort to create an AI moat[vi]. An AI moat is a sustained proprietary advantage derived from a company’s use of artificial intelligence that makes its offerings fundamentally superior, cheaper, or “stickier” than those of rivals, and which is hard to be replicated by rivals.

    Even the most historically bullish institutional investors, like James Anderson[vii], formerly of Baillie Gifford, have turned bearish on Nvidia and pure-play LLM offerings.

    To meet the needs of these services, development of an extra 1,500 data centres has been announced – only a quarter of which are under construction at the time of writing.[viii]

    It is a time reminiscent of the mid-2010s when venture capitalists subsidised the cost of services like Uber and Lyft[ix] to grow markets from the ground up. Going back further to the dot-com era, Amazon took a similar approach with its business. 

    Valuations for the Magnificent 10:  Apple, Alphabet, Amazon, AMD, Broadcom, Meta, Microsoft, Nvidia, Palantir and Tesla — are high. The 24-month forward P/E ratio of the Magnificent 10 is 35 times. By comparison the S&P 500’s equivalent P/E ratio at the peak of the dot-com boom approached 33[x], with a brief peak at the market top of 44.[xi]

    Built into these Magnificent 10 valuations, is an assumption that LLMs will help them cut costs and or drive revenue growth by $1 – 4 trillion in the next two years.[xii]

    Like the dot-com era[xiii], the dot LLM era is spawning several businesses that are likely to be considered weird gizmos or bad business ideas that will be mocked in the future. The dot-com analogues included the likes of proto-digital currencies Beenz and Flooz[xiv], CueCat[xv] – a bar code scanner that allowed web users to scan codes on magazines to get more pages online or the short-lived[xvi] 3Com Audrey[xvii][xviii] and Sony eVilla[xix] internet appliances. 

    (Disclosure: in my first agency-side role, I worked on 3Com’s consumer products and the Palm device business that was spun off as palmOne[xx] to give space for the Ergo connected home internet appliance range. Audrey’s ability to sync with two Palm devices[xxi], despite Palm being seen as an internal competitor, gives you an idea of how disjointed and chaotic internal planning was in companies like 3Com when they were trying to move at ‘internet speed’. One of the last 3Com projects I worked on was the launch of Audrey in October 2000.) 

    Bubbles don’t kill technology from moving forwards

    Like the dot-com era, the dot LLM era is likely to move through two separate cycles: one financial and the other technological. While the financial bubble destroyed a lot of shareholder value, the underlying web technology cycle and use cases became commonplace and evolved. Email became part of our culture[xxii] in the same way that social media became cultural fabric a decade later. LLMs or their successors (such as nested models[xxiii] and world models[xxiv]) are likely to be influential and change the nature of work, life, business and culture. 

    Already we can see the dot LLM era playing out on social media as over half of content is estimated to be produced with generative AI. 

    Human vs AI articles

    This relentless forward progress for technological adoption and refinement was likened to an organic being by author Kevin Kelly in a phenomenon he called the ‘technium’.[xxv]

    Believing that AI is undergoing a dot LLM bubble isn’t the same as not believing that the technology won’t have an ongoing impact. 

    A Tale of Three Bubbles

    When we talk about the dot LLM era we are conflating a number of related bubbles bursting. 

    The bubbles were based around a common conceit: prior experience counted for naught because the internet changed everything. 

    This resulted in three distinct historical bubbles:

    • Online business bubble
    • Open-source bubble
    • Telecommunications bubble

    The one that most people recall is the dot-com boom where online businesses went under.

    Online businesses

    Iconic ones included technically ambitious clothing retailer Boo.com, pet care supplies firm Pets.com and many more. 

    Boo.com burned through $135 million in just 18 months[xxvi]. And they weren’t the only ones. In March 2000, Pegasus Research put out a research paper[xxvii] outlining the burn rates of each online business. The report went under-reported at the time, but took a clear-eyed look at the sector. 

    Successful business people failed. Podcaster and academic Scott Galloway[xxviii] founded RedEnvelope[xxix], an online commerce site that sold gifts including personalised items and experiences.  Bob Geldof’s online travel site deckchair.com[xxx] doesn’t even merit a mention in most profiles of the famous musician. 

    Back when I worked at Yahoo! long-time employees said that only a pivot to provide dating services had kept the rest of Yahoo! Europe afloat during the dot-com bust of 2001/ 2002. Online advertising revenues at the time dropped more than 30% over a 12-month period. The difference between success and failure was a very narrow gap.

    Amazon survived and eventually thrived as it managed to convince its shareholders to defer profitability for a decade to garner growth. That move and the company’s nascent web services business (AWS) led to the online juggernaut that Amazon is today[xxxi]. While Amazon was founded in 1994 and first went online in 1995, it didn’t make its first quarterly profit until the end of 2001[xxxii] of $5 million on revenue of $1.12 billion[xxxiii] and the first annual profit in 2003[xxxiv]. Uber and Lyft learned from the example that Amazon had set a decade earlier. 

    Open-source bubble

    The second bubble was the ‘open-source’ bubble. The rise of the commercial web (and the millennium bug[xxxv]) disrupted existing technology stacks and opened up new opportunities to sell enterprise computing hardware and software. Several companies were launched to support the rollout of open-source software that threatened Microsoft’s and Unix operating system duopoly. 

    My former client VA Linux Systems built web servers and workstations optimised for Linux users[xxxvi]. Now VA Linux Systems is remembered more for its IPO, which valued the company at $30 and opened for trading at $299[xxxvii]. Red Hat[xxxviii] and SuSE[xxxix] provided commercially supported versions of Linux for corporate enterprises. Like their online business counterparts, few of the open-source business bubble companies could be considered ‘successful’, the outlier being Red Hat which eventually sold to IBM in 2019 for $34 billion[xl]

    The winner, Red Hat, didn’t sell the open-source software (Linux) as its business model; it sold enterprise-grade support, integration, and services.

    While the open-source bubble was the smallest of the three bubbles, it had an outsized impact with Linux being the foundation for everything from the Android mobile OS to the largest data centres. 

    Telecoms bubble

    The telecoms bubble was the least visible, yet most spectacular bubble and the one that is most instructive about the dot LLM era. 

    There are three places where you could start the telecoms bubble. April 30, 1995, when the NSFnet was decommissioned[xli], the Telecommunications Act of 1996, or 1984. 

    I am going to go with 1984[xlii]. While the internet was growing in academic and military circles in the US and there were nascent computer networks elsewhere like the UK[xliii] – the real revolution was happening on the London Stock Exchange. The UK government under prime minister Margaret Thatcher looked to get the government out of businesses. A programme of privatisation took place to sell-off numerous nationalised businesses; plans to privatise British Telecom were proposed in 1982.  1984 saw the IPO of British Telecom plc, the previously government owned telecoms provider[xliv]. The UK government also licensed the first competitor Mercury Communications[xlv]

    From a technological perspective the IPO seemed to be a catalyst[xlvi] for wider telecoms deregulation in western Europe[xlvii] and around the world. In 1985, the Japanese government privatised NTT and opened the Japanese telecommunications market up to competition[xlviii]. The European Commission began developing a regulatory framework to open up national telecoms markets in 1987[xlix], Europe and Japan would spend the next decade opening up their markets for alternative telecommunications services. 

    It was into this global landscape that the US overhauled its telecommunications regulations with the Telecommunications Act of 1996[l]. The stated intention of the act was to “let anyone enter any communications business – to let any communications business compete in any market against any other.”[li] The act incentivised the expansion of networks and new services across the US.[lii] Early US netizens rejected the act as a way to regulate cyberspace[liii]

    The following year 69 members of the World Trade Organisation (WTO) agreed to open their basic telecoms markets to competition[liv]

    In parallel with the wider atmosphere of telecommunications liberalisation, was the rise of the internet. The rise of home computers in US households between 1990 and 1997 grew from 15% to 35%[lv]. At that time, a small percentage of people would be dialling directly into work, nascent online services like CompuServe or AOL, dialling into their Charles Schwab account and bulletin boards. 

    Outside the US, it was more likely that your computer was a standalone machine with a spreadsheet, word processing application, maybe design software allowing you to write the document from home and bring it in to work on floppy drive, or possibly an Iomega diskette[lvi] of some sort. 

    Private long distance optical fibre networks together with free local telephone calls were the infrastructure for internet connectivity. The web the way we know it now was not a surefire winner[lvii]. Much speculation was on the internet superhighway – digital cable television with value added services like online shopping.[lviii] Bill Gates at the peak of his power as CEO of Microsoft was convinced that the digital cable TV was the way forward.[lix] The next edition was edited to reflect the reality of the web instead. The open interoperable nature of the web proved to be more attractive than walled garden digital services envisaged by cable TV companies.[lx]

    Investment in telecoms infrastructure increased to meet the future needs of digital services, based on a misreading of internet data traffic growth[lxi]. US telecoms providers invested $500 billion between 1996 and 2001 – mostly on optical fibre networks.[lxii]Much of this spending was done by new entrants including Global Crossing, WorldCom, Enron, Qwest and Level 3. There was a corresponding scale up by equipment makers like Lucent to supply the telecoms providers.[lxiii] Telecommunications equipment companies Lucent and Nortel[lxiv] both provided vendor financing for their dot-com era client base – engineered in such a way to inflate sales figures and their share price.[lxv]

    • Lucent lent customers the money to purchase their equipment. They then booked the loan value as revenue, even though the repayment risk remained and the debt was held as an asset on the Lucent balance sheet. 
    • Nortel used its own shares as financing for its customers. It is believed that Nortel lent $7 billion+ to help start-up telecommunications carriers make equipment purchases. Many of these were unsecured loans, interest-free and tied to future purchases. 

    Carriers engaged in ‘round-tripping’. Global Crossing would ‘sell’ network capacity to Qwest; Qwest would ‘sell’ similar capacity back to Global Crossing for nearly the same amount. Both companies booked the deals as revenue. US regulators found that this was a pre-arranged swap designed to inflate revenue, despite having no commercial purpose.  

    Had the bubble continued into 2005, WorldCom CEO at the time Bernie Ebbers had expected to invest another $100 billion in the company’s network infrastructure that year[lxvi]. Instead, Ebbers left WorldCom investors with a $180 billion loss. When the telecoms bubble imploded, an estimated trillion dollars in debt was owed, much of which was not expected to be recovered.[lxvii]

    In 2002, the telecoms bubble helped change the way business is conducted. In reaction to a number of major corporate and accounting scandals, notably Enron[lxviii] and WorldCom – US lawmakers enacted the Sarbanes-Oxley Act of 2002[lxix]. This Act (SOX as it became known) mandated standards in financial record keeping and reporting for public companies. It covered responsibilities of the board of directors and criminal penalties for certain practices[lxx]. It required the SEC to create regulations for compliance. SOX drove up the cost of a company going public and remaining public due to the administrative burden to remain legally compliant. 

    Technology vendor financing from companies like Cisco and IBM continued to be an issue through the 2008 financial crisis,[lxxi]but was largely kept out of the common discourse by the tsunami of sub-prime mortgage debt defaults. 

    The dot LLM era hinges around service providers and equipment makers, in the same way that the telecoms bubble did. Here are some examples and their dot LLM analogues. 

    Service providersEquipment makers
    Dot-com era
    Enron
    PSINet
    Qwest
    UUNET
    Worldcom
    Dot-com era
    3Com
    Ciena
    Cisco
    Equinix
    Juniper Networks
    Lucent
    Sun Microsystems
    Dot LLM era
    Alphabet
    Amazon
    Anthropic
    OpenAI
    Oracle
    Microsoft
    Salesforce
    Dot LLM era
    AMD
    Applied Materials
    ASML
    Broadcom
    Huawei
    Intel
    Micron
    Nvidia
    Samsung
    TSMC

    Of course, the idea of them being analogues doesn’t line up perfectly. While the excessive build out of optical fibre networks could be considered analogous to hyper-scaled AI infrastructure; it isn’t a perfect match.  

    The acceleration in network and computing capability in hyperscalers show the kind of positive trajectory that Mary Meeker had in her dot-com era analyst presentations[lxxii]

    capex

    Some critics think that the massive acceleration in network and compute investment for LLM purposes represents a Minsky moment in itself[lxxiii] – heralding it as an event that fits Hyman Minsky’s Financial Instability Hypothesis.

    Minsky considered this coming in three parts:

    1. A self-reinforcing boom driven by optimism and easy credit
    2. A shock, that can be minor in nature, has investors re-look at cash-flow shortfalls 
    3. Rapid asset sales and deleveraging / de-risking

    The scale of investment and construction of data centres together with the new electricity generating capacity to power them are orders of magnitude larger than the telecoms boom.  

    Secondly, the LLM infrastructure has a much shorter life. LLM hyperscalers go through GPUs (and TPUs) extremely fast with a useful life of 3 years or so.[lxxiv] Complete technical obsolescence of a given GPU / TPU design has occurred by 5 years from launch.[lxxv]

    Therefore, if there is an AI bust the processors wouldn’t be available to use in the next economic upswing in the tech sector. By comparison the optical fibre networks laid during the dot-com boom had a useful life of 10+ years and the growth of web 2.0 and social startups was largely built on surplus server and networking equipment left over from the dot-com era. The dot LLM era represents a financial and technological amortisation risk.

    There is an added wrinkle in this last point about the useful life of GPUs and TPUs. Company filings of hyperscalers show that they are amortising their network and compute capital expenditure over longer times, by lengthening the assumed useful lives of components in their financial paperwork. 

    useful life

    The economic environment.

    The economic conditions that the dot-com era happened in were very different to the conditions of the dot LLM era. 

    The US had suffered through much of the 1980s and into the early 1990s. Reaganomics had driven a ‘jobless recovery’ as the financial and services sectors took over from manufacturing as the US economic growth engine. In 1989 the Savings and Loan crisis peaked.[lxxvi] This occurred alongside rising interest rates to battle inflation. An oil price spike as a result of the first Gulf War exacerbated economic conditions and the recession ended the ambitions of George H. Bush becoming president for the second time. Under a new government, by spring 1994, jobs and economic growth both picked up. 1996 saw growth continuing and by May 1997 US unemployment dropped below 5% for the first time in 24 years.  

    Other countries had similar recessions in the late 1980s and early 1990s due to restrictive monetary policies, oil prices and the end of the Cold War. By 1994, global GDP growth returned.[lxxvii] Wired magazine talked of the 1980s as a contagious idea:[lxxviii]

    America is in decline, the world is going to hell, and our children’s lives will be worse than our own. The particulars are now familiar: Good jobs are disappearing, working people are falling into poverty, the underclass is swelling, crime is out of control. The post-Cold War world is fragmenting, and conflicts are erupting all over the planet. The environment is imploding—with global warming and ozone depletion, we’ll all either die of cancer or live in Waterworld. As for our kids, the collapsing educational system is producing either gun-toting gangsters or burger-flipping dopes who can’t read.

    In the same article, they thought of the 1990s as the start of ‘The Long Boom’ – 25 years of prosperity freedom and a better environment for the world. 

    By 2000, the US government went from running a budget deficit eight years earlier to running a surplus. This eased the credit markets for businesses and consumers. The US Taxpayer Relief Act lowered marginal capital gains tax and helped fuel stock market investments. Day trading became a thing by 1999,[lxxix] mirroring investors in crypto and stocks in the 2020s.[lxxx]

    By comparison, the current economic climate is more similar to the 1980s than the 1990s. Government debt has reached new heights. Governments have struggled to rein in inflation created by COVID-era supply shocks – which was responsible for several governments including the Biden administration being voted out of office. The high government debt and inflation leave governments with fewer policy tools to manage a systemic shock compared to their 1990s counterparts. The Economist claimed that western countries had government debt levels unseen since Napoleonic times.[lxxxi] There is no US government budget surplus and little ‘headroom’ for monetary policy.

    Wired magazine’s ‘contagious idea’ sounds very familiar:

    • Climate despair has been recognised as a condition by mental health professionals.[lxxxii]
    • Global warming is cited[lxxxiii] as a cause of extreme weather conditions[lxxxiv].
    • Good jobs are disappearing and this is often blamed[lxxxv] on generative AI. 
    • US tariffs, Brexit and the Ukraine war are disrupting global commerce. 

    In conclusion, the dot-com era economy was much more conducive for retail investors than the dot LLM era is. 

    The internet changes everything

    Dot-com businesses had it right in their view that the internet would change business and shopping for consumers and enterprises. Some of them like Amazon made it, many didn’t. The investment bank analysts believed it too.[lxxxvi]

    You see similar things being written about AI now, along with similar looking ‘hockey stick’ charts.[lxxxvii]

    Microsoft research[lxxxviii] suggests that there is a strong link between GDP per capita and AI usage. But also notes that adoption in advanced economies tends to plateau between 25% and 45%, suggesting non-economic factors eventually moderate growth. Suggesting that the dot LLM era may not be the kind of game-changer that it might be believed to be by advocates. I would recommend that the reader keeps an open mind on this rather than automatically thinking that this proves generative AI as being a technological dead-end. More work is required to try and understand why the plateau happens and whether it represents a ceiling or a brief rest before adoption accelerates again. 

    Artificial general intelligence or AGI

    AGI is when the LLM surpasses your average human. The idea of AGI has taken on the similar messianic fervour of people from the dot-com era including George Gilder’s Telecosm. Many executives in the most prominent LLM developers subscribe to an imminent AGI occurring. 

    Elon Musk holds the most aggressive timeline[lxxxix]. He thinks that the main bottlenecks to AGI—specifically power supply and high-end chip availability—are being solved rapidly. Through his company’s xAI’s computing power, he believes that the next generation of models will surpass human intelligence in almost any individual task by early 2026. Anthropic’s CEO Dario Amodei believes that AGI could arrive in 2026/7[xc]. OpenAI’s Sam Altman considers 2027 to be a realistic timeline for the arrival of AGI[xci]. DeepMind co-founder Shane Legg has come up with a notional timeline of 2028. His view is based on the current rate of progress for both computing hardware and LLM algorithms.[xcii] Long time AI advocate Ray Kurzweil has published a series of books about AGI, which he termed the ‘singularity’. The latest of which put 2029 as the year in which AGI is likely to occur[xciii]

    As with any cultural artefact, AGI has become blended with religious thinking, as exemplified by this outlandish quote from podcaster Joe Rogan. 

    “Jesus was born out of a virgin mother. What’s more virgin than a computer? If Jesus does return, you don’t think he could return as artificial intelligence? AI could absolutely return as Jesus.” – Joe Rogan[xciv]

    All of which is reminiscent of Timothy Leary’s infatuation with the early web[xcv] and the Heaven’s Gate Cult[xcvi]

    Despite some prominent advocates, many experts in the field are sceptical about the imminent arrival of AGI. Included in these sceptics are OpenAI co-founder Andrej Karpathy who believes that the nature of LLMs mean that AGI won’t arrive using current techniques and on the timeline that advocates predict[xcvii]. Researchers Rodney Brooks[xcviii] and Yann LeCun[xcix] believe that understanding the physical world is critical for technology to achieve AGI. This work is only starting now. Academic Melanie Mitchell argues that until systems can grasp ‘meaning’ AGI will not happen[c]

    The good bubble

    Some of the most important US business executives of the LLM era admit that we are in some kind of bubble. Here’s what they’ve said in their own words. 

    “When bubbles happen, smart people get overexcited about a kernel of truth … Are we in a phase where investors as a whole are overexcited about AI? My opinion is yes.”[ci]

    “This frenzy gives us pause … The belief in an A.G.I. or superintelligence tipping point flies in the face of the history of technology”[cii]

    “This is a kind of industrial bubble … investors have a hard time in the middle of this excitement, distinguishing between the good ideas and the bad ideas. And that’s also probably happening today.”[ciii]

    “Given the potential of this technology, the excitement is very rational. It is also true when we go through these investment cycles there are moments we overshoot as an industry. We can look back at the internet right now, there was clearly a lot of excess investment, but none of us would question if the internet was profound or did it have a lot of impact it was fundamentally changed how we work digitally as a society. I expect AI to be the same; I think it’s both rational and there are aspects of irrationality to a moment like this.”[civ]  

    “Most other infrastructure buildouts in history, the infrastructure gets built out, people take on too much debt, and then you hit some blip … a lot of the companies wind up going out of business, and then the assets get distressed and then it’s a great opportunity to go buy more … definitely a possibility that something like that would happen here.”[cv]

    The real question is whether the dot LLM era is a ‘good’ bubble or a bad bubble? What does a good bubble look like? And how much will it cost? Most of the quotes above see the dot LLM era as similar in nature to the internet boom and bust. While pioneers may have died society was irrevocably changed. 

    Some of the irrationality in the ‘good bubble’ hypothesis seems to include hubris, for example OpenAI shunned having external advisers to work on its $1.5 trillion worth of data centre deals.[cvi] While OpenAI has relationships with investment banks and corporate law firms – it didn’t make much use of them.

    These explanations assume that there will be a corresponding surplus of infrastructure that will spark new innovation on the backs of dead companies. A concept that most represents the telecoms aspect of the dot-com era. The explanations ignore the financial losses suffered by pension funds and retail investors as these companies went bankrupt. They also ignore that the useful life of AI computing hardware is obsolete faster than railway tracks or laid fibre optic cables.[cvii] Short sellers have accused hyperscalers of estimating unrealistically useful lives for their computer equipment, in particular, the GPUs that power AI model training and inference. The allegations claim that profits are artificially overstated by allowing depreciation of assets over a longer period.[cviii]

    At its peak in March 2000, the NASDAQ index peaked at 5,048. When the dot-com bubble burst the index declined to 1,139. Recovery took 15 years from the peak value. The NASDAQ reached 5,048 again in March 2015.[cix] The risk is arguably greater this time around as the top ten stocks constituting the S&P 500 index constitute 40% of its value.[cx] This implies a vulnerable, brittle market environment prior to any economic bust. So, the idea of ‘good’ is very narrowly defined and asking the term to do a lot of heavy lifting in terms of its language. Predicting the peak of the market[cxi] is challenging too[cxii]

    Can the demand for LLM; grow at the speed implied by invested capital?

    Advertising as a possible use case

    The first use case to consider for how the dot LLM era could meet its full ‘potential’ would be the ongoing disruption of advertising by digital platforms. Depending who you believe the global total market for advertising is close to, or has just exceeded $1 trillion in total value. 

    Globally, advertising represents about 1 percent[cxiii] of global GDP. It usually holds at around that proportion as global economic growth waxes and wanes. In some key markets such as the US, UK and Singapore – it makes up a higher percentage of GDP – as the home of advertising platforms, advertising agencies with international responsibility and technology suppliers to the industry. 

    Advertising isn’t just a cost centre for businesses, but also a driver of economic growth and profit. One Euro of advertising is estimated to generate up to 7 Euros of economic value.[cxiv]

    It took digital advertising over a quarter of a century to go from zero to over half of advertising spend. This hinged around two growth spurts, one in 2000 with the rise of online businesses and the second in 2020 with the COVID-19 lockdown. A factor of the transition of digital advertising growth has been down to the fragmentation of audiences across media platforms and alongside traditional media.

    AI (but not LLMs) has been used in advertising as long as digital advertising has been around. It started to be used for understanding consumer behaviour and delivering targeted advertising.[cxv] Amazon started using AI for its recommendations in 1998.[cxvi]

    Not all economic value in digital advertising accrued from the transfer of ‘traditional’ advertising to digital advertising. There is evidence of a direct correlation between a rise in e-commerce drives a decline in retail properties, given the strong linkage between e-commerce, retail media search advertising – there is part of that value exchange which would accrue to the advertising platforms. 

    … one percent increase in e-commerce sales as a percent of total sales will decrease commercial real estate prices by 7.64%.[cxvii]

    It is worthwhile reading the whole economic paper on the decline in commercial real estate prices to understand the multiple factors that the author tried to take into account to better understand the impact of e-commerce sales. 

    The sales didn’t only shift online, but offshore. For instance, China-based advertisers accounted for around 11% of Meta’s total revenue in 2024[cxviii], which amounted to $18.35 billion. A significant portion of this is believed to come from large e-commerce companies like Temu and Shein[cxix], rather than a large number of small businesses. These companies benefited from the Chinese state support[cxx] covering their international logistics and postage costs and allowed their businesses to be run on razor-thin margins. 

    There has also been a corresponding value transfer from the lost profits of advertising clients to the platforms as well. Advertising industry consultant Michael Farmer made this point in his discussion of large fast-moving consumer goods businesses. 

    …for the fifty years from 1960 to 2010, the combined FMCG sales of P&G, Unilever, Nestle and Colgate-Palmolive grew at about an 8% compounded annual growth rate per year.

    The numbers associated with this long-term growth rate are staggering. P&G alone grew from about $1 billion (1960) to $79 billion in 2010. Throughout this period, P&G was the industry’s advocate for the power of advertising, becoming the largest advertiser in the US, with a focus on traditional advertising — digital / social advertising had hardly begun until 2010. Since 2010, with the advent of digital / social advertising, and massive increases in digital / social spend, P&G, Unilever, Nestle and Colgate-Palmolive have grown, collectively, at less than 1% per year, about half the growth rate of the US economy (2.1% per year).

    They are not the only major advertisers who have grown below GDP rates. At least 20 of the 50 largest advertisers in the US have grown below 2% per year for the past 15 years.

    Digital and social advertising, of course, have come to dominate the advertising scene since 2010, and it represents, today, about 2/3rds of all advertising spend.[cxxi]

    Digital advertising at its heart represents marketing efficiency because of its ability to be created and ‘trafficked’ at a much lower cost and greater speed. But this efficiency comes at the cost of corresponding marketing effectiveness in terms of short-term sales and longer-term preference and purchasing impact. 

    LLMs could undoubtedly further refine marketing efficiency, it could even ‘understand’ the marketing effectiveness challenge. But LLMs are restricted by the way the audience interacts with advertising, limiting their ability to solve the corresponding marketing effectiveness challenge. Marketing conglomerate WPP have launched a performance media platform that looks to further increase marketing efficiency by no longer requiring a traditional client-agency model. WPP Open Pro[cxxii] is the first advertising agency as a software service powered by an LLM. There is some concern that LLMs could destroy the very platforms which serve advertisers to consumers.[cxxiii]

    Based on all these factors, advertising is likely to be only one aspect of a market supporting AI’s growth and is unlikely to contribute more than a small proportion of the implied trillion-dollar payback required in the next two years if the dot LLM era doesn’t turn from boom to economic bust. 

    Business process efficiencies

    A second use case mentioned is deriving business efficiencies. This could be done in a number of ways:

    • Automating white-collar roles
    • Automating blue-collar and pink-collar roles in conjunction with robotics[cxxiv]

    OpenAI recently did research[cxxv] to find out how their service is being used. The sample looked across free, premium and corporate usage of ChatGPT. Some caveats around the research before we delve into it:

    • It ignored the use of API services. 
    • It is worthwhile remembering that ChatGPT may be under-represented for some actions like writing code – as developers are very aware of what is the current best tool for them.[cxxvi]

    Microsoft Worklab research[cxxvii] supports the view of LLM as wingman for white-collar workers. In a story arc that is similar to that of early personal computer adoption, they see LLM use as employee advocated and driven. 

    Actions have consequences

    Economists have models that look at the impact affecting unemployment[cxxviii], inflation and GDP. I have used the Phillips Curve[cxxix] and Okun’s Law[cxxx] in a thought experiment to model the effect on the US economy, if AI managed to provide up to $1 trillion in cost savings through automating jobs. Even with a notional cost savings of $1 trillion, the revenue that would accrue to LLM providers would be a very small proportion of the $1 trillion revenue growth over the next two years implied by current dot LLM era investments. 

    Methodology and assumptions

    • US baseline data.
    • Civilian labour force 170.8 million.
    • Number of unemployed 7.4 million.
    • Baseline unemployment rate 4.3%.[cxxxi]
    • Baseline annual inflation (CPI) 3.0%.
    • Baseline real GDP $23.8 trillion.
    • The average salary is about $94,952 (based on $45.65/hr[cxxxii] x 40 hours/week x 52 weeks/year).
    • $1 trillion in job cuts would represent about 10.53 million unemployed.

    Phillips Curve – used a standard slope where 1% increase in unemployment rate corresponds to a 0.5% decrease in inflation. Okun’s Law – I used a standard co-efficient where a 1% increase in unemployment rate corresponds to a 2% decrease in real GDP.  

    thought experiment

    The degree of economic change, at a time of deflation and drop in GDP would make the environment very hostile for businesses dependent on high growth rates. The economic model of achieving a $1 trillion payback through cost-savings is self-defeating. The very success of automation on that scale would destroy the macroeconomic environment required to sustain the hyperscalers’ growth projections.

    As we have seen in Japan during the lost decades,[cxxxiii] deflation would delay purchases and investments. The reduction in GDP would mean that there would be less money available for purchases and investments – creating a negative economic environment for all parties involved including the hyperscalers who would have precipitated the economic change. This scenario has alternative asset management firm Blackstone concerned that its peers are not considering the level of economic disruption the LLM era will bring.[cxxxiv]

    That is before you even consider the economic shockwave[cxxxv] that would roll around the globe in a similar manner[cxxxvi] to the 2008 financial crisis. All of this means that there is an optimal economic point in increasing productivity through dot LLM era automation without tanking future growth for hyperscalers and their clients. 

    AI optimists would think of the economic shockwave as being short-term in nature, followed by a long-term boom. In this respect, they would draw on examples like the rise of the steam engine, railways or electricity. On balance, I would disagree with these optimists. Economic conditions are very different now. For instance, western economies are now much more ‘financialised’[cxxxvii] and so the ‘short-term’ shockwave could be well over a decade in length, more similar to the great depression.[cxxxviii] Developed economy country governments may not have the headroom[cxxxix] to get out of the depression through a Franklin D. Roosevelt-style New Deal Keynesian stimulus.[cxl]

    Productivity benefits?

    Personally, I have found working with generative AI useful in a number of circumstances, in particular, solving the blank page problem. I have also used it as a research tool, a proof-reader and an editing partner. This article was written with the help of generative AI from an editing perspective. But I have also spent a lot of time looking at the outputs given and ensuring that they accurately reflected the exploration of where I wanted to go. And then there is the issue of hallucinations. 

    So far, the evidence has been mixed. There are a number of factors for this, IT projects are hard to implement successfully. 

    Businesses that have embraced LLMs to improve productivity have been penalised by investors due to the high upfront costs required.[cxli] Some critics claim that US data implies a plateauing of adoption of generative AI tools in companies[cxlii] – I personally think that this data is far from conclusive at the present time. 

    Some AI researchers like DeepSeek’s senior researcher Chen Deli believes that in the short-term AI could be a great assistant to humans, but over a longer period of 5-to-10 years it would threaten job losses as LLMs became good enough to replace humans in some forms of work. 

    “In the next 10-20 years, AI could take over the rest of work (humans perform) and society could face a massive challenge, so at the time tech companies need to take the role of ‘defender’,” he said. “I’m extremely positive about the technology but I view the impact it could have on society negatively.”[cxliii]

    Many of the leading companies in the LLM space such as Nvidia believe that the technology will drive a leap forward in robotics.[cxliv] Companies are currently building training sets on movement that are similar in function to the knowledge training sets used for LLMs. Even for well-known procedures, there are layers of formidable complexity to simple robotics tasks which would tax the most sophisticated process engineers.[cxlv]

    There are limiting factors outside the control of the LLM era ecosystem including power, the degree of control and limitations of mechanical engineering to supply chain challenges wrought by globalisation.[cxlvi] Both of which neither move at, or are related to Moore’s Law speed and scale of innovation. A key component is the strain wave gearing (also known as a harmonic drive)[cxlvii]which are made to standard sizes by very few companies, representing an innovation chokepoint, similar to ASML’s lithography machines in semiconductor manufacturing. The standard sizing limits capabilities from mechanical power to precision and increments of movement, which is one of the reasons why Apple still relies on hand assembly on its iPhones despite P&Ps (‘pick-and-place’[cxlviii] machines or surface mount technology (SMT) component placement machines) being available as far back as the 1980s. This chokepoint is one of the reasons why robotics vendors have focused on software-based differentiation with limited success so far.

    Different LLMs seem to lend themselves to different tasks as show by Anthropic[cxlix] and OpenAI’s[cl] own research into the economic and usage behaviour of their respective tools. 

    The Global environment

    Unlike other technological leaps forward, the LLM era isn’t likely to see American platform domination all around the world outside China. The dot-com era was the high point of American power. Coming out of the cold war, globalisation was benefiting US technology companies. The decline of Russia allowed the Clinton regime to open up the internet to commercial usage. American companies dominated enterprise software, semiconductors, wireless and computer network products. 

    25 years later, the US no longer has pre-eminence. Many of its past champions like Lucent[cli] or Motorola[clii] are either much reduced, or no longer American companies. Globalisation in the technology industries has meant that the concentration of expertise has become interconnected and dissipated to global centres of excellence such as TSMC[cliii], Foxconn[cliv] and Huawei[clv]. China had developed a parallel ecosystem some of which like Bytedance successfully compete head-to-head with large American technology platforms. 

    The LLM era is no longer only American in nature. Chinese companies have compelling offerings. For instance, Chinese hyperscaler Alibaba claim to be able to have models that are comparable to their American counterparts, yet needs 82% fewer Nvidia processors to run.[clvi] Even Silicon Valley companies are using Chinese LLM models over the likes of ChatGPT or Anthropic. The news that Airbnb opted to use Alibaba’s open-source Qwen AI model over ChatGPT was a milestone event.[clvii]US technology sector investors are using the Kimi K2 model because it was ‘way more performant and much cheaper than OpenAI and Anthropic’.[clviii] China benefits from much cheaper model training cost per token. The open-source models can be run on private infrastructure, keeping sensitive data inhouse and ensuring ‘corporate sovereignty.

    In the global south, China’s technology companies have corporate and government business relationships built up over years. Their combination of low cost combined with trusted relationships would reduce American hyperscaler opportunities for global expansion. 

    While US companies have access to more powerful chips, sanctions against Chinese companies aren’t effective with Nvidia chips being smuggled into China and heavy computing work like model training being run in data centres[clix] based in other Asian countries, notably Malaysia.[clx]

    There is one clear parallel between the earlier telecoms bubble and the dot LLM era; demand in the global south seems to be constrained by infrastructure rather than user interest in adopting generative AI tools.[clxi]

    Other bubbles.

    The dot-com era tends to be cited due to it being a technology story as much as an economic story. Many other bubbles were purely financial in nature:

    • The sub-prime mortgage crisis of 2008/9
    • The US savings and loans crisis of the early 1990s
    • 1929 stock market crash
    • Tulpenmanie from 1634 – 1637 

    The 1929 crash has sometimes been described as an electric generation bubble bursting since some 19% of the shares available on the market were from utility companies. But the impact was so widespread that it be hard to argue that it was really a ‘technology bubble’.[clxii]

    The British railway mania of the 1840s is often cited as an analogue of the telecoms bubble a century and a half later. The railway mania rolled out at a slower pace than the dot-com boom. It featured a Minsky moment and resulted in a consolidation of rail companies rather than an outright failure of many businesses. Up to a third of railway companies started during the time collapsed before building their railway line due to poor financial planning.[clxiii]

    The key defining factor for how bad the bust is from a bubble, and how long the bust lasts for is the amount of borrowing (or leverage) involved.[clxiv]

    How might the dot LLM era differ from the dot-com era in terms of the corresponding bust?

    Zero-cost co-ordination

    An economic paradigm shift will have occurred that doesn’t have a clear analogue in history that I am aware of. For instance, there are theoretical writings about how LLMs and agents will change the very nature of economics and the corporation may be changed with the advent of ‘zero-cost co-ordination’[clxv] reducing economic friction. This could upend the very nature of what a company is. 

    Historically one of the reasons given for participating in a firm was that internal coordination costs were cheaper than market coordination (transaction costs). If agentic AI are rational actors that reduce market transaction costs (search, negotiation, contracting) to near-zero, the need for large, hierarchical firms changes and likely diminishes.[clxvi]

    If this theory were true, the excessive capital expenditure would simply be the price paid for creating the world’s first zero-friction economic system. In theory, it’s possible, but it depends on the humans involved being rational decision-makers in a rational culture that doesn’t exhibit risk aversion and that their agents don’t develop similar biases over time. This often isn’t true, even in business-to-business situations, for instance in the past ‘nobody ever got fired for buying IBM’.[clxvii]

    This viewpoint in some ways is similar to Wired magazine’s editorial team circa 1998 and futurist author Kevin Kelly’s ideas on the ‘new economy’.[clxviii] The thesis was that the internet would reduce information friction. The dot-com bust provided a more tempered lens on the ideas of the ‘new economy’. Would efforts to reduce economic friction fare any better than the information friction reduction of the ‘new economy’?

    Google Research economists have asked this same question[clxix] and came back with more open questions than answers. The authors posit that AI systems, being built on optimisation principles, can be modelled as standard “textbook” economic agents. when AI agents deviate from perfect rationality, they may exhibit an “emergent preference” and display behavioural biases similar to those found in humans. They highlight what they termed the “contract” problem. It draws an analogy between the AI alignment problem and the economic theory of ‘incomplete contracts,’ where a designer (the principal) cannot perfectly specify the AI agent’s goals, leading to unpredictable behaviour. The economists were concerned there would be a need for new institutions to govern an AI agent economy to ensure markets remain well-functioning and stable.

    The open questions:

    • Whether AI agents have stable ‘beliefs’?
      • How they update them? 
      • If they can hold ‘higher-order beliefs’ (beliefs about others’ beliefs)?
    • There is a lack of research and benchmarks for evaluating AI performance in complex, multi-agent systems which needs to be addressed. One of the key challenges is that small differences between AI and human behaviour can become magnified in an equilibrium.

    But what if, as Francis Fukuyama argues,[clxx] that transaction friction isn’t the block on economic growth? Instead, it’s resource constraints, social and political considerations that are the brake on how fast economic growth can happen. 

    AI-fuelled breakthroughs

    The infrastructure boom fuels foundational AI research far beyond current capabilities. In this scenario, active engines of scientific discovery. The AI research achieves a breakthrough in a hard-science field like drug discovery (e.g., new classes of effective antibiotics), materials science (e.g., room-temperature superconductors), novel ways of rare earth metal extraction, or sustained controllable nuclear fusion – and facilitates record compression of time to market for these developments. LLMs would not only have to facilitate the breakthrough, but drive mass-accelerated implementation and regulation. 

    In theory, LLMs could: 

    • Optimise experiment and trial design.
    • In- and post-test data analysis. 
    • Drive synthesis of regulatory compliance documents and evidence. 
    • Optimise production and supply chains to facilitate the manufacture and commercialisation of a new break-out product. 

    If all this happened, it would create entirely new sources of economic value, far dwarfing the infrastructure cost. That is a lot of serendipity, of huge scope and massive assumptions: even the NASA Apollo Program[clxxi] took eight years to have its first crewed lunar flight[clxxii] and another year to put the first men on the moon.  

    AI-fuelled breakthroughs are usually linked with progress towards AGI or ‘artificial general intelligence’ or human level intelligence AI.[clxxiii] A research paper from Cornell University that outlined benchmarking for progress to understanding the real world. The paper introduced WorldTest, a new framework for evaluating how AI agents learn and apply internal world models through reward-free exploration and behaviour-based testing in modified environments. Its implementation, revealed that while humans excel at prediction, planning, and change detection tasks, leading AI reasoning models still fall short. Their shortcoming was associated with flexible hypothesis testing and belief updating. The findings suggest that future progress in AI world-modelling depends less on scaling compute and more on improving metacognition, exploration strategy, and adaptive reasoning.

    Platform lock-in and bundling

    Many of the established hyperscalers (Adobe, Alphabet, Amazon, Microsoft, Oracle and Salesforce) have established client relationships in a range of products:

    • CRM.
    • Creative Suite and Marketing Cloud.
    • Office suite or Workspace.
    • Enterprise Cloud services.

    Rather than a disruptive paradigm shift, the LLM payback could come from an instant, embedded non-disruptive increase across existing indispensable products and services. It extracts the value from the existing enterprise wallet, which breaks the historical analogy of relying on new economic value creation. On the face of it, a largely risk-free proposition.

    The US legal environment is very different from the dot-com era. Microsoft would not have to worry about facing an antitrust trial similar to its conflict over bundling with Netscape.[clxxiv]  

    While in the US, antitrust enforcement is considered laxer than during the Biden regime, these technology companies would be concerned about competition regulators in the EU and elsewhere. For example, just this September, Microsoft had to unbundle Teams from its Office software to avoid EU antitrust fines.[clxxv] Alphabet[clxxvi] and Amazon[clxxvii] have had previous bruising run-ins with authorities outside the US which would complicate any decision made to bundle an LLM service. 

    What could dot LLM era outcomes look like?

    I have come up with seven scenarios that range in the kind of impact that generative AI as a sector may provide. These range from being wildly successful to dark failure

    • The breakthrough: total economic transformation due to a post-war breakthrough in science and technology.
    • The ‘new economy’: frictionless co-ordination facilitates more economic activity.
    • The ‘wingman economy’: a managed productivity boom.
    • The ‘Red Hat model’: an open-source foundation driving value-added services.
    • The ‘moral hazard’: major AI players are considered ‘too big to fail’ and backstopped with government loan guarantees.
    • The ‘telecoms bust’: a Minsky moment and amortisation crisis.
    • The ‘weird gizmo’: collapse total bust. 

    How these scenarios map out when thinking about the level of value creation or value saved through increased efficiency.

     Negative / zero net value createdPositive to transformative value creation
    New value creationThe ‘weird gizmo’ collapse (value was illusory)The breakthrough (new science)
    The ‘new economy’ (new coordination)
    Efficiency / existing valueThe ‘telecoms bust’ Capex > value
    The ‘moral hazard’ value is geopolitical rather than financial
    The ‘wingman economy’ (managed productivity)
    The ‘Red Hat’ model (value moves to services)

    The breakthrough: total economic transformation

    What it looks like: The massive capital expenditure on infrastructure is validated because AI achieves a true, hard-science breakthrough. This creates entirely new sources of economic value, such as sustained nuclear fusion, room-temperature superconductors, or new classes of antibiotics. In this outcome, the $1 trillion in implied value is not only met but vastly exceeded. Justifying the “bubble” as the necessary investment for a new industrial revolution.

    What to watch? 

    • Scientific breakthroughs. 

    Metric: 

    • High-impact scientific publications that use AI for novel discovery, NOT just analysis.

    Source: 

    • Track major journals like NatureNew Scientist and Science for breakthroughs in AI-driven drug discovery, materials science, or physics. Recent reports on AI’s role in molecular innovation and even quantum computing show this is a key area to watch.

    The “new economy”: frictionless co-ordination

    What it looks like: Agentic AI successfully reduces market transaction costs (search, negotiation, contracting) to near-zero. This upends the nature of the corporation, as the historical reason for firms (cheaper internal vs. market coordination) diminishes. The massive capital expenditure is seen as the “price paid for creating the world’s first zero-friction economic system”. This is the 1998 Wired “new economy” thesis finally coming true, though it faces challenges like the “Contract problem” and AI alignment.

    What to watch? 

    • Agentic breakthroughs

    Metric: 

    • Demonstrations of “agentic” AI (AI that can independently complete complex, multi-step tasks), particularly in commercial or economic settings.

    Source: 

    • Monitor announcements from leading research labs (DeepMind, FAIR, OpenAI) and market analysis on “agentic AI” to see if it’s moving from theory to reality.

    The ‘wingman’ economy: a managed productivity boom

    What it looks like: The technology finds its “optimal economic point”. LLMs become a powerful “wingman for white-collar workers”, similar to the adoption of early PCs. This drives real productivity gains, but the $1 trillion in cost savings is implemented gradually, avoiding the catastrophic deflationary shock modelled by the Phillips Curve and Okun’s Law. The “Magnificent 10” see steady growth, but the ‘pure play’ LLMs struggle to find profitability on their own.

    What to Watch: 

    • National Productivity Data
    • Enterprise Adoption & AI Mentions in Earnings

    Metrics: 

    • U.S. labour Productivity and unit labour costs. We are looking for a “golden path”: productivity rising faster than unit labour costs, which would suggest companies are becoming more efficient without just slashing jobs en-masse.[clxxviii]
    • The number of S&P 500 companies citing “AI” on their quarterly earnings calls. A high number (e.g., over 40-50%) shows it’s a top-level strategic priority.

    Sources: 

    • U.S. Bureau of Labor Statistics (BLS) – productivity and costs. The quarterly releases from the BLS are the single best macro-indicator for this scenario.
    • FactSet Earnings Insight.[clxxix]  – they regularly publish analyses on the frequency of “AI” mentions in earnings calls, which is a direct proxy for corporate focus and investment.

    The Red Hat analogue: a foundational model

    What it looks like: The “pure play” LLMs like OpenAI and Anthropic, which are losing money, ultimately fail or are acquired for pennies on the dollar. However, open-source and open weight models (like Llama, etc.) proliferate. Alibaba’s Qwen model has already been very successful. Singapore’s national AI programme dropped Meta’s Llama in favour of it.[clxxx] Singapore joins Airbnb as Qwen users;[clxxxi] meanwhile Chinese model DeepSeek has been adopted by European startups.[clxxxii] The long-term winners are not the model creators but the companies that, like Red Hat, sell “enterprise-grade support, integration, and services”. 

    LLM models have an “outsized impact” —becoming the “Linux” for the next generation of applications—but the initial investors see a massive correction.

    What to Watch: 

    • Open-source vs. closed-source momentum

    Metric: 

    • Rate of change in download statistics, new model uploads, and developer activity on open-source AI platforms.

    Source: 

    • Hugging Face Trends.[clxxxiii] This dashboard shows which open-source models are gaining traction. If downloads for open-source models are growing faster than API call revenue for closed-source models (a harder metric to find), it signals a shift toward this “Red Hat” scenario. GitHub’s annual “Octoverse” report is another key source, as it tracks the rise of AI-focused projects.

    The ‘moral hazard’: major dot LLM players are considered ‘too big to fail’ and backstopped with government intervention

    There are elements of a non-bubble, financial crisis aspect to the dot LLM era. Chinese LLM vendors are being given subsidised electricity from local governments,[clxxxiv] alongside preferential rates in data centres. The LLM era in the US could be considered by the government as having become too large a part of the economy to be allowed to fail due to normal market forces. Open AI has recently had to deny rumours[clxxxv] that it sought US government loan guarantees for at least part of the multi-trillion dollar deals it has put in place for data centre infrastructure and hardware. AI sovereignty comes to be seen as taking on a geostrategic and national security imperative as business and investor considerations take a backseat. 

    Hyperscalers are hitting a ‘power wall’ as they cannot get the equivalent electricity generating capacity of 16 Hoover dams. Getting over the wall would require a massive amount of government infrastructure funding.[clxxxvi]

    Major government involvement may impact the speed of development as LLM model providers and supporting infrastructure no longer have to constantly innovate and instead move at the speed of their government clients. 

    What to watch:

    • Shift in rhetoric from commercial to critical: Observe how language from policymakers, military leaders, and national security bodies evolves. A shift from discussing AI in terms of commercial competition (e.g., “market leadership”) to national infrastructure (e.g., “digital sovereignty,” “critical asset,” “geostrategic imperative”) is a primary indicator. This reframes an economic failure as a national security failure.
    • Direct & indirect state support mechanisms: look beyond simple R&D grants. Watch for the creation of new, targeted support instruments:
      • Direct: preferential pricing on energy/compute, state-backed datacentre construction, sovereign wealth fund investments, or direct “national champion” subsidies.
      • Indirect: government-backed loan guarantees for infrastructure (like the rumoured OpenAI deal), strategic procurement (where the government becomes the anchor customer) – Palantir would be an exemplar, and “regulatory moats” that favour incumbents (e.g., high-cost safety/licensing rules that only large, state-backed labs can afford).
    • “Bailout” vs. “investment” framing: monitor how state intervention is publicly justified. A struggling “national champion” AI firm receiving a sudden capital injection from a state-adjacent entity will likely be framed as a “strategic investment in national capability,” not a “bailout.” This framing is key.

    Metrics:

    • Value of state & military contracts: Track the total disclosed value of government contracts (especially from defence and intelligence agencies) awarded to foundational model providers. A rapid increase, or contracts for non-competitive “strategic deployment,” signals TBTF (“Too Big to Fail”) status.
    • Frequency analysis of policy language: quantify the co-occurrence of terms like “AI,” “sovereignty,” “national security,” and “critical infrastructure” in parliamentary/congressional records, national strategy documents, and defence budget justifications. A rising frequency indicates the ideological groundwork for a TBTF policy.
    • State-backed capital flows: monitor announcements from sovereign wealth funds, national investment banks (e.g., UK’s National Security Strategic Investment Fund), or public pension funds. Track the size and frequency of their investments into large, established AI labs, as opposed to a diverse portfolio of early-stage start-ups.
    • Subsidy disclosures: quantify the value of announced subsidies (e.g., tax credits, energy discounts, land grants) specifically earmarked for AI datacentres and R&D hubs associated with the major players.

    Sources:

    • Financial & policy journalism: The Financial TimesBloomberg (especially its Bloomberg Government vertical), and Politico as media sources. Their reporters are often the first to break stories on subsidies, lobbying, and the intersection of tech and state power.
    • Government procurement & grant databases: official portals like USASpending.gov in the US or the UK’s Contracts Finder service. While difficult to navigate, they provide primary evidence of public funds flowing to specific companies.
    • Think tank & national security publications: Reports from organisations like the Center for a New American Security (CNAS) in the US, the Royal United Services Institute (RUSI) in the UK, or the Mercator Institute for China Studies (MERICS). They often analyse and quantify the geostrategic rhetoric and policy shifts. The main challenge with this source might be timeliness of publication in comparison to the previous sources. 
    • Company filings & investor calls: For publicly traded companies (Microsoft, Google, Amazon, Nvidia), annual reports (10-K forms) and quarterly investor calls often mention large government contracts or regulatory risks/opportunities, providing a corporate-side view of this trend.

    The Telecoms Bust: a Minsky moment and amortisation crisis

    What it looks like: The $1 trillion in value fails to materialize from either advertising or business efficiencies. Investors have a Minsky moment and realize the debt and capex are unsustainable. The bubble implodes like the telecoms bubble. The key difference is the financial and technological amortisation risk: the GPUs (with a 2-to-5-year useful life) become obsolete. Unlike the dot-com era’s dark fibre, this infrastructure cannot be repurposed by a “web 2.0”. This leads to trillions in write-offs, analogous to WorldCom’s $180 billion loss.

    What to Watch: 

    • Hyperscaler capital expenditure (Capex)
    • GPU amortisation & resale value

    Metrics: 

    • Quarterly capex announcements from Google (Alphabet), Meta, Microsoft, Oracle and Amazon (AWS). This is made trickier to understand by Meta, Microsoft and Oracle looking at forms of private equity financing. 
    • The rate of change in Nvidia’s[clxxxvii] data centre revenue, Broadcom and AMD’s enterprise / data centre revenue. This is the “equipment maker” side of the equation. As long as this number is growing, the bubble is inflating. A sudden slowdown would be the first sign of a “Minsky Moment.”
    • The resale value of last-generation GPUs (e.g., H100s as B200s/B300s roll out). If these prices collapse, it validates your thesis that the assets cannot be repurposed, and the financial write-downs will be catastrophic.

    Sources: 

    • Hyperscaler capex reports from financial analysts and data centre publications. Recent reports show combined capex is projected to hit hundreds of billions, a clear sign of the infrastructure race.
    • Alphabet, Meta, Microsoft, Oracle and Amazon quarterly results and investor roadshow presentations.
    • NVIDIA, Broadcom and AMD quarterly earnings reports. The Nvidia Q2 2026 report showing data centre revenue at $41.1B is a perfect example of this indicator.
    • Resale value of GPUs is a harder metric to track. Monitor tech hardware forums and eBay listings, or look for analyst reports on the “used GPU market.” A collapse in this secondary market for last generation GPUs is a major red flag.

    The “Weird Gizmo” Collapse: total bust

    What it looks like: The technology is ultimately seen as a novelty. It’s the 2020s version of Boo.com, Beenz and Flooz, or the 3Com Audrey. The argument that “AGI is not imminent[clxxxviii], and LLMs are not the royal road to getting there” wins the day. This bear view of AGI is one that is widely shared by prominent experts[clxxxix] within the machine learning field. Which is why new ways of working like nesting models and world models are being explored, alongside quantum computing. In this scenario, the pure play companies burn through all their cash and vanish. The hyperscalers are left with billions in useless, obsolete silicon, and the “dot LLM era” is remembered as a short-lived period of speculative mania.

    What to Watch: 

    • AI startup burn rates & funding (the “burn Rate” indicator)

    Metric: 

    • Quarterly venture capital funding for AI startups, specifically looking for a rise in “down rounds” (where valuations decrease) or outright failures.

    Source: 

    • Data from firms like CB Insights or Crunchbase.[cxc] Recent reports show that while “mega-rounds” for established players (like Anthropic) are still huge, seed-stage funding is declining, showing a “haves and have-nots” market. A slowdown in the mega-rounds would signal the bust is beginning.

    Personal assessment of likely outcomes by scenario

    ScenarioEstimated likelihood Rationale
    The moral hazard~95%US – China trade disputes and geopolitical strife

    Chinese government investment in startups

    Chinese local government subsidies for operating AI services

    The current position that AI has in driving US GDP growth across sectors including construction and the energy sector

    Likely OpenAI loan guarantees

    Palantir is already deeply embedded in the US government as a vendor and has partnerships with defence contractors like Anduril
    The ‘wingman’ economy~80-90%Some research reports indicate that AI is augmenting knowledge workers in different sectors. 

    Claims of AI replacing workers are more difficult to validate, for example: 
    Klarna moving to automation and then rehiring 

    Clifford Chance offshoring back-office roles to Poland and China while claiming that the job losses were due to AI.
    The ‘Red Hat’ Model~70-80%Airbnb opting to use Alibaba’s open-source Qwen AI model over ChatGPT was a milestone event.[cxci]
    The ‘telecoms bust’~75%Concerns about the size of capital expenditure.

    Rate of growth of supporting infrastructure.

    Uncertainty about length of depreciation affecting overall shareholder trust in hyperscalers. 

    Cheaper alternatives like Qwen.
    The ‘new economy’<15%The uncertain economics of ‘zero friction’ transactions.

    Real-life legal and regulatory issues. 

    Amazon’s dispute with Perplexity using AI agents on its website. 
    The breakthrough<10%A black swan event
    The ‘weird gizmo’<5%It would be unusual for a technology to disappear completely,

    LLMs have been finding some use already.

    The rise of open-source AI models which reduce the cost of operation. 

    Where are we at the moment?

    I worked to put together a diagram to try and assess where we are at the moment given that some of the scenarios outlined are running concurrently with each other. 

    Where are we at the moment?

    Acknowledgements

    Ian Wood (Wireless Foundry),

    Colophon

    The dot LLM era is brought to you with the assistance of:


    [i] Pluralistic: The real (economic) AI apocalypse is nigh (27 Sep 2025) – Pluralistic: Daily links from Cory Doctorow

    [ii] The real (economic) AI apocalypse is nigh | Pluralistic

    [iii] Morning, Ellis (May 7, 2025) The Daily WTF – AI: The Bad, The Worse and The Ugly

    [iv] AMD rallies 24% on OpenAI Deal and Bari Weiss Takes Over CBS News | Prof G Markets Podcast with Ed Elson

    [v] Wong, Matteo (July 24, 2024) The Atlantic – Silicon Valley’s Trillion-Dollar Leap of Faith

    [vi] Agarwal, P. (2025) Moats in AI Startups: Data, Distribution, or Default? (India) Eximius Ventures

    [vii] James Anderson warns Nvidia’s $100bn OpenAI bet echoes dotcom bubble | FT

    [viii] AI’s Power Demand Is Driving Up Your Electricity Bill | Prof G Markets

    [ix] Grabar, Henry (May 18, 2022) Slate – The Decade of Cheap Rides Is Over

    [x] Harris, A. (2025) CAPITAL IDEAS: The market’s valuation rivals the dot-com bubble. What’s next for stock prices? (US) The Berkshire Edge

    [xi] S&P P/E Ratio Is Low, But Has Been Lower (2009) Seeking Alpha

    [xii] How Does the End Begin? – No Mercy / No Malice

    [xiii] Meyer, Katherine (May 3, 2006) The Wall Street Journal – The Best of the Worst

    [xiv] Grant, Conor (July 1, 2018) The Hustle – A decade before crypto, one digital currency conquered the world – then failed spectacularly

    [xv] Stepanek, Marcia (September 28, 2000). “The CueCat Is on the Prowl: This gizmo is on the cutting edge of e-marketing. But with each swipe, it tracks your moves through cyberspace”. Bloomberg Businessweek.

    [xvi] Kanellos, Michael & Wong, Wylie (March 21, 2001) CNet News – Audrey’s life cut short

    [xvii] Cringely, Robert. X (December 2, 2004) PBS I, Cringely – Wishing for Audrey (Now That the World is Finally Ready for Internet Appliances, Where Are They?

    [xviii] Centre for Computing History – Ergo Audrey

    [xix] Chidi Jr, George A (June 18, 2001) CNN – Sony launches Net access device

    [xx] United States Securities and Exchange Commission Form 8-K palmOne, Inc. (October 28, 2003)

    [xxi] PC Mag UK (December 5, 2000) 3Com Ergo Audrey review

    [xxii] Beaumont, M. (2000) e. The novel of liars, lunch and lost knickers & The e Before Christmas (UK) HarperCollins

    [xxiii] Introducing Nested Learning: A new ML paradigm for continual learning | Google Research

    [xxiv] Tett, G. (2025) Behind the AI bubble, another tech revolution could be brewing (UK) Financial Times

    [xxv] Kelly, K. (2010) What Technology Wants (US) Viking Press

    [xxvi] Boo Hoo: A Dot Com Story Paperback by Ernst Malmsten, Erik Portanger and Charles Drazin

    [xxvii] The report was subsequently published in Panic! edited by Michael Lewis

    [xxviii] Scott Galloway CV (PDF)

    [xxix] Crunchbase | redenvelope.com

    [xxx] Rival walks off with Geldof’s Deckchair | Guardian

    [xxxi] The Everything Store: Jeff Bezos and the Age of Amazon (2014) by Brad Stone

    [xxxii] Spector, Robert (2002). Amazon.com: Get Big Fast.

    [xxxiii] CNN Money | Amazon posts first ever profit in 4Q (January 22, 2002)

    [xxxiv] Computerworld | Amazon records first profitable year in its history (January 28, 2004)

    [xxxv] Y2K – renaissance chambara

    [xxxvi] Dell plus Sun equals VA Research | Forbes

    [xxxvii] VA Linux Sets IPO Record – Wired

    [xxxviii] History of Red Hat Linux – Fedora Project Wiki

    [xxxix] SuSE Linux for S/390 Available Today – SuSE

    [xl] IBM Closes Landmark Acquisition of Red Hat for $34 Billion; Defines Open, Hybrid Cloud Future – Red Hat press release

    [xli] Lessons from History: The Rise and Fall of the Telecom Bubble – Fabricated Knowledge by Doug O’Laughlin

    [xlii] British Telecom privatisation share prospectus advert (1984)

    [xliii] 1984-2014 – 30 years of the Janet network (Jisc)

    [xliv] Ellassen, K.A, & From, J. (2009) Deregulation, privatisation and public service delivery: Universal service in telecommunications in Europe (United Kingdom) Policy and Society Vol. 27, issue 3

    [xlv] https://www.instituteforgovernment.org.uk/sites/default/files/british_telecom_privatisation.pdf              

    [xlvi] Herrera-González, F. & Castejón-Martin, L. (2009) The endless need for regulation in telecommunication: An explanation (UK) Telecommunications Policy Volume 22 Issues 10-11

    [xlvii] Morgan, K. Monopolies under Siege in Western Europe – Institute of Development Studies

    [xlviii] Hayashi, K. and Fuke, H. “Changes and deregulation in the Japanese telecommunications market,” in IEEE Communications Magazine, vol. 36, no. 11, pp. 46-53, Nov. 1998.

    [xlix] Tsatsou, P. (2010) EU regulations on telecommunications: The role of subsidiarity and mediation – First Monday volume 16, issue 1

    [l] Clinton signs telecom bill (February 8, 1996) – UPI

    [li] Telecommunications Act of 1996 – Federal Communications Commission (FCC)

    [lii] Brotman, S.N. (2016) Was the 1996 Telecommunications Act successful in promoting competition? – Brookings Institute

    [liii] Barlow, J.P. (1996) A Declaration of the Independence of Cyberspace – Electronic Frontier Foundation

    [liv] https://www.wto.org/english/tratop_e/serv_e/symp_mar02_uk_com_e.pdf            

    [lv] Issues in Labor Statistics – US Department of Labor Bureau of Labor Statics Summary 99-4 March 1999

    [lvi] Manes, S. (1996) Jaz Drive: A Lot of Backup Insurance in a Small Package (US) The New York Times

    [lvii] Elmer-Dewitt, P. (1993) Take A Trip into the Future on the Electronic Superhighway (US) TIME  

    [lviii] On-ramp Prospects for the Information Superhighway Dream by Gordon Bell and Jim Gemmell COMMUNICATIONS OF THE ACM July 1996/Vol. 39, No. 7 

    [lix] Gates, Bill; Myhrvold, Nathan and Rinearson, Peter. (1995). The Road Ahead (1st ed.) (US) Viking

    [lx] Botein, M. (2003) The Demise of The Information Superhighway (US) Digital @NYLS (New York Law School)

    [lxi] Odlyzko, A.M., Internet traffic growth: Sources and implications (US) University of Minnesota

    [lxii] Litan, R.E. (2002) he Telecommunications Crash: What to Do Now? (US) Brookings Institute

    [lxiii] O’Laughlin, D. (2023) Lessons from History: The Rise and Fall of the Telecom Bubble (US) Fabricated Knowledge

    [lxiv] Hunter, D. (2018) Nortel (Canada) The Canadian Encyclopedia

    [lxv] Lebowitz, M. (November 5, 2025) Nvidia Deals: Round Tripping or Vendor Financing? (US) Investing.com

    [lxvi] Claycombe, C. WorldCom/MCI: Massive Accounting Fraud (US) Wichita State University

    [lxvii] Starr, P. (2002) The Great Telecom Implosion (US) American Prospect magazine

    [lxviii] McLean, B & Elkind, P. (2003) The Smartest Guys in the Room: The Amazing Rise and Scandalous Fall of Enron (US) Portfolio

    [lxix] PUBLIC LAW 107–204—JULY 30, 2002116 STAT. 745

    [lxx] The Laws That Govern the Securities Industry – U.S. Securities and Exchange Commission

    [lxxi] Worthen, B., White, B. (October 27, 2008) Tech Finance Defaults Rise; More Makers Offer Loans (US) The Wall Street Journal

    [lxxii] Meeker, M. (1996) Internet Trends (US) Morgan Stanley

    [lxxiii] Kedrosky, P. (2025) Minsky Moments and AI CapEx (US) paulkedrosky.com

    [lxxiv] PS Lee National University of Singapore on LinkedIn

    [lxxv] GPU Life Concerns: Reality And Implications (2024) Beyond the Hype – Looking Past Management & Wall Street Hype

    [lxxvi] Curry, T., Shibut, L. (2000) The Cost of the Savings and Loan Crisis: Truth and Consequences (US) FDIC Banking Review

    [lxxvii] 1993: Recession over – it’s official | BBC

    [lxxviii] Schwartz, P., Leyden, P. (1997) The Long Boom: A History of the Future, 1980 – 2020 (US) Wired magazine

    [lxxix] Kadlec, D. (1999) Day Trading: It’s a Brutal World (US) Time magazine

    [lxxx] Sor, J. (2025) ‘A very lonely sport’: Day traders on the isolating experience of trying to make a living in the stock market (US) Business Insider

    [lxxxi] Are rich countries facing a debt crisis – The Economist on YouTube

    [lxxxii] Climate despair (2023) renaissance chambara

    [lxxxiii] (2025) How many people are already being killed by climate change? (UK) The Economist

    [lxxxiv] Moshiri, A. (2025) Devastation on repeat: How climate change is worsening Pakistan’s deadly floods (UK) BBC

    [lxxxv] Raval, A. (2025) The AI job cuts are accelerating (UK) FT

    [lxxxvi] Meeker, M. (1995 on) Internet reports (US) Bond Capital

    [lxxxvii] Meeker, M., Simons, J., Chae, D., Krey, A. (2025) Trends: Artificial Intelligence (US) Bond Capital

    [lxxxviii] Misra, A., Wang, J., McCullers, S., White, K., and Ferres, J.L. (2025) Measuring AI Diffusion: A Population-Normalized Metric for Tracking Global AI Usage (US) Microsoft

    [lxxxix] (2024) Elon Musk – AI will be smarter than the smartest human | Bloomberg Technology

    [xc] Amodei, D. (2024) Machines of Loving Grace (US) self-published on blog.

    [xci] Pillay, T. (2025) How OpenAI’s Sam Altman Is Thinking About AGI and Superintelligence in 2025 (US) TIME

    [xcii] (2025) The Three-Year Countdown: Inside DeepMind’s AGI Timeline and What It Means for Knowledge Work (US) The Ai Consultancy on Medium

    [xciii] Kurzweil, R. (2024 & updated in 2025) The Singularity is Nearer: When We Merge with AI (US) Random House Publishing

    [xciv] (2025) Joe Rogan: The Truth About Aliens (He Finally Says It) (US) Jesse Michels podcast on YouTube

    [xcv] Ditlea, S. (1996) Leary’s Final Trip, the Web, Realized Multimedia Vision (US) The New York Times Online

    [xcvi] Bearak, B. (1997) Eyes on Glory: Pied Pipers of Heaven’s Gate (US) The New York Times Online

    [xcvii] (2025) Andrej Karpathy — “We’re summoning ghosts, not building animals” (US) Dwarkesh Patel podcast on YouTube

    [xcviii] Brooks, R. (2025) Predictions Scorecard, 2025 January 01 (US) published on personal blog

    [xcix] Leonards, A. Meta chief AI scientist claims AGI will be viable in 3-5 years (UK) National Technology News

    [c] Mitchell, M. (2025) On the Science of “Alien Intelligences”: Evaluating Cognitive Capabilities in Babies, Animals, and AI (US) NeuroIPS

    [ci] Heath, A. (2025) I talked to Sam Altman about the GPT-5 launch fiasco (US) The Verge

    [cii] Schmidt, E., Xu, S. (2025) Silicon Valley Is Drifting Out of Touch with the Rest of America (US) The New York Times

    [ciii] Kharpal, A. (2025) Jeff Bezos says AI is in an industrial bubble but society will get ‘gigantic’ benefits from the tech (US) CNBC

    [civ] Islam, F., Clun, R. (2025) Google boss says trillion-dollar AI investment boom has ‘elements of irrationality’ (UK) BBC

    [cv] (2025) Mark Zuckerberg on the AI bubble and Meta’s new display glasses (US) ACCESS Podcast on YouTube

    [cvi] Kinder, T., Hammond, G. (2025) OpenAI shunned advisers on $1.5tn of deals (UK) Financial Times

    [cvii] Evans, J. (2025) Bubble, bubble, toil and trouble (US) Gradient Ascendant

    [cviii] Li, Y. (2025) ‘Big Short’ investor Michael Burry accuses AI hyperscalers of artificially boosting earnings (US) CNBC

    [cix] Davis, G.B. (2025) 6 Stock Market Lessons from the Dot Com Bubble That Apply in 2025 (US) Yahoo! Finance

    [cx] Galloway, S. (2025) How Does the End Begin? (US) No mercy / no malice

    [cxi] Harnett, I. (2025) The AI capex endgame is approaching (UK) The Financial Times

    [cxii] Galloway, S. (2025) How Does the End Begin? | No Mercy / No Malice (US) Prof G Media

    [cxiii] Kemp, S. (2025) Digital 2025: Global Advertising Trends (Singapore) DataReportal

    [cxiv] The Value of Advertising – World Federation of Advertisers

    [cxv] Hiorns, B. (2023) A Brief History of AI in Advertising #HistoryMonth (UK) Creativepool

    [cxvi] Goldberg, L. (2018) A brief history of artificial intelligence in advertising (UK) Econsultancy

    [cxvii] McGowan, Jacob, “How Has the Growth of E-commerce Sales Affected Retail Real Estate?” (2019). CMC Senior Theses. 2189.

    [cxviii] Merritt, M. (2025) Ad dollars from China are already starting to dry up (US) MorningBrew

    [cxix] Tiprank (2025) Meta Could Face a Massive $7 Billion Ad Revenue Hit from China Tariffs, Warns Analyst (Canada) Globe and Mail

    [cxx] Camille Boullenois, Agatha Kratz and Daniel H. Rosen (2025) Far From Normal: An Augmented Assessment of China’s State Support (US) Rhodium Group

    [cxxi] Farmer, M. (2025) Madison Avenue Media Madness (US) C-Suite Blues

    [cxxii] WPP Open Pro: empowering brands to plan, create and publish campaigns independently (2025) WPP

    [cxxiii] AI may fatally wound web’s ad model, warns Tim Berners-Lee | FT

    [cxxiv] Beltran, M. (2025) Japanese convenience stores are hiring robots run by workers in the Philippines (US) Rest of the World

    [cxxv] Chatterji, A., Cunningham, T., Deming, D., Hitzig, Z., Ong, C., Shan, C., Wadman, K. (2025) How People Use ChatGPT (US) OpenAI, Duke University and Harvard University

    [cxxvi] Coding LLM leaderboard – Vellum.ai

    [cxxvii] AI at Work Is Here. Now Comes the Hard Part 2024 (US) Microsoft Worklab

    [cxxviii] Wessel Vermeulen, Nils Braakmann (2023) How do mass lay-offs affect regional economies? OECD Local Economic and Employment Development (LEED) Papers 2023/01

    [cxxix] How the Unemployment Rate Affects Everybody | Investopedia

    [cxxx] Understanding Okun’s Law: How GDP Growth Affects Unemployment | Investopedia

    [cxxxi] The Employment Situation – August 2025 (US) Bureau of Labor Statistics

    [cxxxii] Employer Costs for Employee Compensation – June 2025 (US) Bureau of Labor Statistics

    [cxxxiii] Jones, R. (2025) A Modern Economic History of Japan: Sho Ga Nai (It Is What It Is) (UK) London Publishing Partnership

    [cxxxiv] Blackstone says Wall Street is complacent about AI disruption | FT

    [cxxxv] Impact of the Global Financial Crisis and Its Implications for the East Asian Economy, Keynote Speech by Mr. Takatoshi Kato, Deputy Managing Director, International Monetary Fund, At the Korea International Financial Association, First International Conference

    [cxxxvi] Andrew Filardo, Jason George, Mico Loretan, Guonan Ma, Anella Munro, Ilhyock Shim, Philip Wooldridge, James Yetman and Haibin Zhu The international financial crisis: timeline, impact and policy responses in Asia and the Pacific. (Bank of International Settlements)

    [cxxxvii] Fontana, G., Dixon, G. (2017) Unlocking the puzzles of financialisation (UK) Applied Institute for Research in Economics

    [cxxxviii] Ross Sorkin, A. (2025) 1929: The Inside Story of The Greatest Crash in Wall Street History (US) Allen Lane

    [cxxxix] Global Debt Report 2025 – OECD

    [cxl] How Keynes Influenced FDR’s New Deal – Future Hindsight

    [cxli] AI’s awfully exciting until companies want to use it: Rightmove edition | FT

    [cxlii] Spencer, M. (2025) Going Short on Generative AI (US) AI Supremacy

    [cxliii] Mo, L., Goh, B. (November 7, 2025) DeepSeek researcher pessimistic over AI’s impact in startup’s first public appearance since success (UK) Reuters

    [cxliv] The Minds of Modern AI: Jensen Huang, Yann LeCun, Fei-Fei Li & the AI Vision of the Future | FT Live – YouTube

    [cxlv] Ford, M. (July 2015) A History of Placement Programming and Optimization (US) Circuits Assembly

    [cxlvi] Is there an end in sight to supply chain disruption? | Financial Times

    [cxlvii] Automata Eve launch | renaissance chambara

    [cxlviii] Component Placement Process – Surface Mount Process

    [cxlix] (2025) Anthropic Economic Index (Anthropic seem to be treating this exercise as a longitudinal research project). 

    [cl] Chatterji, A., Cunningham, T., Deming, D., Hitzig, Z., Ong, C., Shan, C., Wadman, K., (2025) How People Use ChatGPT (US) OpenAI, Duke University & Harvard University

    [cli] (1998 – 2025) AT&T Corporation (US) Encyclopaedia Britannica

    [clii] (1998 – 2023) Motorola, Inc. (US) Encyclopaedia Britannica

    [cliii] Montevirgen, K. (2025) Taiwan Semiconductor Manufacturing Co. (TSMC) (US) Encyclopaedia Britannica

    [cliv] Chinatsu, T. (2025) Foxconn (US) Encyclopaedia Britannica

    [clv] Dou, E. (2025) House of Huawei (UK) Abacus

    [clvi] Chow, V. (2025) Alibaba Cloud claims to slash Nvidia GPU use by 82% with new pooling system (Hong Kong) South China Morning Post

    [clvii] Broersma, M. (2025) Airbnb praises Alibaba’s Open-Source AI model (UK) Silicon

    [clviii] Kynge, J. (2025) Low-cost Chinese AI models forge ahead, even in the US, raising the risks of a US AI bubble (UK) Chatham House

    [clix] Baptista, E., Tang, A., Yong, J.Y. (2025) Malaysia reins in data centre growth, complicating China’s AI chip access (UK) Reuters

    [clx] Jennings, R. (2025) How Malaysia’s data centres became the engine powering China’s AI ambitions (Hong Kong) South China Morning Post

    [clxi] Misra, A., Wang, J., McCullers, S., White, K., and Ferres, J.L. (2025) Measuring AI Diffusion: A Population-Normalized Metric for Tracking Global AI Usage (US) Microsoft

    [clxii] Sorkin, A.R. (2025) 1929: The Inside Story of The Greatest Crash in Wall Street History (US) Allen Lane

    [clxiii] Odlyzko, A. (2010) Collective hallucinations and inefficient markets: The British Railway Mania of the 1840s (US) University of Minnesota

    [clxiv] Sorkin A.R. (2025) Odd Lots: Andrew Ross Sorkin on the Stock Market Crash That Shattered America (US) Bloomberg

    [clxv] Perez, C.E. (2025) The Intelligence Abundance: How Zero-Cost Coordination Solves the Scarcity Problem

    [clxvi] Coase, R.H. (1937) The Nature of the Firm (UK) Economica volume 4, issue 16 published by the London School of Economics

    [clxvii] Melamed, G. (2024) Nobody gets fired for buying IBM (UK) Finextra

    [clxviii] Kelly, K. (1998) New Rules for the New Economy (US) Viking

    [clxix] Hadfield, G.K., Koh, A. (2025) An Economy of AI agents (US) NBER Handbook on the Economics of Transformative AI

    [clxx] Fukuyama, F. (2025) Superintelligence Isn’t Enough (US) Persuasion

    [clxxi] Ostovar, M. (1998) The Decision to Go to the Moon: President John F. Kennedy’s May 25, 1961 Speech before a Joint Session of Congress (US) NASA

    [clxxii] Brooks, C.G., James M. Grimwood, J.M., Swenson, Jr., L.S. (1979) The NASA History Series: Chariots for Apollo: A History of Manned Lunar Spacecraft

    [clxxiii] Warrier, A.,2, Nguyen, T.D., Naim, M., Jain, M., Liang, Y., Schroeder, K., Yang, C., Tenenbaum, J.B., Vollmer, S., Ellis, K., Tavares, Z. (2025) Benchmarking World-Model Learning (US) Cornell University

    [clxxiv] (2000) Microsoft vs the US Justice Dept. Netscape: A history (UK) BBC

    [clxxv] Warren, T. (2025) Microsoft avoids EU fine after Slack complained about Teams bundling (US) The Verge

    [clxxvi] (2020) Google is unbundling Android apps: all the news about the EU’s antitrust ruling (US) The Verge

    [clxxvii] Espinoza, J. (2020) EU accuses Amazon of breaching antitrust rules (UK) FT

    [clxxviii] U.S. Bureau of Labor Statistics (BLS) – Productivity and Costs – quarterly data

    [clxxix] FactSet Insight blog – Search their blog for keywords like “AI” or “earnings.” They regularly publish analyses on the number of S&P 500 companies that cite “AI” on their earnings calls, which is a direct proxy for C-suite focus.

    [clxxx] (2025) Singapore’s national AI program drops Meta model and switches to Alibaba’s Qwen | TechNode

    [clxxxi] Broersma, M. (2025) Airbnb Praises Alibaba’s Open-Source AI Model (UK) Silicon

    [clxxxii] Broersma, M. (2025) European Start-Ups Adopt DeepSeek To Cut Costs (UK) Silicon

    [clxxxiii] Hugging Face models hub – the view can be filtered by ‘trending’ and ‘most downloaded’ to see what the community is using, versus what closed source models are being marketed

    [clxxxiv] Gao, J. (November 8, 2025) How China hits hard to power its AI ambitions post-Nvidia (Taiwan) DigiTimes Asia

    [clxxxv] Sam Altman says OpenAI is not ‘trying to become too big to fail’ | FT

    [clxxxvi] JPMorgan’s Playbook for a 10-15% Correction (or Worse) — ft. Michael Cembalest | Prof G Markets – YouTube

    [clxxxvii] Nvidia investor relations page – The key figure in their quarterly financial reports is ‘Data Center revenue’.

    [clxxxviii] Marcus, G. (2025) Game over. AGI is not imminent, and LLMs are not the royal road to getting there. (US) Marcus on AI

    [clxxxix] The Godmother of AI on jobs, robots & why world models are next | Dr. Fei-Fei Li (US) Lenny’s Podcast on YouTube

    [cxc] Crunchbase News – They provide regular analysis of funding rounds. Watch for ‘down rounds’, M&A consolidation among start-ups or acquihires and slowdowns in $100M+ mega-rounds of fund raising. 

    [cxci] Broersma, M. (2025) Airbnb praises Alibaba’s Open-Source AI model (UK) Silicon

  • May 2025 newsletter

    May 2025 introduction – two little ducks (22) edition

    Welcome to my May 2025 newsletter, this newsletter marks my 22nd issue. 22 is known in bingo halls and the Spanish national lottery as two little ducks.

    Double Duck

    In France, 22 is the equivalent of 5-0 in the English speaking world as slang for the police. 22 is an important number for people who believe in numerology. In Hong Kong, 22 is associated with good fortune. This is down to the number sounding similar to ‘easy’ or ‘bright’ in Cantonese.

    I hope that you are tricked into thinking I am bright based this newsletter, so let’s jump in. Inspired by catching up with my old DJing partner Griff, this month I enjoyed the unashamedly joyous pumped-up sounds of Blackpool’s AZYR at the Boiler Room x TeleTech Festival in 2023. In particular the transition at the end of the set between Frankyeffe – Save me and Infectious! – I need your lovin’. (Extra trainspotter points if you knew that Infectious! is a homage / remake of N.R.G’s The Real Hardcore from a year earlier). Wear your headphones, it might be divisive playing the set out loud in the office. More bangers from AZYR here.

    New reader?

    If this is the first newsletter, welcome! You can find my regular writings here and more about me here

    Strategic outcomes

    Things I’ve written.

    • Predicting market share through share of search volume and what the rise of AI likely means.
    • Reaching a precipice in hydrogen power and trends in Chinese skincare amongst other things.

    Books that I have read.

    • Careless People by Sarah Wynn Williams. Williams account of her time in Facebook had become the most discussed book of the spring in my social circle. I wrote a long review of it here.
    The Road to Conscious Machines
    • The Road to Conscious Machines by Michael Wooldridge examines the profound cultural impact of generative AI, which is currently experiencing a surge in both its cultural influence and practical applications. Drawing parallels to the internet’s transformative impact in the mid-to-late 1990s, where it permeated various aspects of society and fostered rapid adoption, Wooldridge traces the evolution of generative AI as a phenomenon that emerged gradually over the past half-century. Throughout the book, Wooldridge provides a comprehensive historical overview of AI, including the periods of research stagnation known as AI winters. This historical perspective equips readers with a nuanced understanding of the strengths and weaknesses of AI, enabling them to approach AI adoption with a well-informed perspective.
    • As I finish this newsletter during the bank holiday weekend, my light reading is Rogue Asset by Andy McDermott. McDermott comes from a long line of British authors like Jack Higgins, Len Deighton, Frederick Forsyth and Mick Herron who provide novels aimed at a shrinking pool of readers – men. At least, if one is to believe what’s said in the media. Rogue Asset hinges on the premise that the UK has a unit which assassinates the countries enemies on a regular basis. Think somewhere between The Troubles era Det and the modern deep state trope. Our hero is snared into the plot by being discovered on the run thanks to his online behaviour – which is attributed to GCHQ; (but isn’t as mysterious as it sounds because of the programmatic advertising technology stack). So far so good for what it is. I will let know if it goes downhill as a read next month.

    Things I have been inspired by.

    Mmrytok

    Limitations are often the mother of invention. That seems to be the theory behind mmrytok. Mmrytok allows you to do one post a day. It doesn’t support HTML formatting, it doesn’t allow you to link out and doesn’t have a newsfeed. So it’s easy-to-use because it’s less sophisticated than Geocities was. In this respect it is to social media and blogs what Punkt is to smartphones. In an always-on social time, I have found it liberating to use. You can see my page here. I heard of Mmrytok thanks to Matt Muir’s great newsletter Web Curios.

    No, AI isn’t making you dumber

    Australian documentary maker ColdFusion put together an interesting video essay on How AI is making you dumber.

    Yes, you could argue that under certain attributes the population isn’t as smart as they have been in the past. Just last month I shared an article by John Burn-Murdoch. In the article he shared data of a longitudinal trend across countries and age-groups struggling with concentration, declining verbal and numerical reasoning. The problem with Burn-Murdoch’s article vis-a-vis the ColdFusion video is the timeline.

    His article charts a decline further back than the rise of generative AI services. Mia Levitin in an essay for the FT attributed the decline in reading to the quick dopamine hits of social media content.

    A college professor interviewed by The Atlantic put the decline in reading amongst his undergraduate students put it down to a practice in secondary education of atomising content. Pupils in high schools were assigned excerpts, poetry and news articles to read, but not complete books. This has impacted the size of vocabulary and grasp of language that students starting university now have.

    James Gleick

    This isn’t new territory, James Gleick in his book Faster documented the massive acceleration of information through the late 20th century and its effects on the general public. The underlying accelerant was described by Kevin Kelly in What Technology Wants as the technium – a continuous forward progress due to a massively interconnected system of technology.

    There were concerns in research as far back as the late 1980s that television could be adversely affecting children’s reading comprehension and attention spans.

    TL;DR – with generative AI you could become dumber, if you use it unwisely – but the problem lies with all of us and what we chose to do with our personal agency.

    CIA advertise for Chinese spies

    The CIA commissioned a couple of high production value adverts that they’ve been running on social media channels. The adverts are designed to encourage Chinese government employees to come forward as an agent. The sales pitch is about taking control.

    CIA China advert

    A translation of the Chinese tagline: ‘The reason for choosing cooperation: to become the master of (one’s own) destiny‘. More details from the FT about the campaign here, and here’s the two executions currently running on YouTube.

    It remains to be seen if the campaign will be effective. The Chinese Ministry of State Security managed to roll-up the CIA’s spy network back in 2010-2012. Up to 30 informants in China were executed.

    Montirex

    montirex

    Merseyside sports-inspired lifestyle brand Montirex have published a film telling the brand story from its origins to the present day. The brand is expanding beyond its Merseyside roots to get national and international sales.

    Trust, attitudes and use of artificial intelligence

    A 2025 global study covering some 48 countries was conducted by KPMG in association with the University of Melbourne. Some key insights from the report. Consumer generative AI is being used instead of enterprise options by workers. Generative AI adopters still have self-perceived low AI skills but that doesn’t slow their adoption. There is higher adoption and trust rates in emerging markets than in developed markets.

    Pro and anti-trust AI issues solidifying

    Year-on-year we are seeing an increase in both distrust and trust for specific AI use cases, indicating that it is becoming a polarising subject. The lowest trust levels is in tech-savvy Finland. More here.

    Chart of the month. 

    McDonald’s Restaurants saw a decline in sales. This was down to low income consumers spending less, while middle class earners still weren’t going into McDonalds. Normally when there is a recession, McDonalds should benefit from the more well-off trading down to McDonalds. Instead, fortunes have diverged into a ‘k-shaped’ recession. Lower income earners are hit, while middle classes aren’t. What Axios called the ‘McRecession‘.

    McDonald's quarterly sales growth

    Things I have watched. 

    Tony Arzenta (also known as Big Guns). The film is an early 1970s gallo film. French star Alain Delon appears in this classic retribution story based in Milan. As Tony Arzenta, Delon exacts revenge on the former bosses who killed his family by accident in a botched assassination attempt to prevent him from retiring.The film uses a wintry Milan as a good atmospheric backdrop for the action that plays out in a series of shoot-outs and car chases. It’s John Wick before it was even conceived. Delon brings a tension that other stars of the era like Charles Bronson failed to do in similar roles. As Arzenta’s targets flee across Europe, he goes through Germany and Denmark to catch up with them.

    Sansho the Bailiff – as a film Sansho the Bailiff comes encumbered with a weight of praise. It is highly rated by film critics and Martin Scorsese had it as one of his must-watch films for young film makers. Director Kenji Mizoguchi assembled an ensemble cast of Japanese actors to tell a story of family hardship and poverty. Kazuo Miyagawa is key to the the production, providing a signature look to the cinematography. There is a tension between the emotional rollercoaster of the story and the reflective nature of the scenes portrayed – I don’t want to say too more, except that even the character actors like Kikue Môri (who plays a pivotal role in the plot as a priestess) are amazing in the film.

    Warfare – I was a bit leery of watching Alex Garland’s Warfare after watching Civil War which was strong on aesthetics and emotion, but weak in terms of the creative conceits involved in making the story work. Warfare is the collective accounts of a US military unit during a two-hour fire fight. The story is told from multiple perspectives in real-time. The film captures the stress and boredom of inaction as well as what you would normally expect from this kind of film.

    Useful tools.

    Reddit Answers

    Reddit Answers – alternative to Gigabrain that I recommended back in March. Like Gigabrain, Reddit Answers looks like the kind of knowledge search product that we failed to build at Yahoo! twenty years ago (or NORA as Microsoft has been calling the concept for the past few years). Reddit Answers is powered by Google Vertex AI.

    Process online data like its peak web 2.0 all over again

    While WordPress installations come with RSS enabled as standard and is something that can then be disabled, many types of sites aren’t RSS enabled. And where they are the web devs will often disable it just because. RSS app will create an RSS feed for websites that don’t have it. This allows you to pull it into data processing using something like Pipes. RSS app starts at $9.99 per month and goes up to $99.99 a month. Pipes starts at free and goes up to $79 per month.

    The sales pitch.

    I am currently working on a brand and creative strategy engagement at Google’s internal creative agency.

    now taking bookings

    I am now taking bookings for strategic engagements in Q4 (October) – keep me in mind; or discussions on permanent roles. Contact me here.

    More on what I have done here.

    bit.ly_gedstrategy

    The End.

    Ok this is the end of my May 2025 newsletter, I hope to see you all back here again in a month. Be excellent to each other and onward into spring, and I hope you enjoyed the last bank holiday until August.

    Don’t forget to share if you found it useful, interesting or insightful.

    Get in touch if there is anything that you’d like to recommend for the newsletter.

  • Intelligence per watt

    My thinking on the concept of intelligence per watt started as bullets in my notebook. It was more of a timeline than anything else at first and provided a framework of sorts from which I could explore the concept of efficiency in terms of intelligence per watt. 

    TL;DR (too long, didn’t read)

    Our path to the current state of ‘artificial intelligence’ (AI) has been shaped by the interplay and developments of telecommunications, wireless communications, materials science, manufacturing processes, mathematics, information theory and software engineering. 

    Progress in one area spurred advances in others, creating a feedback loop that propelled innovation.  

    Over time, new use cases have become more personal and portable – necessitating a focus on intelligence per watt as a key parameter. Energy consumption directly affects industrial design and end-user benefits. Small low-power integrated circuits (ICs) facilitated fuzzy logic in portable consumer electronics like cameras and portable CD players. Low power ICs and power management techniques also helped feature phones evolve into smartphones.  

    A second-order effect of optimising for intelligence per watt is reducing power consumption across multiple applications. This spurs yet more new use cases in a virtuous innovation circle. This continues until the laws of physics impose limits. 

    Energy storage density and consumption are fundamental constraints, driving the need for a focus on intelligence per watt.  

    As intelligence per watt improves, there will be a point at which the question isn’t just what AI can do, but what should be done with AI? And where should it be processed? Trust becomes less about emotional reassurance and more about operational discipline. Just because it can handle a task doesn’t mean it should – particularly in cases where data sensitivity, latency, or transparency to humans is non-negotiable. A highly capable, off-device AI might be a fine at drafting everyday emails, but a questionable choice for handling your online banking. 

    Good ‘operational security’ outweighs trust. The design of AI systems must therefore account not just for energy efficiency, but user utility and deployment context. The cost of misplaced trust is asymmetric and potentially irreversible.

    Ironically the force multiplier in intelligence per watt is people and their use of ‘artificial intelligence’ as a tool or ‘co-pilot’. It promises to be an extension of the earlier memetic concept of a ‘bicycle for the mind’ that helped inspire early developments in the personal computer industry. The upside of an intelligence per watt focus is more personal, trusted services designed for everyday use. 

    Integration

    In 1926 or 27, Loewe (now better known for their high-end televisions) created the 3NF[i].

    While not a computer, but instead to integrate several radio parts in one glass envelope vacuum valve. This had three triodes (early electronic amplifiers), two capacitors and four resistors. Inside the valve the extra resistor and capacitor components went inside their own glass tubes. Normally each triode would be inside its own vacuum valve. At the time, German radio tax laws were based on the number of valve sockets in a device, making this integration financially advantageous. 

    Post-war scientific boom

    Between 1949 and 1957 engineers and scientists from the UK, Germany, Japan and the US proposed what we’d think of as the integrated circuit (IC). These ideas were made possible when breakthroughs in manufacturing happened. Shockley Semiconductor built on work by Bell Labs and Sprague Electric Company to connect different types of components on the one piece of silicon to create the IC. 

    Credit is often given to Jack Kilby of Texas Instruments as the inventor of the integrated circuit. But that depends how you define IC, with what is now called a monolithic IC being considered a ‘true’ one. Kilby’s version wasn’t a true monolithic IC. As with most inventions it is usually the child of several interconnected ideas that coalesce over a given part in time. In the case of ICs, it was happening in the midst of materials and technology developments including data storage and computational solutions such as the idea of virtual memory through to the first solar cells. 

    Kirby’s ICs went into an Air Force computer[ii] and an onboard guidance system for the Minuteman missile. He went on to help invent the first handheld calculator and thermal printer, both of which took advantage of progress in IC design to change our modern way of life[iii]

    TTL (transistor-to-transistor logic) circuitry was invented at TRW in 1961, they licensed it out for use in data processing and communications – propelling the development of modern computing. TTL circuits powered mainframes. Mainframes were housed in specialised temperature and humidity-controlled rooms and owned by large corporates and governments. Modern banking and payments systems rely on the mainframe as a concept. 

    AI’s early steps 

    Science Museum highlights

    What we now thing of as AI had been considered theoretically for as long as computers could be programmed. As semiconductors developed, a parallel track opened up to move AI beyond being a theoretical possibility. A pivotal moment was a workshop was held in 1956 at Dartmouth College. The workshop focused on a hypothesis ‘every aspect of learning or any other feature of intelligence can be so precisely described that a machine can be made to simulate it’. Later on, that year a meeting at MIT (Massachusetts Institute of Technology) brought together psychologists and linguists to discuss the possibility of simulating cognitive processes using a computer. This is the origin of what we’d now call cognitive science. 

    Out of the cognitive approach came some early successes in the move towards artificial intelligence[iv]. A number of approaches were taken based on what is now called symbolic or classical AI:

    • Reasoning as search – essentially step-wise trial and error approach to problem solving that was compared to wandering through a maze and back-tracking if a dead end was found. 
    • Natural language – where related phrases existed within a structured network. 
    • Micro-worlds – solving for artificially simple situations, similar to economic models relying on the concept of the rational consumer. 
    • Single layer neural networks – to do rudimentary image recognition. 

     By the time the early 1970s came around AI researchers ran into a number of problems, some of which still plague the field to this day:

    • Symbolic AI wasn’t fit for purpose solving many real-world tasks like crossing a crowded room. 
    • Trying to capture imprecise concepts with precise language.
    • Commonsense knowledge was vast and difficult to encode. 
    • Intractability – many problems require an exponential amount of computing time. 
    • Limited computing power available – there was insufficient intelligence per watt available for all but the simplest problems. 

    By 1966, US and UK funding bodies were frustrated with the lack of progress on the research undertaken. The axe fell first on a project to use computers on language translation. Around the time of the OPEC oil crisis, funding to major centres researching AI was reduced by both the US and UK governments respectively. Despite the reduction of funding to the major centres, work continued elsewhere. 

    Mini-computers and pocket calculators

    ICs allowed for mini-computers due to the increase in computing power per watt. As important as the relative computing power, ICs made mini-computers more robust, easier to manufacture and maintain. DEC (Digital Equipment Corporation) launched the first minicomputer, the PDP-8 in 1964. The cost of mini-computers allowed them to run manufacturing processes, control telephone network switching and control labouratory equipment. Mini-computers expanded computer access in academia facilitating more work in artificial life and what we’d think of as early artificial intelligence. This shift laid the groundwork for intelligence per watt as a guiding principle.

    A second development helped drive mass production of ICs – the pocket calculator, originally invented at Texas Instruments.  It demonstrated how ICs could dramatically improve efficiency in compact, low-power devices.

    LISP machines and PCs

    AI researchers required more computational power than mini-computers could provide, leading to the development of LISP machines—specialised workstations designed for AI applications. Despite improvements in intelligence per watt enabled by Moore’s Law, their specialised nature meant that they were expensive. AI researchers continued with these machines until personal computers (PCs) progressed to a point that they could run LISP quicker than LISP machines themselves. The continuous improvements in data storage, memory and processing that enabled LISP machines, continued on and surpassed them as the cost of computing dropped due to mass production. 

    The rise of LISP machines and their decline was not only due to Moore’s Law in effect, but also that of Makimoto’s Wave. While Gordon Moore outlined an observation that the number of transistors on a given area of silicon doubled every two years or so. Tsugio Makimoto originally observed 10-year pivots from standardised semiconductor processors to customised processors[v]. The rise of personal computing drove a pivot towards standardised architectures. 

    PCs and workstations extended computing beyond computer rooms and labouratories to offices and production lines. During the late 1970s and 1980s standardised processor designs like the Zilog Z80, MOS Technology 6502 and the Motorola 68000 series drove home and business computing alongside Intel’s X86 processors. 

    Personal computing started in businesses when office workers brought a computer to use early computer programmes like the VisiCalc spreadsheet application. This allowed them to take a leap forward in not only tabulating data, but also seeing how changes to the business might affect financial performance. 

    Businesses then started to invest more in PCs for a wide range of uses. PCs could emulate the computer terminal of a mainframe or minicomputer, but also run applications of their own. 

    Typewriters were being placed by word processors that allowed the operator to edit a document in real time without resorting to using correction fluid

    A Bicycle for the Mind

    Steve Jobs at Apple was as famous for being a storyteller as he was for being a technologist in the broadest sense. Internally with the Mac team he shared stories and memetic concepts to get his ideas across in everything from briefing product teams to press interviews. As a concept, a 1990 filmed interview with Steve Jobs articulates the context of this saying particularly well. 

    In reality, Jobs had been telling the story for a long time through the development of the Apple II and right from the beginning of the Mac. There is a version of the talk that was recorded some time in 1980 when the personal computer was still a very new idea – the video was provided to the Computer History Museum by Regis McKenna[vi].

    The ‘bicycle for the mind’ concept was repeated in early Apple advertisements for the time[vii] and even informed the Macintosh project codename[viii]

    Jobs articulated a few key concepts. 

    • Buying a computer creates, rather than reduces problems. You needed software to start solving problems and making computing accessible. Back in 1980, you programmed a computer if you bought one. Which was the reason why early personal computer owners in the UK went on to birth a thriving games software industry including the likes of Codemasters[ix]. Done well, there should be no seem in the experience between hardware and software. 
    • The idea of a personal, individual computing device (rather than a shared resource).  My own computer builds on my years of how I have grown to adapt and use my Macs, from my first sit-up and beg Macintosh, to the MacBook Pro that I am writing this post on. This is even more true most people and their use of the smartphone. I am of an age, where my iPhone is still an appendage and emissary of my Mac. My Mac is still my primary creative tool. A personal computer is more powerful than a shared computer in terms of the real difference made. 
    • At the time Jobs originally did the speech, PCs were underpowered for anything but data processing (through spreadsheets and basic word processor applications). But that didn’t stop his idea for something greater. 

    Jobs idea of the computer as an adjunct to the human intellect and imagination still holds true, but it doesn’t neatly fit into the intelligence per watt paradigm. It is harder to measure the effort developing prompts, or that expended evaluating, refining and filtering generative AI results. Of course, Steve Jobs Apple owed a lot to the vision shown in Doug Engelbart’s ‘Mother of All Demos’[x].

    Networks

    Work took a leap forward with office networked computers pioneered by Macintosh office by Apple[xi]. This was soon overtaken by competitors. This facilitated work flow within an office and its impact can still be seen in offices today, even as components from print management to file storage have moved to cloud-based services. 

    At the same time, what we might think of as mobile was starting to gain momentum. Bell Labs and Motorola came up with much of the technology to create cellular communications. Martin Cooper of Motorola made the first phone call on a cellular phone to a rival researcher at Bell Labs. But Motorola didn’t sell the phone commercially until 1983, as a US-only product called the DynaTAC 8000x[xii].  This was four years after Japanese telecoms company NTT launched their first cellular network for car phones. Commercial cellular networks were running in Scandinavia by 1981[xiii]

    In the same way that the networked office radically changed white collar work, the cellular network did a similar thing for self-employed plumbers, electricians and photocopy repair men to travelling sales people. If they were technologically advanced, they may have had an answer machine, but it would likely have to be checked manually by playing back the tape. 

    Often it was a receptionist in their office if they had one. Or more likely, someone back home who took messages. The cell phone freed homemakers in a lot of self-employed households to go out into the workplace and helped raise household incomes. 

    Fuzzy logic 

    The first mainstream AI applications emerged from fuzzy logic, introduced by Lofti A. Zadeh in 1965 mathematical paper. Initial uses were for industrial controls in cement kilns and steel production[xiv]. The first prominent product to rely on fuzzy logic was the Zojirushi Micom Electric Rice Cooker (1983), which adjusted cooking time dynamically to ensure perfect rice. 

    Rice Cooker with Fuzzy Logic 3,000 yen avail end june

    Fuzzy logic reacted to changing conditions in a similar way to people. Through the 1980s and well into the 1990s, the power of fuzzy logic was under appreciated outside of Japanese product development teams. In a quote a spokesperson for the American Electronics Association’s Tokyo office said to the Washington Post[xv].

    “Some of the fuzzy concepts may be valid in the U.S.,”

    “The idea of better energy efficiency, or more precise heating and cooling, can be successful in the American market,”

    “But I don’t think most Americans want a vacuum cleaner that talks to you and says, ‘Hey, I sense that my dust bag will be full before we finish this room.’ “

    The end of the 1990s, fuzzy logic was embedded in various consumer devices: 

    • Air-conditioner units – understands the room, the temperature difference inside-and-out, humidity. It then switches on-and-off to balance cooling and energy efficiency.
    • CD players – enhanced error correction on playback dealing with imperfections on the disc surface.
    • Dishwashers – understood how many dishes were loaded, their type of dirt and then adjusts the wash programme.
    • Toasters – recognised different bread types, the preferable degree of toasting and performs accordingly.
    • TV sets – adjust the screen brightness to the ambient light of the room and the sound volume to how far away the viewer is sitting from the TV set. 
    • Vacuum cleaners – vacuum power that is adjusted as it moves from carpeted to hard floors. 
    • Video cameras – compensate for the movement of the camera to reduce blurred images. 

    Fuzzy logic sold on the benefits and concealed the technology from western consumers. Fuzzy logic embedded intelligence in the devices. Because it worked on relatively simple dedicated purposes it could rely on small lower power specialist chips[xvi] offering a reasonable amount of intelligence per watt, some three decades before generative AI. By the late 1990s, kitchen appliances like rice cookers and microwave ovens reached ‘peak intelligence’ for what they needed to do, based on the power of fuzzy logic[xvii].

    Fuzzy logic also helped in business automation. It helped to automatically read hand-written numbers on cheques in banking systems and the postcodes on letters and parcels for the Royal Mail. 

    Decision support systems & AI in business

    Decision support systems or Business Information Systems were being used in large corporates by the early 1990s. The techniques used were varied but some used rules-based systems. These were used in at least some capacity to reduce manual office work tasks. For instance, credit card approvals were processed based on rules that included various factors including credit scores. Only some credit card providers had an analyst manually review the decision made by system.  However, setting up each use case took a lot of effort involving highly-paid consultants and expensive software tools. Even then, vendors of business information systems such as Autonomy struggled with a high rate of projects that failed to deliver anything like the benefits promised. 

    Three decades on, IBM had a similar problem with its Watson offerings, with particularly high-profile failure in mission-critical healthcare applications[xviii]. Secondly, a lot of tasks were ad-hoc in nature, or might require transposing across disparate separate systems. 

    The rise of the web

    The web changed everything. The underlying technology allowed for dynamic data. 

    Software agents

    Examples of intelligence within the network included early software agents. A good example of this was PapriCom. PapriCom had a client on the user’s computer. The software client monitored price changes for products that the customer was interested in buying. The app then notified the user when the monitored price reached a price determined by the customer. The company became known as DealTime in the US and UK, or Evenbetter.com in Germany[xix].  

    The PapriCom client app was part of a wider set of technologies known as ‘push technology’ which brought content that the netizen would want directly to their computer. In a similar way to mobile app notifications now. 

    Web search

    The wealth of information quickly outstripped netizen’s ability to explore the content. Search engines became essential for navigating the new online world. Progress was made in clustering vast amounts of cheap Linux powered computers together and sharing the workload to power web search amongst them.  As search started to trying and make sense of an exponentially growing web, machine learning became part of the developer tool box. 

    Researchers at Carnegie-Mellon looked at using games to help teach machine learning algorithms based on human responses that provided rich metadata about the given item[xx]. This became known as the ESP game. In the early 2000s, Yahoo! turned to web 2.0 start-ups that used user-generated labels called tags[xxi] to help organise their data. Yahoo! bought Flickr[xxii] and deli.ico.us[xxiii]

    All the major search engines looked at how deep learning could help improve search results relevance. 

    Given that the business model for web search was an advertising-based model, reducing the cost per search, while maintaining search quality was key to Google’s success. Early on Google focused on energy consumption, with its (search) data centres becoming carbon neutral in 2007[xxiv]. This was achieved by a whole-system effort: carefully managing power management in the silicon, storage, networking equipment and air conditioning to maximise for intelligence per watt. All of which were made using optimised versions of open-source software and cheap general purpose PC components ganged together in racks and operating together in clusters. 

    General purpose ICs for personal computers and consumer electronics allowed easy access relatively low power computing. Much of this was down to process improvements that were being made at the time. You needed the volume of chips to drive innovation in mass-production at a chip foundry. While application-specific chips had their uses, commodity mass-volume products for uses for everything from embedded applications to early mobile / portable devices and computers drove progress in improving intelligence-per-watt.

    Makimoto’s tsunami back to specialised ICs

    When I talked about the decline of LISP machines, I mentioned the move towards standardised IC design predicted by Tsugio Makimoto. This led to a surge in IC production, alongside other components including flash and RAM memory.  From the mid-1990s to about 2010, Makimoto’s predicted phase was stuck in ‘standardisation’. It just worked. But several factors drove the swing back to specialised ICs. 

    • Lithography processes got harder: standardisation got its performance and intelligence per watt bump because there had been a steady step change in improvements in foundry lithography processes that allowed components to be made at ever-smaller dimensions. The dimensions are a function wavelength of light used. The semiconductor hit an impasse when it needed to move to EUV (extreme ultra violet) light sources. From the early 1990s on US government research projects championed development of key technologies that allow EUV photolithography[xxv]. During this time Japanese equipment vendors Nikon and Canon gave up on EUV. Sole US vendor SVG (Silicon Valley Group) was acquired by ASML, giving the Dutch company a global monopoly on cutting edge lithography equipment[xxvi]. ASML became the US Department of Energy research partner on EUV photo-lithography development[xxvii]. ASML spent over two decades trying to get EUV to work. Once they had it in client foundries further time was needed to get commercial levels of production up and running. All of which meant that production processes to improve IC intelligence per watt slowed down and IC manufacturers had to start about systems in a more holistic manner. As foundry development became harder, there was a rise in fabless chip businesses. Alongside the fabless firms, there were fewer foundries: Global Foundries, Samsung and TSMC (Taiwan Semiconductor Manufacturing Company Limited). TSMC is the worlds largest ‘pure-play’ foundry making ICs for companies including AMD, Apple, Nvidia and Qualcomm. 
    • Progress in EDA (electronic design automation). Production process improvements in IC manufacture allowed for an explosion in device complexity as the number of components on a given size of IC doubled every 18 months or so. In the mid-to-late 1970s this led to technologists thinking about the idea of very large-scale integration (VLSI) within IC designs[xxviii]. Through the 1980s, commercial EDA software businesses were formed. The EDA market grew because it facilitated the continual scaling of semiconductor technology[xxix]. Secondly, it facilitated new business models. Businesses like ARM Semiconductor and LSI Logic allowed their customers to build their own processors based on ‘blocs’ of proprietary designs like ARM’s cores. That allowed companies like Apple to focus on optimisation in their customer silicon and integration with software to help improve the intelligence per watt[xxx]
    • Increased focus on portable devices. A combination of digital networks, wireless connectivity, the web as a communications platform with universal standards, flat screen displays and improving battery technology led the way in moving towards more portable technologies. From personal digital assistants, MP3 players and smartphone, to laptop and tablet computers – disconnected mobile computing was the clear direction of travel. Cell phones offered days of battery life; the Palm Pilot PDA had a battery life allowing for couple of days of continuous use[xxxi]. In reality it would do a month or so of work. Laptops at the time could do half a day’s work when disconnected from a power supply. Manufacturers like Dell and HP provided spare batteries for travellers. Given changing behaviours Apple wanted laptops that were easy to carry and could last most of a day without a charge. This was partly driven by a move to a cleaner product design that wanted to move away from swapping batteries. In 2005, Apple moved from PowerPC to Intel processors. During the announcement at the company’s worldwide developer conference (WWDC), Steve Jobs talked about the focus on computing power per watt moving forwards[xxxii]

    Apple’s first in-house designed IC, the A4 processor was launched in 2010 and marked the pivot of Makimoto’s wave back to specialised processor design[xxxiii].  This marked a point of inflection in the growth of smartphones and specialised computing ICs[xxxiv]

    New devices also meant new use cases that melded data on the web, on device, and in the real world. I started to see this in action working at Yahoo! with location data integrated on to photos and social data like Yahoo! Research’s ZoneTag and Flickr. I had been the Yahoo! Europe marketing contact on adding Flickr support to Nokia N-series ‘multimedia computers’ (what we’d now call smartphones), starting with the Nokia N73[xxxv].  A year later the Nokia N95 was the first smartphone released with a built-in GPS receiver. William Gibson’s speculative fiction story Spook Country came out in 2007 and integrated locative art as a concept in the story[xxxvi]

    Real-world QRcodes helped connect online services with the real world, such as mobile payments or reading content online like a restaurant menu or a property listing[xxxvii].

    I labelled the web-world integration as a ‘web-of-no-web’[xxxviii] when I presented on it back in 2008 as part of an interactive media module, I taught to an executive MBA class at Universitat Ramon Llull in Barcelona[xxxix]. In China, wireless payment ideas would come to be labelled O2O (offline to online) and Kevin Kelly articulated a future vision for this fusion which he called Mirrorworld[xl]

    Deep learning boom

    Even as there was a post-LISP machine dip in funding of AI research, work on deep (multi-layered) neural networks continued through the 1980s. Other areas were explored in academia during the 1990s and early 2000s due to the large amount of computing power needed. Internet companies like Google gained experience in large clustered computing, AND, had a real need to explore deep learning. Use cases include image recognition to improve search and dynamically altered journeys to improve mapping and local search offerings. Deep learning is probabilistic in nature, which dovetailed nicely with prior work Microsoft Research had been doing since the 1980s on Bayesian approaches to problem-solving[xli].  

    A key factor in deep learning’s adoption was having access to powerful enough GPUs to handle the neural network compute[xlii]. This has allowed various vendors to build Large Language Models (LLMs). The perceived strategic importance of artificial intelligence has meant that considerations on intelligence per watt has become a tertiary consideration at best. Microsoft has shown interest in growing data centres with less thought has been given on the electrical infrastructure required[xliii].  

    Google’s conference paper on attention mechanisms[xliv] highlighted the development of the transformer model. As an architecture it got around problems in previous approaches, but is computationally intensive. Even before the paper was published, the Google transformer model had created fictional Wikipedia entries[xlv]. A year later OpenAI built on Google’s work with the generative pre-trained transformer model better known as GPT[xlvi]

    Since 2018 we’ve seen successive GPT-based models from Amazon, Anthropic, Google, Meta, Alibaba, Tencent, Manus and DeepSeek. All of these models were trained on vast amounts of information sources. One of the key limitations for building better models was access to training material, which is why Meta used pirated copies of e-books obtained using bit-torrent[xlvii]

    These models were so computationally intensive that the large-scale cloud service providers (CSPs) offering these generative AI services were looking at nuclear power access for their data centres[xlviii]

    The current direction of development in generative AI services is raw computing power, rather than having a more energy efficient focus of intelligence per watt. 

    Technology consultancy / analyst Omdia estimated how many GPUs were bought by hyperscalers in 2024[xlix].

    CompanyNumber of Nvidia GPUs boughtNumber of AMD GPUs boughtNumber of self-designed custom processing chips bought
    Amazon196,0001,300,000
    Alphabet (Google)169,0001,500,000
    ByteDance230,000
    Meta224,000173,0001,500,000
    Microsoft485,00096,000200,000
    Tencent230,000

    These numbers provide an indication of the massive deployment on GPT-specific computing power. Despite the massive amount of computing power available, services still weren’t able to cope[l] mirroring some of the service problems experienced by early web users[li] and the Twitter ‘whale FAIL’[lii] phenomenon of the mid-2000s. The race to bigger, more powerful models is likely to continue for the foreseeable future[liii]

    There is a second class of players typified by Chinese companies DeepSeek[liv] and Manus[lv] that look to optimise the use of older GPT models to squeeze the most utility out of them in a more efficient manner. Both of these services still rely on large cloud computing facilities to answer queries and perform tasks. 

    Agentic AI

    Thinking on software agents went back to work being done in computer science in the mid-1970s[lvi]. Apple articulated a view[lvii]of a future system dubbed the ‘Knowledge Navigator’[lviii] in 1987 which hinted at autonomous software agents. What we’d now think of as agentic AI was discussed as a concept at least as far back as 1995[lix], this was mirrored in research labs around the world and was captured in a 1997 survey of research on intelligent software agents was published[lx]. These agents went beyond the vision that PapriCom implemented. 

    A classic example of this was Wildfire Communications, Inc. who created a voice enabled virtual personal assistant in 1994[lxi].  Wildfire as a service was eventually shut down in 2005 due to an apparent decline in subscribers using the service[lxii]. In terms of capability, Wildfire could do tasks that are currently beyond Apple’s Siri. Wildfire did have limitations due to it being an off-device service that used a phone call rather than an internet connection, which limited its use to Orange mobile service subscribers using early digital cellular mobile networks. 

    Almost a quarter century later we’re now seeing devices that are looking to go beyond Wildfire with varying degrees of success. For instance, the Rabbit R1 could order an Uber ride or groceries from DoorDash[lxiii]. Google Duplex tries to call restaurants on your behalf to make reservations[lxiv] and Amazon claims that it can shop across other websites on your behalf[lxv]. At the more extreme end is Boeing’s MQ-28[lxvi] and the Loyal Wingman programme[lxvii]. The MQ-28 is an autonomous drone that would accompany US combat aircraft into battle, once it’s been directed to follow a course of action by its human colleague in another plane. 

    The MQ-28 will likely operate in an electronic environment that could be jammed. Even if it wasn’t jammed the length of time taken to beam AI instructions to the aircraft would negatively impact aircraft performance. So, it is likely to have a large amount of on-board computing power. As with any aircraft, the size of computing resources and their power is a trade-off with the amount of fuel or payload it will carry. So, efficiency in terms of intelligence per watt becomes important to develop the smallest, lightest autonomous pilot. 

    As well as a more hostile world, we also exist in a more vulnerable time in terms of cyber security and privacy. It makes sense to have critical, more private AI tasks run on a local machine. At the moment models like DeepSeek can run natively on a top-of-the-range Mac workstation with enough memory[lxviii].  

    This is still a long way from the vision of completely local execution of ‘agentic AI’ on a mobile device because the intelligence per watt hasn’t scaled down to that level to useful given the vast amount of possible uses that would be asked of the Agentic AI model. 

    Maximising intelligence per watt

    There are three broad approaches to maximise the intelligence per watt of an AI model. 

    • Take advantage of the technium. The technium is an idea popularised by author Kevin Kelly[lxix]. Kelly argues that technology moves forward inexorably, each development building on the last. Current LLMs such as ChatGPT and Google Gemini take advantage of the ongoing technium in hardware development including high-speed computer memory and high-performance graphics processing units (GPU).  They have been building large data centres to run their models in. They build on past developments in distributed computing going all the way back to the 1962[lxx]
    • Optimise models to squeeze the most performance out of them. The approach taken by some of the Chinese models has been to optimise the technology just behind the leading-edge work done by the likes of Google, OpenAI and Anthropic. The optimisation may use both LLMs[lxxi] and quantum computing[lxxii] – I don’t know about the veracity of either claim. 
    • Specialised models. Developing models by use case can reduce the size of the model and improve the applied intelligence per watt. Classic examples of this would be fuzzy logic used for the past four decades in consumer electronics to Mistral AI[lxxiii] and Anduril’s Copperhead underwater drone family[lxxiv].  

    Even if an AI model can do something, should the model be asked to do so?

    AI use case appropriateness

    We have a clear direction of travel over the decades to more powerful, portable computing devices –which could function as an extension of their user once intelligence per watt allows it to be run locally. 

    Having an AI run on a cloud service makes sense where you are on a robust internet connection, such as using the wi-fi network at home. This makes sense for general everyday task with no information risk, for instance helping you complete a newspaper crossword if there is an answer you are stuck on and the intellectual struggle has gone nowhere. 

    A private cloud AI service would make sense when working, accessing or processing data held on the service. Examples of this would be Google’s Vertex AI offering[lxxv]

    On-device AI models make sense in working with one’s personal private details such as family photographs, health information or accessing apps within your device. Apps like Strava which share data, have been shown to have privacy[lxxvi] and security[lxxvii] implications. ***I am using Strava as an example because it is popular and widely-known, not because it is a bad app per se.***

    While businesses have the capability and resources to have a multi-layered security infrastructure to protect their data most[lxxviii]of[lxxix] the[lxxx] time[lxxxi], individuals don’t have the same security. As I write this there are privacy concerns[lxxxii] expressed about Waymo’s autonomous taxis. However, their mobile device is rarely out of physical reach and for many their laptop or tablet is similarly close. All of these devices tend to be used in concert with each other. So, for consumers having an on-device AI model makes the most sense. All of which results in a problem, how do technologists squeeze down their most complex models inside a laptop, tablet or smartphone? 


    [i] Radiomuseum – Loewe (Opta), Germany. Multi-system internal coupling 3NF

    [ii] (1961) Solid Circuit(tm) Semiconductor Network Computer, 6.3 Cubic inches in Size, is Demonstrated in Operation by U.S. Air Force and Texas Instruments (United States) Texas Instruments news release

    [iii] (2000) The Chip that Jack Built Changed the World (United States) Texas Instruments website

    [iv] Moravec H (1988), Mind Children (United States) Harvard University Press

    [v] (2010) Makimoto’s Wave | EDN (United States) AspenCore Inc.

    [vi] Jobs, S. (1980) Presentation on Apple Computer history and vision (United States) Computer History Museum via Regis McKenna

    [vii] Sinofsky, S. (2019) ‘Bicycle for the Mind’ (United States) Learning By Shipping

    [viii] Hertzfeld, A. (1981) Bicycle (United States) Folklore.org

    [ix] Jones, D. (2016) Codemasters (United Kingdom) Retro Gamer – Future Publishing

    [x] Engelbert, D. (1968) A Research Center For Augmenting Human Intellect (United States) Stanford Research Institute (SRI)

    [xi] Hormby, T. (2006) Apple’s Worst business Decisions (United States) OSnews

    [xii] Honam, M. (2009) From Brick to Slick: A History of Mobile Phones (United States) Wired

    [xiii] Ericsson History: The Nordics take charge (Sweden) LM Ericsson.

    [xiv] Singh, H., Gupta, M.M., Meitzler, T., Hou, Z., Garg, K., Solo, A.M.G & Zadeh, L.A. (2013) Real-Life Applications of Fuzzy Logic – Advances in Fuzzy Systems (Egypt) Hindawi Publishing Corporation

    [xv] Reid, T.R. (1990) The Future of Electronics Looks ‘Fuzzy’. (United States) Washington Post

    [xvi] Kushairi, A. (1993). “Omron showcases latest in fuzzy logic”. (Malaysia) New Straits Times

    [xvii] Watson, A. (2021) The Antique Microwave Oven that’s Better than Yours (United States) Technology Connections

    [xviii] Durbhakula, S. (2022) IBM dumping Watson Health is an opportunity to reevaluate artificial intelligence (United States) MedCity News

    [xix] (1998) PapriCom Technologies Wins CommerceNet Award (Israel) Globes

    [xx] Von Ahn, L., Dabbish, L. (2004) Labeling Images with a Computer Game (United States) School of Computing, Carnegie-Mellon University

    [xxi] Butterfield, D., Fake, C., Henderson-Begg, C., Mourachov, S., (2006) Interestingness ranking of media objects (United States) US Patent Office

    [xxii] Delaney, K.J., (2005) Yahoo acquires Flickr creator (United States) Wall Street Journal

    [xxiii] Hood, S., (2008) Delicious is 5 (United States) Delicious blog

    [xxiv] (2017) 10 years of Carbon Neutrality (United States) Google

    [xxv] Bakshi, V. (2018) EUV Lithography (United States) SPIE Press

    [xxvi] Wade, W. (2000) ASML acquires SVG, becomes largest litho supplier (United States) EE Times

    [xxvii] Lammers, D. (1999) U.S. gives ok to ASML on EUV effort (United States) EE Times

    [xxviii] Meade, C., Conway, L. (1979) Introduction to VLSI Systems (United States) Addison-Wesley

    [xxix] Lavagno, L., Martin, G., Scheffer, L., et al (2006) Electronic Design Automation for Integrated Circuits Handbook (United States) Taylor & Francis

    [xxx] (2010) Apple Launches iPad (United States) Apple Inc. website

    [xxxi] (1997) PalmPilot Professional (United Kingdom) Centre for Computing History

    [xxxii] Jobs, S. (2005) Apple WWDC 2005 keynote speech (United States) Apple Inc.

    [xxxiii] (2014) Makimoto’s Wave Revisited for Multicore SoC Design (United States) EE Times

    [xxxiv] Makimoto, T. (2014) Implications of Makimoto’s Wave (United States) IEEE Computer Society

    [xxxv] (2006) Nokia and Yahoo! add Flickr support in Nokia Nseries Multimedia Computers (Germany) Cision PR Newswire

    [xxxvi] Gibson, W. (2007) Spook Country (United States) Putnam Publishing Group

    [xxxvii] The O2O Business In China (China) GAB China

    [xxxviii] Carroll, G. (2008) Web Centric Business Model (United States) Waggener Edstrom Worldwide for LaSalle School of Business, Universitat Ramon Llull, Barcelona

    [xxxix] Carroll, G. (2008) Web of no web (United Kingdom) renaissance chambara

    [xl] Kelly, K. (2018) AR Will Spark the Next Big Tech Platform – Call It Mirrorworld (United States) Wired

    [xli] Heckerman, D. (1988) An Empirical Comparison of Three Inference Methods (United States) Microsoft Research

    [xlii] Sze, V., Chen, Y.H., Yang, T.J., Emer, J. (2017) Efficient Processing of Deep Neural Networks: A Tutorial and Survey (United States) Cornell University

    [xliii] Webber, M. E. (2024) Energy Blog: Is AI Too Power-Hungry for Our Own Good? (United States) American Society of Mechanical Engineers

    [xliv] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L., Polosukhin, I. (2017) Attention Is All You Need (United States) 31st Conference on Neural Information Processing Systems (NIPS 2017)

    [xlv] Marche, S. (2024) Was Linguistic A.I. Created By Accident? (United States) The New Yorker.

    [xlvi] Radford, A. (2018) Improving language understanding with unsupervised learning (United States) OpenAI

    [xlvii] Heath, N. (2025) Authors outraged to discover Meta used their pirated work to train its AI systems (Australia) ABC (Australian Broadcast Corporation)

    [xlviii] Morey, M., O’Sullivan, J. (2024) In-brief analysis: Data center owners turn to nuclear as potential energy source (United States) Today in Energy published by U.S. Energy Information Administration

    [xlix] Bradshaw, T., Morris, S. (2024) Microsoft acquires twice as many Nvidia AI chips as tech rivals (United Kingdom) Financial Times

    [l] Smith, C. (2025) ChatGPT’s viral image-generation upgrade is ruining the chatbot for everyone (United States) BGR (Boy Genius Report)

    [li] Wayner, P. (1997) Human Error Cripples the Internet (United States) The New York Times

    [lii] Honan, M. (2013) Killing the Fail Whale with Twitter’s Christopher Fry (United States) Wired

    [liii] Mazarr, M. (2025) The Coming Strategic Revolution of Artificial Intelligence (United States) MIT (Massachusetts Institute of Technology)

    [liv] Knight, W. (2025) DeepSeek’s New AI Model Sparks Shock, Awe, and Questions from US Competitors (United States) Wired

    [lv] Sharwood, S. (2025) Manus mania is here: Chinese ‘general agent’ is this week’s ‘future of AI’ and OpenAI-killer (United Kingdom) The Register

    [lvi] Hewitt, C., Bishop, P., Steiger, R. (1973). A Universal Modular Actor Formalism for Artificial Intelligence. (United States) IJCAI (International Joint Conference on Artificial Intelligence).

    [lvii] Sculley, J. (1987) Keynote Address On The Knowledge Navigator at Educom (United States) Apple Computer Inc.

    [lviii] (1987) Apple’s Future Computer: The Knowledge Navigator (United States) Apple Computer Inc.

    [lix] Kelly, K. (1995) Out of Control: The New Biology of Machines (United States) Fourth Estate

    [lx] Nwana, H.S., Azarmi, N. (1997) Software Agents and Soft Computing: Towards Enhancing Machine Intelligence Concepts and Applications (Germany) Springer

    [lxi] Rifkin, G. (1994) Interface; A Phone That Plays Secretary for Travelers (United States) The New York Times

    [lxii] Richardson, T. (2005) Orange kills Wildfire – finally (United Kingdom) The Register

    [lxiii] Spoonauer, M. (2024) The Truth about the Rabbit R1 – your questions answered about the AI gadget (United States) Tom’s Guide

    [lxiv] Garun, N. (2019) One year later, restaurants are still confused by Google Duplex (United States) The Verge

    [lxv] Roth, E. (2025) Amazon can now buy products from other websites for you (United States) The Verge

    [lxvi] MQ-28 microsite (United States) Boeing Inc.

    [lxvii] Warwick, G. (2019) Boeing Unveils ‘Loyal Wingman’ UAV Developed In Australia (United Kingdom) Aviation Week Network – part of Informa Markets

    [lxviii] Udinmwen, E. (2025) Apple Mac Studio M3 Ultra workstation can run Deepseek R1 671B AI model entirely in memory using less than 200W, reviewer finds (United Kingdom) TechRadar

    [lxix] Kelly, K. (2010) What Technology Wants (United States) Viking Books

    [lxx] Andrews, G.R. (2000) Foundations of Multithreaded, Parallel, and Distributed Programming (United States) Addison-Wesley

    [lxxi] Criddle, C., Olcott, E. (2025) OpenAI says it has evidence China’s DeepSeek used its model to train competitor (United Kingdom) Financial Times

    [lxxii] Russell, J. (2025) China Researchers Report Using Quantum Computer to Fine-Tune Billion Parameter AI Model (United States) HPC Wire

    [lxxiii] Mistral AI home page (France) Mistral AI

    [lxxiv] (2025) High-Speed Autonomous Underwater Effects. Copperhead (United States) Anduril Industries

    [lxxv] Vertex AI with Gemini 1.5 Pro and Gemini 1.5 Flash (United States) Google Cloud website

    [lxxvi] Untersinger, M. (2024) Strava, the exercise app filled with security holes (France) Le Monde

    [lxxvii] Nilsson-Julien, E. (2025) French submarine crew accidentally leak sensitive information through Strava app (France) Le Monde

    [lxxviii] Arsene, Liviu (2018) Hack of US Navy Contractor Nets China 614 Gigabytes of Classified Information (Romania) Bitdefender

    [lxxix] Wendling, M. (2024) What to know about string of US hacks blamed on China (United Kingdom) BBC News

    [lxxx] Kidwell, D. (2020) Cyber espionage for the Chinese government (United States) U.S. Air Force Office of Special Investigations

    [lxxxi] Gorman, S., Cole, A., Dreazen, Y. (2009) Computer Spies Breach Fighter-Jet Project (United States) The Wall Street Journal

    [lxxxii] Bellan, R. (2025) Waymo may use interior camera data to train generative AI models, but riders will be able to opt out (United States) TechCrunch

  • Car screens and synthesisers

    The current debate over car screens / car as computer design reminded me a lot of the journey that synthesisers have gone through.

    Charging screen

    I went down this train of thought on car screens thanks to a LinkedIn post by Nic Roope, reacting to an article published in Car Design News in praise of push buttons.

    There is a view in car circles that the reliance on screens to mediate so many of the functions of a car can be a bad thing. I can understand it. For enthusiasts driving a car is still a very analogue experience including the haptics of direct steering connectivity and a manual gearbox.

    I would be remiss if I didn’t share the opinion of Doug DeMuro who argued the case for screens in terms of two reasons:

    • Costs. Buttons cost more money and there would be the associated connectors. Modern vehicles offer such a range of controls, that doing them in buttons rather than soft buttons and car screens would be cost and space prohibitive.
    • Technological momentum. DeMuro essentially articulates a position similar to Kevin Kelly’s concept of the technium in his book What Technology Wants. Kelly uses a biological metaphor of progress as an organism or Gaia type metaphor that keeps growing and moving at its own pace. While Kelly has been accused to techno-mysticism, we do know that the development of key technologies like television or the light bulb were happening at the same time in different parts of the world in isolation from each other – there had become a time when they were inevitable.

    the greater, global, massively interconnected system of technology vibrating around us

    Kevin Kelly on the technetium in What Technology Wants

    Colin Chapman versus software engineers.

    DeMuro’s first point is based on the proposition that all this extra control in car screens is a good thing. Do we really need to have car interior mood lighting? And if we do, do we need to have colours that result in night blindness and make the car interior looks like a booth at a bottle service bar in Dubai?

    For some drivers, the answer will be no.

    Different car manufacturers have had different models that do very different things. One of the philosophies articulated most by car enthusiasts is that of Lotus cars founder Colin Chapman “simplify, and add lightness”.

    Chapman’s design ethos was very in-tune with the likes of mid-century thinkers like polymath Buckminster Fuller and those he influenced notably architect Sir Norman Foster.

    Chapman’s world view wasn’t perfect his vehicles were fragile and had quality issues, partly due to his daring use of new materials and techniques influenced by aerospace. It’s also a world away from the Tesla approach, where the vehicle can’t be started up without the screen even as a ‘limp mode’ function.

    Instead the Tesla pickup and car screens are infested with boondoggles including:

    • A video of a fireplace filled with burning logs
    • A game that allows you to break the windows of a virtual CyberTruck
    • Customisable horn sounds including celebrity voices
    • A pre-programmed light show

    Modern car economics.

    Car screens have advanced in tock-step with the move towards an electric car future. A technology transition at the best of times is difficult, but the car industry has other problems that will impact consumer views of vehicles.

    Consumer choice.

    In the 1970s cars cars seldom lasted over a decade, but due to improvements in corrosion treatment and car design that removed water traps the potential life of a car was extended. Given that classic cars are much less damaging to the environment. The average classic emits 563kg of CO2 per year, yet an average passenger car has a 6.8-tonne carbon footprint immediately after production. This means that a new car would need to be run for several years to achieve a similar climate ‘payback’ and older cars can be attractive for consumers, if they meet their needs reliably.

    Vehicle affordability.

    Over the time I have held a driving licence, the secondhand car market went from being the dumping ground for fleet sales to the Alice In Wonderland after effects of the lease agreements that drove new and nearly-new car sales. The financialisation of the car market isn’t without risk and has been considered a possible future risk in the way that consumer finance and home mortgages have been in the past.

    Yamaha DX7II-D

    So what do car touchscreens have to do with synthesisers?

    In order to answer that question, we need to go back in time. Massive steps forward in electronics had inspired research into different ways of creating sounds based on modulation techniques used in radio broadcast signals for decades. In the 1960s digital technology was also moving forward and provided a more stable base for FM synthesis. Stanford University scholars worked with Yamaha technologists to turn FM synthesis into a product.

    The first instrument that it appeared in was the New England Digital Synclavier, who had licensed the technology from Yamaha. The Synclavier, was a couple of racks full of computer storage, a processing unit, cooling and audio interfaces. This was all connected up to a monitor and a keyboard. Over time the Synclavier would evolve into the ancestor of the modern digital audio workstation (DAW) like Apple’s Logic Pro app.

    1983, comes around and Yamaha is finally ready to launch a mainstream product featuring FM synthesis. it also features MIDI, a standard that is still used to control musical instruments (and other studio equipment) remotely. Roland had released a couple of devices that supported the standard.

    But Yamaha’s DX7 proved to be the blockbuster product. At that time electronic music was a niche interest and instrument manufacturers would be very lucky to sell 50,000 units. Yamaha sold over 300,000 units in the first three years of sales over its 7 year life and 10,000s of more devices of the DX and TX families.

    Digital changes the interface

    Analogue synthesisers wer full of switches and dials. This Oberheim synthesiser above, isn’t that different from its analogue predecessors from five decades prior.

    The DX7 was a very different beast, it couldn’t have a dial or button for every parameter, rather like modern car screens with endless settings. So it had a few buttons which changed their function depending on what the synthesiser. A few earlier models had limited sales with a similarly spartan approach, but the DX7 mainstreamed the idea.

    A few things happened that might be instructive for how we now think about car screens:

    • Other synthesiser manufacturers like Roland and Korg copied Yamaha’s approach to interface design. Some of them tried using devices like jog wheels to provide additional intuitive control, in a similar way conceptually to BMW’s iDrive interface for its car screens.
    • Software companies looked to fill the gap to provide a better interface, which eventually begat modern software digital audio workstation applications like Logic Pro. We might see similar developments sold for cars, and this is likely the opportunity that the likes of Apple CarPlay sees. There is consumer demand to support it.
    • Despite the obvious benefit of soft button driven instruments, there still remained a strong demand for analogue controls. Now there is a strong demand for tactile interface controls and old style synthesis. In the car world that would equate to providing car enthusiasts with analogue experiences, while the mainstream goes to Tesla minimalism of the car screen. We can see this in the design of Hyundai’s analogue feeling performance electric cars that try and emulate a manual gear box and Ineos’ switch gear that owes more to aviation than automotive manufacturing.

    You can find similar posts to this here.

    More information

    Average Age of Cars in Great Britain | NimbleFins

    In praise of pushbuttons (and other physical controls) | Car Design News

    Car pollution facts: from production to disposal, what impact do our cars have on the planet? | Auto Express

    MIDI Quest Pro Yamaha DX7 software editor

    Patchbase Yamaha DX7 software editor

  • General Magic

    General Magic has a reputation of being the technology equivalent of the Jordan-era Chicago Bulls, but it ended up going nowhere. I never got to see the device in person, it was only available in Japan and the US. It’s as famous much for its alumni, as it is for its commercial failure.

    Apple "Paradigm" project/General Magic/Sony "Magic Link" PDA

    This is captured in a documentary of the same name. For students of Silicon Valley history and Apple fan boys – the team at General Magic sounds like a who’s who of the great and the good in software development and engineering.

    General Magic started within Apple with a brief that sounds eerily like what I would have expected for the iPhone decades later.

    “A tiny computer, a phone, a very personal object . . . It must be beautiful. It must offer the kind of personal satisfaction that a fine piece of jewelry brings. It will have a perceived value even when it’s not being used… Once you use it you won’t be able to live without it.”

    Sullivan M. (July 26, 2018) “General Magic” captures the legendary Apple offshoot that foresaw the mobile revolution. (United States) Fast Company magazine

    The opening sequence tells you what the documentary is going to lay out. Over carefully curate images of Silicon Valley campuses, Segway riders and the cute bug like Google autonomous vehicle a voice talks about success and failure. That failure is part of the process of development. That General Magic has a legendary status due to its status as precursor to our always-on modern world and while the company failed, the ideas didn’t.

    Autonomous cars aren't nearly as clever as you think, says Toyota exec - Computerworld

    The genesis of the spirit of General Magic goes back to the development and launch of the Macintosh with its vision of making computers accessible. The team looked around the next thing that would have a similar vision and impact of a product. The Mac had got some of these developers on the front cover of Rolling Stone – they were literally rockstars.

    You get a tale of dedication and excitement that revolved around a pied piper type project lead Marc Porat, who managed to come to the table with a pretty complete vision and concept of where General Magic (and the world) would be heading. The archive of footage of the offices with its cool early to mid 1990s Apple Office products still amazes now. The look of the people in the archive footage, make my Yahoo! colleagues a decade later seem corporate and uptight by comparison.

    Veteran journalist Kara Swisher said that she started following the company because it was ‘the start of mobile computing, this is where it leads’.

    What sets the documentary apart is that it tapped into footage shot by film maker David Hoffman who was hired to capture the product development process. The protagonists then provide a voice over of their younger selves. Their idealism reaches back to the spirit of the 1960s. You can see how touch screen screens and the skeuomorphic metaphors were created and even animate emoticons.

    I’ve never known a development process with so much documentary footage. Having been in this process on the inside, the General Magic documentary portrays a process and dynamics that haven’t changed that much.

    The ecosystem that the startup assembled including AT&T, Apple, Motorola and Sony made sense given the ecosystem and power that Microsoft had behind it. It’s hard to explain how dominant and aggressive Microsoft was in the technology space. Newton came out as a complete betrayal and John Sculley, who is interviewed in the documentary comes across worse than he would have liked.

    The documentary also has access to the 1994 promotional film where General Magic publicly discussed the concept of ‘The Cloud’ i.e. the modern web infrastructure – but the documentary doesn’t dwell on this provable claim.

    Goldman Sachs was a key enabler, the idea of the concept IPO set the precedent for Netscape, Uber, WeWork and the 2020s SPAC fever.

    In a time when there is barely one thing changing the technology environment, General Magic were pursuing their walled garden of their private cloud and missed the web for a while. Part of this is down to their relationship with AT&T.

    The documentary covers how project management dogged the project. Part of the problem was perfectionism was winning over the art of the possible and not focusing on the critical items that needed to be done. The panic of having to ship.

    It’s about getting the balance between ‘move fast and break things’ versus crafting a jewel of a product.

    But shipping wasn’t enough, the execution of shopper marketing and sales training was a disaster. The defeat was hard given the grand vision. But the ultimate lesson is that YOU are not representative of the mainstream market.

    The documentary post-mortem featuring thinkers like Kara Swisher and Paul Saffo points out the lack of supporting infrastructure, that would take years to catch up to where General Magic’s Magic Link had gone. Paul Saffo uses a surfing analogy that I had previously read in Bob Cringely’s Accidental Empires about catching the right wave at the right time.

    John Sculley over at Apple made similar mistakes to the General Magic team which resulted in him being fired from Apple. Sculley makes the very human admission that being fired from Apple took him about 15 years to recover from personally.

    IBM Simon

    The documentary gives a lot of the credit (maybe too much of it) to General Magic as the progenitor of what we now think of as smartphones. The reality as with other inventions is that innovation has its time and several possible ‘inventors’; or what author Kevin Kelly would call ‘the technium’. This is the idea that technological progression is inevitable and that it stands on the layers of what has gone before, like fossils found inside rocks several foot deep. For instance, IBM created a device called Simon which was ‘smartphone’ which sold about 50,000 units to BellSouth customers in the six months it was on the market. Motorola – who were a General Magic partner also launched a smartphone version of the Apple Newton called the Motorola Marco in January 1995 and there are more devices around the same time.

    Reality is messy and certainly not like the clean direct line that the General Magic documentary portrays, even the Newton was only part of the story.

    The Wonder Years

    I was thinking about what I liked so much about the General Magic documentary. I immediately thought about it reminding me of my falling in love with the nascent internet and technology, which then bought me to the start of my agency career working with Palm (the company that eventually helped kill off General Magic’s product ambitions) and the Franklin REX which came out of sychronisation pioneers Starfish Software.

    But it was deeper than that. The Silicon Valley portrayed in the General Magic documentary wasn’t the dystopian hellscape of platform firms, generation rent, toxic tech bro culture and ‘churn and burn’ HR culture. Instead the General Magic documentary story represented a halcyon past of Silicon Valley portrayed in books like Where Wizards Stay Up Late, Fire In The Valley and Insanely Great. Where talented people motivated by a fantastic vision thing, with a user centred mission worked miracles. The darkness of fatigue and god knows what else is largely hidden by a Wonder Years TV show feel good nostalgia. Maybe it gives us hope again in the tech sector, despite Peter Thiel, Mark Zuckerberg, Tim Cook and Elon Musk? Maybe that hope might inspire something great again?

    Marc Porat’s personal tragedy and Tony Fadell’s business failure brings a hint of the real world through the door. The documentary uses Fadell’s link with the iPod and iPhone as a point of redemption, resilience, perseverance and vindication for General Magic.

    There’s also a cautionary tale full of lessons learned for new entrepreneurs, who often get the vision thing but forget about the details. More on General Magic here.

    More reviews here.