For the Next Month: Going All-in on The Browser Company's Dia-AI

& we will see where we get, and whether we need… ROADS! From searchbox to sidekick: is the humble web browser the secret to mastering the potential of MAMLMs as natural-language interfaces and as superbig-data, superhigh-dimension, superflexible-function classifiers, rather than being mastered by it? Perhaps I can tune Dia Browser so that it truly remaps the web for my cognitive benefit…

Share

Share Brad DeLong's Grasping Reality

Share Brad DeLong's Grasping Reality

I have decided to embark on a month-long experiment going all-in using Dia Browser <https://www.diabrowser.com/>.

I have decided to explore if I can actually use GPT LLM MAMLM ChatBots—General-Purpose Transformer Large-Language Model Modern Advanced Machine-Learning Model ChatBots—to manage my current very bad case of information overload.

I also hope it will help me critically assessing the real limits and potential of MAMLMs.

Specifically, I hope to gain some insight into how I might be wrong in my current judgment that they can only be of substantial yet strangely limited usefulness—that their simulacrum of intelligence is in fact a place where Sokrates’s observation that no matter how much a statue looks like the person it cannot speak is in fact correct. Think of them as, say, doing for prose roughly what spreadsheets did for accounting—important, mind-blowing, but not Chicxulub.

Give a gift subscription


Back up:

Modern Advanced Machine-Learning Models are, primarily:

  • Very big-data, very high-dimension, very flexible-function classification analyses.

  • Natural-language interfaces to structured and unstructured databases.

And then the rest of the technological and societal knock-on from the coming of General-Purpose Transformer Large-Language Model Modern Advanced Machine-Learning Models—GPT LLM MAMLMs:

  • Platform oligopolists spending money and resources at a scale relative to the then-size of the economy that has never before seen in any General-Purpose Technology build-out…

  • Why? Not to make or to try to make more money, but to buy insurance to guard against Christensenian disruption of their current platform-oligopoly profit flows…

  • Not FILTH—Failed in London? Try Hong Kong!—but FICTA—Failed in Crypto? Try AI!: The post-crypto boom way of separating gullible VC from their money for the benefit of engineers, speculators, utopians, and grifters…

  • Auto-complete for everything on super-steroids…

  • Highly verbal electronic-software therapists, sounding boards, coaches, pets, and… THIS IS A SAFE-FOR-WORK WEBLOG!…

I have written about this a bunch. Most recently: for the Milken Institute Review here: <https://www.milkenreview.org/articles/behind-the-hype?IssueID=58>

But I had more to say that would not fit into that piece! And that got left ln the cutting-room floor!

Specifically, about GPT LLM MAMLM ChatBots:

GPT LLM MAMLM ChatBots merge the first two, the primary aspects of MAMLMs, to create a new summarization cultural technology for the internet, plus whatever else is in their training data.

Leave a comment


I think of GPT LLM MAMLM ChatBots doing for prose roughly what spreadsheets did for accounting—important and mind-blowing, and a destroyer of many not-that-interesting jobs. But not world-shaking. That is what I think. And I think I have good reasons for thinking so.

But I might be wrong.

I think that because the core of a GPT LLM ChatBot, after all, is a huge neural network that is trained to emulate the language-response patterns found in its training data. In practice, that means that its telos is to become a closer and closer approximation of the typical Internet s***poster.

Yes, in addition to neural-network initial training, there is post-training RLHG, RAG, DPO, PEFT, LoRA, PPO, SFT, RM, IFT, and I am sure there will be at least five more such acronyms in common use in a year. But these are all merely (merely!) rough-and-ready ways to try to get the typical internet s***poster that is the ideal GPT LLM ChatBot to behave: to nudge it into a state in which the training data it sees around its attentive core is dense, and in which that training data was constructed by smart, informed people trying to communicate true and important stuff. Actively trying to strip the thing down so that it does not wander off into internet woo-woo land—that is an active area of “research”.

Witness Andrej Karpathy here:

Andrej Karpathy: <https://x.com/karpathy/status/1938626382248149433/>: ‘The race for LLM "cognitive core" - a few billion param model that maximally sacrifices encyclopedic knowledge for capability. It lives always-on and by default on every computer as the kernel of LLM personal computing. Its features are slowly crystalizing…. atively multimodal text/vision/audio at both input and output. - Matryoshka-style architecture…. Reasoning, also with a dial.… On-device finetuning LoRA slots for test-time training, personalization and customization…. It doesn't know that William the Conqueror's reign ended in September 9 1087, but it vaguely recognizes the name and can look up the date. It can't recite the SHA-256 of empty string as e3b0c442..., but it can calculate it quickly…. Super low interaction latency… private access to data and state, offline continuity, sovereignty ("not your weights not your brain")…. Do people *feel* how much work there is still to do. Like wow.

Get 50% off a group subscription

And here:

Andrej Karpathy: <https://x.com/karpathy/status/1937902205765607626>: ‘+1 for "context engineering" over "prompt engineering"…. In every industrial-strength LLM app… the delicate art and science of filling the context window with just the right information…. Science because doing this right involves task descriptions and explanations, few shot examples, RAG, related (possibly multimodal) data, tools, state and history, compacting.… Art because of the guiding intuition around LLM psychology of people spirits…. Vreak up problems just right into control flows - pack the context windows just right - dispatch calls to LLMs of the right kind and capability - handle generation-verification UIUX flows - a lot more…. The term "ChatGPT wrapper" is tired and really, really wrong…

Refer a friend

Thus I tend to dissent from things like Torsten Slok’s claim that “AI” is going to deliver faster measured real GDP growth here in the U.S. over the next five years:

Share Brad DeLong's Grasping Reality

And I have explained why I see the GPT LLM category of MAMLMs as limited to roughly spreadsheet—far below microprocessor—levels of impact:

Give a gift subscription

I see them as such because, at bottom, I see them as what I was taught in third-grade math to call “function machines”: essentially “autocomplete on steroids”. (It is still the case that the best comprehensible introduction to this (for me at least) that I have been pointed to is still Alessandrini & al. (2023).) But the results are very, very impressive indeed across a number of dimensions. And every time I reach this point, I am again unable to avoid quoting Shalizi (2023, 2025):

Cosma Shalizi: ‘Attention’, ‘Transformers’, in Neural Network ‘Large Language Models’ <http://bactra.org/notebooks/nn-attention-and-transformers.html>: ‘"It's Just Kernel Smoothing" vs. "You Can Do That with Just Kernel Smoothing!?!": [That] takes nothing away from the incredibly impressive engineering accomplishment of making the blessed thing work…. nobody [before] achieved anything like the[se] feats…. [We] put effort into understanding… precisely because the results are impressive!…

The neural network architecture here is doing some sort of complicated implicit smoothing across contexts… [that] has evolved (under the selection pressures of benchmark data sets and beating the previous state-of-the-art) to work well for text as currently found online.… Markov models for language are really old…. Nobody, so far as I know, has achieved results anywhere close to what contemporary LLMs can do. This is impressive...

Share

Why are large language models (LLMs) so astonishingly effective? First, because of scale. Trained on vast datasets and operating in thousands of dimensions, they capture the statistical patterns of human language. When you request a summary or script, the output is not true understanding, but a “cultural average”—how people have written such things before.

The high dimensionality of these models’ internal representations is important too. Doing math like “king + woman - man ≈ queen” is not just a clever parlor trick, but evidence of deeper structural regularities and even meaning encoded in the vectors and matrices. There is an emergent flexibility and richness come from the neural network’s ability to classify and interpolate across a vast, multidimensional space.

What does this mean in practice? LLMs excel at compressing, interpolating, and representing the latent structures of human language and knowledge. They are, as the phrase goes, “autocomplete on steroids.” Their outputs are plausible, coherent, and often insightful, even though they lack genuine understanding or grounding in reality.

In my view, LLMs’ “unreasonable effectiveness” is likely to have arisen from five main factors. First, their scale—massive data and computation—enables them to learn what hand-crafted rules cannot. Second, their statistical machinery allows them to interpolate between immense numbers of examples, producing plausible continuations for prompts that bear only a family resemblance to things they have seen. Third, their training uncovers latent structures that make language compressible and predictable, echoing Plato’s dream of “carving nature at its joints.” Fourth, their high-dimensional embeddings allow for analogical reasoning that mimics superficially much the way humans think. Fifth—and perhaps most important—a huge amount of real human thought is preserved in language pattern-correlations.

Still: A great deal of time, energy, and effort has been spent training my 50W-drawing wetware that fits inside the small breadbox-sized bone enclosure at the top of my spine. That had better be better at true cognition—at searching out, absorbing, processing, and then signaling and conveying information. It is, after all, a full Turing-Class entity. If a blank-slate neural network orders-of-magnitude less complex trained to emulate a typical internet s***poster can steam-hammer-to-John-Henry me, that time, energy and effort has been largely wasted. That seems very unlikely. No. If you are actually going to build a Turing-Class sEntity, attention, scale, and a blank-slate neural network to train are highy unlikely to be all that you need.

Thus think of them as the next step in the long tradition of tools that let us offload rote, repetitive, or formulaic parts of knowledge work, freeing up our attention for higher-order thinking. It is the next step in a process that started back when the first task of a scribe in writing a document was mixing the clay for the cuneiform to the proper consistency and smoothness, and the sudden ability to import papyrus to Sumer from Tawy was a true godsend.

Share Brad DeLong's Grasping Reality


And yet, even if GPT LLM ChatBots are merely neural-network Markov-chain without AGI emergence Sub-Turing mimics of human minds, they are powerful cultural technologies for search and summarization above all, and also for link-following, code-snippet generating, ritual, milestone-marking, translation, genre-riffing, outlining, brainstorming, and more.

These models are not reasoning agents. They cannot verify facts and are prone to hallucinations—confidently generating plausible-sounding nonsense. But they do have strengths. They are calculators for language, not philosophers. They are tools for offloading cognitive drudgery, not engines of new insight. This is a real and a big productivity gain, especially for those who already know what “good” looks like and simply need a first draft or a nudge in the right direction. They are not magical.

But, still: They are remarkable. And useful.

GPT LLMs should be considered remixers and redistributors of existing culture, offering highly powerful affordances for access, recombination, and dissemination. There are dangers: the sheer persuasive power of a natural-language interface to humans tuned to human-sounding language primed to over-attribute intentionality. But if we guard against those, all of a sudden we now have a better way of funneling the firehose of information into a form in which we can make sense of it. We can distill the vast, chaotic sprawl of the territory human textual production into a manageable, if inevitably lossy, map. And while the map is not the territory, some maps can be very good and useful indeed.

Thus these things are immensely useful for search and research. I do not know to what degree to blame Google’s pursuit of advertising profits at all cost and to what degree to blame SEO search-arbitrage attention jackals for the degeneration of Google search, but even in its present something like ChatGPT is a better search engine than Google search as long as you tell it to act like a high-quality search engine. Witness Joe Weisenthal:

Joe Weisenthal: ChatGPT Is Becoming My Default Search Engine <https://www.bloomberg.com/news/newsletters/2025-07-17/chatgpt-is-becoming-my-default-search-engine>: ‘The same type of query on Google isn’t nearly as useful…. The actual links that Google turns up (the old, core search business) is not nearly as helpful…. To be fair… Google is better… if I just want to find a Wikipedia page… or if I’m looking for a very specific phrase…. But for a range of queries, o3 from ChatGPT increasingly feels like a strictly better experience…

Leave a comment

The act of using a GPT LLM to query the internet should not be thought of as an act of delegating thought, but rather of curating and steering a process of retrieval from the memory of the Anthology Super-Intelligence that is Humanity’s Collective Mind. collective memory retrieval; the user’s judgment remains indispensable. GPT LLMs promise to democratize access to cultural and intellectual resources, enabling non-experts to interact with and synthesize information in ways previously reserved for the highly literate or technically skilled. They also promise to greatly amplify the cognitive information-surveying and information-acquisition capabilities of those skilled in their use. They are, I think, going to provide us with much better whips and chairs for trying to tame informational overload than we had even five short years ago.

Thus whether or not I am wrong about the limits of these ChatBots, natural-language interfaces to structured and unstructured datastores promise to be very useful indeed to somebody like me who is both a huge user of datastores and a very sophisticated user of natural language. So how can I quickly and easily gain the maximum amount of facility with these tools so I can use them for good, and not either let the potential benefit I might draw from them wither on the vine, or, worse, find somebody else using them to hack my brain for my detriment?

These issues are now even tremendously important because we do now have true information overload. This problem has been building for quite a while. If one is a white-collar worker—if one is a front-end node to and in the Anthology Super-Intelligence of Humanity’s Collective Mind—one’s tasks, whether it is that of trying to enrich or emprestige your employers so they pay you, or enrich your life—fall into three groups:

  1. The business of thinking through an issue once you have gathered your sources of information.

  2. The issue of then getting your conclusions disseminated in a form in which they are neither evanescent nor unpersuasive.

  3. The issue of finding and assembling the information that you need to actually do your thinking.

Think of these as the signal, the process, and the funnel. Someone who wants to succeed and be useful to society in my business needs to be able to do all three, and to be able to do all three very well. I want very much to be able to do a better job in making sure that I read more of what I should be reading and from people who are not in the circle of usual suspects

The problem is the funnel. Dealing with the funnel is becoming an increasingly large part of our jobs. And here is where the hope that our modern information technologies understood as cultural technologies of summarization and classification may of great help.

I do not understand how these things work (but does anybody?). I do not understand how to use them well. There is too much to read. There is too much to experiment with. And yet, as Cosma Shalizi says, we “clearly… need to wrap… [our] head[s] around it, before… [we soon] become technically obsolete…”

Give a gift subscription


And so I am going to try to spend the next month going all-in on The Browser Company’s Dia Browser <https://www.diabrowser.com/>. What is Dia Browser? One of legion coming down the pike:

M. G. Seigler: Begun, the AI Browser Wars Have <https://spyglass.org/ai-browser-wars/?ref=the-spyglass-column-newsletter: ‘Dia ushers in a new day for the web browser.… You could basically just use Dia as a slightly cleaner-looking version of Chrome… and never touch the AI elements…. [But] if Dia is going to succeed, it will be because of AI. As such, they're now in a foot race…. Google… every other browser… dabbling in some level of AI… Perplexity and… OpenAI…. Anyone who uses Dia should see why all of these players will build their own browsers. Extensions aren't going to be good enough here, you have to go deeper to fully control the experience and have full vision into what's going on in that browser…

Dia Browser is, with a “@” denoting a link to another browser tab:

Chat with your tabs… Hey Dia…
Write with your tabs… Make this @JuneNewsletter sound on-brand @BrandVoice…
Learn with your tabs… Explain this to me like I’m five @Singularities…
Plan with your tabs… Compare @TuscanVilla @IlCasone in a table with price, proximity, to best coffee…
Shop with your tabs… Convince me not to buy this @VintageFendiBaguette… <https://www.diabrowser.com/>

The idea is that with every webpage you visit you can then immediately, seemlessly ask a question related to it, also bringing in whatever else you think is potentially relevant to the answer you get. The hope is to make the interaction-feedback loop as tight as possible, and then see what develops.

Dia is The Browser Company’s second project. The first was their power-user browser Arc: <https://arc.net/> “The Chrome Replacement You’ve Been Waiting For!” The Browser Company’s ambition with Arc Browser was never just technical—it was anthropological. Their goal was to make the browser feel like a home on the internet, personal and customizable, letting users shape the technology to fit their lives to extend and adapt their browser into a digital home.

Dia Browser aims at this with a twist: empowering users—even those with no coding experience—to build their own tools and workflows by what is now called “vibe coding”. The bet is that one of the true promises of MAMLMs is in providing natural-language in-the-moment tool-building for simple bespoke software to help us navigate the messy, context-rich decisions of real life—like planning a trip, balancing preferences, or making sense of ambiguity.

Traditionally, browsers have served as “funnels for intent”—the search box where users express what they want—and as platforms for running webapps. But why should the browser omnibox link only to legacy search engines? Why not route intent to other digital experiences, webapps, including those powered by AI? Perhaps most value now lies not in the browser itself, but in what can be built atop it. The goal is a personal assistant—an “intelligence layer”—that sits across all devices, learns from user context, and helps manage tasks, not just tabs. In this new landscape, context, memory, and personalization may be the true differentiators and wisdom amplifiers.

The Browser Company found that users of the alpha version quickly found their way to writing their own simple prompts for the ChatBot in the right pane of the browser window to act on the webpage in the left to automate or streamline their daily tasks: new tools, workflows, and automations, mini-agents and micro-applications that users could build for themselves, or frictionlessly adopt and then modify.

Refer a friend


Perhaps the browser is about to undergo a radical transformation in this age of MAMLMs. Perhaps a good way for it to take advantage of natural-language interface and stupendous classification capatilities is for it to stand in the middle, intercepting and interpreting my intent, turning the searchbox into a launchpad for personalized, context-aware assistance and assistants. Perhaps context, memory, and customization become key. The Browser Company thinks its Dia Browser could soon become my most valuable ally in dealing with information overload.

Is The Browser Company right?

I think I need to experiment enough to see.

Get 50% off a group subscription


References:

Leave a comment

Subscribe now


If reading this gets you Value Above Replacement, then become a free subscriber to this newsletter. And forward it! And if your VAR from this newsletter is in the three digits or more each year, please become a paid subscriber! I am trying to make you readers—and myself—smarter. Please tell me if I succeed, or how I fail…

#dia-why-i-am-going-to-try-to-go-all-in-on-dia-browser-this-summer