AMEN. And that's the Real Issue. They can't give a first page of garbage!
They might have to actually do a Deep Search like it's 1999.
[I work at Microsoft, formerly Bing, now Azure Machine Learning -- speaking on my own behalf, I don't have special knowledge about anything here]
As usual I'm late to the comments party but there's a variety of interesting factors at play here. The per-interaction serving cost of LLMs is, compared to the typical cost of serving results from an inverted index lookup (
grossly oversimplifying web search), astronomical. However, the state of the art in "internet scale generation + serving of a distributed information index" has had about two decades to mature. In contrast, AI model serving is relatively nascent, and internet-scale serving is effectively a toddler. Expect advances here as profit motives incentivize further R&D. Lots of people are working very hard to make this cost less.
However, if you use Bing Chat today, what it's doing is pretty obvious: generating, executing, and synthesizing results from one or more
traditional search queries. So what I find really interesting, as a now outside observer of this new scenario, is that all the extant costs of having that massive knowledge index and search engine infrastructure, are being piled
on top of the new costs of hosting an LLM to synthesize the data. If I'm a search engine, I'm not necessarily ecstatic about that. If I'm a search engine with 97% worldwide share, I'm certainly not excited about scaling that out immediately in response to a competitor's threat. Now I have to do all the work of being a search engine in addition to this new thing. However, I'm confident Google can figure it out. They got YouTube working for them when, at one time, the costs of owning and operating the platform seemed likely to dwarf any conceivable profit they could wring from it.
I think, also, it might be a little bit early to declare that "10x cost" is a reasonable factor. I suspect the real number will end up being lower, because the incentives are just so good to make that the case. Not to mention, it may force a pivot in fundamental assumptions about search engines, knowledge indexing, etc. As AI-generated content proliferates, it seems likely to me that our ability to sift out genuine, human-created content will need to improve. That genuine human content, particularly the quality stuff, is a tiny fraction of the garbage that ends up in search indices today. If the profit motives shift to fewer, higher-quality data sources, it seems plausible that all that godawful content farm spam that has made the traditional/current search experience so increasingly awful over the last decade, may dry up, in which case the cost of the existing service naturally drops.
There's also the real possibility that these additive AI assistant style data synthesizers will, in fact, end up too expensive or too impractical to give away in an ad-supported manner, and will end up being bundled with some sort of subscription service or pay-per-use model, further upending the ad-supported internet search-driven mechanism generating so much traffic today. I can't say, as a user, that I would be sad to see the current advertisement-based revenue model of most of the internet be severely disrupted in favor of a pay-for-use mechanism.