By chance I saw a question on Superuser today. Someone wanted to know if the IP addresses 192.168.1.50/20 and 192.168.2.200/20 could communicate directly, without a router. The answer is yes, as they are both in the same subnet, but before asking Superuser, they asked ChatGPT. ChatGPT very unhelpfully told them: "that 192.168.1.x and 192.168.2.x cant communicate with 20 subnet mask." When you think about it, of course that's the answer ChatGPT would return, the 192.168.0.0/16 IP space is typically used for what was formerly Class C private IP networks, now known as /24 subnets. But nothing really forces you to use a /24 in that IP space. It just is the most likely use of it, so that's what ChatGPT responded.Exactly what I was thinking. AI has the potential to destroy SEO as we know it and show us the information that we're actually looking for, and I can see that having a very negative impact on Google's profits, especially since sites pay Google for preferential treatment.
Bard may be a major step forward for customer usability, but Bard2 will be a shameless cash-whore.
I also don't see what's the problem to tack on ads on the side of the chat page, driven by the chat's content by traditional algorithms. I don't like it, but Google's board won't be asking my opinion.
It will be an interesting experiment for them to have the AI handle ad placement. Tell advertisers "pay for clicks/impressions, but no AdWords control, you let the AI decide where it thinks it should place your ads"… and see what happens.
I finally started using ChatGPT this week, and have been underwhelmed. My main beef is the severe hallucination problems, making it great for fiction and propaganda, and bad for most anything else (without very careful human review). Most importantly, OpenAI (and other LLM developers) say they're training new versions of the AI to reduce hallucinations considerably. I don't want them reduced ! I want a different AI model which by design, if it doesn't have something good to say, says nothing. (Instead of making things up at all costs.) It's a fundamental flaw in this type of model.
[I work at Microsoft, formerly Bing, now Azure Machine Learning -- speaking on my own behalf, I don't have special knowledge about anything here]AMEN. And that's the Real Issue. They can't give a first page of garbage!
They might have to actually do a Deep Search like it's 1999.
If they were to open source the model, we can prune it and it'd be able to run just fine, even if it's not fully featured.
This is why AI will be so disruptive to Google's business model.
We'll have models that are very capable, that can run locally from a 200MB file.
They'll probably solve this the same way they solved the YouTube problem, with bespoke hardware. Neurally-designed silicon is a thing and Google can afford to spin up a fab.
Not really good news though, as others have mentioned it will also lead to increased carbon emissionsReally pretty much the only good news about ChatGPT I’ve heard so far.
People with personal assistants are quite likely to be utterly amazed by ChatGPT and the like.That was just one example. Basically chatgpt could do anything I would ask a moderately competent personal assistant to do, for a fraction of the price and 100x the speed
People with personal assistants are quite likely to be utterly amazed by ChatGPT and the like.
ChatGPT-style search represents a 10x cost increase for Google, Microsoft
Most modern data centers either are carbon neutral or working towards that and pretty close to the finish line.Of course, more compute equals more power consumption which equals more carbon.
More hardware to meet the compute requirements also means more carbon.
I very much agree with your assessment.Text chats do have the advantage in that you can design them to display multiple profit-generating links at once in the result. Voice chat is limited by the medium in which it exists to present you with one result at a time.
Still, once the novelty wears off, are people really going to be that much more enthusiastic about chatting at a computer than speaking to one? I'm not sure. I think the headline grabbing tech is not going to be at the forefront of things, but relegated to the background. That is, improving things (e.g. customer service bots, code completion tools) that already exist and have narrower and more easily defined use cases. I will (again) warn that one of the best use cases for this tech as it exists today is propaganda, which requires content that is merely plausible sounding, a lower bar than all the other use cases that require results that are also true.
Having said that, the image bots seem to be very useful as a concept-phase tool as-is too. This is all going to be very complicated and will affect different industries at different rates.
Entries don't get "ranked and categorized" when you search. That happens BEFORE you search and it certainly doesn't take less than a second.Today Google search works by building a huge index of the web, and when you search for something, those index entries gets scanned and ranked and categorized, with the most relevant entries showing up in your search results. Google's results page actually tells you how long all of this takes when you search for something, and it's usually less than a second.
But they’d still have to embed non-sponsored links, at which point you’ve kinda just replicated … Google search results.Bing Search actually seems to do a pretty good job of embedding links to sources in its answers. If Google can do something similar with Bard, it could embed, e.g. sponsored links to its answer, i'm sure advertisers would gladly pay for that.
"trusting the results" is indeed the problem. CNET fired up some ChatGPT reporters apparently, and they sounded great. Well, the article read well I should say. The problem is, there were several statements (of fact) in the article and each of them was wrong.Another Reuters report says that Microsoft has already met with advertisers to detail its plan of "inserting [ads] into responses generated by the Bing chatbot," but it's unclear how awkward that would be or if consumers will react when a chatbot suddenly kicks over into an ad break.
The problem with chat-interface search is going to be trusting the results. That's two-fold, one side is that some people won't trust the results, a problem which will be magnified by including ads which indicate the answer is biased. The other side is that some people will trust the results even when they shouldn't, and presenting ads as facts could easily drift over towards fraud -- especially if the ads are driven through an automated system where no human is checking their appropriateness.
Entries don't get "ranked and categorized" when you search. That happens BEFORE you search and it certainly doesn't take less than a second.
Who will retrain that multiple times every day for each employee? May cost more than a few bucks to do this.Would companies be willing to pay a few bucks more per user per month to have access to an AI assistant that knows everything contained in your company's Sharepoint, your OneDrive, and your mailbox, and can provide you with information on demand or compose emails and letters for you with minimal input?
More likely it would be a standard model that can run searches in your environment and digest the results, same as the Bing AI does now. In fact if you are logged in to an Office 365 account, your OneDrive, SharePoint, Outlook and Teams are already included in your Bing search results.Who will retrain that multiple times every day for each employee? May cost more than a few bucks to do this.
Google, Bing, etc face an interesting problem here, it seems to me. Their language models require sources of knowledge to learn from. So, blogs, news sites, academic writing, artists, etc. Those all have to exist on separate pages. But if they are successful in essentially hoovering up all the knowledge from those sites, and then using their language models to re-write the ideas in a way that doesn't technically violate copyright, they might kill of the very sources of knowledge that their language models depend on to learn from.From my perspective, it would increase the time spent by users on each search engine as it becomes more interactive and you can get the answers without leaving Google or Bing.
So, yes more compute, but I would expect a higher user engagement and time spent on the platform, so potentially more revenue.
Weirdly, if Google wanted to return solid results in less time, all they would have to do is re-enable Boolean search and actually honor limiters (like the - they seem to now ignore, or the + that was removed with G+ launch) as an advanced option. For most general searches Google works fine, but the second you start to get technical or try to be very, VERY specific you’re immediately going to get irrelevant or content farm results, and nothing more.
Sounds good on the surface, but how practical is this going to be? The processing that is needed is gargantuan, is not "real time" right now, and does one company's data even remotely approach the SCALE that these LLMs NEED to start being useful? I'm not sure all or even any of those issues are trivial to solve.Microsoft already has another obvious path to profitability with this, by adding capabilities to Office 365. Would companies be willing to pay a few bucks more per user per month to have access to an AI assistant that knows everything contained in your company's Sharepoint, your OneDrive, and your mailbox, and can provide you with information on demand or compose emails and letters for you with minimal input? I'm guessing that would be an easy sell for most. Forget New Bing, New Clippy could be huge.
Learning to instinctively recognize the hype cycle is a good thing. It'll help you avoid the snake oil. I've been using self-driving cars as my go-to for the hype around all of this but 3d TV is a good analogy too.A couple of weeks ago I speculated on here that we might be looking at the "3D TV of the early-to-mid-20s".
It's starting to look more and more that that wild guess based on a gut feeling hunch might serve me pretty well in 10 years. At least there's hope!
Well, even if they are " green", they use up green energy that could have been used for potentially more urgent/useful needs. It's not like we have all the time in the world to decarbonize, it's a fucking race against time!!Most modern data centers either are carbon neutral or working towards that and pretty close to the finish line.
They say it's to be green, but I'm pretty sure they most just care about greenbacks. Zero emission energy is cheaper.
Online search is crappy. It prioritizes producing lots of hits over the right hits. It turns up pages that lack search terms, or even any relation to the search terms. There is still a lot that can be improved.Google needs to realise that search is a solved problem, and has been for 20 years. Not a single person on the planet who doesn't work for Google wants Google search to be any "better", except for removing the ads. It works perfectly. Just give me Wikipedia, StackOverflow, YouTube... maybe IMDB... etc. It works fine. Just forget about it. The fact that it makes so much money has misled them into thinking that continuing to "improve" it is the way forward.
And as you mention, online search has a problem in that it includes a lot of results that are automatically generated spam content (a problem which AI will make worse) rather than original content. I don't need to see a site that ran Wikipedia's article through some filters to make it look like their own content.Actually I would like them to blacklist cplusplus.com, because that's the bad one that's always wrong/out of date. cppreference.com is the correct one. Sometimes I'm not paying attention and end up at the wrong site. But that and the ads are literally my only problems with Google search.
While I'm not on board with your second sentence here, I agree that having a conversation is not my goal. If I do a search, I just want a quick and correct answer.I do NOT want to have to have a "conversation" with a computer. Ever. Avoiding conversations is probably a significant part of the reason why I'm using a computer in the first place.
You could consider something like backstage (https://backstage.io/). My company uses it for internal stuff and all the API docs are synced from github to backstage for searching.I'm not going to lie, being able to ask eg "what does the /v1/frob/buzz API response look like" and not have to trawl our sprawling confluence wiki/JIRA system (or the Google/Microsoft/whatever equivalents) would be incredibly useful. However, it does sound like it would be expensive to build and maintain such a specially trained model, so who knows when that would be a viable product
If you have any search engine skills, you can ask it follow up questions to continue giving more and more in-depth answers, and keep in mind we're on an old version without internet access.By chance I saw a question on Superuser today. Someone wanted to know if the IP addresses 192.168.1.50/20 and 192.168.2.200/20 could communicate directly, without a router. The answer is yes, as they are both in the same subnet, but before asking Superuser, they asked ChatGPT. ChatGPT very unhelpfully told them: "that 192.168.1.x and 192.168.2.x cant communicate with 20 subnet mask." When you think about it, of course that's the answer ChatGPT would return, the 192.168.0.0/16 IP space is typically used for what was formerly Class C private IP networks, now known as /24 subnets. But nothing really forces you to use a /24 in that IP space. It just is the most likely use of it, so that's what ChatGPT responded.
It's really not anywhere ready for prime time as a search engine. So far all we've managed to do is string together words that have the highest probability of following the words that came before them, in a manner that is a semi-convincing simulacrum of human language. But the highest probability answer isn't the correct answer, but the chatbot will return it authoritatively so people will treat it as so.
I just added an 18TB drive for $250. I don't think space is a concern. What's the cost in CPU/GPU cycles?Unless Bard is very different from ChatGPT, running it on a consumer grade machine, even a high-end gaming PC does not sound practical.
According to Wikipedia, GPT-3's parameters take up 800 gigs.
https://en.wikipedia.org/wiki/GPT-3