I'm not going to lie, being able to ask eg "what does the /v1/frob/buzz API response look like" and not have to trawl our sprawling confluence wiki/JIRA system (or the Google/Microsoft/whatever equivalents) would be incredibly useful. However, it does sound like it would be expensive to build and maintain such a specially trained model, so who knows when that would be a viable productMicrosoft already has another obvious path to profitability with this, by adding capabilities to Office 365. Would companies be willing to pay a few bucks more per user per month to have access to an AI assistant that knows everything contained in your company's Sharepoint, your OneDrive, and your mailbox, and can provide you with information on demand or compose emails and letters for you with minimal input? I'm guessing that would be an easy sell for most. Forget New Bing, New Clippy could be huge.
The principled thing to do is to decentralise the whole compute calculation to users. For the betterment of society as a whole, less inequality and resiliency against cyber attacks. NO WAY THAT THIS IS GOING TO HAPPEN ... unless the big four make custom silicon and let Nvidia to dry ...Unless Bard is very different from ChatGPT, running it on a consumer grade machine, even a high-end gaming PC does not sound practical.
According to Wikipedia, GPT-3's parameters take up 800 gigs.
https://en.wikipedia.org/wiki/GPT-3
I wouldn't be surprised if there were a lot of low-hanging fruit to optimize in these chat bots, they were basically research papers not very long ago. My gut feeling is the models could get away with being more sparse than they are, but that work is way above my pay grade.Yes, the whole idea that ML is free/cheap is extremely temporary. Even these $15 a month subs are not cutting it. I even expect the likes of nvidia too to have, dlss 4 or something on subscription.
I was involved in a small scale ML project with ~20000 users, and the amazon bill was going higher and higher. At least as long as we wanted to do a good job. If Microsoft and google have a race for the best ML datacentres the costs will skyrocket. Google wanted to develop their own custom silicon though.... which will be doomed to failure because they microdose so much that they can't keep any project for more than 13 months.
The principled thing to do is to decentralise the whole compute calculation to users. For the betterment of society as a whole, less inequality and resiliency against cyber attacks. NO WAY THAT THIS IS GOING TO HAPPEN ... unless the big four make custom silicon and let Nvidia to dry ...
Of course consumer electronics will have to change, mostly in terms of memory ... parallelism .. and networking.
Maybe i am a bit to deep in the world of serve the home, but you know you have your oven, boiler and your rack !
It won’t, and it won’t reduce garbage, it will be all garbage, just more sophistcated and convincing. Sadly.How much will it impact profits to use AI to reduce the amount of garbage returned in a search?
The thing with the large language GPTs is the magic comes from the giant size of the models and all the fine-tuning, reinforcement learning that is done after the pre-training. And that's a lot of human labor involved.Thankfully, there are open source projects that are being released to the public. I've been running Stable Diffusion for months on my desktop. And stability AI is working on everything, so eventually they'll release their (less capable) version of ChatGPT.
A lot of LLM already is open source. The problem is that simply generating the content from the model requires ridiculous amounts of GC power. To then regularly retrain the model with up to date datasets goes from 'ridiculous' to 'batshit fucking insane' at the scales we're talking about here. You can't scale the necessary processing easily with off the shelf hardware, which is exactly the opposite of their current search technology backend.Oh it's too expensive google?
Then just open source the model, so people can run them locally. We'll save you a ton of money, you can thank us later ;-)
On the cheap no, I think the problem is the GPU memory ... ideally you want a few GPUs with a total 800Gb of ultra fast ram, to train the most top notch models. At the moment top consumer GPUs have 24GB and pro have 48 GB GDDR6 (ECC) but you can connect them to appear as one; the pro ones cost an insane amount of money ... the consumer ones too. So the market has to change a lot for consumer electronics if there is a chance to decentralise processing power.If they were to open source the model, we can prune it and it'd be able to run just fine, even if it's not fully featured.
This must be the Joke of the Day:This is a huge and very strange assumption to make right off the bat:
You're implying that they lost $100B of market cap purely because they executed a demo poorly - and not the much more reasonable assumption that the stock market thinks the whole thing (large language model based chat search) is a bad idea. Google has spent 20 years honing their search model, figuring out how to execute their core product well. It makes much more sense that investors want them to stick with what they're good at (even if being good at running a good business around search is independent from making the product good for the user).
The thing with the large language GPTs is the magic comes from the giant size of the models and all the fine-tuning, reinforcement learning that is done after the pre-training. And that's a lot of human labor involved.
That’s what I’d like to know. I don’t want to chat with an AI. I don't want to ask an AI to answer a question because there’s an excellent chance it made up the answer.How much will it impact profits to use AI to reduce the amount of garbage returned in a search?
Generating an alias email can be done in Android. You can have Android make up an email address for you.
Ad
Ad
Generating an email alias is a great way to hide your email address.
Ad
A big button looking thing that says “Next” on it, but takes you to an ad
You can generate an email alias in Android and have it forward to you email address. Did you see this ad yet?…
Chat isn't a very good match for search, especially because getting it right matters. But search is something that a very large number of people do. Generating a list of random names etc is a perfect use, but it's also a niche case. Are there really enough GMs and writers to make that a business case for something as expensive as the training for these systems? I doubt it.
The reinforcement learning is really important, that's the "magic" I'm referring to. I've played around with a number of GPT2 and GPT3 models and they are not nearly as convincing as ChatGPT or Text-Davinci-003.Actually, transformer-based language models like GPT are typically trained using unsupervised learning, not reinforcement learning. ChatGPT, like the original GPT models, uses unsupervised pre-training and supervised fine-tuning without the use of reinforcement learning. While pre-training large language models can require significant computational resources and expertise, it does not necessarily require a lot of human labor.
This is one instance where I could definitely see paying for a service -- I would happily spend $10/mo for a search service that gave me the correct, well formed answer. I do searches dozens of times a day and don't have time to wait for interstitials.Unfortunately, if they're really keen on pushing adverts, then they'll likely move to interstitials which you must look at before getting your results.
[...]
Maybe there will be two interfaces, so you can chat with gchat or perform a traditional search.
And you would know this, how? Your post is entirely speculation. You don't know the first thing about ML, you know you don't, but talked out of your ass anyway.If they were to open source the model, we can prune it and it'd be able to run just fine, even if it's not fully featured.
This is why AI will be so disruptive to Google's business model.
We'll have models that are very capable, that can run locally from a 200MB file.
Actually, transformer-based language models like GPT are typically trained using unsupervised learning, not reinforcement learning. ChatGPT, like the original GPT models, uses unsupervised pre-training and supervised fine-tuning without the use of reinforcement learning. While pre-training large language models can require significant computational resources and expertise, it does not necessarily require a lot of human labor.
This must be the Joke of the Day:
"Google has spent 20 years honing their search model"
Those days are Long Gone, Google search has degraded over the last 7 years to almost worst available search engine.
That's because Amazon doesn't have search. They have a recommendation engine that is seeded by the keywords you type in the box. It may have a little magnifying glass icon but that doesn't mean anything. Their recommendation engine routinely tosses in items that don't meet your search criteria, because they estimate there is a chance you'll buy those things. Also, it's half paid advertisements.Edit: The only thing worse is Amazon Store Search which is a Disaster of Junk and results that have got nothing to do with your search phrase.
I read that BingChat gives you an answer and a source link.
The hit search engine DuckDuckGo already does that for certain types of searches. eg:
director fern gully
returns "Bill Kroyer" at the top above all results.
Unfortunately, if they're really keen on pushing adverts, then they'll likely move to interstitials which you must look at before getting your results.
Or are they envisioning this scenario?
Fred > Who directed FernGully?
gchat > That is a good question, which reminds me. Did you know there are bargains to be had on smart home tech? Oh, to answer your question, Bill Moyers.
Fred > Thanks.
gchat > Anytime, meat bag.
Maybe there will be two interfaces, so you can chat with gchat or perform a traditional search.
If they were to open source the model, we can prune it and it'd be able to run just fine, even if it's not fully featured.
This is why AI will be so disruptive to Google's business model.
We'll have models that are very capable, that can run locally from a 200MB file.
Modeled on what? That's a pretty huge exaggeration.A ChatGPT-style search engine would involve firing up a huge neural network modeled on the human brain
And you would know this, how? Your post is entirely speculation. You don't know the first thing about ML, you know you don't, but talked out of your ass anyway.
Why is that?
Yes, and that will make these people use gpt-like AIs to create whole website that look legitimate and praise their product. Imagine a site like ars, complete with stories and user comments, continuously updated, all AI generated, just to promote something. There is definitely a possible future where the internet is 99% made-up by AIs. This could be wild.I think the safer assumption is that LLM's may trigger a shift in SEO. There is an inherent arms race between search tools and tools that clog results with bullshit. There are constant financial incentives for each side to improve their techniques. New tech doesn't make that go away. It just further incentivizes the other side to up their game.
I think 200MB models that are useful is absurdly optimistic.
So what we did is, you know, the collaboration with various entities which we paid an important part in stable diffusion to, is kind of led by us. We took a hundred thousand gigabytes of image label past two billion images and created a 1.6 gigabyte file that can run offline in your MacBook and create a 2.6 gigabyte file which is relatively you can transmit it over over the phone network.
Yes, and that will make these people use gpt-like AIs to create whole website that look legitimate and praise their product. Imagine a site like ars, complete with stories and user comments, continuously updated, all AI generated, just to promote something. There is definitely a possible future where the internet is 99% made-up by AIs. This could be wild.
Yes, the whole idea that ML is free/cheap is extremely temporary. Even these $15 a month subs are not cutting it. I even expect the likes of nvidia too to have, dlss 4 or something on subscription.
I was involved in a small scale ML project with ~20000 users, and the amazon bill was going higher and higher. At least as long as we wanted to do a good job. If Microsoft and google have a race for the best ML datacentres the costs will skyrocket. Google wanted to develop their own custom silicon though.... which will be doomed to failure because they microdose so much that they can't keep any project for more than 13 months.