I used to work in Alexa and I just don't see it being converted to an LLM solution at all. A core problem is that user expectation is that responses are quite reliable outside of general knowledge, and the currently hyped LLMs just don't provide that. Also, general knowledge has never even been a big draw for average consumers for a voice assistant. Having Alexa work off of an LLM is just not going to be particularly reliable anytime soon and is ripe with risk for consumer harm in various ways. Another massive problem that LLMs are facing for voice assistants is in time of response, both with time to response start and length of response. If a voice response isn't highly reactive and to the point, users get pretty pissed off and decrease engagement.
We just aren't anywhere near having LLMs be reliable enough with low enough latency and to-the-point responses to launch as a voice assistant replacement for Alexa/Siri/Google Assistant.
Yeah the vast majority of what we use Alexa for is smart device control and then timers.I think they have seriously overestimated how much people want that.
I have Amazon Echos in my home....the majority of use is home automation activation, weather, and timers. My wife and I also use it to communicate between our two home offices, which are on different floors.
The problem with that in the space of voice assistants is: that's not enough to launch anything. Look at what Apple has been doing so far with Apple Intelligence that isn't even within the voice assistant space. It's all so minimal and therefore not particularly useful. Now Apple is building a reputation of having near-zero value with Apple Intelligence, so consumers are beginning to conclude Apple can't do anything with AIFunny how Apple’s caution seems to be borne out by Amazon’s experience.
Personally, I’d rather have tiny, incremental, predictable improvements in on-device agents with abilities to do real-world things.
There’s always ChatGPT if you want to…chat.
I think you are missing the point of what Alexa is aiming to achieve with using LLMs for voice interactions, specifically regarding Alexa responses. It wants to leverage the general knowledge successes of things like ChatGPT in a way that customers perceive it as having enough value to pay for on a monthly basis, likely through offering 1 to 3 month free trial of it when it becomes available. In a perfect world, that would be a killer revenue stream. Do I think they'll succeed? Absolutely not.If regular Alexa couldn't generate any meaningful subscription revenue or sales commissions, does the team seriously expect GenAI Alexa (which is presumably far more expensive per query) to be different? It seems like Amazon is doubling down on a failed monetization strategy.
The thing is: there are a massive amount of applications for 'AI' because it's an extremely broad technical space, including the example you gave using visual models. 'AI' has been a thing for a very long time, but most recently it is getting a lot of attention with large language models (LLMs), which is a smaller but still huge subset of 'AI'.I would have agreed with you but we've crossed the threshold where 'AI' means absolutely anything the marketing department decides which may or may not be a step in any direction whatsoever.
E.g. any sort of eye tracking for camera autofocus now seems to be called AI; we had lesser versions of eye tracking autofocus at least a decade ago and nobody felt the need to call it AI but at some point the tag got attached. Including by Sony, which people generally regard as having the best eye-tracking autofocus in the industry.
I really think they just kept improving what they had and put a new label to it.
No one wants to be the first to hook up AI to things that affect the real world even in small ways. Wonder why.The problem with that in the space of voice assistants is: that's not enough to launch anything. Look at what Apple has been doing so far with Apple Intelligence that isn't even within the voice assistant space. It's all so minimal and therefore not particularly useful. Now Apple is building a reputation of having near-zero value with Apple Intelligence, so consumers are beginning to conclude Apple can't do anything with AI
I don't think they can solve this problem, but I disagree that it will block implementation. Alexa is a money-loser, and there is a very significant part of the American public that does not care about accuracy. Yes, it hallucinates, yes, it's going to be slower than the current system, yes, it's going to actually make responses to simple tasks less accurate. And they will release it anyway and we will get to see videos on your choice of social media about the most amusing results.I used to work in Alexa and I just don't see it being converted to an LLM solution for responses at all. A core problem is that user expectation is that responses are quite reliable outside of general knowledge, and the currently hyped LLMs just don't provide that. Also, general knowledge has never even been a big draw for average consumers for a voice assistant. Having Alexa work off of an LLM is just not going to be particularly reliable anytime soon and is ripe with risk for consumer harm in various ways. Another massive problem that LLMs are facing for voice assistants is in time of response, both with time to response start and length of response. If a voice response isn't highly reactive and to the point, users get pretty pissed off and decrease engagement.
We just aren't anywhere near having LLMs be reliable enough with low enough latency and to-the-point responses to launch as a voice assistant replacement for Alexa/Siri/Google Assistant.
Yeah, my impression of the way AI is done is that there's no way to avoid hallucinations. It's literally baked into the processes used to create responses. And input matters wrt output in subtle ways that won't necessarily parse properly the first time, resulting in some odd response that's a dead giveaway that the response came from an AI, and not a human.On one hand, I’m glad that at least one massive company seems to care about its products producing nonsense.
On the other hand… isn’t avoiding hallucinations basically impossible, given how LLMs are mainly very clever autocomplete? Even with some effort to provide sources like Google tries to, they’re terrible at picking up tone and cite sarcasm.
We got an Echo Dot for $5 on a punt circa 2018 because it was bundled with something or another. My pronunciation is about as straightforward and uninflected as southern England can produce and my history in publishing means that I chide myself if I even so much as split an infinitive. That's despite having spent now almost half my adult life in the US; people even ask when my natural-born-American child moved here because he's done such an excellent job of learning my British accent.
Alexa couldn't even understand the prompt for a cookie recipe that I recited directly from its literature. Due to its general uselessness we passed on the Echo Dot to somebody else shortly thereafter.
So either the speech recognition is calibrated by geolocation or Alexa was having a really bad few days.
When was Alexa not an AI product?At this point, whenever I see a new product release and the main feature is "now includes AI" I'm 100% not interested in that product. It's a step backwards IMO. Sure hope this trend of cramming AI into everything ends sooner rather than later.
Never. A pattern-matching state-machine with an unreliable voice-recognition interface is not AI.When was Alexa not an AI product?
'Your plastic pal that's fun to be with' was sarcasm. NOT a road map.
I am unsure what that has slipped past the tech bros. Of course, there is the whole problem with '1984' so I suppose I shouldn't be surprised.
Yeah no kidding, who knew shopping for a TV would end up being so complicated.This! I'm currently shopping for a new TV but many manufacturers have ruled themselves out with this insane BS.
It struggles with "Turn the reading lamp on" in my standard american dialect.Damn, if it can't handle South Anglian then it definitely can't handle Tihdewottah's 17th century South Anglian.
Personally I'm looking forward to Alexa AI ordering a give gallon drum of sexual lube for someone by accident. The prudishness of American society will make that significantly worse for Amazon to deal with than just an accidental order for too many widgets. Might even cause them to pull the plug on Alexa AI overnight.If history is any guide, they will invest heavily, still have hallucinations, give into the sunk cost fallacy, hype the shit out of their new "Alexa AI", and ignore the inevitable tsunami of complaints from people whose Alexa routinely gives false information and accidentally orders 10000 widgets.
"Alexa, turn on the lights"Personally I'm looking forward to Alexa AI ordering a give gallon drum of sexual lube for someone by accident. The prudishness of American society will make that significantly worse for Amazon to deal with than just an accidental order for too many widgets. Might even cause them to pull the plug on Alexa AI overnight.
I’ve set my various routines to recognize a few different phrases / word order I might use for a particular action.You know what, Alexa is so bad at just understanding something simple, how bad could the hallucinations be comparatively?
Alexa cant understand "turn off the downstairs lights" is functionally the same as "turn off the lights downstairs". Yes, i know one is grammatically better than the other, but functionally this is something simple
Whatever its faults, in my experience, Alexa is miles ahead of Apple in terms of voice control as it stands right now. I’ve had no problems at all controlling a myriad of smart device brands with Alexa/Habitat but Apple/Homekit is very fussy. Half the time Siri cannot understand what I’m saying about a specific device but that’s rarely an issue with Alexa.Apple and Amazon are two of the biggest players in home voice assistants (hardware and software). Both have struggled for years to develop a (arguable) compelling voice product. Each is taking a very different approach.
It's still very early in the AI wars. One can argue that there is no market for more advanced products, but these companies seem to believe otherwise and they have the resources to continue to evolve.
It was marketed as such though..Never. A pattern-matching state-machine with an unreliable voice-recognition interface is not AI.
The same thing wrong with Blade Runner or the Torment Nexus, Musk et Al look at wild wealth disparity in a cyberpunk dystopia as a preferable outcome.Just out of curiosity... What whole problem with 1984? Be specific.
I mean, I've read an awful lot of dystopias and near-dystopias and cyberpunk grim futures, and, when this comes up, I'm always curious specifically what is meant.
Thanks in advance for your thoughtful reply.
The same thing wrong with Blade Runner or the Torment Nexus, Musk et Al look at wild wealth disparity in a cyberpunk dystopia as a preferable outcome.
It doesn’t help when the majority of idiots who worship them also reference 1984 as a reason to oppose universal healthcare but have never read or critically analyzed the actual story.
Different LLMs produce different numbers and types of hallucinations. And various post-training and prompting techniques fudge those numbers around as well... So who knows, maybe they will settle on some threshold, like 0.3% of some sample of queries, and then also try to disable it from doing and saying the kinds of things that get them sued.On one hand, I’m glad that at least one massive company seems to care about its products producing nonsense.
On the other hand… isn’t avoiding hallucinations basically impossible, given how LLMs are mainly very clever autocomplete?
I am SICK to death with everyone calling all this stuff AI. It is NOT, I repeat NOT AI if you just write software that is more and more complex, capable, large, and (to repeat) capable. You can make it as complex and sophisticated and capable as you want, but it is still just deterministic SOFTWARE, doing what it is programmed to do. It may be appear to be intelligent and appear to go off in some strange humanistic type directions sometime, but it is still just very, very, very, very complex and capable software. "AI" is just a buzzword to make something appear to be more important, or different than it is. Remember the other famous buzzword/acronym, Y2K.