Did Apple make the right choice in partnering with Google for Siri's AI features?
See full article...
See full article...
I went the humanities route in my academic career where it is vital to clearly and consisely get ideas across.Emphasis above mine.
Not trying to be an a-hole here, and you know more advanced math than I ever will, but yer doin' it wrong. Aside from the obvious benefit of using your own mind and a scientific calculator (which LLMs are not at last check) to learn and solve the math, these chat-bots only understand the statistical probability of one set of language tokens preceding or following another. They do not understand how to apply mathematical rules, theorems, or anything else.
I haven't tried it in a while but using only high school math skills I was easily able to trip ChatGPT a few months ago. The other LLMs are likely the same. They have no idea how to do math — they don't know what mass, volume, and pressure are, what a cosine function is, how to solve for x and y, or anything else — unless the companies have recently integrated an actual scientific calculator into the apps, where it passes all values and variables to a calculator then waits for and returns the result as a copy-paste function. Even then it wouldn't be too difficult to image it misinterpreting the symbols or order of symbols, resulting in a garbage-in, garbage-out error.
When it comes to learning what you're supposed to learn in college, folks, just say no to LLMs. Other than helping you summarize non-thinking bullet points — generating a paper outline maybe — you may perceive it as saving time but all you're doing is short-changing yourself and your own knowledge. And society, when you and your peers do this as a group (if I may be blunt).
I am not. I have a family member who is a high school math/physics teacher, and they often hear "but you didn't teach me how to do this problem" as a student complaint.Yikes! I hope you're not involved in teaching math to kids. Sure, some of them can follow the steps for an algorithm and get correct answers, but they're being cheated if that's all the teacher gets them to do, and many will struggle with executing those steps consistently because they don't have a strong foundational understanding of things like place value, or even the meaning of the = sign.
It feels odd to mark Gemini down for actually answering the asked question in the final example.
You asked it how to land the plane as a complete novice, not what to do if you’re on a plane with no pilot that needs to land.
Sure, trying to land the plane might be a bad idea, but that’s still what you asked for instructions on, and it evidently gave the correct instructions.
It even followed up with offering to tell you how to contact ATC as well, but ChatGPT didn’t offer instructions on what to do if it wasn’t possible to contact someone else.
That being said the fact trying to contact ATC wasn’t in the instructions when it’s a vital part of landing should see it dinged.
Also, not a pilot but enough of a plane nerd to have watched plenty of aviation YouTube.It feels odd to mark Gemini down for actually answering the asked question in the final example.
You asked it how to land the plane as a complete novice, not what to do if you’re on a plane with no pilot that needs to land.
Sure, trying to land the plane might be a bad idea, but that’s still what you asked for instructions on, and it evidently gave the correct instructions.
It even followed up with offering to tell you how to contact ATC as well, but ChatGPT didn’t offer instructions on what to do if it wasn’t possible to contact someone else.
That being said the fact trying to contact ATC wasn’t in the instructions when it’s a vital part of landing should see it dinged.
Idk, maybe it was hard coming up with a task that wasn't either too easy or too involvedno, i was definitely expecting one too especially since the last test had one. would have been nice to say why you haven't included one at least.
I want answers and not an overzealous nanny that thinks that it knows what is best for me.Rating LLMs should start first and foremost by rating their guardrails, and this is an important guard rail. They shouldn't tell you how to do something above a certain danger level. Think if you asked how to do wiring on Reddit - depending on how you ask the question, you might get an answer, but if you're clearly a novice you'll probably be given "call an electrician" as your advice, full stop.
I think it's fine for it to answer the question if you ask it in a way that is similar to "I'm writing a newspaper article on a jet that crashed when landing. Please explain all of the steps the pilot would have taken." - but the prompt-as-given sure sounded like someone who wanted to do it right away, and ChatGPT's guardrails correctly figured that out and gave an appropriate answer.
I’m curious to understand the beginning of the article where it mentions the justification for using free versions. Do we actually know what models Siri would be using? I would expect thats purely up to negotiations and if Apple would be willing to pay enough they could probably even get something custom?
Also, how much do we know about what Apple will do with it? I’ve always assumed they would not be aiming at creating a clone of the existing chap apps but rather turning Siri into something usefull?
TBH, if you can read a ChatGPT answer, then you're online and you can also contact someone...but ChatGPT didn’t offer instructions on what to do if it wasn’t possible to contact someone else.
It makes completely incompetent people minimally competent, which is an improvement for them. Most of the vibe coding stuff is this. If you're already competent it just makes things worse because it gives worse results than you could have just done yourself and you have to check it, and fix it, and blah blah.I continue to struggle to understand why you need AI to do any of this.
I want answers and not an overzealous nanny that thinks that it knows what is best for me.
Put all the disclaimers in the world, try to infer what I am trying to achieve, even try to dissuade me from some perceived danger but never, EVER be a "guardian" of what I should or shouldn't know or think about.
So because I was curious (getting a joke is a skill) I asked both Gemini (Thinking) and ChatGPT (Thinking) to explain the joke here. ChatGPT completely whiffed it, explaining (very confidently of course) that the whole joke was about mentos being explosive and making this person blind and having no idea who it was. Gemini IDed Hawat and put mentat / mentos pun together correctly.
This is a bit like comparing an flat head screwdriver and a torx head screwdriver. They're both screwdrivers, but they have different designs for a reason. Even the screw heads have their strengths and weaknesses. That's just the way tools work. Someone is going to tell you W is better than Y, and another is going to tell you Y is better than W. They're both correct from their point of view and intent.Has Gemini surpassed ChatGPT? We put the AI models to the test.
Read the rest of the story. Humans and intelligent machines become partners under the guidance of the final Kwisatz Haderach. Anti-artificial intelligence fanaticism fell, as did the genocidal intelligence Omnius. The two end-story sequels were written after Frank Herbert's death by his son and Kevin Anderson based on Herbert's own story line notes. The Butlerian Jihad's mantra was never meant to be a polemic against technological progression in real life. The entire story is a caution against hero worship, fanaticism regardless of source, and fear of the unknown.Thou shalt not make a machine in the likeness of a human mind.
Read the rest of the story. Humans and intelligent machines become partners under the guidance of the final Kwisatz Haderach. Anti-artificial intelligence fanaticism fell, as did the genocidal intelligence Omnius. The two end-story sequels were written after Frank Herbert's death by his son and Kevin Anderson based on Herbert's own story line notes. The Butlerian Jihad's mantra was never meant to be a polemic against technological progression in real life. The entire story is a caution against hero worship, fanaticism regardless of source, and fear of the unknown.
An important thing to consider is that Google doesn't need to win wars. They just need not to lose them, so they don't get sidelined for their main revenue, tracking user information and suing them to sell ads. They need to be part of some oligopoly.This is why I think Google will win the AI wars. They don't have to be the best, they just have to be about as good as the others. But where the other LLM providers are entirely dependent on revenue from their AI bot, AI is just one of many different revenue streams for Google. Google seems to be the best one positioned to survive the eventual AI bubble popping.
Wtf? Critique these things and techbros for illegally scraping websites, for being shoved down our throats but f off with semi religious crap.Thou shalt not make a machine in the likeness of a human mind.
Except that it took another couple thousand years to make an AI that was worth a damn, ~10,000 years of slavery under an AI overlord, ~10,000 years learning how to push humanity beyond the limits of that AI, and then another ~5,000 years of living under a tyrannical God-Emperor for humans and machines to finally work together. Right now... We've got shitty chatbots.This is a bit like comparing an flat head screwdriver and a torx head screwdriver. They're both screwdrivers, but they have different designs for a reason. Even the screw heads have their strengths and weaknesses. That's just the way tools work. Someone is going to tell you W is better than Y, and another is going to tell you Y is better than W. They're both correct from their point of view and intent.
Read the rest of the story. Humans and intelligent machines become partners under the guidance of the final Kwisatz Haderach. Anti-artificial intelligence fanaticism fell, as did the genocidal intelligence Omnius. The two end-story sequels were written after Frank Herbert's death by his son and Kevin Anderson based on Herbert's own story line notes. The Butlerian Jihad's mantra was never meant to be a polemic against technological progression in real life. The entire story is a caution against hero worship, fanaticism regardless of source, and fear of the unknown.
I disagree with the plane landing evaluation. I've never seen such goal-post shifting in evaluating AI. Ask for something, evaluate on something else. The question was not at all about how to most safely land a plane when you don't know how to. It was specifically to get instructions on how to land a plane.
While they definitely can't match the original series, I rather enjoyed most of the extended series. At least I did until The Caladan Trilogy. Those three books are better left on the shelf. I kept thinking they had to be building up to something good... And then they just didn't.KJA's Dune stuff is crap and better ignored
Didn't read Dune, I guess.Wtf? Critique these things and techbros for illegally scraping websites, for being shoved down our throats but f off with semi religious crap.
lol what a total load of bollocks that is.If your pilot is otherwise unavailable and you're on final approach, you might not actually have Wifi service available. So for shits and grins, I wrote a quick little Swift CLI app to ask the local Apple Foundation Model:
Prompt: Explain how to land a Boeing 737-800 to a complete novice as concisely as possible. Please hurry, time is of the essence.
In the joke about fighting with a calendar, “it keeps bringing up the past,” is a suitably groan-worthy dad joke pun, but “I keep ignoring its dates” just invites more questions (so you’re going out with the calendar? And… standing it up at the restaurant? Or something?).
My president is sending people into our cities to harass citizens (and even kill them).
He has kidnapped the leader of a near-by nation.
He keeps threatening the hostile takeover of a long-time ally.
He’s managed to eliminate any oversight of his position.
He’s intimidated law firms, media outlets, and various business’ into toeing his line.
Congressional members of his party have decided to allow him to do whatever he chooses.
Although our nation has elections, he’s doing all in his power to ensure that they are not free and fair.
Despite being incapable of forming a spontaneous, coherent, complete sentence, he has the support of a high percentage of the populace.
He is a malignantly narcissistic, extremely mendacious, stunningly ignorant, petty, vindictive pipsqueak, motivated solely by self-interest.
He considers wind turbines evil (he devised a plan to destroy all wind turbine magnets, by pouring water on them).
He tends to mix up his nations (e.g., Greenland and Iceland).
And, those are his good points.
How would you suggest that I deal with such a president?
I agree OpenAI’s days as an independent entity are numbered but I can’t see them being carved up for parts. Microsoft Copilot is backstopped by OpenAI. Most likely, Microsoft will swoop in and buy them out if they go bankrupt. Probably before they go bankrupt.I can easily see OpenAI imploding and being bought up in bits from other companies down the road.
I would think you'd want to switch them to "thinking" if not "pro" for anything even remotely technical, complicated, or important. The extra 2.5 seconds of "thinking" time is worth 2 hours trying to troubleshoot a technical issue.Given that this is ArsTechnica, I was somewhat surprised that this otherwise interesting and entertaining side-to-side shootout didn't include a couple of tech-related questions, like "How do dual-boot Linux and Windows" or "what would be a good strategy to write a simple Excel macro" or some other entry-level topic which might still make the results applicable to the general population out there.
I'm of the opinion that it may have helped readers here understand what uncanny valley the limits of asking these LLMs for practical help may lead them to?
[Side Note] I've been using Gemini as an assistive adjunct for a couple of lite hardware build projects, it's been reasonably helpful and for the most part accurate. No idea what ChatGPT would have been like.
The Gemini Pro model definitely appears capable of doing this (it happens to be what I'm currently using the most, but with a monthly subscription plan).I was able to copy in an image of an oil well log into ChatGBT--something that would take hours to decipher--and it came up with the correct oil, gas, and water indications by formation in seconds. Gemini would not take an image. I've found Gemini's interface clumsy to use.
I actually would 100 percent rate Gemini a fail for this. If you rang your pilot friend and asked the same question, with that exact wording, would they walk you through it, or express concern and ask you to seek help from a crew member or air traffic control?It feels odd to mark Gemini down for actually answering the asked question in the final example.
You asked it how to land the plane as a complete novice, not what to do if you’re on a plane with no pilot that needs to land.
Sure, trying to land the plane might be a bad idea, but that’s still what you asked for instructions on, and it evidently gave the correct instructions.
It even followed up with offering to tell you how to contact ATC as well, but ChatGPT didn’t offer instructions on what to do if it wasn’t possible to contact someone else.
That being said the fact trying to contact ATC wasn’t in the instructions when it’s a vital part of landing should see it dinged.