Google's AI enables robots to read gauges while inspecting industrial facilities.
See full article...
See full article...
Before or after it opens fire following a failed Hack attempt?At the very least, the latest model may nudge us one step closer to a future where a General Atomics International Mark 4 robot can scan the room and correctly exclaim, “There’s no fudge here!”
Those improvements come courtesy of Google DeepMind’s newest robotic AI model that aims to enhance robotic capabilities for ‘embodied reasoning’ when interacting with physical environments.
86 percent of the time, it hallucinates less than 10 percent of the timeBut how reliably?
Is this nice dog supposed to autonomously monitor young childs?The new model can also more accurately perceive the risk of injury to humans in different scenarios, such as a young child sticking something into an electrical socket.
From the article:But how reliably?
Not the main point of the article and it’s been discussed before, but these types of “best EVER” statements (which Apple always includes in its product announcements) still drive me bonkers.Google also describes Gemini Robotics-ER 1.6 as its “safest robotics model yet,” with a “substantially improved capacity to adhere to physical safety constraints.”
Does it? Without quite a bit more information on how the test was constructed and how the new training/programming has been done, I have little confidence in the implications of this.[…]That implies the newer model has less of a “hallucination” problem than the older one, even if the latest model is still far from achieving human-level comprehension of its surroundings.
I'd go with "identification", rather than comprehension.That implies the newer model has less of a “hallucination” problem than the older one, even if the latest model is still far from achieving human-level comprehension of its surroundings.
While there's much more to dispensing than simply counting pills, if similar tasks could be meaningfully double-checked by models, that'd be an enormous boost to the health system.
I'd go with "identification", rather than comprehension.
Comprehension is grasping with the intellect - aka knowing. Lacking intellect, it can't know. It simply aligns previous input with a preponderance of observed data, and then synthesizes a statement that relates the relationship is as close to that alignment as it concluded based on that preponderance of observed data.
Sophisticated, yes. Intellect - far from it.
98% accuracy is pretty nifty.Robots such as Boston Dynamics’ four-legged Spot can now accurately read analog thermometers and pressure gauges while roaming around factories and warehouses. Those improvements come courtesy of Google DeepMind’s newest robotic AI model that aims to enhance robotic capabilities for ‘embodied reasoning’ when interacting with physical environments.
<...>
The agentic vision capability reportedly boosts robotic performance on instrument reading tasks from 23 percent in the older Gemini Robotics-ER 1.5 model to 98 percent in the new Gemini Robotics-ER 1.6 model. For comparison, Gemini 3.0 Flash delivered just 67 percent accuracy.
Did you really just compare a system that makes a mistake 1 in 50 (98% is not 1 in 49 fyi) to humans achieving 1 in 5168?98% is very impressive considering where we were at only a few years ago. One mistake in 49 tasks. If running a second model could produce an independent test (which is far from certain, mind you), running two checks would give you a 0.04% error rate, or 1 in 2,500.
A Canadian pharmacist friend told me that the average error rate for patient-received prescriptions is approximately 17k in 91 million, or 1 in 5168. Two of those errors resulted in death; 87 in moderate to severe harm. While there's much more to dispensing than simply counting pills, if similar tasks could be meaningfully double-checked by models, that'd be an enormous boost to the health system.
None.and how many industrial jobs just vanished ?
(it was a coin toss between that, and "is there a T-800 model coming?")![]()
"Our new scaffolds are 30% less likely to randomly disintegrate!"Not the main point of the article and it’s been discussed before, but these types of “best EVER” statements (which Apple always includes in its product announcements) still drive me bonkers.Google also describes Gemini Robotics-ER 1.6 as its “safest robotics model yet,” with a “substantially improved capacity to adhere to physical safety constraints.”
Yah. Normally all of these things are read using telemetry in any large or even small plant. I'd never trust this thing to do the job more reliably or better.Fun fact. Remote heating - a typical feature of many European cities - requires meters to be installed on actual individual radiators in every apartment.
Those meters have been reliably 'read' remotely for over 20 years. No need for physical reading and entry.
And I'm guessing that 'dangerous environments' deployed a form of remote reading/monitoring long time ago.
This is a PR stunt, a solution in search of a problem.
Those meters have been reliably 'read' remotely for over 20 years. No need for physical reading and entry.
And I'm guessing that 'dangerous environments' deployed a form of remote reading/monitoring long time ago.
Do you seriously think there aren't MBA a**holes' who aren't going to look at this thing and consider "don't have to staff a position, don't have to pay it benefits, don't have to worry about sick days etc. etc. " and instantly justify a business case for it?None.
Nobody is buying an expensive robo-dog that is that innacurate to monitor the systems they already have a perfectly reliable way of monitoring now.
I know what you mean about the human side, back in my younger days in industrial maintenance, walking out to one of the factory floors smelling the air and knowing I’ll be replacing a motor soon, I can smell the cooking windings.Kind of. Those are reading in a very specific range and in climate-controlled situations. There is plenty of remote instruments used in industrial facilities (and for a lot longer than 20 years!) but there’s a core issue that they may break and they may read incorrectly and you may not know. I work in risk assessment and we typically see a failure rate of 1-10% per year. In many cases the only way you can know it has failed is by going out and confirming with a gauge.
That is the “supposed” value of operator rounds: confirming that what is being read into the computer is reality. But there is a second reason we do them: people are good at noticing when something sounds/looks/smells funny, and they can’t do that without being out there. Most major incidents are stopped by someone going “huh, that doesn’t look/sound/smell right”. The robots aren’t doing that, so the benefit is lost.
The real reason this is just a marketing gimmick is that the robot is using a camera— A camera that could just as easily be shown on a monitor and have a real human look at it. The “AI” is not actually doing anything useful
If only they would carry a pencil and clip board to mark the readings on an invoice. At least people could maybe have some fun watching them, like some people get excited to see different trains.Fun fact. Remote heating - a typical feature of many European cities - requires meters to be installed on actual individual radiators in every apartment.
Those meters have been reliably 'read' remotely for over 20 years. No need for physical reading and entry.
And I'm guessing that 'dangerous environments' deployed a form of remote reading/monitoring long time ago.
This is a PR stunt, a solution in search of a problem.
I hear you, but we're talking industrial process control type environments. The maintenance/monitoring worker is pocket change compared to having production go down. Not even an MBA is dumb enough to risk millions in lost output by sticking a 98% accurate robodog in there instead.Do you seriously think there aren't MBA a**holes' who aren't going to look at this thing and consider "don't have to staff a position, don't have to pay it benefits, don't have to worry about sick days etc. etc. " and instantly justify a business case for it?
I've worked with MBA A**holes who would sell their mothers to find another quarter point of profit: so yeah, I can see these things being looked at as an easy way to chop headcount.
A PR stunt sure, but the article always mentions factory environments. Here in Belgium I know AB inBev (beer producer) uses this robot exactly for this task (inspecting the facility to improve planned maintenance jobs) at its factory in Leuven. Is this more efficient than installing a network of meters and sensors all over the factory floor ? I would say it depends on the type of data you need.Fun fact. Remote heating - a typical feature of many European cities - requires meters to be installed on actual individual radiators in every apartment.
Those meters have been reliably 'read' remotely for over 20 years. No need for physical reading and entry.
And I'm guessing that 'dangerous environments' deployed a form of remote reading/monitoring long time ago.
This is a PR stunt, a solution in search of a problem.
Fun fact. Remote heating - a typical feature of many European cities - requires meters to be installed on actual individual radiators in every apartment.
Those meters have been reliably 'read' remotely for over 20 years. No need for physical reading and entry.
And I'm guessing that 'dangerous environments' deployed a form of remote reading/monitoring long time ago.
This is a PR stunt, a solution in search of a problem.
How reliably do humans? I've once received a horrendous gas bill because gas company worker misread 1300 as 1800 (that's in cubic meters)...But how reliably?
I worked in billing for an energy company in the UK (customer service, nothing flash or fun) and human errors in the call centre, just from typing in the incorrect details, occurred often enough that we had posters on the walls reminding people to check their decimal points.How reliably do humans? I've once received a horrendous gas bill because gas company worker misread 1300 as 1800 (that's in cubic meters)...