AI finds value in motorsport, multiplying limited computational fluid dynamics resources.
See full article...
See full article...
Sounds like they're really using this for final tuning and fast comparisons on data gained from models that've run through the traditional means already.“It sounds magical, but the reality is that the accuracy of the model is only guaranteed within specific range of situations that are not too far from what you have already explored,” Baqué told me. “So all the trick and the gap from the idea to the value is to find what are the right workflows, what kind of data do I need to generate to be able to explore what kind of configurations afterward in which type of setting, and how often do I need to retrain my model, all the data hygiene around the design workflow.”
Because it stays in the "multi-dimensional" interpolation domain. And I agree that it is a good use of it. There are plenty of good uses of AI. Just not all what is sold to us is a good use.Begore anyone starts complaining about AI using so much power
The reason folks use AI here is because it is orders of magnitude less computationally expensive for these usecases.
Meaning faster results and more iterations.
AI is not always the heavier option![]()
I noticed that and I have a question: Why? What's special about this use case and this type of "AI" that makes it so much less computationally expensive than other options?Begore anyone starts complaining about AI using so much power
The reason folks use AI here is because it is orders of magnitude less computationally expensive for these usecases.
Meaning faster results and more iterations.
AI is not always the heavier option![]()
The cost is front-loaded in training with this tech, then each end result is very cheap.I noticed that and I have a question: Why? What's special about this use case and this type of "AI" that makes it so much less computationally expensive than other options?
The other option is to iteratively evaluate the Navier-Stokes equations on every cell in a grid of a few million cells that describe the flow field around the car. Since the flow conditions in any one cell depend on the flow conditions in all the cells around it, you end up with millions of coupled equations and there's no way to solve that in one shot. You have to start with some assumed initial conditions, propagate that through the whole grid, then initialize again with those conditions, propagate that through the grid, repeat ad nauseum until you either (a) reach an iteration that has no significant difference from the previous iteration and is therefore probably close enough to what the real fluid flow would be, or (b) run out of compute credits and give up. The finer the grid, the more accurate the result, and the more computing power you need for each iteration.I noticed that and I have a question: Why? What's special about this use case and this type of "AI" that makes it so much less computationally expensive than other options?
The other option is to iteratively evaluate the Navier-Stokes equations on every cell in a grid of a few million cells that describe the flow field around the car. Since the flow conditions in any one cell depend on the flow conditions in all the cells around it, you end up with millions of coupled equations and there's no way to solve that in one shot. You have to start with some assumed initial conditions, propagate that through the whole grid, then initialize again with those conditions, propagate that through the grid, repeat ad nauseum until you either (a) reach an iteration that has no significant difference from the previous iteration and is therefore probably close enough to what the real fluid flow would be, or (b) run out of compute credits and give up. The finer the grid, the more accurate the result, and the more computing power you need for each iteration.
For bonus fun, you can have spinny bits (turbomachinery), hot bits (combustors), changes in fluid composition (combustors again), compressible / supersonic flow, mixed-phase flow, several kinds of heat transfer..... a CFD sim of the guts of an airplane engine will tie up a team of engineers and a high-end supercomputer for an absurd amount of time.
CFD is one of a handful of problem domains that will happily consume ALL the computing power you are willing to throw at them, for as long as payments drawn on your bank account will clear. Weather forecasting and seismic data analysis are two others.
If you can give an AI model a few such simulations and say "now take your best guess at what happens if we try a case that's right here in the middle of the parameter space between them" and it can give you a close-enough answer, you're creating significant value for the engineers who need to do this kind of work.
Using the model saves a lot of power for the constrained teams.Begore anyone starts complaining about AI using so much power
The reason folks use AI here is because it is orders of magnitude less computationally expensive for these usecases.
Meaning faster results and more iterations.
AI is not always the heavier option![]()
Dude, literally nobody is complaining about ML and other specialised and often quite helpful AI approaches using so much power. It's the generic useless LLMs that people complain about…Begore anyone starts complaining about AI using so much power
The reason folks use AI here is because it is orders of magnitude less computationally expensive for these usecases.
Meaning faster results and more iterations.
AI is not always the heavier option![]()
@MMarsh's answer is spot on for the specific details of what this is replacing.I noticed that and I have a question: Why? What's special about this use case and this type of "AI" that makes it so much less computationally expensive than other options?
Yes, agreed, this is the old school deep learning stuff from early on before they slapped chat bots into it and told it to start writing and answering questions. This is the stuff it excels at, the stuff we saw early reports on that looks so promising. It's repeatedly mutating and iterating towards a solution and then saving the results.This isn't the LLM/image generator/agentic bullshit that we mostly have a problem with though, right? This is more what we used to just call "machine learning" - training specialized models on specific datasets for a particular purpose.
I have no problem with that. That's not what's pumping the nonsense bubble, except a little tangentially via the tendency to lump everything together under the broad heading of "AI".
I have had some very tiresome arguments with LLM boosters who enjoy conflating these things to construct strawmen.
I can only find the sporting regulations from 2025, but appendix 7 lays out the regulations for aero testing restrictions. Section 4 is titled "Restricted CFD (RFCD) Simulations".Which bring me to the important point : Are the teams kinda cheating by outsourcing the hard computing cycles (training), so that they can run more sims under the cap?
For the avoidance of doubt, if any CFD simulation (other than the power unit simulation defined above) reveals information to a Competitor or to an Associate of the Competitor whether directly, via a contracted party or via an external entity working on behalf of a Competitor or for its own purposes and subsequently providing the results of its work to a Competitor, about flows that are gaseous on a F1 car then it is a RCFD simulation.
Gotta love Ars. You guys complain when the general public doesn't understand the nuances of AI, and then downvote a great question like this which serves as a perfect way to help explain some of those nuances to an interested party.I noticed that and I have a question: Why? What's special about this use case and this type of "AI" that makes it so much less computationally expensive than other options?
By that logic, could someone also feed a bunch if wind tunnel test data into the same or similar transformer model and do the same work?I can only find the sporting regulations from 2025, but appendix 7 lays out the regulations for aero testing restrictions. Section 4 is titled "Restricted CFD (RFCD) Simulations".
It starts by talking about what is and isn't an RFCD, then goes into the details of how computing time is actually defined and measured.
This bit at the beginning stands out (emphasis mine):
So, by definition, if it's not a CFD simulation, it can't be a restricted CFD, and so doesn't come under a team's limited CFD time.
Compute time for other tasks doesn't appear to be covered, and since taking the results of existing CFDs and then training models on them doesn't incur another CFD run, that would appear to be fair game. Training the models is undoubtedly very compute-intensive, but there's nothing in the (2025) rules that bans that.
So, typical behaviour in F1 - if the rules don't say you can't do that, then you damn well investigate whether doing that will result in an advantage. It might be this will be something that future rulesets will clamp down on, as it's probably going to become very widespread.
“It sounds magical, but the reality is that the accuracy of the model is only guaranteed within a specific range of situations that are not too far from what you have already explored,” Baqué told me. “So all the trick and the gap from the idea to the value is to find what are the right workflows, what kind of data do I need to generate to be able to explore what kind of configurations afterward in which type of setting, and how often do I need to retrain my model, all the data hygiene around the design workflow.”
I think you gain the ability to try many permutations quickly, and importantly, you can optimize very fast once you are "in the ball park". The use case is limited for the reasons you highlight, but making these optimizations quickly are apparently valuable to this industry.This is an interesting approach but in practice only works well when trying many small variations within an otherwise unchanged problem.
You need to generate a significant training set specific to the problem setup (which you do by running a bunch of actual CFD sims) and do the actual training. If you don’t plan on running significantly more sims than the size of the training set then you are actually much slower than just doing the CFD. Curiously, the paper does not report the time spent generating the dataset or on the training.
This is an issue because it’s really difficult to generalize the CFD outside of the training set geometries. They point to a slightly larger wing angle going outside the bounds of the data as being a big achievement here which is telling. A completely different body type would be a no go. If most of the wing angles are in the training set already, what are you gaining?
But that's the entire point of loss-minimization gradient descent. When you have an incredibly large search space (massive number of dimensions and coefficients), estimation can be valuable to help you concentrate on "neighborhoods" of likely returns, rather than searching the entire space sequentially (or stochastically). On some level this was also what genetic algorithms gave us, but with far less impact and far more responsibility to encode solution sub-spaces to hybridize against. NNs are all empirical, but that doesn't inherently make their results less actionable.I am anti-LLM, but separately I've been anti-ML for less moral and more scientific reasons. It's because of stuff like this:
…
That is not true with these AI interpolations. These are all ad hoc and empirical, with no theoretical basis. And it's probably true that for most cases its good enough, especially if you're going to validate it against a wind tunnel. But you're always losing accuracy when you take these shortcuts, and you'll never know how wrong you are unless you do it properly.
The critical meaning of that bit at the beginning very much depends on how "CFD simulation" is defined though, so you've kind of skipped the important part there.It starts by talking about what is and isn't an RFCD, then goes into the details of how computing time is actually defined and measured.
This bit at the beginning stands out (emphasis mine):
Yeah but... see above. It could be characterised as CFD. It's just not approved COTS CFD software.So, by definition, if it's not a CFD simulation, it can't be a restricted CFD, and so doesn't come under a team's limited CFD time.
Solver refers to the program or programs that compute the solution of the equations describing the flow including any extension of the simulation or simulations involving additional numerical computation (for example but not limited, to adjoint computation).
The drafters would have been fearful of accidentally banning spreadsheets I think, but as I said above, it's nowhere near as narrow as they could have made it.Compute time for other tasks doesn't appear to be covered
I'm not sure I'm actually suggesting "simulations involving additional numerical computation (for example but not limited, to adjoint computation)" blocks this but I'm definitely highlighting that this clause exists and is up for discussion..., and since taking the results of existing CFDs and then training models on them doesn't incur another CFD run, that would appear to be fair game.
Yup. And of course we don't know what this year's rules say anyway, but good sleuthing on last year's.So, typical behaviour in F1 - if the rules don't say you can't do that, then you damn well investigate whether doing that will result in an advantage. It might be this will be something that future rulesets will clamp down on, as it's probably going to become very widespread.
Next, you're gonna tell me that drinking green tea and cleansing my chakras isn't as scientific and precise as getting an mRNA vaccine, or that my snowball doesn't disprove climate change!I am anti-LLM, but separately I've been anti-ML for less moral and more scientific reasons. It's because of stuff like this:
The reason classical computational fluid dynamics works for almost any non-quantum fluid, even complex non-Newtonian fluids, is that the Navier-Stokes equation being solved is a theoretically robust model that truly describes the motion; it is very complicated mathematically but physically it is just F=dp/dt applied to fluid (not quite F=ma since "m" is changing).
That is not true with these AI interpolations. These are all ad hoc and empirical, with no theoretical basis. And it's probably true that for most cases its good enough, especially if you're going to validate it against a wind tunnel. But you're always losing accuracy when you take these shortcuts, and you'll never know how wrong you are unless you do it properly.
I’ve written about yacht racing a couple of times but like motorbikes it’s pretty much outside my wheelhouse.@Dr Gitlin, have you ever considered grand prix level sailing? Sail GP and Americas Cup seem to be right up your alley. The nerdom of the technical aspects of the sport are F1 level, completely at the bleeding edge of speed technology with all of the technical rules weaving through the craft like a motorsport, without the motor but with more spectacular crashes. And if you decide to burn a few hours trying to understand it all, like F1 there are forums of former racers, builders, designers, and too pedantic nerds diving deep into the details.
Good article, thank you as always!
People are really glossing over the most damning point here: "the reality is that the accuracy of the model is only guaranteed within a specific range of situations that are not too far from what you have already explored,"This is an interesting approach but in practice only works well when trying many small variations within an otherwise unchanged problem.
You need to generate a significant training set specific to the problem setup (which you do by running a bunch of actual CFD sims) and do the actual training. If you don’t plan on running significantly more sims than the size of the training set then you are actually much slower than just doing the CFD. Curiously, the paper does not report the time spent generating the dataset or on the training.
This is an issue because it’s really difficult to generalize the CFD outside of the training set geometries. They point to a slightly larger wing angle going outside the bounds of the data as being a big achievement here which is telling. A completely different body type would be a no go. If most of the wing angles are in the training set already, what are you gaining?
I just want to show my appreciation for the term "data hygiene", as a former medical research assistant, science teacher, and skeptic of all statistics (especially those used in advertising). Analysis is a waste of time without clean data.
[edit: That is, I like the implication that unclean data is kind of gross and embarrassing.]
hygiene and similar words (clean, dirty, etc) a great deal across the discipline.clean working directory. Sometimes even pristine! Similarly, ugly code that technically works but is unsatisfactory aesthetically may be called a dirty hacks. And then there's code smells as a whole category of indicators of underlying problems with quality, architecture, etc.