The cynists will comment that I've just been sucked in by the PR. However, I know this team and have been using these techniques for other problems. I know they are so close to a computationally-assisted proof of counterexample that it is virtually inevitable at this point. If they don't do it, I'm pretty sure I could take a handful of people and a few years and do it myself. Mostly a lot of interval arithmetic with a final application of Schauder that remains; tedious and time-consuming, but not overly challenging compared to the parts already done.
This is not just PR and is very interesting. However, in my view, (and from a quick read of the paper) this is actually a very classical method in applied math work:
- Build a complex intractable mathematical model (here, Navier-Stokes)
- Approximate it with a function approximator (here, a Physics Informed Neural Network)
- Use the some property of function approximator to search for more solutions to the original model (here, using Gauss-Newton)
In a sense, this is actually just the process of model-based science anyway: use a model for the physical world and exploit the mathematics of the model for real-world effects.
This is very very good work, but this heritage goes back to polynomial approximation even from Taylor series, and has been the foundation of engineering for literal centuries. Throughout history, the approximator keeps getting better and better and hungrier and hungrier for data (Taylor series, Chebyshev + other orthogonal bases for polynomials, neural networks, RNNs, LSTMs, PINNs, <the future>).
You didn't say anything to the contrary, and neither did the original video, but it's very different than what some other people are talking about in this thread ("run an LLM in a loop to do science the way a person does it"). Maybe I'm just ranting at the overloading of the term AI to mean "anything on a GPU".
This is absolutely true, but it still makes use of the advantages and biases of neural networks in a clever way. It has to, because computationally-assisted proofs for PDEs with singularities is incredibly difficult. To me, this is not too similar from using them as heuristics to find counterexamples, or other approaches where the implicit biases pay off. I think we do ourselves a disservice to say that "LLMs replacing people" = "applications of AI in science".
I also wouldn't say this is entirely "classical". Old, yes, but still unfamiliar and controversial to a surprising number of people. But I get your point :-).
> I know they are so close to a computationally-assisted proof of counterexample that it is virtually inevitable at this point.
That's a strong claim. Is it based on more than the linked work on some model problems from fluid mechanics?
I will say that I dread the discourse if it works out, since I don't believe enough people will understand that using a PINN to get new solutions of differential equations has substantially no similarity to asking ChatGPT (or AlphaProof etc) for a proof of a conjecture. And there'll be a lot of people trying to hide the difference.
It's based on knowledge of the related estimates, applying similar techniques to geometric problems, knowledge of all the prior works that lead to the current work, and speaking with members of the team themselves. They are much further along than it appears at first glance. All of the major bottlenecks have fallen; the only concern was whether double precision accuracy is good enough. The team seems to have estimates that are strong enough for this, but obviously keep them close to their chest.
PINNs are different in concept, yes, but clearly no less important, so the additional attention will be appreciated. Asking LLMs for proofs is a different vein of research, often involving Lean. It is much further behind, but still making ground.
> PINNs are different in concept, yes, but clearly no less important
If anything I think they're more important! Whether or not it works out for Navier-Stokes, this kind of thing is an extremely plausible avenue of approach and could yield interesting singularities for other major equations. I am however extremely concerned about public understanding. I know you are well aware that this is worlds away from the speculative technologies like 'mathematical superintelligence' but, if it works out, it'll be like a nuclear bomb of misinformation about AI and math.
They are building a formally-defined counter example to this? Am I understanding correctly?
> In three space dimensions and time, given an initial velocity field, there exists a vector velocity and a scalar pressure field, which are both smooth and globally defined, that solve the Navier–Stokes equations.
I think the PR is making it seem that Deepmind is not standing on the shoulder of giants, when in fact it very much is. The paper itself makes this clear. I wish them luck!
> The cynists will comment that I've just been sucked in by the PR
You can just ignore them. I see a lot of science-literate folks try to meet the anti-science folks as if they're on equal footing and it's almost always a waste of time. Imagine if every time you talked about biology you had to try to address the young earth creationists in the room and try to pre-rebut their concerns.
I'll share something as a former solar researcher.
Scientific progress is heavily influenced by how many bodies you can throw at a problem.
The more experiments you can run, with more variety and angles the more data you can get, the higher the likelihood of a breakthrough.
Several huge scientist are famous not because they are geniuses, but because they are great fundraisers and can have 20/30/50 bodies to throw at problems every year.
This is true in virtually any experimental field.
If LLMs can be de facto another body then scientific progress is going to sky rocket.
Robots also tend to be more precise than humans and could possibly lead to better replication.
But given that LLMs cannot interact with the real world I don't see that happening anytime soon.
> But given that LLMs cannot interact with the real world
What type of interaction do you envision? Could a non-domain-expert, but somewhat trained person provide a bridge? If the LLM comes up with the big ideas and tells a human technical assistant to execute (put the vial here, run the 3D printer with this file, put the object there, drive in a screw), would that help? But dexterous robots are getting more and more advanced, see CoRL demos right now.
Someone needs to evaluate the big ideas spat out by the llm is the big issue. Lab work can already be automated. And bs holders are even cheaper than an automated machine.
In my computational niche the bottleneck was always writing up the results :) And wow does AI help there.... It's not hard to get a decent first draft written by AI based on my existing results.
Computational can be a huge bottleneck. Some steps they really do take dozens of hours to run on cluster. And you are not the main character, others might be using the cluster and your jobs might be waiting in a queue. You might not be able to appreciate parameters need adjustment until the run is over and you evaluate output.
Another delay point is getting collaborators schedules to align for meetings on progress or potential directions.
Placing the results in context takes some time but not so much as you might guess if you are constantly reading and writing sourced paragraphs and skeleton papers needing only results plopped in when they are ready and some exposition in the discussion section.
Writing the code might be the fastest step in the process already.
I am reposting something along the lines of a flagged and dead comment: This would be lend more credibility to the premise AI is revolutionizing scientific discovery if it came from someone who's Nobel (or work in general) were in a non-AI-centered domain. This is not a critique of his speech or points, but I think the lead implied by the (especially Youtube) title would hit harder if it came from someone whose work wasn't AI-centered.
Jumper's work is the poster child of AI success in science; this isn't about a new domain being revolutionized by it.
I will throw out an idea I've been thinking about recently about a far less ambitious idea, but related: Amber (MD package) provides Force Field names and partial charges for a number of small organic molecules in their GeoStd set. I believe these come from its Antechamber program. Would it be possible to infer useful FF name and Partial charge for arbitrary organic molecules using AI instead, trained on the GeoStd set data?
> This would be lend more credibility to the premise AI is revolutionizing scientific discovery if it came from someone who's Nobel (or work in general) were in a non-AI-centered domain.
No it wouldn't. I've seen anti-AI people try to make this sort of argument repeatedly and it doesn't make any sense.
It's an attempt to smuggle in an ad-hominem. It's relying on the fact that people who hate AI also hate people who work in AI.
Hey! I think I misused words, or was otherwise unclear. Sorry about that. I'm not Anti-AI and don't hate AI or people who work on it. I don't mean to conflate my post with someone who is Anti-AI. No ad-hominem intended.
Hey, sorry I don't mean to imply any of those things about you.
Before commenting I did check out your profile and saw that you weren't AI hater. My intention was the opposite -- I was trying to point out that some arguments don't survive on logic, but on emotional appeal. When enough smart people repeat something, it can slip past our skepticism. But of course my comment was terse and you had no way of knowing that.
In my view, among all people who make claims about what is on the forefront of science or knowledge, the most credible claims come from people whose research is on the forefront of science or knowledge. That's why the claims of experts are considered more credible in their fields than non-experts, even when the non-experts are experts in nearby fields.
So, applying that general rule to this particular case, we would expect the people to most credibly talk about the application of AI to science to be people who know the most about the application of AI to science.
There are other scientists whom I really respect but whose opinions on this particular topic I would find less weighty. For example, Terry Tao is a brilliant mathematician and has consulted on using AI to do research level math. But he was just a few months ago figuring out how to set up ChatGPT with VS Code, so I wouldn't expect him to be the most up to date on how AI is impacting science.
On the other hand, a Nobel laureate who has shown an aptitude for making important scientific discoveries is exactly the sort of person I believe would be able to talk knowledgeably about how to carve up science problems and about how to apply the current generation of AI to solve them. Especially if they've seen the internal world of a company that has a top tier foundational model and a track record for making scientific discoveries. Because in that case they see how science is being done if you have unlimited funds and access to some of the smartest scientists on the planet.
In contrast, I would be much more skeptical of a similar claim made by a company like OpenAI or Microsoft that doesn't have the same track record of producing new science.
For those reasons I don't think it's true that the claim would be more credible from someone with more distance (and hence less expertise). And I think similar claims made in other contexts would strike most people as bizarre. For example, if I said that claims about medicine are more credible if they come from people who aren't medical doctors.
It is not at all an ad hominem. Disclosure of interests and conflicts for interest were assumed to be declared in the open even a decade ago.
Carter sold his peanut farm to avoid conflicts of interest, Trump launched a coin pump & dump on his first day.
If a Nobel Prize winner works for a corporation, that should be disclosed (the original title contained "Nobel Prize Laureate" instead of "DeepMind Director").
But I suppose that in the current age where everyone just wants to get rich these courtesies no longer matter.
I would argue that we need an effective alternative to benchmarks entirely given how hard they are to obtain in scientific disciplines. Classical statistics has gone very far by getting a lot out of limited datasets, and train-test splits are absolutely unnecessary there.
I kind of dislike the benchmarkification of AI for science stuff tbh. I've encountered a LOT of issues with benchmark datasets that just aren't good...
In a lot of cases they are fine and necessary, but IMO, the standard for legit "success" in a lot of ML for science applications should basically be "can this model be used to make real scientific or engineering insights, that would have been very difficult and/or impossible without the proposed idea."
Even if this is a super high bar, I think more papers in ML for science should strive to be truly interdisciplinary and include an actual science advancement... Not just "we modify X and get some improvement on a benchmark dataset that may or may not be representative of the problems scientists could actually encounter." The ultimate goal of "ml for science" is science, not really to improve ML methods imo
As someone who works in the field, it really doesn't feel like more money (proportionally speaking) is going to this. A little bit is done here and there for PR. The number that are working on net positive applications for AI is still shockingly low compared to everything else.
I have a mildly psychotic friend who think that he uncovered the secrets to everything with AI. Quantum theory and Jungian archetypes, together with 4 dimensions - great mix
First jump that computers gave us : speed. With excess of speed came the ability to brute force many problems.
Next jump given by AI (not LLMs specifically, I mean “machine learned systems” in general) is navigation. Even with large amounts of speed some problems are still impractically large, we are using AI to better explore that space, by navigating it smarter, rather than just speeding through it combinatorially.
Not published just yet are experiments for finding solutions to mathematical problems traditionally found with SAT solvers, at much larger scale than was previously possible.
Um, okay? Isn't that how most optimisation involving AI is supposed to go?
Perhaps some much needed context. Mathematicians are not stupid; we are very much aware of all the existing forms of genetic optimisation algorithms, cross entropy method, etc. Nothing works on these problems, or at least not well at scale. As I said, state of the art for many of these was SAT-related. The problem is that the heuristic used for exploring new solutions always required very careful consideration as the naive ones rarely worked well.
Here, the transformer is proving effective at searching for good heuristics, far more so than any other existing technique. In this sense, it is achieving far, far. far better performance in optimisation than previous approaches. That is a breakthrough, at least for us mathematicians. If this doesn't constitute improvements in optimisation, I don't know what does.
Saying it's "just meta heuristic relying on local search" is akin to saying these tasks are "just optimisation". If it's so procedural, why weren't we making ground on these things before?
Also, by the way, a :facepalm: is not exactly the pinnacle of academic rebuttal, no matter how wrong I could have been.
It’s just that the paper cited is no different than any other paper in the meta-heuristic community.
Some idea for guiding the local search. Some limited sample results. No promises on bounds or generalizability of the method.
If this is ground breaking, then every legitimate meta heuristic paper in the past 50 years was also ground breaking.
I will change my mind if I see a wide set of benchmark results where it consistently beats or is even head-to-head with the SoTA. Then we would know that we have a game changer.
So for reference, I am a mathematician, I work in probability theory. I don't work in optimisation theory, but saw a talk a few days ago by a prominent member in that field ranting about the obsession with pointless bounds when industry does not care, Gurobi team does not care, and for OR, NN-centric heuristics coupled with massive GPU compute are eating the lunch of academia. But I digress.
The key difference is that this one approach (and the improvements that follow up on it) seems to be generating counterexamples and sharper estimates at such a rapid speed that they simply can't publish them fast enough. It almost seems pointless to do so given how easy they now are to find. Mathematicians are not stupid, we know about meta-heuristics. It takes forever to get a single result with existing techniques, so long that each result is often its own paper. Getting everything in the PatternBoost paper took a few months, as I understand. Local search is problem-specific, but the point is that the NN is doing a lot of heavy-lifting so the choice of local search is not as critical.
Here is a frivolous example mentioned in the paper itself: largest subset of a 2D lattice with no embedded isosceles triangle. For larger lattices (I don't remember the exact size under consideration), SoTA are SAT solvers which work up to n=32. Previous meta-heuristics cannot even do this, of course, you can try if you don't believe me. PatternBoost has just recently gotten to ~96% of the expected size for n=64 and is still improving by the day. Once it reaches 100%, there are techniques to show local optimality. Does this work as a benchmark? There are plenty more in there and unpublished.
For more serious cases that are difficult to explain here, the group in Sydney have counterexamples to hundreds of graph-theory and rep. theory conjectures that have stood for many decades. I also disagree on the "no different" aspect; Geordie Williamson is a very strong mathematician, and does not tend to jump on trivial things. He is very receptive to discussion on these matters, so you can ask him yourself how this is actually a game-changer downstream.
Yes, it is a meta-heuristic. But almost all meta-heuristics have been useless so far for these problems. This one is not, and for the people downstream, that's really all that matters.
Google DeepMind Director John Jumper. Literally no one who is not connected to the "AI" industrial complex praises "AI". In any video or blog post there is a link.
Something else to add is mathematical discovery. There is a team that is very close to solving the Navier-Stokes Millenium Prize problem: https://deepmind.google/discover/blog/discovering-new-soluti...
The cynists will comment that I've just been sucked in by the PR. However, I know this team and have been using these techniques for other problems. I know they are so close to a computationally-assisted proof of counterexample that it is virtually inevitable at this point. If they don't do it, I'm pretty sure I could take a handful of people and a few years and do it myself. Mostly a lot of interval arithmetic with a final application of Schauder that remains; tedious and time-consuming, but not overly challenging compared to the parts already done.
This is not just PR and is very interesting. However, in my view, (and from a quick read of the paper) this is actually a very classical method in applied math work:
- Build a complex intractable mathematical model (here, Navier-Stokes)
- Approximate it with a function approximator (here, a Physics Informed Neural Network)
- Use the some property of function approximator to search for more solutions to the original model (here, using Gauss-Newton)
In a sense, this is actually just the process of model-based science anyway: use a model for the physical world and exploit the mathematics of the model for real-world effects.
This is very very good work, but this heritage goes back to polynomial approximation even from Taylor series, and has been the foundation of engineering for literal centuries. Throughout history, the approximator keeps getting better and better and hungrier and hungrier for data (Taylor series, Chebyshev + other orthogonal bases for polynomials, neural networks, RNNs, LSTMs, PINNs, <the future>).
You didn't say anything to the contrary, and neither did the original video, but it's very different than what some other people are talking about in this thread ("run an LLM in a loop to do science the way a person does it"). Maybe I'm just ranting at the overloading of the term AI to mean "anything on a GPU".
This is absolutely true, but it still makes use of the advantages and biases of neural networks in a clever way. It has to, because computationally-assisted proofs for PDEs with singularities is incredibly difficult. To me, this is not too similar from using them as heuristics to find counterexamples, or other approaches where the implicit biases pay off. I think we do ourselves a disservice to say that "LLMs replacing people" = "applications of AI in science".
I also wouldn't say this is entirely "classical". Old, yes, but still unfamiliar and controversial to a surprising number of people. But I get your point :-).
> I know they are so close to a computationally-assisted proof of counterexample that it is virtually inevitable at this point.
That's a strong claim. Is it based on more than the linked work on some model problems from fluid mechanics?
I will say that I dread the discourse if it works out, since I don't believe enough people will understand that using a PINN to get new solutions of differential equations has substantially no similarity to asking ChatGPT (or AlphaProof etc) for a proof of a conjecture. And there'll be a lot of people trying to hide the difference.
It's based on knowledge of the related estimates, applying similar techniques to geometric problems, knowledge of all the prior works that lead to the current work, and speaking with members of the team themselves. They are much further along than it appears at first glance. All of the major bottlenecks have fallen; the only concern was whether double precision accuracy is good enough. The team seems to have estimates that are strong enough for this, but obviously keep them close to their chest.
PINNs are different in concept, yes, but clearly no less important, so the additional attention will be appreciated. Asking LLMs for proofs is a different vein of research, often involving Lean. It is much further behind, but still making ground.
> PINNs are different in concept, yes, but clearly no less important
If anything I think they're more important! Whether or not it works out for Navier-Stokes, this kind of thing is an extremely plausible avenue of approach and could yield interesting singularities for other major equations. I am however extremely concerned about public understanding. I know you are well aware that this is worlds away from the speculative technologies like 'mathematical superintelligence' but, if it works out, it'll be like a nuclear bomb of misinformation about AI and math.
They are building a formally-defined counter example to this? Am I understanding correctly?
> In three space dimensions and time, given an initial velocity field, there exists a vector velocity and a scalar pressure field, which are both smooth and globally defined, that solve the Navier–Stokes equations.
I think the PR is making it seem that Deepmind is not standing on the shoulder of giants, when in fact it very much is. The paper itself makes this clear. I wish them luck!
> The cynists will comment that I've just been sucked in by the PR
You can just ignore them. I see a lot of science-literate folks try to meet the anti-science folks as if they're on equal footing and it's almost always a waste of time. Imagine if every time you talked about biology you had to try to address the young earth creationists in the room and try to pre-rebut their concerns.
I'll share something as a former solar researcher.
Scientific progress is heavily influenced by how many bodies you can throw at a problem.
The more experiments you can run, with more variety and angles the more data you can get, the higher the likelihood of a breakthrough.
Several huge scientist are famous not because they are geniuses, but because they are great fundraisers and can have 20/30/50 bodies to throw at problems every year.
This is true in virtually any experimental field.
If LLMs can be de facto another body then scientific progress is going to sky rocket.
Robots also tend to be more precise than humans and could possibly lead to better replication.
But given that LLMs cannot interact with the real world I don't see that happening anytime soon.
> But given that LLMs cannot interact with the real world
Pair LLMs with machines and robotics and you are getting closer
> But given that LLMs cannot interact with the real world
What type of interaction do you envision? Could a non-domain-expert, but somewhat trained person provide a bridge? If the LLM comes up with the big ideas and tells a human technical assistant to execute (put the vial here, run the 3D printer with this file, put the object there, drive in a screw), would that help? But dexterous robots are getting more and more advanced, see CoRL demos right now.
Someone needs to evaluate the big ideas spat out by the llm is the big issue. Lab work can already be automated. And bs holders are even cheaper than an automated machine.
Can these robots move a chess piece from one square to another?
I agree that there is power in numbers for science, but not all science is lab work. Sometimes the bottleneck is purely computational.
In my computational niche the bottleneck was always writing up the results :) And wow does AI help there.... It's not hard to get a decent first draft written by AI based on my existing results.
Computational can be a huge bottleneck. Some steps they really do take dozens of hours to run on cluster. And you are not the main character, others might be using the cluster and your jobs might be waiting in a queue. You might not be able to appreciate parameters need adjustment until the run is over and you evaluate output.
Another delay point is getting collaborators schedules to align for meetings on progress or potential directions.
Placing the results in context takes some time but not so much as you might guess if you are constantly reading and writing sourced paragraphs and skeleton papers needing only results plopped in when they are ready and some exposition in the discussion section.
Writing the code might be the fastest step in the process already.
I liked this proof of concept:
https://arxiv.org/abs/2509.06503
They set up scoreable computational science problems and do search over solutions.
I am reposting something along the lines of a flagged and dead comment: This would be lend more credibility to the premise AI is revolutionizing scientific discovery if it came from someone who's Nobel (or work in general) were in a non-AI-centered domain. This is not a critique of his speech or points, but I think the lead implied by the (especially Youtube) title would hit harder if it came from someone whose work wasn't AI-centered.
Jumper's work is the poster child of AI success in science; this isn't about a new domain being revolutionized by it.
I will throw out an idea I've been thinking about recently about a far less ambitious idea, but related: Amber (MD package) provides Force Field names and partial charges for a number of small organic molecules in their GeoStd set. I believe these come from its Antechamber program. Would it be possible to infer useful FF name and Partial charge for arbitrary organic molecules using AI instead, trained on the GeoStd set data?
> This would be lend more credibility to the premise AI is revolutionizing scientific discovery if it came from someone who's Nobel (or work in general) were in a non-AI-centered domain.
No it wouldn't. I've seen anti-AI people try to make this sort of argument repeatedly and it doesn't make any sense.
It's an attempt to smuggle in an ad-hominem. It's relying on the fact that people who hate AI also hate people who work in AI.
Hey! I think I misused words, or was otherwise unclear. Sorry about that. I'm not Anti-AI and don't hate AI or people who work on it. I don't mean to conflate my post with someone who is Anti-AI. No ad-hominem intended.
Hey, sorry I don't mean to imply any of those things about you.
Before commenting I did check out your profile and saw that you weren't AI hater. My intention was the opposite -- I was trying to point out that some arguments don't survive on logic, but on emotional appeal. When enough smart people repeat something, it can slip past our skepticism. But of course my comment was terse and you had no way of knowing that.
In my view, among all people who make claims about what is on the forefront of science or knowledge, the most credible claims come from people whose research is on the forefront of science or knowledge. That's why the claims of experts are considered more credible in their fields than non-experts, even when the non-experts are experts in nearby fields.
So, applying that general rule to this particular case, we would expect the people to most credibly talk about the application of AI to science to be people who know the most about the application of AI to science.
There are other scientists whom I really respect but whose opinions on this particular topic I would find less weighty. For example, Terry Tao is a brilliant mathematician and has consulted on using AI to do research level math. But he was just a few months ago figuring out how to set up ChatGPT with VS Code, so I wouldn't expect him to be the most up to date on how AI is impacting science.
On the other hand, a Nobel laureate who has shown an aptitude for making important scientific discoveries is exactly the sort of person I believe would be able to talk knowledgeably about how to carve up science problems and about how to apply the current generation of AI to solve them. Especially if they've seen the internal world of a company that has a top tier foundational model and a track record for making scientific discoveries. Because in that case they see how science is being done if you have unlimited funds and access to some of the smartest scientists on the planet.
In contrast, I would be much more skeptical of a similar claim made by a company like OpenAI or Microsoft that doesn't have the same track record of producing new science.
For those reasons I don't think it's true that the claim would be more credible from someone with more distance (and hence less expertise). And I think similar claims made in other contexts would strike most people as bizarre. For example, if I said that claims about medicine are more credible if they come from people who aren't medical doctors.
It is not at all an ad hominem. Disclosure of interests and conflicts for interest were assumed to be declared in the open even a decade ago.
Carter sold his peanut farm to avoid conflicts of interest, Trump launched a coin pump & dump on his first day.
If a Nobel Prize winner works for a corporation, that should be disclosed (the original title contained "Nobel Prize Laureate" instead of "DeepMind Director").
But I suppose that in the current age where everyone just wants to get rich these courtesies no longer matter.
You are making a completely different point than the person I'm responding to. An equally bad point, but a different point nonetheless.
His title is in the video thumbnail.
> But I suppose that in the current age where everyone just wants to get rich these courtesies no longer matter.
Surely you agree this is a veiled ad hominem
Awful title, great video.
Three points jumped out
1) "really when you look at these machine learning breakthroughs they're probably fewer people than you imagine"
In a world of idiots, few people can do great things.
2) External benchmarks forced people upstream to improve
We need more of these.
3) "the third of these ingredients research was worth a hundredfold of the first of these ingredients data."
Available data is 0 for most things.
> We need more of these.
> Available data is 0 for most things.
I would argue that we need an effective alternative to benchmarks entirely given how hard they are to obtain in scientific disciplines. Classical statistics has gone very far by getting a lot out of limited datasets, and train-test splits are absolutely unnecessary there.
I kind of dislike the benchmarkification of AI for science stuff tbh. I've encountered a LOT of issues with benchmark datasets that just aren't good... In a lot of cases they are fine and necessary, but IMO, the standard for legit "success" in a lot of ML for science applications should basically be "can this model be used to make real scientific or engineering insights, that would have been very difficult and/or impossible without the proposed idea."
Even if this is a super high bar, I think more papers in ML for science should strive to be truly interdisciplinary and include an actual science advancement... Not just "we modify X and get some improvement on a benchmark dataset that may or may not be representative of the problems scientists could actually encounter." The ultimate goal of "ml for science" is science, not really to improve ML methods imo
NVIDIA published the Illustrated Evo2 a few days ago, walking through the architecture of their genetics foundation model:
https://research.nvidia.com/labs/dbr/blog/illustrated-evo2/
It's nice to see more and more labs using ai for drug discovery, something truly net positive for society.
As someone who works in the field, it really doesn't feel like more money (proportionally speaking) is going to this. A little bit is done here and there for PR. The number that are working on net positive applications for AI is still shockingly low compared to everything else.
I have a mildly psychotic friend who think that he uncovered the secrets to everything with AI. Quantum theory and Jungian archetypes, together with 4 dimensions - great mix
I see this sort of work as a natural extension of Combinatorial Chemistry or bootstrapping and Monte Carlo methods in stats.
https://en.wikipedia.org/wiki/Combinatorial_chemistry
First jump that computers gave us : speed. With excess of speed came the ability to brute force many problems.
Next jump given by AI (not LLMs specifically, I mean “machine learned systems” in general) is navigation. Even with large amounts of speed some problems are still impractically large, we are using AI to better explore that space, by navigating it smarter, rather than just speeding through it combinatorially.
No evidence so far that "AI" has improved our general optimization capabilities. At all.
Still at the top of the benchmarks of integer optimization by huge margin are the traditional usual suspects. Same in constraint programming and SAT.
Here is some evidence for you then: https://arxiv.org/abs/2411.00566
Not published just yet are experiments for finding solutions to mathematical problems traditionally found with SAT solvers, at much larger scale than was previously possible.
This is just meta heuristic relying on local search :facepalm:
You could call it artificial ant colony optimization.
People come up with such ideas all the time. Sorry, but nothing groundbreaking here.
Um, okay? Isn't that how most optimisation involving AI is supposed to go?
Perhaps some much needed context. Mathematicians are not stupid; we are very much aware of all the existing forms of genetic optimisation algorithms, cross entropy method, etc. Nothing works on these problems, or at least not well at scale. As I said, state of the art for many of these was SAT-related. The problem is that the heuristic used for exploring new solutions always required very careful consideration as the naive ones rarely worked well.
Here, the transformer is proving effective at searching for good heuristics, far more so than any other existing technique. In this sense, it is achieving far, far. far better performance in optimisation than previous approaches. That is a breakthrough, at least for us mathematicians. If this doesn't constitute improvements in optimisation, I don't know what does.
Saying it's "just meta heuristic relying on local search" is akin to saying these tasks are "just optimisation". If it's so procedural, why weren't we making ground on these things before?
Also, by the way, a :facepalm: is not exactly the pinnacle of academic rebuttal, no matter how wrong I could have been.
Apologies I didn’t mean to be cocky / dismissive.
It’s just that the paper cited is no different than any other paper in the meta-heuristic community.
Some idea for guiding the local search. Some limited sample results. No promises on bounds or generalizability of the method.
If this is ground breaking, then every legitimate meta heuristic paper in the past 50 years was also ground breaking.
I will change my mind if I see a wide set of benchmark results where it consistently beats or is even head-to-head with the SoTA. Then we would know that we have a game changer.
So for reference, I am a mathematician, I work in probability theory. I don't work in optimisation theory, but saw a talk a few days ago by a prominent member in that field ranting about the obsession with pointless bounds when industry does not care, Gurobi team does not care, and for OR, NN-centric heuristics coupled with massive GPU compute are eating the lunch of academia. But I digress.
The key difference is that this one approach (and the improvements that follow up on it) seems to be generating counterexamples and sharper estimates at such a rapid speed that they simply can't publish them fast enough. It almost seems pointless to do so given how easy they now are to find. Mathematicians are not stupid, we know about meta-heuristics. It takes forever to get a single result with existing techniques, so long that each result is often its own paper. Getting everything in the PatternBoost paper took a few months, as I understand. Local search is problem-specific, but the point is that the NN is doing a lot of heavy-lifting so the choice of local search is not as critical.
Here is a frivolous example mentioned in the paper itself: largest subset of a 2D lattice with no embedded isosceles triangle. For larger lattices (I don't remember the exact size under consideration), SoTA are SAT solvers which work up to n=32. Previous meta-heuristics cannot even do this, of course, you can try if you don't believe me. PatternBoost has just recently gotten to ~96% of the expected size for n=64 and is still improving by the day. Once it reaches 100%, there are techniques to show local optimality. Does this work as a benchmark? There are plenty more in there and unpublished.
For more serious cases that are difficult to explain here, the group in Sydney have counterexamples to hundreds of graph-theory and rep. theory conjectures that have stood for many decades. I also disagree on the "no different" aspect; Geordie Williamson is a very strong mathematician, and does not tend to jump on trivial things. He is very receptive to discussion on these matters, so you can ask him yourself how this is actually a game-changer downstream.
Yes, it is a meta-heuristic. But almost all meta-heuristics have been useless so far for these problems. This one is not, and for the people downstream, that's really all that matters.
If you only know how to use a hammer, everything looks like a nail.
Combinatorics will always be a (tough) nail regardless of what a random HN commentator thinks.
Bring any tool you wish, but the problem is very well defined and very real.
Google DeepMind Director John Jumper. Literally no one who is not connected to the "AI" industrial complex praises "AI". In any video or blog post there is a link.
> Literally no one who is not connected to the "AI" industrial complex praises "AI".
Why lying when one can easily find examples of exactly that?
1) https://www.ecmwf.int/en/about/media-centre/news/2025/ecmwfs...
2) https://pmc.ncbi.nlm.nih.gov/articles/PMC11510778/
But I’m sure you could find the connections to the AI industry complex somehow.
Thank you for this context, should be the title IMO..
In the same vein, any link to Gary Marcus’s blog posts should be labeled “AI hater”, don’t you think?
I'm a researcher in AI and I haven't met anyone who has gotten substantial help from AI.
Many people have tried, many people have been let down.
in the scientific domain?
Follow the money…