The headline may make it seem like AI just discovered some new result in physics all on its own, but reading the post, humans started off trying to solve some problem, it got complex, GPT simplified it and found a solution with the simpler representation. It took 12 hours for GPT pro to do this. In my experience LLM’s can make new things when they are some linear combination of existing things but I haven’t been to get them to do something totally out of distribution yet from first principles.
Humans have worked out the amplitudes for integer n up to n = 6 by hand, obtaining very complicated expressions, which correspond to a “Feynman diagram expansion” whose complexity grows superexponentially in n. But no one has been able to greatly reduce the complexity of these expressions, providing much simpler forms. And from these base cases, no one was then able to spot a pattern and posit a formula valid for all n. GPT did that.
Basically, they used GPT to refactor a formula and then generalize it for all n. Then verified it themselves.
> I think this was all already figured out in 1986 though
They cite that paper in the third paragraph...
Naively, the n-gluon scattering amplitude involves order n! terms. Famously, for the special case of MHV (maximally helicity violating) tree amplitudes, Parke and Taylor [11] gave a simple and beautiful, closed-form, single-term expression for all n.
It also seems to be a main talking point.
I think this is a prime example of where it is easy to think something is solved when looking at things from a high level but making an erroneous conclusion due to lack of domain expertise. Classic "Reviewer 2" move. Though I'm not a domain expert and so if there was no novelty over Parke and Taylor I'm pretty sure this will get thrashed in review.
You're right. Parke & Taylor showed the simplest nonzero amplitudes have two minus helicities while one-minus amplitudes vanish (generically). This paper claims that vanishing theorem has a loophole - a new hidden sector exists and one-minus amplitudes are secretly there, but distributional
My comment was in response to the claim I responded to. Any inference you have made about my feelings about OpenAI are that of your own. You can search my comment history if you want to verify or reject your suspicion. I don't think you'll be able to verify it...
I feel for you because you kinda got baited into this by the language in the first couple comments. But whatever’s going on in your comment is so emotional that it’s hard to tell what you’re asking for that you haven’t been able to read already, tl;dr proof stuck at n=4 for years is now for arbitrary n
Yeah I kind of fell for it. I was hoping to be pleasantly surprised by a particle physicist in the openai victory lap thread or someone with insight into what “GPT 5.2 originally conjectured this” means exactly because the way it’s phrased in the preprint makes it sound like they were all doing bongrips with chatgpt and it went “man do you guys ever think about gluon tree amplitudes?” but uh, my empty post getting downvoted hours after being made empty makes it pretty clear that this is a strictly victory-lap-only thread
Fwiw I'm not trying to celebrate for OpenAI. The press piece definitely makes bolder claims than the paper.
I was just stating the facts and correcting a reaction that went too far in the other direction. By taking my comment as supporting or validating OpenAI's claim is just as bad. An error of the same magnitude.
I feel like I've been quoting Feynman a lot this week: The first principle is to not fool yourself, and you're the easiest person to fool. You're the easiest person for you to fool because you're as smart as yourself and deception is easier than proving. We all fall for these traps and the smartest people in the world (or history) are not immune to it. But it's interesting to see on a section of the internet that prides itself for its intelligence. I think we just love blinders, which is only human
It bears repeating that modern LLMs are incredibly capable, and relentless, at solving problems that have a verification test suite. It seems like this problem did (at least for some finite subset of n)!
This result, by itself, does not generalize to open-ended problems, though, whether in business or in research in general. Discovering the specification to build is often the majority of the battle. LLMs aren't bad at this, per se, but they're nowhere near as reliably groundbreaking as they are on verifiable problems.
> modern LLMs are incredibly capable, and relentless, at solving problems that have a verification test suite.
Feel like it's a bit what I tried to expressed few weeks ago https://news.ycombinator.com/item?id=46791642 namely that we are just pouring computational resources at verifiable problems then claim that astonishingly sometimes it works. Sure LLMs even have a slight bias, namely they do rely on statistics so it's not purely brute force but still the approach is pretty much the same : throw stuff at the wall, see what sticks, once something finally does report it as grandiose and claim to be "intelligent".
> throw stuff at the wall, see what sticks, once something finally does report it as grandiose and claim to be "intelligent".
What do we think humans are doing? I think it’s not unfair to say our minds are constantly trying to assemble the pieces available to them in various ways. Whether we’re actively thinking about a problem or in the background as we go about our day.
Every once in a while the pieces fit together in an interesting way and it feels like inspiration.
The techniques we’ve learned likely influence the strategies we attempt, but beyond all this what else could there be but brute force when it comes to “novel” insights?
If it’s just a matter of following a predefined formula, it’s not intelligence.
If it’s a matter of assembling these formulas and strategies in an interesting way, again what else do we have but brute force?
See what I replied just earlier https://news.ycombinator.com/item?id=47011884 namely the different regimes, within paradigm versus challenging it by going back to first principles. The ability to notice something is off beyond "just" assembling existing pieces, to backtrack within the process when failures get too many and actually understand the relationship is precisely different.
So I don’t really see why this would be a difference in kind. We’re effectively just talking about how high up the stack we’re attempting to brute force solutions, right?
How many people have tried to figure out a new maths, a GUT in physics, a more perfect human language (Esperanto for ex.) or programming language, only to fail in the vast majority of their attempts?
Do we think that anything but the majority of the attempts at a paradigm shift will end in failure?
If the majority end in failure, how is that not the same brute force methodology (brute force doesn’t mean you can’t respond to feedback from your failed experiments or from failures in the prevailing paradigms, I take it to just fundamentally mean trying “new” things with tools and information available to you, with the majority of attempts ending in failure, until something clicks, or doesn’t and you give up).
While I don't think anyone has a plausible theory that goes to this level of detail on how humans actually think, there's still a major difference. I think it's fair to say that if we are doing a brute force search, we are still astonishingly more energy efficient at it than these LLMs. The amount of energy that goes into running an LLM for 12h straight is vastly higher than what it takes for humans to think about similar problems.
In the research group I am, we have usually try a few approach to each problem, let's say we get a:
Method A) 30% speed reduction and 80% precision decrease
Method B) 50% speed reduction and 5% precision increase
Method C) 740% speed reduction and 1% precision increase
and we only publish B. It's not brute force[1], but throw noodles at the wall, see what sticks, like the GP said. We don't throw spoons[1], but everything that looks like a noodle has a high chance of been thrown. It's a mix of experience[1] and not enough time to try everything.
The field of medicine - pharmacology and drug discovery, is an optimized version of that. It works a bit like this:
Instead of brute-forcing with infinite options, reduce the problem space by starting with some hunch about the mechanism. Then the hard part that can take decades: synthesize compounds with the necessary traits to alter the mechanism in a favourable way, while minimizing unintended side-effects.
Then try on a live or lab grown specimen and note effectiveness. Repeat the cycle, and with every success, push to more realistic forms of testing until it reaches human trials.
Many drugs that reach the last stage - human trials - often end up being used for something completely other than what they were designed for! One example of that is minoxidil - designed to regular blood pressure, used for regrowing hair!
Yes, this is where I just cannot imagine completely AI-driven software development of anything novel and complicated without extensive human input. I'm currently working in a space where none of our data models are particularly complex, but the trick is all in defining the rules for how things should work.
Our actual software implementation is usually pretty simple; often writing up the design spec takes significantly longer than building the software, because the software isn't the hard part - the requirements are. I suspect the same folks who are terrible at describing their problems are going to need help from expert folks who are somewhere between SWE, product manager, and interaction designer.
Even more generally than verification, just being tied to a loss function that represent something we actually care about. E.g. compiler and test errors, LEAN verification in Aristotle, basic physics energy configs in AlphaFold, or win conditions in e.g. RL, such as in AlphaGo.
RLHF is an attempt to push LLMs pre-trained with a dopey reconstruction loss toward something we actually care about: imagine if we could find a pre-training criterion that actually cared about truth and/or plausibility in the first place!
There's been active work in this space, including TruthRL: https://arxiv.org/html/2509.25760v1. It's absolutely not a solved problem, but reducing hallucinations is a key focus of all the labs.
That paper from the 80s (which is cited in the new one) is about "MHV amplitudes" with two negative-helicity gluons, so "double-minus amplitudes". The main significance of this new paper is to point out that "single-minus amplitudes" which had previously been thought to vanish are actually nontrivial. Moreover, GPT-5.2 Pro computed a simple formula for the single-minus amplitudes that is the analogue of the Parke-Taylor formula for the double-minus "MHV" amplitudes.
It's hard to get someone to do literature first when they get free publicity by not doing literature search and claiming some major AI assisted breakthrough...
Heck, it's hard to get authors to do literature search, period: never mind not thoroughly looking for prior art, even well known disgraced papers get citated continue to get possitive citations all the time...
After last month’s Erdos problems handling by LLMs at this point everyone writing papers should be aware that literature checks are approximately free, even physicists.
Sounds somehow similar to the groundbreaking application of a computer to prove the 4 color theorem. Then the researchers wrote a program to find and formally prove the numerous particular cases. Here the computer finds a simplifying pattern.
I'm not sure if GPTs ability goes beyond a formal math package's in this regard or its just its just way more convienient to ask ChatGPT rather than using these software.
> but I haven’t been to get them to do something totally out of distribution yet from first principles
Can humans actually do that? Sometimes it appears as if we have made a completely new discovery. However, if you look more closely, you will find that many events and developments led up to this breakthrough, and that it is actually an improvement on something that already existed. We are always building on the shoulders of giants.
From my reading yes, but I think I am likely reading the statement differently than you are.
> from first principles
Doing things from first principles is a known strategy, so is guess and check, brute force search, and so on.
For an llm to follow a first principles strategy I would expect it to take in a body of research, come up with some first principles or guess at them, then iteratively construct and tower of reasonings/findings/experiments.
Constructing a solid tower is where things are currently improving for existing models in my mind, but when I try openai or anthropic chat interface neither do a good job for long, not independently at least.
Humans also often have a hard time with this in general it is not a skill that everyone has and I think you can be a successful scientist without ever heavily developing first principles problem solving.
"Constructing a solid tower" from first principles is already super-human level. Sure, you can theorize a tower (sans the "solid") from first principles; there's a software architect at my job that does it every day. But the "solid" bit is where things get tricky, because "solid" implies "firm" and "well anchored", and that implies experimental grounds, experimental verification all the way, and final measurable impact. And I'm not even talking particle physics or software engineering; even folding a piece of paper can give you surprising mismatches between theory and results.
Even the realm of pure mathematics and elegant physic theories, where you are supposed to take a set of axioms ("first principles") and build something with it, has cautionary tales such as the Russel paradox or the non-measure of Feymann path integrals, and let's not talk about string theory.
You could nitpick a rebuttal, but no matter how many people you give credit, general relativity was a completely novel idea when it was proposed. I'd argue for special relatively as well.
I am not a scientific historian, or even a physicist, but IMO relativity has a weak case for being a completely novel discovery. Critique of absolute time and space of Newtonian physics was already well underway, and much of the methodology for exploring this relativity (by way of gyroscopes, inertial reference frames, and synchronized mechanical clocks) were already in parlance. Many of the phenomena that relativity would later explain under a consistent framework already had independent quasi-explanations hinting at the more universal theory. Poincare probably came the closest to unifying everything before Einstein:
> In 1902, Henri Poincaré published a collection of essays titled Science and Hypothesis, which included: detailed philosophical discussions on the relativity of space and time; the conventionality of distant simultaneity; the conjecture that a violation of the relativity principle can never be detected; the possible non-existence of the aether, together with some arguments supporting the aether; and many remarks on non-Euclidean vs. Euclidean geometry.
Now, if I had to pick a major idea that seemed to drop fully-formed from the mind of a genius with little precedent to have guided him, I might personally point to Galois theory (https://en.wikipedia.org/wiki/Galois_theory). (Ironically, though, I'm not as familiar with the mathematical history of that time and I may be totally wrong!)
Right on with special relativity—Lorentz also was developing the theory and was a bit sour that Einstein got so much credit. Einstein basically said “what if special relativity were true for all of physics”, not just electromagnetism, and out dropped e=mc^2. It was a bold step but not unexplainable.
As for general relativity, he spent several years working to learn differential geometry (which was well developed mathematics at the time, but looked like abstract nonsense to most physicists). I’m not sure how he was turned on to this theory being applicable to gravity, but my guess is that it was motivated by some symmetry ideas. (It always come down to symmetry.)
If people want to study this, perhaps it makes more sense to do like we used to: don't include the "labels" of relativity into the training set and see if it comes up with it.
> Critique of absolute time and space of Newtonian physics was already well underway
This only means Einstein was not alone, it does not mean the results were in distribution.
> Many of the phenomena that relativity would later explain under a consistent framework already had independent quasi-explanations hinting at the more universal theory.
And this comes about because people are looking at edge cases and trying to solve things. Sometimes people come up with wild and crazy solutions. Sometimes those solutions look obvious after they're known (though not prior to being known, otherwise it would have already been known...) and others don't.
Your argument really makes the claim that since there are others pursuing similar directions that this means it is in distribution. I'll use a classic statistics style framing. Suppose we have a bag with n red balls and p blue balls. Someone walks over and says "look, I have a green ball" and someone else walks over and says "I have a purple one" and someone else comes over and says "I have a pink one!". None of those balls were from the bag we have. There are still n+p balls in our bag, they are still all red or blue despite there being n+p+3 balls that we know of.
> I am not a [...] physicist
I think this is probably why you don't have the resolution to see the distinctions. Without a formal study of physics it is really hard to differentiate these kinds of propositions. It can be very hard even with that education. So be careful to not overly abstract and simplify concepts. It'll only deprive you of a lot of beauty and innovation.
To be clear, I don't think coming up with relativity was "in distribution" based on the results of the time. I would be exceedingly surprised if an LLM trained on all of the physics up until that point and nothing else would come up with the framework that Einstein did, from such elegant first principles at that. Without handholding from a prompter, I expect an LLM (or non-critical human thinker) would only parrot the general consensus of confusion and non-uniformity that predominated in that era.
I only believe that (1) if it hadn't been Einstein, it would very soon have been someone else using very similar concepts and evidence, (2) "completely novel idea" is a stricter criterion than "not in distribution," and (3) better examples of completely novel ideas from history exist as a benchmark for this sort of things.
> Without a formal study of physics it is really hard to differentiate these kinds of propositions. It can be very hard even with that education. So be careful to not overly abstract and simplify concepts. It'll only deprive you of a lot of beauty and innovation.
I agree, but with the caveat that I think ancestor worship is also an impediment to understanding our intellectual and cultural heritage. Either all of human creativity deserves to be treated sacredly, or none of it does.
> To be clear, I don't think coming up with relativity was "in distribution" based on the results of the time.
This is difficult to infer from the context of the conversation.
> only believe that (1) if it hadn't been Einstein, it would very soon have been someone else
I also agree, but am unsure of your point.
> (2) "completely novel idea" is a stricter criterion than "not in distribution,"
Sorry, I used a looser word. If you have a strong definition of what "in distribution" means I'll be happy to adapt.
> (3) better examples of completely novel ideas from history exist
Sure. Maybe? I can't judge. I think determining how novel something is really requires domain expertise. I only have an undergraduate degree in physics so I am not really qualified on determining the novelty of relativity, but it appears fairly novel to me fwiw. (And I am an enjoyer of scientific history. I'd really recommend Cropper's The Quantum Physicists: And an Introduction to Their Physics as it teaches QM in a more historical progression. I'd also recommend the An Opinionated History of Mathematics podcast which goes through a lot of interesting stuff, including Galileo)
> I think ancestor worship is also an impediment to understanding our intellectual and cultural heritage
I'm in full agreement here (I have past comments on HN to support this too tbh. Probably best to search for things related to Schmidhuber since that's when ancestor worship frequently happens in those topics). It's good to recognize people, but we over emphasize some and entirely forget most. I don't think this is malicious but more logistical. Even Cropper's work misses many people but I think it is still a good balance considering the audience.
I think the best way to avoid the problem is to remember "my understanding is limited" and always will be. At least until we somehow become omniscient, but I'm not counting on that ever happening.
> The quintic was almost proven to have no general solutions by radicals by Paolo Ruffini in 1799, whose key insight was to use permutation groups, not just a single permutation.
Thing is, I am usually the kind of person who defends the idea of a lone genius. But I also believe there is a continuous spectrum, no gaps, from the village idiot to Einstein and beyond.
Let me introduce, just for fun, not for the sake of any argument, another idea from math which I think it came really out of the blue, to the degree that it's still considered an open problem to write an exposition about it, since you cannot smoothly link it to anything else: forcing.
At least Einstein didn't just suddenly turn around and say:
```ai-slop
But wait, this equation is too simple, I need to add more terms or it won't model the universe. Let me think about this again. I have 5 equations and I combined them and derived e=mc^2 but this is too simple. The universe is more complicated. Let's try a different derivation. I'll delete the wrong outputs first and then start from the input equations.
<Deletes files with groundbreaking discovery>
Let me think. I need to re-read the original equations and derive a more complex formula that describes the universe.
<Re-reads equation files>
Great, now I have the complete picture of what I need to do. Let me plan my approach. I'm ready. I have a detailed plan. Let me check some things first.
I need to read some extra files to understand what the variables are.
<Reads the lunch menu for the next day>
Perfect. Now I understand the problem fully, let me revise my plan.
<Writes plan file>
Okay I have written the plan. Do you accept?
<Yes>
Let's go. I'll start by creating a To Do list:
- [ ] Derive new equation from first principles making sure it's complex enough to describe reality.
- [ ] Go for lunch. When the server offers tuna, reject it because the notes say I don't like fish.
```
(You know what's really sad? I wrote that slop without using AI and without referring to anything...)
You need to differentiate between special and general relativity when making these statements.
It is absolutely true that someone else would have come up with special relativity very soon after Einstein. All that would be necessary is someone else to have the wherewithal to say "perhaps the aether does not need to exist" for the equations already known at the time by others before Einstein to lead to the general theory.
General relativity is different. Witten contends that it is entirely possible that without Einstein, we may have had to wait for the early string theorists of the 1960s to discover GR as a classical limit of the first string theories in their quest to understand the strong nuclear force.
As opposed to SR, GR is one of the most singular innovative intellectual achievements in human history. It's definitely "out of distribution" in some sense.
Yes, the principle of relativity was known to Newton, but the other idea, that the speed of light is the same in all reference frames, was new, counterintuitive, and what makes special relativity the way it is.
General relativity was a completely novel idea. Einstein took a purely mathematical object (now known as the Einstein tensor), and realized that since its coveriant derivative was zero, it could be equated (apart fron a constant factor) to a conserved physical object, the energy momentum tensor (except for a constant factor). It didn't just fall out of Riemannian geometry and what was known about physics at the time.
Special relativity was the work of several scientists as well as Einstein, but it was also a completely novel idea - just not the idea of one person working alone.
I don't know why anyone disputes that people can sometimes come up with completely novel ideas out of the blue. This is how science moves forward. It's very easy to look back on a breakthrough and think it looks obvious (because you know the trick that was used), but it's important to remember that the discoverer didn't have the benefit of hindsight that you have.
Even if I grant you that, surely we’ve moved the goal posts a bit if we’re saying the only thing we can think of that AI can’t do is the life’s work of a man who’s last name is literally synonymous with genius.
It isn't an anteceent, it's part of special relativity, discovered by Lorentz. It's well known that special relativity is the work of several people as well as Einstein.
Not really. Pretty sure I read recently that Newton appreciated that his theory was non-local and didn't like what Einstein later called "spooky action at a distance". The Lorentz transform was also known from 1887. Time dilation was understood from 1900. Poincaré figured out in 1905 that it was a mathematical group. Einstein put a bow on it all by figuring out that you could derive it from the principle of relativity and keeping the speed of light constant in all inertial reference frames.
I'm not sure about GR, but I know that it is built on the foundations of differential geometry, which Einstein definitely didn't invent (I think that's the source of his "I assure you whatever your difficulties in mathematics are, that mine are much greater" quote because he was struggling to understand Hilbert's math).
And really Cauchy, Hilbert, and those kinds of mathematicians I'd put above Einstein in building entirely new worlds of mathematics...
Newton wrote, "That one body may act upon another at a distance through a vacuum without the mediation of anything else, by and through which their action and force may be conveyed from one another, is to me so great an absurdity that, I believe, no man who has in philosophic matters a competent faculty of thinking could ever fall into it."
This quote itself must be taken in the context of Newton's own aspirations. Newton was specifically searching for force capable of moving distant objects when he realised the essence of gravity. No apple really fell on his head - that story was likely invented by those who could not stand Newton (he was famously brash) and meant simply that his personality was a result of getting hit on the head.
And Newton was famously interested in dark religous interference in worldly affairs - what today we would call The Occult. When he did finally succeed in finding his force for moving objects at a distance, without need for an intervening body, he gave credit to these supernatural entities - at least that is how this quote was taken in his day. This religious context is not well known today, nor is Newton's difficult character, so today it is easy to take the quote out of context. Newton was (likely) not disputing the validity of his discovery, rather, he was invoking one of his passions (The Occult) in the affairs of one of his successful passions (finding a force to move distant objects).
It should be noted that some of Newton's successful religious work is rarely attributed to him. For a prominent example, it was Newton that calculated Jesus's birth to be 4 BC, not 1 AD as was the intention of the new calendar.
The process you’re describing is humans extending our collective distribution through a series of smaller steps. That’s what the “shoulders of giants” means. The result is we are able to do things further and further outside the initial distribution.
So it depends on if you’re comparing individual steps or just the starting/ending distributions.
A discovery by a giant is in some sense a new base vector in the space of discoveries. The interesting question is if a statistical machine can only perform a linear combination in the space of discoveries, or if a statistical machine can discover a new base vector in the space of discoveries.. whatever that is.
For sure we know modern LLMs and AIs are not constrained by anything particularly close to simple linear combinations, by virtue of their depth and non-linear activation functions.
But yes, it is not yet clear to what degree there can be (non-linear) extrapolation in the learned semantic spaces here.
Arguably it's precisely a paradigm shift. Continuing whatever worked until now is within the paradigm, our current theories and tools works, we find few problems that don't fit but that's fine the rest is still progress, we keep on hitting more problems or those few pesky unsolved problems actually appear to be important. We then go back to the theory and its foundations and finally challenge them. We break from the old paradigm and come up with new theories and tools because the first principles are now better understood and we iterate.
So that's actually 2 different regimes on how to proceed. Both are useful but arguably breaking off of the current paradigm is much harder and thus rare.
The tricky part is that LLMs aren't just spewing outputs from the distribution (or "near" learned manifolds), but also extrapolating / interpolating (depending on how much you care about the semantics of these terms https://arxiv.org/abs/2110.09485).
There are genuine creative insights that come from connecting two known semantic spaces in a way that wasn't obvious before (e.g, novel isomorphism). It is very conceivable that LLMs could make this kind of connection, but we haven't really seen a dramatic form of this yet. This kind of connection can lead to deep, non-trivial insights, but whether or not it is "out-of-distribution" is harder to answer in this case.
I mean, there’s just no way you can take the set of publicly known ideas from all human civilizations, say, 5,000 years ago, and say that all the ideas we have now were “in the distribution” then. New ideas actually have to be created.
If that were true then science should have accelerated a lot faster. Science would have happened differently and researchers would have optimized to trying to ingest as many papers as they can.
Dig deep into things and you'll find that there are often leaps of faith that need to be made. Guesses, hunches, and outright conjectures. Remember, there are paradigm shifts that happen. There are plenty of things in physics (including classical) that cannot be determined from observation alone. Or more accurately, cannot be differentiated from alternative hypotheses through observation alone.
I think the problem is when teaching science we generally teach it very linearly. As if things easily follow. But in reality there is generally constant iterative improvements but they more look like a plateau, then there are these leaps. They happen for a variety of reasons but no paradigm shift would be contentious if it was obvious and clearly in distribution. It would always be met with the same response that typical iterative improvements are met with "well that's obvious, is this even novel enough to be published? Everybody already knew this" (hell, look at the response to the top comment and my reply... that's classic "Reviewer #2" behavior). If it was always in distribution progress would be nearly frictionless. Again, with history in how we teach science we make an error in teaching things like Galileo, as if The Church was the only opposition. There were many scientists that objected, and on reasonable grounds. It is also a problem we continually make in how we view the world. If you're sticking with "it works" you'll end up with a geocentric model rather than a heliocentric model. It is true that the geocentric model had limits but so did the original heliocentric model and that's the reason it took time to be adopted.
By viewing things at too high of a level we often fool ourselves. While I'm criticizing how we teach I'll also admit it is a tough thing to balance. It is difficult to get nuanced and in teaching we must be time effective and cover a lot of material. But I think it is important to teach the history of science so that people better understand how it actually evolves and how discoveries were actually made. Without that it is hard to learn how to actually do those things yourself, and this is a frequent problem faced by many who enter PhD programs (and beyond).
> We are always building on the shoulders of giants.
And it still is. You can still lean on others while presenting things that are highly novel. These are not in disagreement.
It's probably worth reading The Unreasonable Effectiveness of Mathematics in the Natural Sciences. It might seem obvious now but read carefully. If you truly think it is obvious that you can sit in a room armed with only pen and paper and make accurate predictions about the world, you have fooled yourself. You have not questioned why this is true. You have not questioned when this actually became true. You have not questioned how this could be true.
When chess engines were first developed, they were strictly worse than the best humans. After many years of development, they became helpful to even the best humans even though they were still beatable (1985–1997). Eventually they caught up and surpassed humans but the combination of human and computer was better than either alone (~1997–2007). Since then, humans have been more or less obsoleted in the game of chess.
Five years ago we were at Stage 1 with LLMs with regard to knowledge work. A few years later we hit Stage 2. We are currently somewhere between Stage 2 and Stage 3 for an extremely high percentage of knowledge work. Stage 4 will come, and I would wager it's sooner rather than later.
There's a major difference between chess and scientific research: setting the objectives is itself part of the work.
In chess, there's a clear goal: beat the game according to this set of unambiguous rules.
In science, the goals are much more diffuse, and setting those in the first place is what makes a scientist more or less successful, not so much technical ability. It's a very hierarchical field where permanent researchers direct staff (postdocs, research scientists/engineers), direct grad students. And it's at the bottom of the pyramid where the technical ability is the most relevant/rewarded.
Research is very much a social game, and I think replacing it with something run by LLMs (or other automatic process) is much more than a technical challenge.
The evolution was also interesting: first the engines were amazing tactically but pretty bad strategically so humans could guide them. With new NN based engines they were amazing strategically but they sucked tactically (first versions of Leela Chess Zero). Today they closed the gap and are amazing at both strategy and tactics and there is nothing humans can contribute anymore - all that is left is to just watch and learn.
With a chess engine, you could ask any practitioner in the 90's what it would take to achieve "Stage 4" and they could estimate it quite accurately as a function of FLOPs and memory bandwidth. It's worth keeping in mind just how little we understand about LLM capability scaling. Ask 10 different AI researchers when we will get to Stage 4 for something like programming and you'll get wild guesses or an honest "we don't know".
That is not what happened with chess engines. We didn’t just throw better hardware at it, we found new algorithms, improved the accuracy and performance of our position evaluation functions, discovered more efficient data structures, etc.
People have been downplaying LLMs since the first AI-generated buzzword garbage scientific paper made its way past peer review and into publication. And yet they keep getting better and better to the point where people are quite literally building projects with shockingly little human supervision.
Chess grandmasters are living proof that it’s possible to reach grandmaster level in chess on 20W of compute. We’ve got orders of magnitude of optimizations to discover in LLMs and/or future architectures, both software and hardware and with the amount of progress we’ve got basically every month those ten people will answer ‘we don’t know, but it won’t be too long’. Of course they may be wrong, but the trend line is clear; Moore’s law faced similar issues and they were successively overcome for half a century.
> With a chess engine, you could ask any practitioner in the 90's what it would take to achieve "Stage 4" and they could estimate it quite accurately as a function of FLOPs and memory bandwidth.
And the same practitioners said right after deep blue that go is NEVER gonna happen. Too large. The search space is just not computable. We'll never do it. And yeeeet...
We are at level 2.5 for software development, IMO. There is a clear skill gap between experienced humans and LLMs when it comes to writing maintainable, robust, concise and performant code and balancing those concerns.
The LLMs are very fast but the code they generate is low quality. Their comprehension of the code is usually good but sometimes they have a weightfart and miss some obvious detail and need to be put on the right path again. This makes them good for non-experienced humans who want to write code and for experienced humans who want to save time on easy tasks.
> The LLMs are very fast but the code they generate is low quality.
I think the latest generation of LLM with claude code is not low quality. It's better than the code that pretty much every dev on our team can do outside of very narrow edge cases.
"GPT did this". Authored by Guevara (Institute for Advanced Study), Lupsasca (Vanderbilt University), Skinner (University of Cambridge), and Strominger (Harvard University).
Probably not something that the average GI Joe would be able to prompt their way to...
I am skeptical until they show the chat log leading up to the conjecture and proof.
I'm a big LLM sceptic but that's… moving the goalposts a little too far. How could an average Joe even understand the conjecture enough to write the initial prompt? Or do you mean that experts would give him the prompt to copy-paste, and hope that the proverbial monkey can come up with a Henry V? At the very least posit someone like a grad student in particle physics as the human user.
I would interpret it as implying that the result was due to a lot more hand-holding that what is let on.
Was the initial conjecture based on leading info from the other authors or was it simply the authors presenting all information and asking for a conjecture?
Did the authors know that there was a simpler means of expressing the conjecture and lead GPT to its conclusion, or did it spontaneously do so on its own after seeing the hand-written expressions.
These aren't my personal views, but there is some handwaving about the process in such a way that reads as if this was all spontaneous involvement on GPTs end.
But regardless, a result is a result so I'm content with it.
Hi I am an author of the paper. We believed that a simple formula should exist but had not been able to find it despite significant effort. It was a collaborative effort but GPT definitely solved the problem for us.
Oh that's really cool, I am not versed in physics by any means, can you explain how you believed there to be a simple formula but were unable to find it? What would lead you to believe that instead of just accepting it at face value?
There are closely related "MHV amplitudes" which naively obey a really complicated formula, but for which there famously also exists a much simpler "Parke-Taylor formula". Alfredo had derived a complicated expression for these new "single-minus amplitudes" and we were hoping we could find an analogue of the simpler "Parke-Taylor formula" for them.
In this case there certainly were experts doing hand-holding. But simply being able to ask the right question isn't too much to ask, is it? If it had been merely a grad student or even a PhD student who had asked ChatGPT to figure out the result, and ChatGPT had done that, even interactively with the student, this would be huge news. But an average person? Expecting LLMs to transcend the GIGO principle is a bit too much.
The Average Joe reads at an 8th grade level. 21% are illiterate in the US.
LLMs surpassed the average human a long time ago IMO. When LLMs fail to measure up to humans, it's that they fail to measure up against human experts in a given field, not the Average Joe.
they probably also acknowledge pytorch, numpy, R ... but we don't attribute those tools as the agent who did the work.
I know we've been primed by sci-fi movies and comic books, but like pytorch, gpt-5.2 is just a piece of software running on a computer instrumented by humans.
I don't see the authors of those libraries getting a credit on the paper, do you ?
>I know we've been primed by sci-fi movies and comic books, but like pytorch, gpt-5.2 is just a piece of software running on a computer instrumented by humans.
And we are just a system running on carbon-based biology in our physics computer run by whomever. What makes us special, to say that we are different than GPT-5.2?
> And we are just a system running on carbon-based biology in our physics computer run by whomever. What makes us special, to say that we are different than GPT-5.2?
Do you really want to be treated like an old PC (dismembered, stripped for parts, and discarded) when your boss is done with you (i.e. not treated specially compared to a computer system)?
But I think if you want a fuller answer, you've got a lot of reading to do. It's not like you're the first person in the world to ask that question.
You misunderstood, I am prohumanism. My comment was about challenging the believe that models cant be as intelligent as we are, which cant be answered definitely, though a lot of empirical evidence seems to point to the fact, that we are not fundamentally different intelligence wise. Just closing our eyes will not help in preserving humanism, so we have to shape the world with models in a human friendly way, aka alignment.
It's always a value decision. You can say shiny rocks are more important than people and worth murdering over.
Not an uncommon belief.
Here you are saying you personally value a computer program more than people
It exposes a value that you personally hold and that's it
That is separate from the material reality that all this AI stuff is ultimately just computer software... It's an epistemological tautology in the same way that say, a plane, car and refrigerator are all just machines - they can break, need maintenance, take expertise, can be dangerous...
LLMs haven't broken the categorical constraints - you've just been primed to think such a thing is supposed to be different through movies and entertainment.
I hate to tell you but most movie AIs are just allegories for institutional power. They're narrative devices about how callous and indifferent power structures are to our underlying shared humanity
Their point is, would you be able to prompt your way to this result? No. Already trained physicists working at world-leading institutions could. So what progress have we really made here?
No it’s like saying: New expert drives new results with existing experts.
The humans put in significant effort and couldn’t do it. They didn’t then crank it out with some search/match algorithm.
They tried a new technology, modeled (literally) on us as reasoners, that is only just being able to reason at their level and it did what they couldn’t.
The fact that the experts were a critical context for the model, doesn’t make the models performance any less significant. Collaborators always provide important context for each other.
I don't want to be rude but like, maybe you should pre-register some statement like "LLMs will not be able to do X" in some concrete domain, because I suspect your goalposts are shifting without you noticing.
We're talking about significant contributions to theoretical physics. You can nitpick but honestly go back to your expectations 4 years ago and think — would I be pretty surprised and impressed if an AI could do this? The answer is obviously yes, I don't really care whether you have a selective memory of that time.
It's a nontrivial calculation valid for a class of forces (e.g. QCD) and apparently a serious simplification to a specific calculation that hadn't been completed before. But for what it's worth, I spent a good part of my physics career working in nucleon structure and have not run across the term "single minus amplitudes" in my memory. That doesn't necessarily mean much as there's a very broad space work like this takes place in and some of it gets extremely arcane and technical.
One way I gauge the significance of a theory paper are the measured quantities and physical processes it would contribute to. I see none discussed here which should tell you how deep into math it is. I personally would not have stopped to read it on my arxiv catch-up
I never said LLMs will not be able to do X. I gave my summary of the article and my anecdotal experiences with LLMs. I have no LLM ideology. We will see what tomorrow brings.
> We're talking about significant contributions to theoretical physics.
Whoever wrote the prompts and guided ChatGPT made significant contributions to theoretical physics. ChatGPT is just a tool they used to get there. I'm sure AI-bloviators and pelican bike-enjoyers are all quite impressed, but the humans should be getting the research credit for using their tools correctly. Let's not pretend the calculator doing its job as a calculator at the behest of the researcher is actually a researcher as well.
If this worked for 12 hours to derive the simplified formula along with its proof then it guided itself and made significant contributions by any useful definition of the word, hence Open AI having an author credit.
How much precedence is there for machines or tools getting an author credit in research? Genuine question, I don't actually know. Would we give an author credit to e.g. a chimpanzee if it happened to circle the right page of a text book while working with researchers, leading them to a eureka moment?
Would it? I think there's a difference between "the researchers used ChatGPT" and "one of the researchers literally is ChatGPT." The former is the truth, and the latter is the misrepresentation in my eyes.
I have no problem with the former and agree that authors/researchers must note when they use AI in their research.
> now you are debating exactly how GPT should be credited. idk, I'm sure the field will make up some guidance
In your eyes maybe there's no difference. In my eyes, big difference. Tools are not people, let's not further the myth of AGI or the silly marketing trend of anthropomorphizing LLMs.
>How much precedence is there for machines or tools getting an author credit in research?
Well what do you think ? Do the authors (or a single symbolic one) of pytorch or numpy or insert <very useful software> typically get credits on papers that utilize them heavily?
Well Clearly these prominent institutions thought GPT's contribution significant enough to warrant an Open AI credit.
>Would we give an author credit to e.g. a chimpanzee if it happened to circle the right page of a text book while working with researchers, leading them to a eureka moment?
Cool Story. Good thing that's not what happened so maybe we can do away with all these pointless non sequiturs yeah ? If you want to have a good faith argument, you're welcome to it, but if you're going to go on these nonsensical tangents, it's best we end this here.
> Well what do you think ? Do the authors (or a single symbolic one) of pytorch or numpy or insert <very useful software> typically get credits on papers that utilize them heavily ?
I don't know! That's why I asked.
> Well Clearly these prominent institutions thought GPT's contribution significant enough to warrant an Open AI credit.
Contribution is a fitting word, I think, and well chosen. I'm sure OpenAI's contribution was quite large, quite green and quite full of Benjamins.
> Cool Story. Good thing that's not what happened so maybe we can do away with all these pointless non sequiturs yeah ? If you want to have a good faith argument, you're welcome to it, but if you're going to go on these nonsensical tangents, it's best we end this here.
It was a genuine question. What's the difference between a chimpanzee and a computer? Neither are humans and neither should be credited as authors on a research paper, unless the institution receives a fat stack of cash I guess. But alas Jane Goodall wasn't exactly flush with money and sycophants in the way OpenAI currently is.
If you don't read enough papers to immediately realize it is an extremely rare occurrence then what are you even doing? Why are you making comments like you have the slightest clue of what you're talking about? including insinuating the credit was what...the result of bribery?
You clearly have no idea what you're talking about. You've decided to accuse prominent researchers of essentially academic fraud with no proof because you got butthurt about a credit. You think your opinion on what should and shouldn't get credited matters ? Okay
Do I need to be credentialed to ask questions or point out the troubling trend of AI grift maxxers like yourself helping Sam Altman and his cronies further the myth of AGI by pretending a machine is a researcher deserving of a research credit? This is marketing, pure and simple. Close the simonw substack for a second and take an objective view of the situation.
If a helicopter drops someone off on the top of Mount Everest, it's reasonable to say that the helicopter did the work and is not just a tool they used to hike up the mountain.
Who piloted the helicopter in this scenario, a human or chatgpt? You'd say the pilot dropped them off in a helicopter. The helicopter didn't fly itself there.
“They have chosen cunning instead of belief. Their prison is only in their minds, yet they are in that prison; and so afraid of being taken in that they cannot be taken out.”
> In my experience LLM’s can make new things when they are some linear combination of existing things but I haven’t been to get them to do something totally out of distribution yet from first principles.
What's the distinction between "first principles" and "existing things"?
I'm sympathetic to the idea that LLMs can't produce path-breaking results, but I think that's true only for a strict definition of path-breaking (that is quite rare for humnans too).
Hmm feels a bit trivializing, we don't know exactly how difficult it was to come up with the generic set of equations mentioned from the human starting point.
I can claim some knowledge of physics from my degree, typically the easy part is coming up with complex dirty equations that work under special conditions, the hard part is the simplification into something elegant, 'natural' and general.
Also
"LLM’s can make new things when they are some linear combination of existing things"
Doesn't really mean much, what is a linear combination of things you first have to define precisely what a thing is?
Serious questions, I often hear about this "let the LLM cook for hours" but how do you do that in practice and how does it manages its own context? How doesn't it get lost at all after so many tokens?
I’m guessing, would love someone who has first hand knowledge to comment. But my guess is it’s some combination of trying many different approaches in parallel (each in a fresh context), then picking the one that works, and splitting up the task into sequential steps, where the output of one step is condensed and is used as an input to the next step (with possibly human steering between steps)
From what I've seen is a process of compacting the session once it reaches some limit, which basically means summarizing all the previous work and feeding it as the initial prompt for the next session.
the annoying part is that with tool calls, a lot of those hours is time spent on netowrk round trips.
over long periods of time, checklists are the biggest thing, so the LLM can track whats already done and whats left. after a compact, it can pull the relevant stuff back up and make progress.
having some level or hierarchy is also useful - requirements, high level designs, low level designs, etc
> I haven’t been to get them to do something totally out of distribution yet from first principles.
Agree with this. I’ve been trying to make LLMs come up with creative and unique word games like Wordle and Uncrossy (uncrossy.com), but so far GPT-5.2 has been disappointing. Comparatively, Opus 4.5 has been doing better on this.
But it’s good to know that it’s breaking new ground in Theoretical Physics!
$200/month would cover many such sessions every month.
The real question is, what does it cost OpenAI? I'm pretty sure both their plans are well below cost, at least for users who max them out (and if you pay $200 for something then you'll probably do that!). How long before the money runs out? Can they get it cheap enough to be profitable at this price level, or is this going to be "get them addicted then jack it up" kind of strategy?
Surely higher level math is just linear combinations of the syntax and implications of lower level math. LLMs are taught syntax of basically all existing math notation, I assume. Much of math is, after all, just linguistic manipulation and detection of contradiction in said language with a more formal, a priori language.
I intended to imply this with "detection of contradiction". Coherence seems to me to be the only a priori meaning. Most of the meaning of "meaning" seems to me to be a posteriori. After all, what is the point of an a priori floating signifier?
Setting the framework (what I short-handed by "definitions") precludes the exploration of results (or at least an efficient one) that would yield to some framework-defining analysis.
The search space is much too rich to be explored anything but greedily in timid steps off the trodden path, and the frameworks (arbitrary) set both the highways and the vehicles by which we move along and out of them.
Now, the argument can be made that the "meta-mathematical" (but outright mathematical, really) setting of frameworks follows the same structure, and LLMs could also explore that space.
Even assuming that, a major roadblock remains: mathematics should remain understandable by humans, and yield fast progress in desirable (by whom? until now, by humans) directions, so the constraints on admissible frameworks are not as simple as "yields to coherent results".
Also, to take a step back, I wonder what the pertinence of using numerical math is to derive analytical math when we can already solve a great deal of problems through numerical methods. For instance, is it worth spending however many MWh on LLMs to derive an analytical solution to an optimization problem, which might itself be very expensive to compute (human-derived expressions tend to be particularly cheap to evaluate precisely because we are so limited; machines (formal calculus for instance) will happily give you multi-page formulae with thousands of operations to evaluate), when there's a vast array of algorithms at the ready to provide arbitrarily precise solutions?
What remains is the kind of math that, arguably, is much more precious to understand than to derive.
My physics professor once claimed that imagination is just mental manipulation of past experiences. I never thought it was true for human beings but for LLMs it makes perfect sense.
I must be a Luddite, how do you have a model working for 12 hours on a problem. Mine is ready with an answer and always interrupts to ask confirmation or show answer
That's on the harness - the device actually sending the prompt to the model. You can write a different harness that feeds the problem back in for however long you want. Ask Claude Code or Codex to build it for you in as minimal a fashion as possible and you'll see that a naïve version is not particularly more complex than `while true; do prompt $file >> file; done` (though it's not that precisely, obviously).
You're assuming there aren't "new things" latent inside currently existing information. That's definitely false, particulary for math/physics.
But it's worth thinking more about this. What gives humans the ability to discover "new things"? I would say it's due to our interaction with the universe via our senses, and not due to some special powers intrinsic to our brains that LLMs lack. And the thing is, we can feed novel measurements to LLMs (or, eventually, hook them up to camera feeds to "give them senses")
No it isn't false. If it is new it is novel, novel because it is known to some degree and two other abstracted known things prove the third. Just pattern matching connecting dots.
The vast majority of work by mathematicians uses n abstracted known things to prove something that is unproven. In fact, there is a view in philosophy that all math consists only of this.
In my experience humans can make new things when they are some linear combination of existing things but I haven’t been able to get them to do something totally out of distribution yet from first principles[0].
Is every new thing not just combinations of existing things? What does out of distribution even mean? What advancement has ever made that there wasn’t a lead up of prior work to it? Is there some fundamental thing that prevents AI from recombining ideas and testing theories?
For example, ever since the first GPT 4 I’ve tried to get LLM’s to build me a specific type of heart simulation that to my knowledge does not exist anywhere on the public internet (otherwise I wouldn’t try to build it myself) and even up to GPT 5.3 it still cannot do it.
But I’ve successfully made it build me a great Poker training app, a specific form that also didn’t exist, but the ingredients are well represented on the internet.
And I’m not trying to imply AI is inherently incapable, it’s just an empirical (and anecdotal) observation for me. Maybe tomorrow it’ll figure it out. I have no dogmatic ideology on the matter.
> Is every new thing not just combinations of existing things?
If all ideas are recombinations of old ideas, where did the first ideas come from? And wouldn't the complexity of ideas be thus limited to the combined complexity of the "seed" ideas?
I think it's more fair to say that recombining ideas is an efficient way to quickly explore a very complex, hyperdimensional space. In some cases that's enough to land on new, useful ideas, but not always. A) the new, useful idea might be _near_ the area you land on, but not exactly at. B) there are whole classes of new, useful ideas that cannot be reached by any combination of existing "idea vectors".
Therefore there is still the necessity to explore the space manually, even if you're using these idea vectors to give you starting points to explore from.
All this to say: Every new thing is a combination of existing things + sweat and tears.
The question everyone has is, are current LLMs capable of the latter component. Historically the answer is _no_, because they had no real capacity to iterate. Without iteration you cannot explore. But now that they can reliably iterate, and to some extent plan their iterations, we are starting to see their first meaningful, fledgling attempts at the "sweat and tears" part of building new ideas.
Well, what exactly an “idea” is might be a little unclear, but I don’t think it clear that the complexity of ideas that result from combining previously obtained ideas would be bounded by the complexity of the ideas they are combinations of.
Any countable group is a quotient of a subgroup of the free group on two elements, iirc.
There’s also the concept of “semantic primes”. Here is a not-quite correct oversimplification of the idea: Suppose you go through the dictionary and one word at a time pick a word whose definition includes only other words that are still in the dictionary, and removing them. You can also rephrase definitions before doing this, as long as it keeps the same meaning. Suppose you do this with the goal of leaving as few words in it as you can. In the end, you should have a small cluster of a bit over 100 words, in terms of which all the other words you removed can be indirectly defined.
(The idea of semantic primes also says that there is such a minimal set which translates essentially directly* between different natural languages.)
I don’t think that says that words for complicated ideas aren’t like, more complicated?
There are in fact ways to directly quantify this, if you are training e.g. a self-supervised anomaly-detection model.
Even with modern models not trained in that manner, looking at e.g. cosine distances of embeddings of "novel" outputs could conceivably provide objective evidence for "out-of-distribution" results. Generally, the embeddings of out-of-distribution outputs will have a large cosine (or even Euclidean) distance from the typical embedding(s). Just, most "out-of-distribution" outputs will be nonsense / junk, so, searching for weird outputs isn't really helpful, in general, if your goal is useful creativity.
Very very few human individuals are capable of making new things that are not a linear combination of existing things. Even such things as special relativity were an application of two previous ideas. All of special relativity is deriveable from the principles of relative motion (known into antiquity) and the constant speed of light (which was known to Einstein). From there it is a straightforwards application of the Pythagorean theorem to realize there is a contradiction and the lorentz factor falls out naturally via basic algebra.
My issue with any of these claims is the lack of proof. Just share the chat and now it got to the discovery. I'll believe it when I can see it for myself at this point. It's too easy to make all sorts of claims without proof these days. Elon Musk makes them all the time.
I'm pretty sure it is widely known that the early 5.x series were built from 4.5 (unreleased). It seems more plausible the 5.x series is still in that continuation.
For some extra context, pre-training is ~1/3 of the training, where it gains the basic concepts of how tokens go together. Mid & late training are where you instill the kinds of anthropic behaviors we see today. I expect pre-training to increasingly become a lower percentage of overall training, putting aside any shifts of what happens in each phase.
So to me, it is plausible they can take the 4.x pre-training and keep pushing in the later phases. There is a lot of results out there to show scaling laws (limits) have not peaked yet. I would not be surprised to learn that Gemini 3 Deep Research had 50% late-training / RL
Okay I see what you mean, and yeah that sounds reasonable too. Do you have any context on that first part? I would like to know more about how/why they might not have been able to pursue more training runs.
I have not done it myself (don't have the dinero), but my understanding is that there are many runs, restarts, and adjustments at this phase. It's surprisingly more fragile than we know aiui
If you already have a good one, it's not likely much has changed since a year ago that would create meaningful differences at this phase (in data, arch is diff, I know less here). If it is indeed true, it's a datapoint to add to the others singling internal (everybody has some amount of this, not good when it makes the headlines)
Distillation is also a powerful training method. There are many ways to stay with the pack without having new pre-training runs. It's pretty much what we see from all of them with the minor versions. So coming back to it, the speculation is that OpenAi is still on their 4.x pre-train, but that doesn't impede all progress
Well one could make the argument that Musk is a short or medium term problem and that in 4 years when Trump is gone everyone will forget about hating on Tesla and it will be a great car company again. Musk is in his 50s and won’t be CEO forever. So if your investment horizon is 10+ years and you don’t predict total company collapse then it might be a bargain time to buy.
Tesla market cap is currently $959B, shipped 0.5M vehicles in 2024.
Toyota is second with $244B and shipped 11M vehicles.
Ford shipped 4.4M vehicles and has a market cap of $41B.
If I look at [1] then Tesla is worth more than the next 12 or so combined.
Nothing about this makes any sense. Maybe ten or fifteen years ago you could have made the argument based on future prospects, but that case is exceedingly hard to make today when there's tons more competition on the EV market.
That's not how it works for Tesla. Totally different buying experience. You order online, get a notification some days or weeks later that it's ready, go to the store sign a couple of documents and drive off with the car. Maybe 10 minutes total. (Source: bought a Tesla a few years ago). I just recently got a Kia EV9 and it was a 3-4 hour marathon of paperwork, talking, more paperwork. Really terrible buying experience. Nice car though.
Yeah that's one of the reasons my dad went with Tesla for his recent car purchase. My wife helped him get a RAV4 (Toyota) a few years ago and the three of us lost four hours on a weekend because of the song and dance that dealerships put you through. Never again. From now on it'll be Tesla, Rivian, Polestar, or Lucid instead.
That's the old way of thinking -- they're trying to do just that across the government and without some enforcement mechanism to make them send the checks, the practical result is that the President can indeed cancel pieces of legislation via impoundment.
Example 3 - CHIPS act would have had funding withheld if a Federal Court hadn't stepped in, but it's unclear what enforcement mechanism can force the funding to resume: https://archive.is/BxjHw
Sure, but a lot of people at NIST who were in charge of implementing the CHIPS act have been fired. He definitely seems to be doing all he can to sabotage the CHIPS act without needing any congressional action.
He sure seems to be able to just terminate legislation signed into law. He already did it with USAID, and is in the process of doing it to many other departments.
You like this rule being broken? Great, good for you.
What about the other rules? The ones protecting you and the country? Is due process not valuable any more? What protects us from people with bad or selfish intentions?
Yeah, kids starving and dying of cholera.. fuck em /s
> The Inspector General also warned that $489 million in humanitarian food aid was at risk of spoiling due to staff furloughs and unclear guidance. The Office of Presidential Personnel fired the Inspector General the next day, despite a law requiring 30 days notice to Congress before firing an Inspector General.
The correct way for the government to reel in USAID would be for congress to give them less funding and to tell them specifically what they want funded. Regardless if it offends you personally, those are all lawful uses of the money and the only illegal thing that's happened here is the funding being stopped by the President.
First, I would not trust the current USAID disbursement personnel not to piss the money into the wind. I want them gone.
And it's not a question of being offended personally - these are just ridiculous expenses that cannot possibly be justified. But I am indeed offended that the amount 4x of my real estate taxes that I can barely scrape off the bottom of a barrel is being wasted on some opera abroad. If you are wondering why people vote for Trump, this is one of those reasons.
Regarding legality of funding being stopped by the President, I am not a lawyer (and I am guessing neither are you), so I am not going to take your legal opinion on this and will wait for the courts to issue the final ruling.
That they are senseless enough to their their personal opinions on budgeting should run the entire government, and that their little agendas are the reason everything should burn to the ground? Yes, that is why people voted for trump (they are stupid and vindictive).
You can't berate or threaten people into thinking your voting or political opinions are smart/well founded. It either is or it isn't.
Watching Trump illegally destroy institutions that collectively use <5% of the federal budget, while increasing the defect, and rationalizing it as "At least Trump is trying to do something about the runaway government spending" is stupid. Straight up stupid.
The fact that there’s a specific law called the Impoundment Control Act where the specific actions Trump is trying were made illegal should give you a hint to which way the court cases are going to go..
Why are you confident the Supreme Court will not declare the Impoundment Control Act an unconstitutional restriction of executive power? Or declare the only recourse is impeachment? Who do you expect would enforce the ruling you predicted?
That is surely the elephant in the room.. every time it’s been litigated before the court in the past, the Act has been found constitutional but who knows with this specific set of justices and their obsequiousness to Trump and his executive branch.
Those numbers are for the wrong line items, and the WH press secretary was wrong about the source of those funds. Both of those were out of the state department budget, which (putting aside the present murky status) did not oversee USAID at the time.
What can I say... if you are correct about this (there are a lot of claims from both sides but no proofs), I hope DOGE gets its hands on the State Department, too. We have enough worthy causes to take care of inside the US.
But DOGE has been trying to do effectively that for the past month, and has been distressingly successful at it. (For all that conservatives whined about the existence of an unaccountable deep state override elected officials making laws, that's basically an accurate description of DOGE.)
I would argue that the main benefit of democracy is not electing "the right people" by a free and fair election, but by having a mechanism to remove bad leaders without violence. So a propaganda machine influencing elections is not ideal, but if it results in bad leaders, then that will become obvious to people at some point and they will vote them out. Elections will always have some random factors. Not everyone is going to vote. There will be fads. So election isn't the important part, UN-election is.
I mean, technically, the US does it every four years. At least, it used to. But that was just a blip in the historical record. We'll see what happens in four years from now to see which way the trend moves. Here's hoping it was truly a blip and not the start of a trend
This is not medical advice, but I have used rapamycin for an as of yet undiagnosed autoimmune condition (likely psoriatic arthritis vs rheumatoid arthritis) and it almost completely cures the condition while taking it, at the cost of mild-moderate hair loss and acne (which recovers on cessation of the drug). I cycle on and off rapamycin every few weeks to minimize side effects and whatever unknown long-term risks.
Summary of article: A single NPR link got automatically flagged by X to display a warning after NPR changed the URL for an unknown reason. It was reported to X who said it was a false positive and corrected it.
The details in the summary given by the previous poster were from the article. I don't know what distinction you think exists between a "hit piece" and "fact based reporting", but this article falls into the fact based reporting category.
I called it a hit piece because instead of saying users recently found a bug affecting 1 link to NPR. They say X "Caught" blocking links like it was something intentional and not an erroneous error. Any other site not connected to Elon with the same issue and it wouldn't of been writen I suspect.
> instead of saying users recently found a bug affecting 1 link to NPR
This is a good example of bias. It is not a fact that this was a bug. The fact is Twitter displayed the warning. Twitter has told NPR it was a false positive, but a journalist shouldn’t take an uncorroborated secondhand statement from the party being accused of potential wrongdoing as proof of anything.
Imagine for a second that this was done maliciously. Do you think Twitter would immediately admit that? Of course not, their statement would likely be identical to the one we got. Therefore the “fact based reporting” thing to do is present the facts, present Twitter’s response, and let the reader come to their own conclusions.
>It's a fact that a pattern of behavior was accused, but a pattern of behavior was not shown.
Yes, that is exactly what both the article and I said. The warning was applied to the link. Some people accused Twitter of doing that nefariously. Twitter denied it and claimed it was a mistake. Those are the facts of the situation.
The motivation for applying the warning or whether it was a mistake are not facts that can be confirmed by an independent journalist. They are speculation regardless of which side of the issue you come down on.
What are you talking about specifically? Can you point to a quote from the article that you think crosses a line journalistically? Because it seems like you’re equating reporting on the existence of accusations with actively making accusations and that type of thinking comes from either bias or a lack of media literacy.
> Who needs to know X erroneously blocked 1 link to NPR?
The appearance of a conflict of interest is a story. This appears to be a conflict of interest because a site is making it more difficult to read a negative story about a candidate endorsed by the company’s owner.
> Is it one link or many?
This is a rather pedantic complaint but multiple instances of the same link qualify as “links”. This is easier to see if you imagine them physically, multiple copies of the same book would be referred to as “books”.
reply