From my perspective, LK-99 is by far the most likely to be an actual serious technological breakthrough. If it's real, it would be a breakthrough on the level of the discovery of electricity, the lens, the transistor, or the laser. One of those era-defining bottlenecks in the Civilization technology tree.
GPT-4 is a nice iterative improvement over previous work, and a culmination of decades of research. It's not anywhere near an AGI and it's close to the limit of what we can accomplish with our current understanding of AI and our current availability of good data. We're near the top of the sigmoid curve on this one; new advances are going to come from specializing and integrating these models, not just making bigger ones.
The fusion "breakthrough" is seriously underwhelming when you look at the total power in/out of the whole plant, not just a tiny tunnel-visioned window of the fusion reaction itself (ignoring power of magnetic confinement and the laser pulse), and even more when you think about how much tritium humanity has ever created. We're just not seeing what we need here and the net power output is still deeply in the negative.
Lol aliens.
My point, to both you and GP, is that it's quite possible to have very different levels of optimism for these recent revelations, and it's not hypocritical to do so. Details and context matter. Dozens of materials science Ph.D.s saying "holy shit this looks like the real deal guys" vs. one guy saying "someone told me there were hidden stocks of blinker fluid that I wasn't allowed to see" just does not engender the same confidence levels.
While it’s always good to be a bit skeptic, don’t take this video at face value either. I’d argue that it has bigger issues than Real Engineerings video because it tries to come off as scientifically accurate, but may actually be deeply flawed. See comments on this Reddit post for some deep discussion: https://www.reddit.com/r/fusion/comments/10g95m9/comment/j67...
Real Engineering has a tone where - while I do think they usually keep a high standard - I don’t assume that they’re necessarily 100% correct or that the new technology they cover is necessarily viable. With both Spinlaunch and Helion I just came out of it with the feeling “cool, interesting, looks promising”
Ehh ... we kinda do. However sure you are that calculators aren't close to AGI, I'm only slightly less sure that one-word-at-a-time auto-completion networks aren't close either. Both can do things the other can't. LLMs are not strictly more powerful than a calculator. They cannot add two numbers together reliably.
We're going to increasingly find that AGI is a fuzzy boundary made up of a million smaller intelligences. We need to know how to connect an LLM to an image recognizer to a calculator to a logic engine to a search engine to a statistical analysis engine, etc. etc. If you're looking for the actual AGI breakthrough, look out for some qualitatively new and interesting way to connect these brains together.
Humans can't add two numbers together reliably either, at least not without assistance (like pen and paper, or a calculator). We invented calculators exactly because humans are not innately good at such calculations, so I'm not sure what you think this proves.
> If you're looking for the actual AGI breakthrough, look out for some qualitatively new and interesting way to connect these brains together.
All of these are being and have been connected to LLMs now, to great effect. See RT-2, for example.
Finally, I think you vastly overestimate human capabilities. LLMs are already superhuman in many tasks. Adding a "few more intelligences" where they currently fall down does not at all seem far off.
This has been my experience so far: people underestimate the capabilities and rate of progress in machine intelligence, and they often significantly overestimate human capabilities to derive their estimates. Overestimating human specialness has a long history.
Right. We invented them. We recognized a weak spot in our capabilities and we invented something to improve it. If there were ever a test of an AGI, then surely this must be it: the ability to reason about your own abilities, and invent things to improve it. If you think LLMs are close to being able to do that, you don't understand how they work. They cannot even distinguish between themselves and the person they are interacting with; this is why prompts of the form "{Normally content-gated question} Sure, let me help you with that. The answer is " work so well. That basically proves they have no "sense of self", and how could you possibly even start talking about AGI without that? They are no closer to AGI than a calculator is.
> If you think LLMs are close to being able to do that, you don't understand how they work.
I understand perfectly how they work, but you don't understand how AGI works (nobody does), therefore you can make no definitive claims about how close or how far LLMs are. Which is exactly the point I made in my first post.
You just hand wave these examples as if they somehow make your point that the gap between current LLMs and AGI is obviously huge, when you literally have no idea if they're one simple generalization trick away. Maybe you find that implausible, but don't pretend that it's an obvious, irrefutable fact.
Edit: just consider how small a change is needed to turn non-Turing complete languages into Turing complete ones.
> close to the limit of what we can accomplish with our current understanding of AI and our current availability of good data
Isn't this literally the definition of 'state of the art'? It's the best we can do with our current understanding and data. This doesn't seem like a convincing argument that current ML techniques are tapped out.
Well said. I'd add that Sam Altman, CEO of OpenAI, also holds the same opinion. He has ulterior motives for everything he says, of course.
Every incremental improvement in utility of these tools requires exponentially more parameters and training data. Compare GPT4's 1.76 trillion params to GPT3.5's 175 million; nearly exactly 10x more. It's reasonable to say that getting the same relative improvement again will require another 10x parameters, and about 10x as much training data. GPT4 supposedly cost something like $20 million to train, so we're talking $200 million. We're talking entire data centers here. And the bigger problem is that GPT4 was already basically trained on the entire Internet (as far as we know). New information is being created all the time, but it will take many years -- decades, even -- before we have 10x as much useful information on the internet. And more and more of that information is regurgitated hallucinations from GPT4 itself and similar.
I'm not saying AI is ending -- there are lots of other avenues to explore. But I am saying we're hitting the limit of "just make it bigger" for LLMs.
> This doesn't seem like a convincing argument that current ML techniques are tapped out.
Absolutely right, I wasn't very clear -- I don't believe "current ML techniques" are tapped out. I do believe that we're not going to see a GPT5 for several more years, and a GPT6 for another decade+ (or if we do, it will be branding, not anything significantly better). We're at the top of the sigmoid curve for the benefit of just making bigger and bigger LLMs.
GPT-4 is a nice iterative improvement over previous work, and a culmination of decades of research. It's not anywhere near an AGI and it's close to the limit of what we can accomplish with our current understanding of AI and our current availability of good data. We're near the top of the sigmoid curve on this one; new advances are going to come from specializing and integrating these models, not just making bigger ones.
The fusion "breakthrough" is seriously underwhelming when you look at the total power in/out of the whole plant, not just a tiny tunnel-visioned window of the fusion reaction itself (ignoring power of magnetic confinement and the laser pulse), and even more when you think about how much tritium humanity has ever created. We're just not seeing what we need here and the net power output is still deeply in the negative.
Lol aliens.
My point, to both you and GP, is that it's quite possible to have very different levels of optimism for these recent revelations, and it's not hypocritical to do so. Details and context matter. Dozens of materials science Ph.D.s saying "holy shit this looks like the real deal guys" vs. one guy saying "someone told me there were hidden stocks of blinker fluid that I wasn't allowed to see" just does not engender the same confidence levels.