116
submitted 1 week ago by JRepin@lemmy.ml to c/technology@beehaw.org

cross-posted from: https://lemmy.ml/post/20858435

Will AI soon surpass the human brain? If you ask employees at OpenAI, Google DeepMind and other large tech companies, it is inevitable. However, researchers at Radboud University and other institutes show new proof that those claims are overblown and unlikely to ever come to fruition. Their findings are published in Computational Brain & Behavior today.

top 50 comments
sorted by: hot top controversial new old
[-] kbal@fedia.io 39 points 1 week ago

Meh. It's not a problem of scale. It's a problem of we have no idea how the fuck to do that. Scaling up existing techniques is neither necessary nor sufficient.

[-] JayDee@lemmy.ml 5 points 1 week ago

Right on the money. One of the big things with AI safety is "we have no fucking clue how AGI can originate so we are constantly in the dark." If we ever did create it, we likely would not immediately know it was AGI, and that creation could go very terribly in a number of ways.

[-] TranquilTurbulence@lemmy.zip 24 points 1 week ago* (last edited 1 week ago)

Sounds really counterintuitive to say that it’s impossible.

The article says that we would run out of computing power, and that’s definitely true for current hardware and software. It’s just that they are being developed all the time, so I think we need to leave that door open. Who knows how efficient things can get within the next decade or century. The article didn’t even mention any fundamental obstacle that would make AGI completely impossible. It’s not like AGI would be violating the laws of physics.

[-] JackOverlord@beehaw.org 16 points 1 week ago

Whenever I hear someone say that something is impossible with current technology, I think about my grandma. When she was a kid, only some important people had telephones. Doctors, police, etc.

In her lifetime we went from that to today, and, since she's still alive, even further into the future.

Whenever someone calls something impossible, I think about how far technology will progress in my own lifetime and I know that they've got no idea what they're talking about. (Unless, like you said, it's against the laws of physics. But sometimes even then I'm not so sure, cause it's not like we understand those entirely. )

[-] Saganaki@lemmy.one 13 points 1 week ago

That’s not an apt comparison.

More like “we’ll have flying cars 50 years from now.”

[-] megopie@beehaw.org 10 points 1 week ago

I love the flying car example because it reveals a huge issue with the whole “tech will get better” idea. People are still trying to make flying cars happen but it’s running in to the same fundamental issues; large things that are mechanically complex, energy intensive, and moving at high speeds in a crowded urban environments are just too expensive and dangerous.

There is no way around the physical realities, no clever trick or efficiency that will push it over some threshold of practicality.

[-] DdCno1@beehaw.org 13 points 1 week ago

The thing is, we have no idea where technological progress is taking us. So far, most predictions have been wrong. 50 to 60 years ago, people thought we would already be colonizing other planets by now. Barely anyone was able to predict the Internet, smartphones, social media, etc. - the kind of technology that is actually shaping our civilization's future right now.

Another aspect that I feel is often neglected is the assumption that technological progress will continue forever or at least continue at this current rapid pace. This wasn't true in the past and we might simply be experiencing a historical anomaly right now, one that could correct itself very soon in the future, either towards stagnation or even regression.

[-] d0ntpan1c@lemmy.blahaj.zone 6 points 1 week ago

The space example is extremely apt. Its possible we could have had tons of space stations, a moon colony, maybe even some other stuff going on around the solar system, asteroid mining, etc. But thay would have at least required the space race to continue longer and for spending to grow to create a big enoigh industry to ensure thay outcome, assuming no capacity or time issue. Alas, we took another path.

Something that seems important to us might not matter in even 10 years, or at least, not have a monetary and/or societal incentive to keep advancing.

[-] DdCno1@beehaw.org 6 points 1 week ago* (last edited 1 week ago)

I was also based on the assumption that the rapid progress of aerospace technology that happened in the 1920s to 1960s would continue onward at the same pace, whereas what actually happened was that barriers emerged that nobody was able to circumvent, like for example engineering things to withstand incredibly abrasive Moon dust (or really do anything productive on that lifeless rock), how to deal with the endless pitfalls of a long Mars journey, how to bring down the cost of launch vehicles so that grand projects like giant space stations would even be remotely possible (von Braun was already thinking about huge space stations all the way back in 1945). Many of these issues couldn't simply be solved by throwing more money at them, which is important. Deciders, both in Washington and Moscow, were smart enough to realize this in the 1970s, for the most part at least (the Space Shuttle and its Soviet clone, each a gigantic waste of money, are major counter example from this era).

The point I'm making here is that everyone assumed linear progress in this area, just like there are people currently making many billion dollar bets on linear progress in regards to computer technology in general and AI in particular, but at least, with the benefit of hindsight given past examples, there's a reasonable amount of doubt this time around.

[-] sydneybrokeit@beehaw.org 4 points 1 week ago

This wasn’t true in the past and we might simply be experiencing a historical anomaly right now

While our exact pacing might be slightly different from the pure extrapolation, human history has been a long, steady increase in the rate of invention. Access to education has meant that more people are making things, and then the next generations build on top of their work to make even bigger things.

[-] TranquilTurbulence@lemmy.zip 3 points 1 week ago

In addition, technological development can take unexpected twists and turns. For a while, it looked like analogue technology involving gears was going to solve every problem… until transistors were developed and mechanical calculators were soon forgotten. Also, the development of fertilizers revolutionized farming and and food production, which changed the world more than anyone even realized.

[-] massive_bereavement@fedia.io 11 points 1 week ago

Let's put it this way: If in our lifetime we can simulate the intelligence of a vinegar fly as general intelligence, that would be a monumental landmark in AGI. And we're far, far, far away from it.

As far as the iron age was from the metal alloys used in the Space Shuttle.

Talking about AGI simulating higher intelligence at the level of a dog or a cat, dear I say a pigeon or a crow is as far fetched as expecting ancient Egyptians to harness the power of the atom.

[-] JackOverlord@beehaw.org 6 points 1 week ago* (last edited 1 week ago)

Let's put it this way: If in our lifetime we can simulate the intelligence of a vinegar fly as general intelligence, that would be a monumental landmark in AGI. And we're far, far, far away from it.

I get what you mean here and I agree with it, if we're talking about current "AI", which isn't anywhere close. I know, because I've programmed some simple "AIs" (Mainly ML models) myself.

But your comparison to ancient egypt is somewhat lacking, considering we had the aptly named dark ages between then and now.

Lot's of knowledge got lost all the time during humanity's history, but ever since the printing press, and more recently the internet, came into existence, this problem has all but disappeared. As long as humanity doesn't nuke itself back to said dark ages, I recon we aren't that far away from AGI, or at least something close to it. Maybe not in my lifetime, but another ~2000 years seems a little extreme.

[-] GiveMemes@jlai.lu 5 points 1 week ago
load more comments (1 replies)
[-] princessnorah@lemmy.blahaj.zone 6 points 1 week ago

Actually, we do already know that we're close to a theoretical limit of increasing computing power as we currently know it. The transistor can't really get that much smaller, before it stops working.

Also, if you're talking about the article as linked, that is a mere introduction to a much longer paper.

[-] ContrarianTrail@lemm.ee 5 points 1 week ago

The fact that human brain is capable of general intelligence tells us everything we need to know about the processing power needed to run one.

[-] emr@lemmy.sdf.org 7 points 1 week ago

Well it sets an upper bound on compute requirements at 'simulate 10^27 atoms for thirty years' remains to be seen if what we can optimize away ever converges with what's feasible to build.

[-] orcrist@lemm.ee 4 points 1 week ago

The article did mention a fundamental obstacle. It said quite clearly that we would run out of resources before we had enough computing power. I suppose you could counter that by arguing that we could discover magic, or magical technology, or a lot of new resources through space exploration.

Of course things get more efficient. But in the past few decades they've gotten efficient in predictable, and mostly predicted, ways. It's certainly possible that totally unexpected things can happen. I could win the lottery next week. Is that the standard? Are you pushing the stance that says AGI is somewhat less likely than winning the lottery or getting struck by lightning, but by golly it's more than zero, how dare you suggest that it's anywhere close to zero?

load more comments (1 replies)
[-] ChairmanMeow@programming.dev 21 points 1 week ago

The actual paper is an interesting read. They present an actual computational proof, stating that even if you have essentially infinite memory, a computer that's a billion times faster than what we have now, perfect training data that you can sample without bias and you're only aiming for an AGI that performs slightly better than chance, it's still completely infeasible to do within the next few millenia. Ergo, it's definitely not "right around the corner". We're lightyears off still.

They prove this by proving that if you could train an AI in a tractable amount of time, you would have proven P=NP. And thus, training an AI is NP-hard. Given the minimum data that needs to be learned to be better than chance, this results in a ridiculously long training time well beyond the realm of what's even remotely feasible. And that's provided you don't even have to deal with all the constraints that exist in the real world.

We perhaps need some breakthrough in quantum computing in order to get closer. That is not to say that AI won't improve or anything, it'll get a bit better. But there is a computationally proven ceiling here, and breaking through that is exceptionally hard.

It also raises (imo) the question of whether or not we can truly consider humans to have general intelligence or not. Perhaps we're not as smart as we think we are either.

[-] BarryZuckerkorn@beehaw.org 10 points 1 week ago

The paper's scope is to prove that AI cannot feasibly be trained, using training data and learning algorithms, into something that approximates human cognition.

The limits of that finding are important here: it's not that creating an AGI is impossible, it's just that however it will be made, it will need to be made some other way, not by training alone.

Our squishy brains (or perhaps more accurately, our nervous systems contained within a biochemical organism influenced by a microbiome) arose out of evolutionary selection algorithms, so general intelligence is clearly possible.

So it may still be the case that AGI via computation alone is possible, and that creating such an AGI will not require solution of an NP-hard problem. But this paper closes one potential pathway that many believe is a viable pathway (if the paper's proof is actually correct, I definitely am not the person to make that evaluation). That doesn't mean they've proven there's no pathway at all.

[-] ChairmanMeow@programming.dev 4 points 1 week ago

Our squishy brains (or perhaps more accurately, our nervous systems contained within a biochemical organism influenced by a microbiome) arose out of evolutionary selection algorithms, so general intelligence is clearly possible.

That's assuming that we are a general intelligence. I'm actually unsure if that's even true.

That doesn't mean they've proven there's no pathway at all.

True, they've only calculated it'd take perhaps millions of years. Which might be accurate, I'm not sure to what kind of computer global evolution over trillions of organisms over millions of years adds up to. And yes, perhaps some breakthrough happens, but it's still very unlikely and definitely not "right around the corner" as the AI-bros claim (and that near-future thing is what the paper set out to disprove).

load more comments (3 replies)

A breakthrough in quantum computing wouldn't necessarily help. QC isn't faster than classical computing in the general case, it just happens to be for a few specific algorithms (e.g. factoring numbers). It's not impossible that a QC breakthrough might speed up training AI models (although to my knowledge we don't have any reason to believe that it would) and maybe that's what you're referring to, but there's a widespread misconception that Quantum computers are essentially non-deterministic turing machines that "evaluate all possible states at the same time" which isn't the case.

[-] ChairmanMeow@programming.dev 8 points 1 week ago

I was more hinting at that through conventional computational means we're just not getting there, and that some completely hypothetical breakthrough somewhere is required. QC is the best guess I have for where it might be but it's still far-fetched.

But yes, you're absolutely right that QC in general isn't a magic bullet here.

Yeah thought that might be the case! It's just a thing that a lot of people have misconceptions about so it's something that I have a bit of a knee jerk reaction to.

[-] ChairmanMeow@programming.dev 5 points 1 week ago

Haha it's good that you do though, because now there's a helpful comment providing more context :)

[-] Umbrias@beehaw.org 3 points 1 week ago

the limitation is specifically using the primary machine learning technique, same one all chatbots use at places claiming to pursue agi, which is statistical imitation, is np-hard.

load more comments (2 replies)
load more comments (2 replies)
[-] ContrarianTrail@lemm.ee 17 points 1 week ago* (last edited 1 week ago)

AGI is inevitable unless:

  1. General intelligence is substrate independent and what the brain does cannot be replicated in silica. However, since both are made of matter, and matter obeys the laws of physics, I see no reason to assume this.

  2. We destroy ourselves before we reach AGI.

Other than that, we will keep incrementally improving our technology and it's only a matter of time untill we get there. May take 5 years, 50 or 500 but it seems pretty inevitable to me.

[-] rysiek@mstdn.social 13 points 1 week ago* (last edited 1 week ago)

@ContrarianTrail @JRepin well I guess somebody would first need to clearly define what "AGI" is. Currently it's just "whatever the techbro hypers want it to be".

And then there's the matter (ha!) of your assumption that we understand all laws of physics necessary that "matter obeys", or that we can reasonably understand them. That's a pretty strong assumption: individual human minds are pretty limited and communication adds overhead, and we might reach a point where we're stuck.

[-] ContrarianTrail@lemm.ee 11 points 1 week ago

A chess engine is intelligent in one thing: playing chess. That narrow intelligence doesn’t translate to any other skill, even if it's sometimes superhuman at that one task, like a calculator.

Humans, on the other hand, are generally intelligent. We can perform a variety of cognitive tasks that are unrelated to each other, with our only limitations being the physical ones of our "meat computer."

Artificial General Intelligence (AGI) is the artificial version of human cognitive capabilities, but without the brain's limitations. It should be noted that AGI is not synonymous with AI. AGI is a type of AI, but not all AI is generally intelligent. The next step from AGI would be Artificial Super Intelligence (ASI), which would not only be generally intelligent but also superhumanly so. This is what the "AI doomers" are concerned about.

[-] rysiek@mstdn.social 6 points 1 week ago* (last edited 1 week ago)

@ContrarianTrail

> A chess engine is intelligent in one thing: playing chess

No. That's not how the adjective "intelligent" works, outside of marketing drivel of course ("intelligent washing machine" etc).

> Artificial General Intelligence (AGI) is the artificial version of human cognitive capabilities

Can you give a definition of "intelligence" or "human cognitive abilities" that would allow us to somehow unequivocably establish that "X is intelligent" or "X has human cognitive abilities"?

[-] JayDee@lemmy.ml 4 points 1 week ago

IIRC, within computer science, which is the field most heavily driving AI design and research forward, an 'intelligent agent' is essentially defined as any 'agent' which takes external stimulai from a collection of sensors in some form of environment, processes that stimulai in a dynamic fashion (one of the criteria IIRC is a branching decision tree based on the stimulai), and then applies that processing to a collection of affectors in the environment.

Yes, this definition is an extremely low bar and includes a massive amount of code, software and scripts. It also includes basic natural intelligences such as worms, ants, amoeba, and even viruses. One example of mechanical AI are some of Theo Jansen's StrandBeasts

load more comments (8 replies)
load more comments (3 replies)
load more comments (2 replies)
[-] vrighter@discuss.tchncs.de 11 points 1 week ago

incremental improvements on a dead end, still gets you to the dead end.

[-] ContrarianTrail@lemm.ee 5 points 1 week ago

Then you need to give me an explanation for why it's a dead end

[-] vrighter@discuss.tchncs.de 5 points 1 week ago

because, having coded them myself, I am under no illusions as to their capabilities. They are not magic. "just" some matrix multiplications that generate a probability distribution for the next token, which is then randomly sampled.

[-] ContrarianTrail@lemm.ee 5 points 1 week ago* (last edited 1 week ago)

You seem to be talking about LLMs now and I'm not. LLMs being a dead end is perfectly compatible with what I just said. We'll just try a different approach next then. Even the fact of realising they're a dead end is yet another step towards AGI.

[-] vrighter@discuss.tchncs.de 5 points 1 week ago

yeah, so that means that it's not incremental improvement on what we have that we need. That will get us nowhere. We need a (as yet unknown) completely different approach. Which is the opposite of incremental improvement.

[-] ContrarianTrail@lemm.ee 4 points 1 week ago

I didn't say we need to improve on what we have. We just need to keep making better technology which we will keep doing unless we destroy ourselves first.

load more comments (6 replies)

Another possibility is that humans just aren't smart enough to figure out AGI. While I'm sure that we will continue incrementally improving technology in some form, it's not at all self-evident that these improvements will eventually add up to AGI.

load more comments (3 replies)
[-] Steve@communick.news 9 points 1 week ago

Will AI soon surpass the human brain?
If you ask employees at OpenAI, Google DeepMind and other large tech companies, it is inevitable.

That doesn't answer the question.
If it will happen is unrelated to When it will happen.
I'd expect we'll see AGI some time between the next 20 and 200 years. I think that's pretty soon. You may not.

[-] ContrarianTrail@lemm.ee 6 points 1 week ago

If there were a giant asteroid hurling toward Earth, set to impact sometime in the next 20 to 200 years, I’d say there’s definitely a need for urgency. A true AGI is somewhat of an asteroidal impact in itself.

[-] webghost0101@sopuli.xyz 3 points 1 week ago

A single AGI would not be to different from a human. But it may not take long for AGI to develop ASI, superior to human intelligence.

Thats not an astronaut impact but alien contact

load more comments (1 replies)
load more comments (1 replies)
[-] Banzai51@midwest.social 6 points 1 week ago

The steam engine won't replace John Henry!!!

[-] Vodulas@beehaw.org 6 points 1 week ago

Not really a good comparison. The steam engine was an extant technology at that point. AGI is not, and we really no idea if/when it will be. One thing is clear though, it is not as close on the horizon as tech bros want us to think it is.

[-] Aggravationstation@feddit.uk 6 points 1 week ago

Possible or not I don't think we'll get to the point of AGI. I'm pretty sure at some point someone will do something monumentally stupid with AI that will wipe out humanity.

[-] drwho@beehaw.org 9 points 1 week ago

Like wrecking the biosphere in its persuit.

load more comments (3 replies)
[-] SplashJackson@lemmy.ca 4 points 1 week ago

I like SCUMM but AGI is okay I just don't like typing commands

load more comments
view more: next ›
this post was submitted on 07 Oct 2024
116 points (100.0% liked)

Technology

37664 readers
464 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS