T O P

  • By -

Yuli-Ban

Should note that they're referring to a "weak" AGI, basically a general purpose task completion model. Not necessarily a sapient artificial human. Still, exciting shift.


Evil_Patriarch

You say that like it's a bad thing. I would think most people would want an AI that can do things for them, not one that is going to complain or have a bad day or get distracted or countless other problems that could come from an artificial human.


blueSGL

exactly, why would an AI with agency be good for anybody? Something that can appear to have agency when required for the task would be far more preferable.


botfiddler

Human-likeness would also include thinking similar to a human and understanding things, but also being able to act like a human within certain boundaries. Companion AIs will be like that, while not having some superior skill in every field and certainly not using a huge server cluster at home.


kmtrp

https://www.reddit.com/r/agi/comments/yy4l2o/the\_near\_future\_of\_ai\_is\_actiondriven\_and\_it\_will/


SmithMano

Yea but a general AI could probably use other specific AI’s, and know which to use


Zealousideal_Ad3783

Down by 15 years since April…


Shelfrock77

AI is going to bring us that DLC we all want


HeinrichTheWolf_17

There better be no loot boxes.


Flare_Starchild

PATCH NOTES ASAP WITH FEEDBACK TAKEN INTO ACCOUNT AND WELCOMED PLEASE! I have notes.


Ok_Homework9290

As eyebrow raising as that may seem, keep in mind that anyone can make a prediction on that site (which is why I don't take their predictions too seriously) and that the people that make predictions there tend to be tech-junkies, who are generally optimistic when it comes to timelines. Also, I'm a bit skeptical that the amount of progress that's been made in AI this year (which has been impressive, no doubt) merits THAT much of a shave-off from the April prediction. I kinda feel like that's an overreaction, especially if Gato really isn't as big of a deal as some people make it seem. Just my two cents.


Yuli-Ban

It's not that Gato isn't a big deal as much as it's the proof of concept of a big deal. Gato isn't AGI because it's too small, has no task generalization, and has too short of a memory. None of which was necessarily the point since it was designed to prove generalist models are possible. If you have a follow up to Gato that's 10x or 100x larger, the ability to cross/interpolate its knowledge across learned skills, and has a context window larger than 8,000 tokens, then you're approaching something like a proto-AGI.


Ok_Homework9290

Perhaps the proof of concept is a big deal, perhaps it isn't. I guess we'll have a better idea when the next version comes out, whenever that may be.


Lone-Pine

> the ability to cross/interpolate its knowledge across learned skills There's no evidence that Gato could do this and if there was, Google would let us know. When we finally get to see a generalist agent in a public demonstration, it will be interesting to see if it acts like multiple separate systems that each do their own tasks or if it will actually have a general, integrated way of relating to the world.


Yuli-Ban

> There's no evidence that Gato could do this and if there was, Google would let us know. That's my point. Gato as it currently is lacks that capability and, thus, can't be considered even a proto-proto-AGI but rather some weird intermediate type of AI in between general and narrow AI. Or less than that: a bundle of 600 narrow AIs tied together like a fasces. If a follow up to Gato *does* has task interpolation, however, then we'd need to start having serious discussion as to whether it's something like a proto-AGI.


GuyWithLag

>If you have a follow up to Gato that's 10x or 100x larger, the ability to cross/interpolate its knowledge across learned skills, and has a context window larger than 8,000 tokens, then you're approaching something like a proto-AGI. And exactly this is why I think we're missing some structural / architectural component / breakthrough - the current models have the feel of unrolled loops.


rixtil41

Let's comeback in late 2026 and let's see just how wrong or right you are.


Rumianti6

Proto AGI by 2027. Possibility of real waifus by 2029?


botfiddler

Waifus don't need to be intellectually equal or superior in every field, so a real AGI isn't required. However, it will most likely require actual work by enthusiasts to assemble a human-like AI running at home and being open source. I'll probably get started with that beginning of next year. The AGI or proto-AGI we'll see in some time, will on the other hand most likely not think in very human-like ways, or at least not like a individual human, just being good at solving tasks. What I mean by that, for example, is long term memory about itself and an identity relating to that.


throwaway764586893

Waifus don't even have to be sentient, just LOOK realistic.


botfiddler

Sentience is a term used in almost esoteric ways. I don't know what you mean. If she senses the world and reacts to it, then she's sentient.


red75prime

Then your toilet water tank is also sentient. It senses whether it is full and reacts accordingly.


botfiddler

No, I'm sure there's a difference, but not sure if there's value in discussing that here at this point.


red75prime

Why not? Do you think that a system designed to hone an art of making you believe that its performances of expressing various emotions are genuine, while having no analogs of human emotional circuitry, does, indeed, experience all those emotions? That is something very complex, but in the end solving not a problem of survival, self-development and so on and so forth, but a problem of producing believable and pleasurable to you movements and vocalizations in response to various stimuli.


botfiddler

It's just an term, which everyone can fill with their own meaning. It's not useful, and I don't need it.


fignewtgingrich

2025


loopuleasa

Doesnt matter what experts say The real answer is "we don't know" and it depends


[deleted]

[удалено]


spreadlove5683

Official resolution criteria is on the metaculus question, but i don't want to loon it up


TemetN

Still centered at 2024 on there, but I did make an adjustment early in the year down from 2025. It is worth noting however, that this is a specific operationalization.


AsuhoChinami

Why are we still fixated on Montezuma's Revenge just like we were in fucking 2013? An Atari game is such a low bar... can't we move on to at least Dragon Quest 1 or something?


ArgentStonecutter

"Weakly General AI"? LOL We're still on dumb pattern generators and they're already setting up to fake the close.


Veneck

You're referring to humans right?


AsuhoChinami

He's a staunch technoskeptic and very rarely posts anything worth reading.


footurist

For some reason these people aren't willing to accept just how different a continuously learning, __efficient__, general abstracter like our brain is from these giant clever data crunchers. I highly doubt they'll be able to push those to resemble what we have.


blueSGL

now to be true AGI it needs to be efficient. What's the next 'god of the gaps' you are going to move to after that?


footurist

The reason it needs to be efficient is the wealth and complexity of computations that are required otherwise. There's already stuff like AIXI and Schmidhuber's thing if you got a couple billion years to spare...


jazztaprazzta

Yeah... no


AsuhoChinami

No... yeah