Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

Planlagt Fastgjort Låst Flyttet Ikke-kategoriseret
machinelearningllmresearch
86 Indlæg 57 Posters 0 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne tråd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • lunadragofelis@void.lgbtL lunadragofelis@void.lgbt
    @devsimsek I think AGI and self-improvement is possible. But definitely not with the technology (neural LLMs) that is being marketed as "AI" today.

    I think that AGI needs to be able to think logically.
    A This user is from outside of this forum
    A This user is from outside of this forum
    aoeuidhtns@app.wafrn.net
    wrote sidst redigeret af
    #58

    @devsimsek@universeodon.com @LunaDragofelis@void.lgbt

    if you make agi able to think logically then the world ends.
    we need to stop all ai research. if you are researching ai, and are not actively trying to sabotage it, then everyone's going to die.

    1 Reply Last reply
    0
    • dpiponi@mathstodon.xyzD dpiponi@mathstodon.xyz

      @Quantensalat @devsimsek There's a setup around equations (9) and (10) where the distribution used for training the next generation is a linear combination of the distribution your current generation generates and external data. As the amount of external data goes to zero, you expect model collapse. This is hardly surprising. I don't know anyone who expects you can just keep training based on previous results and expect something radically new to happen. (Though something *useful* can happen - eg. you may improve performance this way. See "rectification" in flow-matching.)

      Note that this doesn't rule out all forms of self-training - just one kind. As a concrete example, an LLM trained to generate code can learn from the output of the generated code. Such output is, in some sense, exogenous.

      rootwyrm@weird.autosR This user is from outside of this forum
      rootwyrm@weird.autosR This user is from outside of this forum
      rootwyrm@weird.autos
      wrote sidst redigeret af
      #59

      @dpiponi @Quantensalat @devsimsek that part, that is ultimately a rehash of well-known theory. THAT part IIRC goes back to like the 1940's or 1950's.

      And it absolutely rules out all forms of 'self-training.' It is not just mathematically impossible but a total logical fallacy. How can a system with no reference make correct determinations? Simple: it can't.

      resuna@ohai.socialR 1 Reply Last reply
      0
      • rootwyrm@weird.autosR This user is from outside of this forum
        rootwyrm@weird.autosR This user is from outside of this forum
        rootwyrm@weird.autos
        wrote sidst redigeret af
        #60

        @anne_twain @devsimsek this requires two components LLMs do not, cannot, and will not ever have. Intent and originality.
        Researchers have done self-modifying targeted things. It takes no time at all for things to become impossible for humans to understand. This does not mean they are better. Usually they weren't. Even when hyper-focused with specific controls.

        1 Reply Last reply
        0
        • huxley@furry.engineerH huxley@furry.engineer

          @devsimsek this is one of those things that seemed intuitive to us skeptics but it's great to see it proven

          lioh@social.anoxinon.deL This user is from outside of this forum
          lioh@social.anoxinon.deL This user is from outside of this forum
          lioh@social.anoxinon.de
          wrote sidst redigeret af
          #61

          @huxley @devsimsek doesn't scepticism and intuation mitigate each other?

          1 Reply Last reply
          0
          • alahmnat@woof.techA alahmnat@woof.tech

            @aka_quant_noir @devsimsek Oh I think we've achieved billionaire intelligence already. I just have a much dimmer view of billionaires.

            aka_quant_noir@hcommons.socialA This user is from outside of this forum
            aka_quant_noir@hcommons.socialA This user is from outside of this forum
            aka_quant_noir@hcommons.social
            wrote sidst redigeret af
            #62

            @alahmnat @devsimsek
            I think we're in the billionaire intelligence decline phase. They're going nuts.

            1 Reply Last reply
            0
            • devsimsek@universeodon.comD devsimsek@universeodon.com

              Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

              Not "we think it's unlikely." Not "it seems hard." Formally proved.

              The model doesn't climb toward AGI — it slowly forgets what reality looks like. They call it model collapse. The math calls it inevitable.
              I wrote about it 👇

              https://smsk.dev/2026/04/26/ai-cannot-self-improve-and-math-behind-proves-it/

              #AI #MachineLearning #LLM #Research

              paul@notnull.spaceP This user is from outside of this forum
              paul@notnull.spaceP This user is from outside of this forum
              paul@notnull.space
              wrote sidst redigeret af
              #63

              @devsimsek excellent. Thanks for the overview!

              1 Reply Last reply
              0
              • devsimsek@universeodon.comD devsimsek@universeodon.com

                Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

                Not "we think it's unlikely." Not "it seems hard." Formally proved.

                The model doesn't climb toward AGI — it slowly forgets what reality looks like. They call it model collapse. The math calls it inevitable.
                I wrote about it 👇

                https://smsk.dev/2026/04/26/ai-cannot-self-improve-and-math-behind-proves-it/

                #AI #MachineLearning #LLM #Research

                hermlon@yuustan.spaceH This user is from outside of this forum
                hermlon@yuustan.spaceH This user is from outside of this forum
                hermlon@yuustan.space
                wrote sidst redigeret af
                #64

                @devsimsek isn't the idea of self-improving AI that the AI modifies its code, so the underlying algorithm / architecture?

                1 Reply Last reply
                0
                • devsimsek@universeodon.comD devsimsek@universeodon.com

                  Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

                  Not "we think it's unlikely." Not "it seems hard." Formally proved.

                  The model doesn't climb toward AGI — it slowly forgets what reality looks like. They call it model collapse. The math calls it inevitable.
                  I wrote about it 👇

                  https://smsk.dev/2026/04/26/ai-cannot-self-improve-and-math-behind-proves-it/

                  #AI #MachineLearning #LLM #Research

                  lorxus@yiff.lifeL This user is from outside of this forum
                  lorxus@yiff.lifeL This user is from outside of this forum
                  lorxus@yiff.life
                  wrote sidst redigeret af
                  #65

                  @devsimsek @qualia I think you claim too much here. As I understand it, this result deals only with the intrinsic failures of RL-flavored approaches and not things like self-play, let alone problems that might arise from merely very good AI that still outdoes humans economically.

                  And I largely agree! I'm glad that someone's finally formalized the intuition that synthetic data is sawdust to bulk out real-world data with and more carefully investigated catastrophic forgetting and the general weaknesses of gradient descent.

                  That said... to what extent did you have Claude write this post? Because the format is... distinctive.

                  1 Reply Last reply
                  0
                  • dpiponi@mathstodon.xyzD dpiponi@mathstodon.xyz

                    @Quantensalat @devsimsek There's a setup around equations (9) and (10) where the distribution used for training the next generation is a linear combination of the distribution your current generation generates and external data. As the amount of external data goes to zero, you expect model collapse. This is hardly surprising. I don't know anyone who expects you can just keep training based on previous results and expect something radically new to happen. (Though something *useful* can happen - eg. you may improve performance this way. See "rectification" in flow-matching.)

                    Note that this doesn't rule out all forms of self-training - just one kind. As a concrete example, an LLM trained to generate code can learn from the output of the generated code. Such output is, in some sense, exogenous.

                    dpiponi@mathstodon.xyzD This user is from outside of this forum
                    dpiponi@mathstodon.xyzD This user is from outside of this forum
                    dpiponi@mathstodon.xyz
                    wrote sidst redigeret af
                    #66

                    @Quantensalat @devsimsek For something more formal on this subject see

                    https://arxiv.org/abs/2601.03220

                    The abstract starts "Can we learn more from data than existed in the generating process itself?"

                    1 Reply Last reply
                    0
                    • devsimsek@universeodon.comD devsimsek@universeodon.com

                      Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

                      Not "we think it's unlikely." Not "it seems hard." Formally proved.

                      The model doesn't climb toward AGI — it slowly forgets what reality looks like. They call it model collapse. The math calls it inevitable.
                      I wrote about it 👇

                      https://smsk.dev/2026/04/26/ai-cannot-self-improve-and-math-behind-proves-it/

                      #AI #MachineLearning #LLM #Research

                      rednikki@toot.bostonR This user is from outside of this forum
                      rednikki@toot.bostonR This user is from outside of this forum
                      rednikki@toot.boston
                      wrote sidst redigeret af
                      #67

                      @devsimsek “slowly forgets what reality looks like.” Sort of like billionaires.

                      1 Reply Last reply
                      0
                      • devsimsek@universeodon.comD devsimsek@universeodon.com

                        Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

                        Not "we think it's unlikely." Not "it seems hard." Formally proved.

                        The model doesn't climb toward AGI — it slowly forgets what reality looks like. They call it model collapse. The math calls it inevitable.
                        I wrote about it 👇

                        https://smsk.dev/2026/04/26/ai-cannot-self-improve-and-math-behind-proves-it/

                        #AI #MachineLearning #LLM #Research

                        troed@masto.sangberg.seT This user is from outside of this forum
                        troed@masto.sangberg.seT This user is from outside of this forum
                        troed@masto.sangberg.se
                        wrote sidst redigeret af
                        #68

                        @devsimsek The existence of humans disprove the paper.

                        resuna@ohai.socialR 1 Reply Last reply
                        0
                        • devsimsek@universeodon.comD devsimsek@universeodon.com

                          Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

                          Not "we think it's unlikely." Not "it seems hard." Formally proved.

                          The model doesn't climb toward AGI — it slowly forgets what reality looks like. They call it model collapse. The math calls it inevitable.
                          I wrote about it 👇

                          https://smsk.dev/2026/04/26/ai-cannot-self-improve-and-math-behind-proves-it/

                          #AI #MachineLearning #LLM #Research

                          aburka@hachyderm.ioA This user is from outside of this forum
                          aburka@hachyderm.ioA This user is from outside of this forum
                          aburka@hachyderm.io
                          wrote sidst redigeret af
                          #69

                          @devsimsek did an LLM write this toot or do LLMs just write like you 😅

                          1 Reply Last reply
                          0
                          • devsimsek@universeodon.comD devsimsek@universeodon.com

                            Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

                            Not "we think it's unlikely." Not "it seems hard." Formally proved.

                            The model doesn't climb toward AGI — it slowly forgets what reality looks like. They call it model collapse. The math calls it inevitable.
                            I wrote about it 👇

                            https://smsk.dev/2026/04/26/ai-cannot-self-improve-and-math-behind-proves-it/

                            #AI #MachineLearning #LLM #Research

                            anyia@lgbtqia.spaceA This user is from outside of this forum
                            anyia@lgbtqia.spaceA This user is from outside of this forum
                            anyia@lgbtqia.space
                            wrote sidst redigeret af
                            #70

                            @devsimsek "Don't worry bro, we can totally fix this by adding a committee of expert LLMs to reason about what training data to select, another committee of LLMs to plan the optimal training order, and then a larger one to evaluate the training output. We just need you to sign this cheque for our next three hyperscale GPU data centres..."

                            1 Reply Last reply
                            0
                            • rootwyrm@weird.autosR rootwyrm@weird.autos

                              @dpiponi @Quantensalat @devsimsek that part, that is ultimately a rehash of well-known theory. THAT part IIRC goes back to like the 1940's or 1950's.

                              And it absolutely rules out all forms of 'self-training.' It is not just mathematically impossible but a total logical fallacy. How can a system with no reference make correct determinations? Simple: it can't.

                              resuna@ohai.socialR This user is from outside of this forum
                              resuna@ohai.socialR This user is from outside of this forum
                              resuna@ohai.social
                              wrote sidst redigeret af
                              #71

                              @rootwyrm @dpiponi @Quantensalat @devsimsek

                              "How can a system with no reference make correct determinations? Simple: it can't."

                              Especially since it has no model of "correctness" other than "similar to the symbol streams the neural net weights were initialized from".

                              1 Reply Last reply
                              0
                              • troed@masto.sangberg.seT troed@masto.sangberg.se

                                @devsimsek The existence of humans disprove the paper.

                                resuna@ohai.socialR This user is from outside of this forum
                                resuna@ohai.socialR This user is from outside of this forum
                                resuna@ohai.social
                                wrote sidst redigeret af
                                #72

                                @troed @devsimsek

                                Large language models are fundamentally different from mammals on every level. They do not build models or reason about them. A rat is more "intelligent".

                                troed@masto.sangberg.seT 1 Reply Last reply
                                0
                                • rootwyrm@weird.autosR rootwyrm@weird.autos

                                  @devsimsek and this is old math, old theory, old knowledge. Gods do I wish I'd kept the various papers.

                                  We've literally known for over two decades that LLMs are dead-ends. It's why IBM spent billions hyper-focusing Watson. We already knew more context just made it worse, regardless of compute or method. It's not 'intelligence,' it's a bad search function. There's shit demonstrating that back to the 1980's.

                                  resuna@ohai.socialR This user is from outside of this forum
                                  resuna@ohai.socialR This user is from outside of this forum
                                  resuna@ohai.social
                                  wrote sidst redigeret af
                                  #73

                                  @rootwyrm @devsimsek

                                  Mark V. Shaney.

                                  1 Reply Last reply
                                  0
                                  • quantensalat@scicomm.xyzQ quantensalat@scicomm.xyz

                                    @devsimsek Is that a thing people believe, that LLMs generate themselves towards the singularity simply by eating their own output and no other feedback?

                                    wronglang@bayes.clubW This user is from outside of this forum
                                    wronglang@bayes.clubW This user is from outside of this forum
                                    wronglang@bayes.club
                                    wrote sidst redigeret af
                                    #74

                                    @Quantensalat @devsimsek the main issue is that unless you maintain an external signal (so human input in the form of token sequences that are actually carefully curated for coherence) the models become more and more incoherent. Sounds like you're on board with that. The next step is that we're quickly devaluing money spent on human creativity and the world is awash in LLM garbage. So the human signal *is* disappearing.

                                    quantensalat@scicomm.xyzQ 1 Reply Last reply
                                    0
                                    • quantensalat@scicomm.xyzQ quantensalat@scicomm.xyz

                                      @musicman @devsimsek As with all mathematical theorems, there's probably a not too far-fetched loophole circumventing some of their assumptions, doesn't mean skynet is becoming self-aware any time soon once that is the case.

                                      wronglang@bayes.clubW This user is from outside of this forum
                                      wronglang@bayes.clubW This user is from outside of this forum
                                      wronglang@bayes.club
                                      wrote sidst redigeret af
                                      #75

                                      @Quantensalat @musicman @devsimsek depends on what you mean by far fetched, certainly nothing as easy as "their more compute at it' which is what made this jump in investment so dramatic.

                                      quantensalat@scicomm.xyzQ 1 Reply Last reply
                                      0
                                      • devsimsek@universeodon.comD devsimsek@universeodon.com

                                        Researchers just mathematically proved that AI can't recursively self-improve its way to superintelligence.

                                        Not "we think it's unlikely." Not "it seems hard." Formally proved.

                                        The model doesn't climb toward AGI — it slowly forgets what reality looks like. They call it model collapse. The math calls it inevitable.
                                        I wrote about it 👇

                                        https://smsk.dev/2026/04/26/ai-cannot-self-improve-and-math-behind-proves-it/

                                        #AI #MachineLearning #LLM #Research

                                        emma@orbital.horseE This user is from outside of this forum
                                        emma@orbital.horseE This user is from outside of this forum
                                        emma@orbital.horse
                                        wrote sidst redigeret af
                                        #76

                                        @devsimsek so it doesn't get stuck in a local optimum, it hill-climbs a non-existent one?

                                        1 Reply Last reply
                                        0
                                        • musicman@mastodon.socialM musicman@mastodon.social

                                          @Quantensalat @devsimsek tech bros have been claiming their AIs are alive for years so if the average person who knows nothing about computers thinks we already have AGI, who can really blame them. Anthropic all but claims to have invented Terminator.

                                          Maybe something like this will stop the panic.

                                          Which is not to say people shouldn't be concerned in general and very specifically about environmental impacts

                                          M This user is from outside of this forum
                                          M This user is from outside of this forum
                                          mike805@noc.social
                                          wrote sidst redigeret af
                                          #77

                                          @musicman @Quantensalat @devsimsek Anyone who ever copied an audio tape (or worse a VHS tape) knows that the copy is always worse than the original. And in the video case, soon unwatchable.

                                          Ever heard a repeating echo on a video meeting that just turns to a buzz? Same phenomenon.

                                          So what you need is an AI that can perform experiments in the real world to learn how to do better whatever it is you want it to do.

                                          Inbreeding animals doesn't work too well either.

                                          1 Reply Last reply
                                          0
                                          Svar
                                          • Svar som emne
                                          Login for at svare
                                          • Ældste til nyeste
                                          • Nyeste til ældste
                                          • Most Votes


                                          • Log ind

                                          • Har du ikke en konto? Tilmeld

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          Graciously hosted by data.coop
                                          • First post
                                            Last post
                                          0
                                          • Hjem
                                          • Seneste
                                          • Etiketter
                                          • Populære
                                          • Verden
                                          • Bruger
                                          • Grupper