Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

Planlagt Fastgjort Låst Flyttet Ikke-kategoriseret
40 Indlæg 32 Posters 35 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne tråd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • gossithedog@cyberplace.socialG This user is from outside of this forum
    gossithedog@cyberplace.socialG This user is from outside of this forum
    gossithedog@cyberplace.social
    wrote sidst redigeret af
    #1

    Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

    If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
    https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

    scottgal@hachyderm.ioS troed@swecyb.comT cigitalgem@sigmoid.socialC aqunt@piipitin.fiA carpetbomberz@mastodon.onlineC 26 Replies Last reply
    1
    0
    • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

      Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

      If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
      https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

      scottgal@hachyderm.ioS This user is from outside of this forum
      scottgal@hachyderm.ioS This user is from outside of this forum
      scottgal@hachyderm.io
      wrote sidst redigeret af
      #2

      @GossiTheDog Local models are getting good enough now (and uncensored) to make this trivial even for the inept pervert. Pandora's personal paedophillia producers' box is already open sadly.

      gossithedog@cyberplace.socialG 1 Reply Last reply
      0
      • scottgal@hachyderm.ioS scottgal@hachyderm.io

        @GossiTheDog Local models are getting good enough now (and uncensored) to make this trivial even for the inept pervert. Pandora's personal paedophillia producers' box is already open sadly.

        gossithedog@cyberplace.socialG This user is from outside of this forum
        gossithedog@cyberplace.socialG This user is from outside of this forum
        gossithedog@cyberplace.social
        wrote sidst redigeret af
        #3

        @scottgal that doesn't mean using child sexual abuse material images to train AI is okay.

        drhyde@fosstodon.orgD scottgal@hachyderm.ioS 2 Replies Last reply
        0
        • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

          Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

          If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
          https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

          troed@swecyb.comT This user is from outside of this forum
          troed@swecyb.comT This user is from outside of this forum
          troed@swecyb.com
          wrote sidst redigeret af
          #4

          @GossiTheDog this sounds pretty unbelievable tbh. LAION having "thousands" was a big public thing forcing re-release of the dataset. Others just piling on after this was discovered with no detection algorithms having been used??

          Amazon should really publish this information.

          https://petapixel.com/2024/09/03/major-ai-image-dataset-is-back-online-after-being-pulled-over-csam-laion-5b/

          wall_e@ioc.exchangeW 1 Reply Last reply
          0
          • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

            Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

            If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
            https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

            cigitalgem@sigmoid.socialC This user is from outside of this forum
            cigitalgem@sigmoid.socialC This user is from outside of this forum
            cigitalgem@sigmoid.social
            wrote sidst redigeret af
            #5

            @GossiTheDog reminder that recursive pollution remains a HUGE open problem with ML models.

            https://berryvilleiml.com/2026/01/10/recursive-pollution-and-model-collapse-are-not-the-same/

            1 Reply Last reply
            0
            • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

              @scottgal that doesn't mean using child sexual abuse material images to train AI is okay.

              drhyde@fosstodon.orgD This user is from outside of this forum
              drhyde@fosstodon.orgD This user is from outside of this forum
              drhyde@fosstodon.org
              wrote sidst redigeret af
              #6

              @GossiTheDog @scottgal they say they're not training on it, it was detected before training. But that's not the point. Amazon got the stuff from somewhere, and a decent person would report where it came from so that the rozzers can trace it back upstream. I flat out don't believe Amazon's claim to not know where it came from, they must know, because they must have got copyright clearance for making a derivative work from all that content 😉

              scottgal@hachyderm.ioS atlovato@mastodon.socialA 2 Replies Last reply
              0
              • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                aqunt@piipitin.fiA This user is from outside of this forum
                aqunt@piipitin.fiA This user is from outside of this forum
                aqunt@piipitin.fi
                wrote sidst redigeret af
                #7

                @GossiTheDog Can’t read the article so this is speculation: Amazon admitted having lots of CSAM but refuses to tell where they downloaded from? I thought holding on to CSAM is a crime in self, but as usual rules do not apply to big tech. And where did the material came from? Secret access to customer data they refuse to disclose?

                1 Reply Last reply
                0
                • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                  Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                  If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                  https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                  carpetbomberz@mastodon.onlineC This user is from outside of this forum
                  carpetbomberz@mastodon.onlineC This user is from outside of this forum
                  carpetbomberz@mastodon.online
                  wrote sidst redigeret af
                  #8

                  @GossiTheDog AI = CSAM

                  1 Reply Last reply
                  0
                  • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                    Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                    If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                    https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                    jonly@mastodon.socialJ This user is from outside of this forum
                    jonly@mastodon.socialJ This user is from outside of this forum
                    jonly@mastodon.social
                    wrote sidst redigeret af
                    #9

                    @GossiTheDog wasnt that confirmed to be the case years ago when all this ai bullshit started.
                    Like even if you just scrape the clear web youll likely scrape some of that shit

                    1 Reply Last reply
                    0
                    • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                      Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                      If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                      https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                      thirstybear@agilodon.socialT This user is from outside of this forum
                      thirstybear@agilodon.socialT This user is from outside of this forum
                      thirstybear@agilodon.social
                      wrote sidst redigeret af
                      #10

                      @GossiTheDog Non-paywall version here

                      https://archive.is/20260129113044/https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                      1 Reply Last reply
                      0
                      • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                        Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                        If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                        https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                        gossithedog@cyberplace.socialG This user is from outside of this forum
                        gossithedog@cyberplace.socialG This user is from outside of this forum
                        gossithedog@cyberplace.social
                        wrote sidst redigeret af
                        #11

                        As an aside, Microsoft had a publicly reported security incident a year or so ago where petabytes of data was left in a public Azure Storage Blob.

                        What they didn't say - that petabytes of data was customer photos of animals they'd classified and taken for AI work, t'was some grads just exporting stuff. Good job everybody is preaching about Responsible AI(tm).

                        masek@infosec.exchangeM 1 Reply Last reply
                        0
                        • drhyde@fosstodon.orgD drhyde@fosstodon.org

                          @GossiTheDog @scottgal they say they're not training on it, it was detected before training. But that's not the point. Amazon got the stuff from somewhere, and a decent person would report where it came from so that the rozzers can trace it back upstream. I flat out don't believe Amazon's claim to not know where it came from, they must know, because they must have got copyright clearance for making a derivative work from all that content 😉

                          scottgal@hachyderm.ioS This user is from outside of this forum
                          scottgal@hachyderm.ioS This user is from outside of this forum
                          scottgal@hachyderm.io
                          wrote sidst redigeret af
                          #12

                          @DrHyde @GossiTheDog Oh yeah I get that, sorry. I don't understand the ramifications of their possession, the originator's (presumably continued possession) of now identified CSAM material...which means they would be legally required to remove and report the user.
                          NO IDEA how they wouldn't have ANY moral qualms about NOT doing that nevermind what should be OBVIOUS legal liability (but corps are 'special' etc...)!

                          1 Reply Last reply
                          0
                          • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                            @scottgal that doesn't mean using child sexual abuse material images to train AI is okay.

                            scottgal@hachyderm.ioS This user is from outside of this forum
                            scottgal@hachyderm.ioS This user is from outside of this forum
                            scottgal@hachyderm.io
                            wrote sidst redigeret af
                            #13

                            @GossiTheDog BUT certain types of AI it would be obviously. THOSE need to exist in a regulated way and made open source. Like current PII scrubbing models it's a public good but I don't know any commercial company who COULD do it. Orthogonal sorry but just occurred to me...how do you get those models?

                            atlovato@mastodon.socialA 1 Reply Last reply
                            0
                            • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                              Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                              If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                              https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                              moses_izumi@fe.disroot.orgM This user is from outside of this forum
                              moses_izumi@fe.disroot.orgM This user is from outside of this forum
                              moses_izumi@fe.disroot.org
                              wrote sidst redigeret af
                              #14
                              In my country, the abbreviation CP only means cerebral palsy.

                              Med andra ord är GenAI-branschen fullständigt CP-skadad.

                              RE: https://cyberplace.social/@GossiTheDog/115978385132170439
                              1 Reply Last reply
                              0
                              • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                                Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                                If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                                https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                                mrundkvist@archaeo.socialM This user is from outside of this forum
                                mrundkvist@archaeo.socialM This user is from outside of this forum
                                mrundkvist@archaeo.social
                                wrote sidst redigeret af
                                #15

                                @GossiTheDog
                                Another headline here might be "Amazon admits in public to possessing a huge volume of child pornography".

                                1 Reply Last reply
                                0
                                • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                                  Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                                  If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                                  https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                                  jmcrookston@mastodon.socialJ This user is from outside of this forum
                                  jmcrookston@mastodon.socialJ This user is from outside of this forum
                                  jmcrookston@mastodon.social
                                  wrote sidst redigeret af
                                  #16

                                  @GossiTheDog

                                  What? Hand curation of trillions of issues didn't work?

                                  I'm shocked ayes tell ya, shocked!

                                  1 Reply Last reply
                                  0
                                  • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                                    As an aside, Microsoft had a publicly reported security incident a year or so ago where petabytes of data was left in a public Azure Storage Blob.

                                    What they didn't say - that petabytes of data was customer photos of animals they'd classified and taken for AI work, t'was some grads just exporting stuff. Good job everybody is preaching about Responsible AI(tm).

                                    masek@infosec.exchangeM This user is from outside of this forum
                                    masek@infosec.exchangeM This user is from outside of this forum
                                    masek@infosec.exchange
                                    wrote sidst redigeret af
                                    #17

                                    @GossiTheDog I would expect that they harvest open (no auth, indexable) S3 buckets for AI training.

                                    And you probably know what you find there ....

                                    imbrium_photography@mastodon.socialI 1 Reply Last reply
                                    0
                                    • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                                      Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                                      If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                                      https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                                      driusan@doomscroller.socialD This user is from outside of this forum
                                      driusan@doomscroller.socialD This user is from outside of this forum
                                      driusan@doomscroller.social
                                      wrote sidst redigeret af
                                      #18

                                      @GossiTheDog@cyberplace.social Sounds like police should be arresting and charging people at Amazon, then.

                                      1 Reply Last reply
                                      0
                                      • masek@infosec.exchangeM masek@infosec.exchange

                                        @GossiTheDog I would expect that they harvest open (no auth, indexable) S3 buckets for AI training.

                                        And you probably know what you find there ....

                                        imbrium_photography@mastodon.socialI This user is from outside of this forum
                                        imbrium_photography@mastodon.socialI This user is from outside of this forum
                                        imbrium_photography@mastodon.social
                                        wrote sidst redigeret af
                                        #19

                                        @masek @GossiTheDog But have they plundered Amazon S3 customer data, that the customers had set as private ?

                                        masek@infosec.exchangeM atlovato@mastodon.socialA 2 Replies Last reply
                                        0
                                        • gossithedog@cyberplace.socialG gossithedog@cyberplace.social

                                          Amazon have reported "hundreds of thousands" of pictures of child sexual abuse material found in shared AI training data... but is refusing to tell regulators which data sets.

                                          If you're using generative AI tools, there's a pretty good chance you're generating imagery with child porn training data behind the scenes.
                                          https://www.bloomberg.com/news/features/2026-01-29/amazon-found-child-sex-abuse-in-ai-training-data

                                          sassinake@mastodon.socialS This user is from outside of this forum
                                          sassinake@mastodon.socialS This user is from outside of this forum
                                          sassinake@mastodon.social
                                          wrote sidst redigeret af
                                          #20

                                          @GossiTheDog

                                          well there's your Epstein files right there!

                                          corax42@mastodon.socialC 1 Reply Last reply
                                          0
                                          Svar
                                          • Svar som emne
                                          Login for at svare
                                          • Ældste til nyeste
                                          • Nyeste til ældste
                                          • Most Votes


                                          • Log ind

                                          • Har du ikke en konto? Tilmeld

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          Graciously hosted by data.coop
                                          • First post
                                            Last post
                                          0
                                          • Hjem
                                          • Seneste
                                          • Etiketter
                                          • Populære
                                          • Verden
                                          • Bruger
                                          • Grupper