Skip to content
  • Hjem
  • Seneste
  • Etiketter
  • Populære
  • Verden
  • Bruger
  • Grupper
Temaer
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Kollaps
FARVEL BIG TECH
  1. Forside
  2. Ikke-kategoriseret
  3. AI is not inevitable.

AI is not inevitable.

Planlagt Fastgjort Låst Flyttet Ikke-kategoriseret
48 Indlæg 10 Posters 17 Visninger
  • Ældste til nyeste
  • Nyeste til ældste
  • Most Votes
Svar
  • Svar som emne
Login for at svare
Denne tråd er blevet slettet. Kun brugere med emne behandlings privilegier kan se den.
  • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

    @olivia @apostolis I don’t think I said you are against different analyses?

    the point I was trying to make is simply that what is breaking things right now is a confluence of forces and actors. If we are going to counter the destructive effects we need a systemic analysis of how these forces are interacting.

    I don’t take you to be someone who would object to that in principle 😉

    I suspect what we do have disagreements on is what the relative importance of these different forces and actors are, and what’s required to push back as a result (even in principle)

    olivia@scholar.socialO This user is from outside of this forum
    olivia@scholar.socialO This user is from outside of this forum
    olivia@scholar.social
    wrote sidst redigeret af
    #24

    @UlrikeHahn @apostolis

    "Most importantly of all, resistance can and should take on many forms. Remember to rest and take care of yourself and your community. If talking to friends and colleagues is easy, then try to engage them on these issues. If it is not possible to do so, you can instead (or in addition) seek out allies online."

    https://olivia.science/before/#can

    1 Reply Last reply
    0
    • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

      @apostolis @olivia the reason why this ultimately matters that pushing back against the real driver (the “organic” adoption of these tools by individuals) requires me to understand and engage with the perceived value and function these tools have for them…

      …and that means trying to understand both what they can and what they can’t do. Simply declaring that these tools are garbage (“semantically meaningless random text generator”) isn’t useful for actually productively countering AI use in this configuration…(if they genuinely were meaningless random text generators I wouldn’t be faced with the negative effects in the first place).

      the Fodor quote doesn’t feel like it’s aimed at that kind of understanding

      abucci@buc.ciA This user is from outside of this forum
      abucci@buc.ciA This user is from outside of this forum
      abucci@buc.ci
      wrote sidst redigeret af
      #25
      @UlrikeHahn@fediscience.org @apostolis@social.coop @olivia@scholar.social There is very little that could be credibly called organic adoption when it comes to AI. It is being fiercely pushed in support of multiple hundreds of billion dollar investment. People are being told repeatedly, in every channel, that AI is inevitable, is here to stay, etc. It is disingenuous to place this responsibility at the feet of students, throw up your hands, or ask someone else to tell you what to do about it. That kind of behavior from people empowered to know and do better is the problem.
      ulrikehahn@fediscience.orgU lednabm@stranger.socialL 2 Replies Last reply
      0
      • olivia@scholar.socialO olivia@scholar.social

        AI is not inevitable. Nothing in human societies is inevitable because we design them. Healthcare can be free for the public. Books can be bought instead of bombs. Universities can be free for students, and they can even receive a stipend to live off. Don't let companies dictate the future.

        Read more in section 3.2 here https://doi.org/10.5281/zenodo.17065099

        yhancik@thereisno.computerY This user is from outside of this forum
        yhancik@thereisno.computerY This user is from outside of this forum
        yhancik@thereisno.computer
        wrote sidst redigeret af
        #26

        @olivia this feels good to read again just the day after we (finally) had a first meeting in school to discuss AI (and of course I shared the paper!).

        But god, the belief in inevitability is so deeply engrained.

        1 Reply Last reply
        0
        • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

          @apostolis @olivia no disagreement with that!

          lednabm@stranger.socialL This user is from outside of this forum
          lednabm@stranger.socialL This user is from outside of this forum
          lednabm@stranger.social
          wrote sidst redigeret af
          #27

          @UlrikeHahn @apostolis @olivia

          Wow.... interesting discussion, folks. Thank you. I'm a long way from university level experience, being an engineer in the electronic design industry for over 40 years. We've gone from one computer to share among engineers thru now to AI assistance across our individual computers. IMHO, we need to separate what AI can do from what they do. Humans, almost instinctively anthropomorphise everything. FFS... people still worship an imaginary AI in the sky and.... 1/2

          lednabm@stranger.socialL 1 Reply Last reply
          0
          • lednabm@stranger.socialL lednabm@stranger.social

            @UlrikeHahn @apostolis @olivia

            Wow.... interesting discussion, folks. Thank you. I'm a long way from university level experience, being an engineer in the electronic design industry for over 40 years. We've gone from one computer to share among engineers thru now to AI assistance across our individual computers. IMHO, we need to separate what AI can do from what they do. Humans, almost instinctively anthropomorphise everything. FFS... people still worship an imaginary AI in the sky and.... 1/2

            lednabm@stranger.socialL This user is from outside of this forum
            lednabm@stranger.socialL This user is from outside of this forum
            lednabm@stranger.social
            wrote sidst redigeret af
            #28

            @UlrikeHahn @apostolis @olivia 2/2 ... call it their god(s). I think the best resistance is to cooperate. After all, no matter how human these things can seem, they will never be more than tools. As humans, we "feel" a lot. We need to not let our feelings blind us to what these new tools can do. I'm no teacher. I've found that my method of communication doesn't do well explaining to others how to think, instead of what to think. I just know the tools we use evolve all the time....

            1 Reply Last reply
            0
            • abucci@buc.ciA abucci@buc.ci
              @UlrikeHahn@fediscience.org @apostolis@social.coop @olivia@scholar.social There is very little that could be credibly called organic adoption when it comes to AI. It is being fiercely pushed in support of multiple hundreds of billion dollar investment. People are being told repeatedly, in every channel, that AI is inevitable, is here to stay, etc. It is disingenuous to place this responsibility at the feet of students, throw up your hands, or ask someone else to tell you what to do about it. That kind of behavior from people empowered to know and do better is the problem.
              ulrikehahn@fediscience.orgU This user is from outside of this forum
              ulrikehahn@fediscience.orgU This user is from outside of this forum
              ulrikehahn@fediscience.org
              wrote sidst redigeret af
              #29

              @abucci @apostolis @olivia I’m going to point you toward the scare quotes around the word “organic” in my post, which are there for precisely those reasons.

              I am also going to push back against the notion that I am “placing the responsibility at the feet of students”: I am simply describing the (widely documented) problem in higher education that students are using AI tools in significant volumes *even where there use is explicitly sanctioned and forbidden*.

              That is the concrete problem of AI now undermining higher education. Asking what “resisting AI” is supposed to mean for me in that context seems legitimate to me, and if it’s not, Olivia (who I’ve known for a long time as an academic colleague) is more than capable of telling me that herself.

              abucci@buc.ciA 1 Reply Last reply
              0
              • abucci@buc.ciA abucci@buc.ci
                @UlrikeHahn@fediscience.org @apostolis@social.coop @olivia@scholar.social There is very little that could be credibly called organic adoption when it comes to AI. It is being fiercely pushed in support of multiple hundreds of billion dollar investment. People are being told repeatedly, in every channel, that AI is inevitable, is here to stay, etc. It is disingenuous to place this responsibility at the feet of students, throw up your hands, or ask someone else to tell you what to do about it. That kind of behavior from people empowered to know and do better is the problem.
                lednabm@stranger.socialL This user is from outside of this forum
                lednabm@stranger.socialL This user is from outside of this forum
                lednabm@stranger.social
                wrote sidst redigeret af
                #30

                @abucci @UlrikeHahn @apostolis @olivia

                Wouldn't an approach where the AIs have to pass the class as students, be better. After all, regurgitating data is not the way to learn how to think. As for the pollitical/economics of the whole mess, well, that's on us to some extent. It's a problem educated people deal with all the time, even among each other. IMHO, humanity is still growing up. We've not abandoned our superstitions for the hard real wonder of actual nature. Is AI part of our nature?

                ulrikehahn@fediscience.orgU abucci@buc.ciA teledyn@mstdn.caT 3 Replies Last reply
                0
                • jwcph@helvede.netJ jwcph@helvede.net shared this topic
                • lednabm@stranger.socialL lednabm@stranger.social

                  @abucci @UlrikeHahn @apostolis @olivia

                  Wouldn't an approach where the AIs have to pass the class as students, be better. After all, regurgitating data is not the way to learn how to think. As for the pollitical/economics of the whole mess, well, that's on us to some extent. It's a problem educated people deal with all the time, even among each other. IMHO, humanity is still growing up. We've not abandoned our superstitions for the hard real wonder of actual nature. Is AI part of our nature?

                  ulrikehahn@fediscience.orgU This user is from outside of this forum
                  ulrikehahn@fediscience.orgU This user is from outside of this forum
                  ulrikehahn@fediscience.org
                  wrote sidst redigeret af
                  #31

                  @lednaBM @abucci @apostolis @olivia if I understand you correctly, you are suggesting we, in an sense, embrace AI and treat it in such a way that makes it better (ie accept it as students)? if yes, I don’t personally really want to make AI systems ‘better’ - they are causing huge damage and disruption at current levels of performance. I’d personally rather put a brake on that.

                  lednabm@stranger.socialL 1 Reply Last reply
                  0
                  • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

                    @lednaBM @abucci @apostolis @olivia if I understand you correctly, you are suggesting we, in an sense, embrace AI and treat it in such a way that makes it better (ie accept it as students)? if yes, I don’t personally really want to make AI systems ‘better’ - they are causing huge damage and disruption at current levels of performance. I’d personally rather put a brake on that.

                    lednabm@stranger.socialL This user is from outside of this forum
                    lednabm@stranger.socialL This user is from outside of this forum
                    lednabm@stranger.social
                    wrote sidst redigeret af
                    #32

                    @UlrikeHahn @abucci @apostolis @olivia

                    I understand what you're saying, and maybe language is not serving us well. You seemed to have juxtaposed helping it be better versus creating a disruption. And again, maybe I fully don't understand the dilemma. When I need very technical information that I can not recall or need help with, I would go to a book or a specification. Now, I can ask AI, check its results, and decide whether I can rely upon what's being presented. It's a tool... 1/2

                    lednabm@stranger.socialL 1 Reply Last reply
                    0
                    • lednabm@stranger.socialL lednabm@stranger.social

                      @UlrikeHahn @abucci @apostolis @olivia

                      I understand what you're saying, and maybe language is not serving us well. You seemed to have juxtaposed helping it be better versus creating a disruption. And again, maybe I fully don't understand the dilemma. When I need very technical information that I can not recall or need help with, I would go to a book or a specification. Now, I can ask AI, check its results, and decide whether I can rely upon what's being presented. It's a tool... 1/2

                      lednabm@stranger.socialL This user is from outside of this forum
                      lednabm@stranger.socialL This user is from outside of this forum
                      lednabm@stranger.social
                      wrote sidst redigeret af
                      #33

                      @UlrikeHahn @abucci @apostolis @olivia 2/2 tools generally need calibration. Is it possible to use the disruption you speak of as a teaching moment? I don't know. Am I being foolish about the political/economic consequences of those benefitting from the disruption? Maybe. I agree with the original poster. We should have free education, health care, and representation in the way we govern ourselves. Problem there is, IMHO, the white elephant that is religion working against secular human values..

                      ulrikehahn@fediscience.orgU 1 Reply Last reply
                      0
                      • lednabm@stranger.socialL lednabm@stranger.social

                        @UlrikeHahn @abucci @apostolis @olivia 2/2 tools generally need calibration. Is it possible to use the disruption you speak of as a teaching moment? I don't know. Am I being foolish about the political/economic consequences of those benefitting from the disruption? Maybe. I agree with the original poster. We should have free education, health care, and representation in the way we govern ourselves. Problem there is, IMHO, the white elephant that is religion working against secular human values..

                        ulrikehahn@fediscience.orgU This user is from outside of this forum
                        ulrikehahn@fediscience.orgU This user is from outside of this forum
                        ulrikehahn@fediscience.org
                        wrote sidst redigeret af
                        #34

                        @lednaBM @abucci @apostolis @olivia I think one of the problems, particularly in the context of education, lies in the idea that “now I can use AI to give me an answer and check the results”. It is precisely the “ability to check the results” in a particular scientific or academic discipline that higher education degrees are trying to provide. Leaning on AI to “find” answers by students is undermining the learning of the skills that underpin “the ability to check”.

                        lednabm@stranger.socialL 1 Reply Last reply
                        0
                        • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

                          @lednaBM @abucci @apostolis @olivia I think one of the problems, particularly in the context of education, lies in the idea that “now I can use AI to give me an answer and check the results”. It is precisely the “ability to check the results” in a particular scientific or academic discipline that higher education degrees are trying to provide. Leaning on AI to “find” answers by students is undermining the learning of the skills that underpin “the ability to check”.

                          lednabm@stranger.socialL This user is from outside of this forum
                          lednabm@stranger.socialL This user is from outside of this forum
                          lednabm@stranger.social
                          wrote sidst redigeret af
                          #35

                          @UlrikeHahn @abucci @apostolis @olivia
                          That's a great point. Teaching youth only to rely upon AI sounds like a mistake. I guess I have trouble with the notion that AI is anything more than a tool. Its applications threaten a lot, probably a lot beyond its scope, but not beyond its profit scam. Hopefully, some applications are identified as misapplications. I'm reminded of Huxleys Brave New World. Will AI be the soma drug to placate the masses, even though they were designed to be placated.

                          aoanla@hachyderm.ioA 1 Reply Last reply
                          0
                          • lednabm@stranger.socialL lednabm@stranger.social

                            @UlrikeHahn @abucci @apostolis @olivia
                            That's a great point. Teaching youth only to rely upon AI sounds like a mistake. I guess I have trouble with the notion that AI is anything more than a tool. Its applications threaten a lot, probably a lot beyond its scope, but not beyond its profit scam. Hopefully, some applications are identified as misapplications. I'm reminded of Huxleys Brave New World. Will AI be the soma drug to placate the masses, even though they were designed to be placated.

                            aoanla@hachyderm.ioA This user is from outside of this forum
                            aoanla@hachyderm.ioA This user is from outside of this forum
                            aoanla@hachyderm.io
                            wrote sidst redigeret af
                            #36

                            @lednaBM @UlrikeHahn @abucci @apostolis @olivia At the risk of butting into this conversation, I think the problem here is that you think that "just a tool" is a neutral concept.

                            Tools, by their very nature, change the way we interact with the world. Cars are "just a tool", but dependence on cars for transport has both positive and negative effects, because of how their use changes how we behave (and what other things we want to change about the world now "we" want to use cars all the time). Is "car-using humanity" healthier than "pre-car humanity"?

                            In this sense, even if "AI is just a tool", the existence of cognitive tools *clearly* implies that use of them will change the way people behave - *regardless* of any concept of "applications being identified as misapplications". Dependence on a tool for *thinking* feels inherently more problematic than dependence on a tool for travelling distances...

                            ulrikehahn@fediscience.orgU olivia@scholar.socialO abucci@buc.ciA 3 Replies Last reply
                            0
                            • aoanla@hachyderm.ioA aoanla@hachyderm.io

                              @lednaBM @UlrikeHahn @abucci @apostolis @olivia At the risk of butting into this conversation, I think the problem here is that you think that "just a tool" is a neutral concept.

                              Tools, by their very nature, change the way we interact with the world. Cars are "just a tool", but dependence on cars for transport has both positive and negative effects, because of how their use changes how we behave (and what other things we want to change about the world now "we" want to use cars all the time). Is "car-using humanity" healthier than "pre-car humanity"?

                              In this sense, even if "AI is just a tool", the existence of cognitive tools *clearly* implies that use of them will change the way people behave - *regardless* of any concept of "applications being identified as misapplications". Dependence on a tool for *thinking* feels inherently more problematic than dependence on a tool for travelling distances...

                              ulrikehahn@fediscience.orgU This user is from outside of this forum
                              ulrikehahn@fediscience.orgU This user is from outside of this forum
                              ulrikehahn@fediscience.org
                              wrote sidst redigeret af
                              #37

                              @aoanla @lednaBM @abucci @apostolis @olivia that’s very well put!

                              1 Reply Last reply
                              0
                              • lednabm@stranger.socialL lednabm@stranger.social

                                @abucci @UlrikeHahn @apostolis @olivia

                                Wouldn't an approach where the AIs have to pass the class as students, be better. After all, regurgitating data is not the way to learn how to think. As for the pollitical/economics of the whole mess, well, that's on us to some extent. It's a problem educated people deal with all the time, even among each other. IMHO, humanity is still growing up. We've not abandoned our superstitions for the hard real wonder of actual nature. Is AI part of our nature?

                                abucci@buc.ciA This user is from outside of this forum
                                abucci@buc.ciA This user is from outside of this forum
                                abucci@buc.ci
                                wrote sidst redigeret af
                                #38
                                @lednaBM@stranger.social @UlrikeHahn@fediscience.org @apostolis@social.coop @olivia@scholar.social The fact that you can selectively ignore the strings of a marionette does not mean it is alive, part of our nature, or able to attend and pass a course. I suspect this is even obvious to AI!
                                1 Reply Last reply
                                0
                                • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

                                  @abucci @apostolis @olivia I’m going to point you toward the scare quotes around the word “organic” in my post, which are there for precisely those reasons.

                                  I am also going to push back against the notion that I am “placing the responsibility at the feet of students”: I am simply describing the (widely documented) problem in higher education that students are using AI tools in significant volumes *even where there use is explicitly sanctioned and forbidden*.

                                  That is the concrete problem of AI now undermining higher education. Asking what “resisting AI” is supposed to mean for me in that context seems legitimate to me, and if it’s not, Olivia (who I’ve known for a long time as an academic colleague) is more than capable of telling me that herself.

                                  abucci@buc.ciA This user is from outside of this forum
                                  abucci@buc.ciA This user is from outside of this forum
                                  abucci@buc.ci
                                  wrote sidst redigeret af
                                  #39
                                  @UlrikeHahn@fediscience.org You stated you were pushing back against the characterization of your stance that you were laying responsibility at the feet of your students, and then immediately placed responsibility at the feet of the students! Are you really unable to see this in your own post?

                                  @fediscience.org @apostolis@social.coop @olivia@scholar.social
                                  ulrikehahn@fediscience.orgU ncf@types.plN 2 Replies Last reply
                                  0
                                  • aoanla@hachyderm.ioA aoanla@hachyderm.io

                                    @lednaBM @UlrikeHahn @abucci @apostolis @olivia At the risk of butting into this conversation, I think the problem here is that you think that "just a tool" is a neutral concept.

                                    Tools, by their very nature, change the way we interact with the world. Cars are "just a tool", but dependence on cars for transport has both positive and negative effects, because of how their use changes how we behave (and what other things we want to change about the world now "we" want to use cars all the time). Is "car-using humanity" healthier than "pre-car humanity"?

                                    In this sense, even if "AI is just a tool", the existence of cognitive tools *clearly* implies that use of them will change the way people behave - *regardless* of any concept of "applications being identified as misapplications". Dependence on a tool for *thinking* feels inherently more problematic than dependence on a tool for travelling distances...

                                    olivia@scholar.socialO This user is from outside of this forum
                                    olivia@scholar.socialO This user is from outside of this forum
                                    olivia@scholar.social
                                    wrote sidst redigeret af
                                    #40

                                    @aoanla @lednaBM @UlrikeHahn @abucci @apostolis

                                    Indeed! FWIW I touch on tools versus technologies in this context here if useful. https://scholar.social/@olivia/114937376930475208

                                    1 Reply Last reply
                                    0
                                    • aoanla@hachyderm.ioA aoanla@hachyderm.io

                                      @lednaBM @UlrikeHahn @abucci @apostolis @olivia At the risk of butting into this conversation, I think the problem here is that you think that "just a tool" is a neutral concept.

                                      Tools, by their very nature, change the way we interact with the world. Cars are "just a tool", but dependence on cars for transport has both positive and negative effects, because of how their use changes how we behave (and what other things we want to change about the world now "we" want to use cars all the time). Is "car-using humanity" healthier than "pre-car humanity"?

                                      In this sense, even if "AI is just a tool", the existence of cognitive tools *clearly* implies that use of them will change the way people behave - *regardless* of any concept of "applications being identified as misapplications". Dependence on a tool for *thinking* feels inherently more problematic than dependence on a tool for travelling distances...

                                      abucci@buc.ciA This user is from outside of this forum
                                      abucci@buc.ciA This user is from outside of this forum
                                      abucci@buc.ci
                                      wrote sidst redigeret af
                                      #41
                                      @aoanla@hachyderm.io @lednaBM@stranger.social @UlrikeHahn@fediscience.org @apostolis@social.coop @olivia@scholar.social The "just a tool" framing also does a great deal of heavy lifting for the political project that AI represents and forwards. What saddens me most is that this project is nearly transparent, its actors almost totally honest about what they are attempting to accomplish even as they dissemble about it. Yet we go around and around in circles about whether these things are "just" tools, or wring our hands about what to do about students using them, or waffle about whether the tools are useful or have this or that impact on productivity. These things are symptoms, not causes.
                                      1 Reply Last reply
                                      0
                                      • abucci@buc.ciA abucci@buc.ci
                                        @UlrikeHahn@fediscience.org You stated you were pushing back against the characterization of your stance that you were laying responsibility at the feet of your students, and then immediately placed responsibility at the feet of the students! Are you really unable to see this in your own post?

                                        @fediscience.org @apostolis@social.coop @olivia@scholar.social
                                        ulrikehahn@fediscience.orgU This user is from outside of this forum
                                        ulrikehahn@fediscience.orgU This user is from outside of this forum
                                        ulrikehahn@fediscience.org
                                        wrote sidst redigeret af
                                        #42

                                        @abucci @apostolis @olivia let me say this then: I find your original reply to me, someone you have never met, aggressive and inflammatory.

                                        One of the main benefits of exchange on platforms like this, to me, lies in being able to talk things through with others whose opinion and expertise I value but who disagree with me - that allows me to learn things and clarify my thoughts, and I’ve found this exchange with Olivia really helpful in that regard.

                                        Trying to navigate disagreement in a way that it doesn’t lead to conflict is incredibly hard. In a context like this thread where people are investing significant effort in trying to navigate disagreement in a constructive way, I don’t personally have time, energy, or interest in exchanges with people who aren’t making that effort. The world is fraught enough as it is.

                                        1 Reply Last reply
                                        0
                                        • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

                                          @olivia Olivia, what would it mean for me to “refuse adoption” in universities when it is students who are the drivers for my courses and they are widely using AI in ways that are already forbidden?

                                          I feel like the “resistance” and critique of inevitability talk isn’t quite connecting with my reality on the ground

                                          mycotropic@beige.partyM This user is from outside of this forum
                                          mycotropic@beige.partyM This user is from outside of this forum
                                          mycotropic@beige.party
                                          wrote sidst redigeret af
                                          #43

                                          @UlrikeHahn @olivia

                                          I teach those students as well. We see them coming in with 85%+ self reported past-week AI use so we started surveying them on K,A&B regarding that use. They return the party line; it's everywhere and I have to use it in order to succeed. We switched to a Write To Learn model and switched our assessments away from format, grammar and spelling and into content, personal insight and creativity. That's after a module on the history and mechanics of LLMs, not just the problem with hallucinations but how they form as AI constructs its responses.

                                          The feedback was and remains positive for short, low threat assignments. If I can personally generate similar bursts of dopamine compared to a chatbots injection of "great question" and other disingenuous slop then perhaps I can actually engage with the learner.

                                          One last point, my institution just bought a one year instance of openai.edu and the students are HAMMERING leadership over the expense, environmental impacts and stolen creativity. Our shared governance organization is pushing back citing this industry interaction as a failure of shared decision making articulated in our governing constitution. AI is pedagogy and that's faculty business and not the job of leadership.

                                          ulrikehahn@fediscience.orgU 1 Reply Last reply
                                          0
                                          Svar
                                          • Svar som emne
                                          Login for at svare
                                          • Ældste til nyeste
                                          • Nyeste til ældste
                                          • Most Votes


                                          • Log ind

                                          • Har du ikke en konto? Tilmeld

                                          • Login or register to search.
                                          Powered by NodeBB Contributors
                                          Graciously hosted by data.coop
                                          • First post
                                            Last post
                                          0
                                          • Hjem
                                          • Seneste
                                          • Etiketter
                                          • Populære
                                          • Verden
                                          • Bruger
                                          • Grupper