Authors using a new tool to search a list of 183,000 books used to train AI are furious to find their works on the list.

  • Shurimal@kbin.social
    link
    fedilink
    arrow-up
    15
    arrow-down
    17
    ·
    1 year ago

    These are machines, though, not human beings.

    What’s the difference? On the most fundamental level it’s all the same.

    • brygphilomena@lemmy.world
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      4
      ·
      1 year ago

      A human, regardless of how many books they read, will have personal experiences that are undeniably unique to themselves. They will interpret the works they read differently from each other based on their worldly experiences. Their writing, no matter how many books they read and get inspired on, will always be influenced by their own personal lives. They can experience love, hate, heartbreak, empathy, sadness, and happiness.

      This is something a LLM does not have, and in my opinion, is a massive distinguishing factor. So on a “fundamental” level, it is not the same. It is no where near the same.

      • lloram239@feddit.de
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        1 year ago

        A human, regardless of how many books they read, will have personal experiences that are undeniably unique to themselves.

        So will every AI. ChatGPT will give you different answers than Bard or WizardLM, since they are all trained on different books. And every StableDiffusion model creates different images, different styles, different topics, etc. It’s all in the data they “experienced”.

      • originalucifer@moist.catsweat.com
        link
        fedilink
        arrow-up
        2
        arrow-down
        4
        ·
        1 year ago

        do you really think we are that far off… from giving a foundational memory and motivation layers to these LLMs, that could mimic… or even… generate the generic thoughts youre indicating?

        i dont think so. you seem to imply its impossibility, i expect its inevitability. the human brain will not be a black box forever… it still exists in a world of physics we can emulate, even if rudimentary.

    • Wander@kbin.social
      link
      fedilink
      arrow-up
      15
      arrow-down
      5
      ·
      1 year ago

      Unless you think theres no difference between killing a person and closing a program, I think we can agree they should be treated differently in the eyes of the law.

      And so theres a difference between a person reading a book and being inspired by it, and someone writing a program that automatically transforms the book in data that can create new books.

    • AnonStoleMyPants@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      17
      arrow-down
      7
      ·
      1 year ago

      The same thing as with tooooooons of things: scale.

      Nobody cares if one dude steals office supplies at work. Now, if everyone stats doing it, or if the single guy steals everything, then action is taken.

      Nobody cares if a random person draws in the same style and with same characters as you, but if they start to sell them, or god forbid, out-sell you, then there is a problem.

      Nobody cares (except police I guess) if a random driver drives double the speed limit and annoys people living next to the road on the weekends, but when tons of people do it, you get speed bumps.

      Nobody cares if few people pirate movies, but when it gets to mainstream and companies notice that there might be money being lost. Then you get whatever we have now.

      Nobody cares if the mudhill behind your house erodes a bit and you get mud on your shoes. Have a bunch of that erode and you realise the danger…

      You have been fine-tuning your own writing style for a decade and random schmuck starts to write similarly, you probably don’t care. No harm done. Now, get an AI to write 10 000 books in a weekend and someone starts to sell them… well now you have a completely different problem.

      On a fundamental level the exact same thing is happening, yet action is only taken after a certain threshold is step over.