• Liz@midwest.social
    link
    fedilink
    English
    arrow-up
    103
    ·
    8 months ago

    Ask a man his salary. Do it. How else are you supposed to learn who is getting underpaid? The only way to rectify that problem is to learn about it in the first place.

  • kernelle@0d.gs
    link
    fedilink
    arrow-up
    64
    arrow-down
    4
    ·
    8 months ago

    “Publicly available data” - I wonder if that includes Disney’s catalogue? Or Nintendo’s IP? I think they are veeery selective about their “Publicly available data”, it also implies the only requirement for such training data is that it is publicly available, which almost every piece of media ever? How an AI model isn’t public domain by default baffles me.

    • Even_Adder@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      16
      ·
      8 months ago

      You should check out this article by Kit Walsh, a senior staff attorney at the EFF, and this one by Katherine Klosek, the director of information policy and federal relations at the Association of Research Libraries.

      • kernelle@0d.gs
        link
        fedilink
        arrow-up
        9
        ·
        8 months ago

        Great articles, first is one of the best I’ve read about the implications of fair use. I argue that because of the broadness of human knowledge that is interpreted through these models, everyone is entitled to have unrestricted access to them (not the servers or algorithms used, the models). I’ll dub it “the library of the digital age” argument.

    • Programmer Belch@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      8 months ago

      The problem is that if copyrighted works are used, you could generate a copyrighted artwork that would be made into public domain, stripping its protection. I would love this approach, the problem is the lobbyists don’t agree with me.

      • kernelle@0d.gs
        link
        fedilink
        arrow-up
        3
        arrow-down
        3
        ·
        8 months ago

        Not necessarily, if a model is public domain, there could still be a lot of proprietary elements used in interpreting that model and actually running it. If you own the hardware and generate something using AI, I’d say the copyright goes to you. You use AI as the brush to paint your painting and the painting belongs to you, but if a company allows you to use their canvas and their painting tools, it should go to them.

        • Programmer Belch@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          6
          ·
          8 months ago

          I think that if I paint with my own brush a mario artwork that isn’t to Nintendo’s standart, they have the legal power to take it down from wherever I upload

          • kernelle@0d.gs
            link
            fedilink
            arrow-up
            3
            arrow-down
            2
            ·
            8 months ago

            Really? Even if your artwork isn’t used in a commercial way?

            • Programmer Belch@lemmy.dbzer0.com
              link
              fedilink
              English
              arrow-up
              6
              ·
              8 months ago

              I’m really not in the know abput these things but I have seen free fangames taken down because they used copyrighted property even though the creators don’t receive a penny.

              • kernelle@0d.gs
                link
                fedilink
                arrow-up
                2
                arrow-down
                2
                ·
                8 months ago

                I’ll compare it with the recent takedown of the Switch emulator Yuzu. It’s my understanding they actively solicited donations and piracy, both of which could be seen as commercial activities. Which in a project of that scale the latter was their downfall, meanwhile Ryujinx is still up and running. But we’ll see if that remains true.

                • Grimy@lemmy.world
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  edit-2
                  8 months ago

                  Copyrights and IP laws don’t only come into effect if profit is made. Fan products are usually tolerated by companies because it’s free advertising and fans get angry when it does get taken down.

                  When a fan product starts making money, it’s usually because it directly competes with the original IP and then they act. Even then, Etsy has thousands of shops with copyrighted content but the small profit loss doesnt justify the loss of reputation for the companies.

                  That being said, it’s the user who uploads it who is at fault and not the tool used to create it.

                  Ultimately, I think it’s the platforms that let users upload copyrighted content and celebrity likenesses that should be at fault. Take for example the Taylor Swift debacle. An image generator was used to create the images sure but twitter chose to let it float on their website for a whole day even though it was most likely reported in the first 5 minutes.

                  There’s also the fact that if we start demanding AI doesn’t use copyrighted content, it kills the open source scene and cements Google and Microsoft’s grip on our economy as we move towards an AI driven society.

            • the_artic_one@programming.dev
              link
              fedilink
              English
              arrow-up
              2
              ·
              8 months ago

              Yes, fanart is almost certainly copyright infringement unless the copyright holder grants a license. Many companies have an official license for non-commercial fanart and generally nobody cares about it but if someone really wanted to they could absolutely file takedown requests against all fanart of their work.

        • Natanael@slrpnk.net
          link
          fedilink
          arrow-up
          3
          ·
          8 months ago

          The existing legal precedence in most places is that most use of ML doesn’t count as human expression and doesn’t have copyright protection. You have to have significant control over the creation of the output to have copyright (the easiest workaround is simply manually modifying the ML output and then only releasing the modified version)

          • kernelle@0d.gs
            link
            fedilink
            arrow-up
            2
            arrow-down
            1
            ·
            8 months ago

            The existing legal precedence

            I know that’s how law works, but there is no precedent for AI at this scale and will only get worse. What if AI gains full sentience? Are they a legally recognised person? Do they have rights and do they not own the copyright themselves? All very good questions with no precedent in law.

        • jaybone@lemmy.world
          link
          fedilink
          arrow-up
          6
          arrow-down
          4
          ·
          8 months ago

          If you rent a brush to paint with, is the painting not yours? If you rent a musical instrument to record an original song with, is the song not yours?

          • kernelle@0d.gs
            link
            fedilink
            arrow-up
            2
            arrow-down
            2
            ·
            8 months ago

            Exactly! When you pay for a service you own the copyright, like having a photoshop license. I meant in other situations where it’s free or provided as research tools to engineers under a company.

  • zinderic@programming.dev
    link
    fedilink
    arrow-up
    37
    arrow-down
    2
    ·
    8 months ago

    It’s almost impossible to audit what data got into an AI model. Until this is true companies could scrape and use whatever they like and no one would be the wiser to what data got used or misused in the process. That makes it hard to make such companies accountable to what and how they are using.

    • po-lina-ergi@kbin.social
      link
      fedilink
      arrow-up
      41
      arrow-down
      3
      ·
      8 months ago

      Then it needs to be on companies to prove their audit trail, and until then require all development to be open source

      • zinderic@programming.dev
        link
        fedilink
        arrow-up
        7
        ·
        8 months ago

        That would be amazing. But it won’t happen any time soon if ever… I mean - just think about all that investment in GPU compute and the need to realize good profit margins. Until there are laws and legislation that requires AI companies to open their data pipelines and make public all details about the data sources I don’t think much would happen. They’ll just keep feeding any data they get their hands on and nothing can stop that today.

        • ipkpjersi@lemmy.ml
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          8 months ago

          Until there are laws and legislation that requires AI companies to open their data pipelines and make public all details about the data sources I don’t think much would happen.

          I don’t expect those laws to ever happen. They don’t benefit large corporations so there’s no reason those laws would ever be prioritized or considered by lawmakers, sadly.

        • InputZero@lemmy.ml
          link
          fedilink
          arrow-up
          1
          arrow-down
          1
          ·
          edit-2
          8 months ago

          Maybe not today and maybe not every AI but maybe some AI in the near future will have it’s data sources made explainable. There are a lot of applications where deploying AI would be an improvement over what we have. One example I can bring up is in-silico toxicology experiments. There’s been a huge desire to replace as many in-vivo experiments with in-vitro or even better in-silico to minimize the number of live animals tested on. Both for ethical reasons and cost savings. AI has been proposed as a new tool to accomplish this but it’s not there yet. One of the biggest challenges to overcome is making the AI models used in-silico to be explainable, because we can not regulate effectively what we can not explain. Regardless there is a profits incentive for AI developers to make at least some AI explainable. It’s just not where the big money is. To which end that will apply to all AI I haven’t the slightest idea. I can’t imagine OpenAI would do anything to expose their data.

    • po-lina-ergi@kbin.social
      link
      fedilink
      arrow-up
      4
      arrow-down
      4
      ·
      8 months ago

      Then it needs to be on companies to prove their audit trail, and until then require all development to be open source

    • Natanael@slrpnk.net
      link
      fedilink
      arrow-up
      19
      ·
      8 months ago

      Training from scratch and retraining is expensive. Also, they want to avoid training on ML outputs as samples, they want primarily human made works as samples, and after the initial public release of LLMs it has become harder to create large datasets without ML stuff in them

      • Scrubbles@poptalk.scrubbles.tech
        link
        fedilink
        English
        arrow-up
        13
        ·
        edit-2
        8 months ago

        There was a good paper that came out recently saying that training on ml data will result in a collapse of cohesion. It’s going to be real interesting, I don’t know if they’ll be able to train as easily ever again

      • Iron Lynx@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        8 months ago

        I recall spotting a few things about Image Generators having their training data contaminated using generated images, and the output becoming significantly worse. So yeah, I guess LLMs and IGA’s need natural sources, or it gets more inbred than the Habsburgs.

      • TurtleJoe@lemmy.world
        link
        fedilink
        arrow-up
        3
        arrow-down
        3
        ·
        8 months ago

        I think it’s telling that they acknowledge that the stuff their bots churn out is often such garbage that training their bots on it would ruin them.

    • can@sh.itjust.works
      link
      fedilink
      arrow-up
      3
      ·
      8 months ago

      Hey, did you know your profile is set to appear as a bot and as a result many may be filtering your posts and comments? You can change this in your Lemmy settings.

      Unless you are a bot… In which case where did you get your data?

    • potustheplant@feddit.nl
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      8 months ago

      Where do you get that from? At least ChatGPT isn’t limited to data from 2021. I haven’t researched about other models.

  • Hjalmar@feddit.nu
    link
    fedilink
    arrow-up
    5
    ·
    8 months ago

    I love of it isn’t just a image of the open ai logo but also a sad person besides it

  • uis
    link
    fedilink
    arrow-up
    1
    ·
    8 months ago

    Wat de fuq with not asking salary?

    • MutilationWave@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      8 months ago

      It’s an old “rule” of being polite. Since the measure of a man is how much he makes of course.

      Really though you need to know at least what your coworkers and boss make to ensure no one is getting underpaid. Really 90% of people are underpaid but that’s a different conversation.

      There are companies that will tell you you’re not allowed to discuss your salary, which is a bullshit line. There are employees who believe the bullshit who will refuse to disclose their salary.