• Gaywallet (they/it)@beehaw.orgOP
    link
    fedilink
    arrow-up
    0
    ·
    edit-2
    6 months ago

    You’re absolutely correct, yet ask someone who’s very pro AI and they might dismiss such claims as “needing better prompts”. Also many people may not be as tech informed as you are, and bringing light to algorithmic bias can help them understand and navigate the world we now live in. Dismissing the article just because you already know the answer doesn’t really encourage people to participate in a discussion.

    • helenslunch@feddit.nl
      link
      fedilink
      arrow-up
      0
      ·
      6 months ago

      Dismissing the article just because you already know the answer doesn’t really encourage people to participate in a discussion.

      If the author doesn’t know the answer, then it is helpful to provide it. If they know the answer, then why are they phrasing the title as a question?

      • MBM@lemmings.world
        link
        fedilink
        arrow-up
        0
        ·
        6 months ago

        If you genuinely don’t know: because it’s an attention-grabbing title (which isn’t inherently bad)

    • Even_Adder@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      6 months ago

      It’s really hard getting dark skin sometimes. A lot of the time it’s not even just the model, LoRAs and Textual Inversions make the skin lighter again so you have to try even harder. It’s going to take conscious effort from people to tune models that are inclusive. With the way media is biased right now, I feel like it’s going to take a lot of effort.

      • jarfil@beehaw.org
        link
        fedilink
        arrow-up
        0
        ·
        6 months ago

        “Inclusive models” would need to be larger.

        Right now people seem to prefer smaller quantized models, with whatever set of even smaller LoRAs on top, that make them output what they want… and only include more generic elements in the base model.

          • jarfil@beehaw.org
            link
            fedilink
            arrow-up
            0
            ·
            6 months ago

            Are you ready to run a 100B FP64 parameter model? Or even a 10B FP32 one?

            Over time, I wouldn’t be surprised if 500B INT8 models became commonplace with neuromorphic RAM, but there’s still some time for that to happen.

              • jarfil@beehaw.org
                link
                fedilink
                arrow-up
                0
                ·
                6 months ago

                For more inclusive models, or for current ones? In order to add something, either the size has to grow, or something would need to get pushed out (content, or quality). 4GB models are already at the limit of usefulness, both DALLE3 and SDXL run at about 12B parameters, so to make them “more inclusive” they’d have to grow.

                • Even_Adder@lemmy.dbzer0.com
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  6 months ago

                  I’m saying SD 1.5 and SDXL capture the concepts just fine, it’s just during fine-tuning people train away some of the diversity.

                  • jarfil@beehaw.org
                    link
                    fedilink
                    arrow-up
                    0
                    ·
                    6 months ago

                    Wait, by “fine-tuning”… do you mean LoRAs? Because those are more like brain surgery with a sledgehammer, rather the opposite of “fine”. I don’t think it’s possible for LoRAs to avoid having undesirable side effects… and I don’t think people even want that.

                    Actual “fine” tuning, would be adding the LoRA’s training data to the original set, then training the whole model from scratch… and that would require increasing the model’s size to encode the increased amount of data for the same output quality.