• xmunk@sh.itjust.works
    link
    fedilink
    arrow-up
    1
    arrow-down
    3
    ·
    3 months ago

    Is a kid just a 60% reduction by volume of an adult? And these are generative algorithms… nobody really understands how it perceives the world and word relations.

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      3 months ago

      It understands young and old. That means it knows a kid is not just a 60% reduction by volume of an adult.

      We know it understands these sorts of things because of the very things this whole kerfuffle is about - it’s able to generate images of things that weren’t explicitly in its training set.

      • xmunk@sh.itjust.works
        link
        fedilink
        arrow-up
        1
        ·
        3 months ago

        But it doesn’t fully understand young and “naked young person” isn’t just a scaled down “naked adult”. There are physiological changes that people go through during puberty which is why the “It understands young vs. old” is a clearly vapid and low effort comment. Yours has more meaning behind it so I’d clarify that just being able to have a vague understanding of young and old doesn’t mean it can generate CSAM.

    • GBU_28@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      3 months ago

      Just go ask a model to show you, with legal subject matter