• kromem@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    1
    ·
    1 year ago

    Because it has five billion images?

    The potentially at issue images comprise less than one percent of one percent of one percent of the total.

      • kromem@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 year ago

        No, it’s not manually labeled. It connects the text to the image based on things like alt text or the comment next to it in a social media post, and then ran them through a different AI (CLIP) which rated how well the text description matched the image and they filter out the ones with a low score.

        The point of the OP research is that they should add another step to check CSAM databases and not rely on social media curation to have avoided illegal material (which they should, even though it’s a very very small portion of the overall dataset).

        But at no time was a human reviewing CSAM, labeling it, and including it in the data.