Ich mag Pflanzen und hab ein extra Regal dafür. Support your FOSS (Free and OpenSource Software) projects!🫶🏼

  • 0 Posts
  • 18 Comments
Joined 2 years ago
cake
Cake day: September 12th, 2023

help-circle


















  • This isn’t possible as of now, at least not reliably. Yes, you can tailor a model to one specific generative model, but because we have no reliable outlier detection (to train the “AI made detector”), a generative model can always be trained with the detector model incorporated in the training process. The generative model (or a new model only designed to perturb output of the “original” generative model) would then learn to create outliers to the outlier detector, effectively fooling the detector. An outlier is everything that pretends to be “normal” but isn’t.

    In short: as of now we have no way to effectively and reliably defend against adversarial examples. This implies, that we have no way to effectively and reliably detect AI generated content.

    Please correct me if I’m wrong, I might be mixing up some things.