Just 250 malicious training documents can poison a 13B parameter model - that’s 0.00016% of a whole dataset Poisoning AI models might be way easier than previously thought if an Anthropic study is anything to go on. …

  • IMALlama@lemmy.world
    link
    fedilink
    arrow-up
    4
    ·
    4 days ago

    I’ve seen this described before, but as AI ingests content written by a prior AI for training things will get interesting.