New tool lets creators check if AI trained on their work

New tool lets creators check if AI trained on their work

TL;DR

An open-source platform helps creators check if their content was used to train large language models, without heavy compute or opaque workflows.

What it does

  • Verifies if a specific work appears in LLM training datasets.
  • Improves similarity matching and dataset validation.
  • Cuts compute by 10-30% via efficient API calls.
  • Offers an intuitive UI and scalable backend.

Why it matters

As legal scrutiny grows, accessible, transparent tools can support ethical AI development and empower artists, writers, and developers to protect their rights.

Paper: "Copyright Detection in Large Language Models: An Ethical Approach to Generative AI Development" by David Szczecina, Senan Gaffori, and Edmond Li — https://arxiv.org/abs/2511.20623v1

Paper: https://arxiv.org/abs/2511.20623v1

Register: https://www.AiFeta.com

AI Copyright LLM GenAI OpenSource Transparency CreatorsRights Ethics

Read more