New tool lets creators check if AI trained on their work
TL;DR
An open-source platform helps creators check if their content was used to train large language models, without heavy compute or opaque workflows.
What it does
- Verifies if a specific work appears in LLM training datasets.
- Improves similarity matching and dataset validation.
- Cuts compute by 10-30% via efficient API calls.
- Offers an intuitive UI and scalable backend.
Why it matters
As legal scrutiny grows, accessible, transparent tools can support ethical AI development and empower artists, writers, and developers to protect their rights.
Paper: "Copyright Detection in Large Language Models: An Ethical Approach to Generative AI Development" by David Szczecina, Senan Gaffori, and Edmond Li — https://arxiv.org/abs/2511.20623v1
Paper: https://arxiv.org/abs/2511.20623v1
Register: https://www.AiFeta.com
AI Copyright LLM GenAI OpenSource Transparency CreatorsRights Ethics