Sign the manifesto: protect our art and data from AI companies

Sign the manifesto: protect our art and data from AI companies

Started
February 8, 2023
Signatures: 9,720Next Goal: 10,000
Support now

Why this petition matters

Started by Lorenzo Ceccotti

We are a group of artists, creatives, publishers and associations from all over Europe united in bringing to the public attention how our data and intellectual properties are being exploited without our consent, on a scale never seen before. Such an unprecedented situation has led us to join our forces to reach out to the European Institutions and have our voices heard. If you believe that your data and creative work should not be exploited with impunity for profit by a handful of corporations, join us in supporting this battle.

Summer 2022 has seen the rise of a new, incredible technology: AI text-to-image (TTI) or image-to-image. Every time a groundbreaking technology comes to life, our society has to oversee its deployment in order to avoid any harm or infringing of human rights. This hasn’t happened with AI technology yet. These forms of artificial intelligence can generate images or other contents from textual prompts or other media given by the user. To do so, an AI needs to be trained on a dataset of media.

The quality of a generative AI is defined by the quality of its dataset – for example, in regard to images, the more pictures and illustrations an AI learns on, the more styles the AI is able to replicate and the more things it can do. Therefore, the products sold by AI companies are the result of operations on datasets, which contain all sorts of data, including millions of copyrighted images, private pictures and other sensitive material. These files were collected by indiscriminately scraping the internet without the consent of the owners and people portrayed in them and are currently being used by AI companies for profit. Furthermore, some of these companies are using the names of some of the artists whose work they have scraped to advertise their AIs and the styles they can replicate. This exploitation of our work and datas not only does not meet the minimum requirements imposed by basic human rights that are foundational to our society: it is also severely damaging the art market, potentially scarring it forever. We see this as only the beginning of a crisis that will afflict all sorts of jobs and occupation, whether they are creative jobs or not. The art market is the first one to be affected only because of its structural vulnerabilities, which make it an easy prey. It is time to change this.

These are the key points we want to enforce:

1) Any data related to people or works, in any form, be it digital data – such as text files, audios, videos or images – or captured from reality by camera, microphones or any other mean of registration, shall not be used to train AI model without the explicit and informed consent of its owner. We ask for an extension to the AIs of the principles protecting personal data previously introduced by the GDPR and the introduction of a new form of protection specifically for this kind of exploitation: the “training right”.

2) Using the names of people, stage names or titles of works not covered by a license to exploit for AI training shall be prohibited for those software that allows the use of textual or vocal prompt to generate images, videos, texts or audio.

3) Using videos, images, audios and texts not covered by a license to exploit for AI training shall be prohibited for those software that allow to upload media contents to generate an image, a video, a text or an audio.

4) A “human and machine readable” indexing and certification system shall be established, reporting all AIs’ activities and the full content of their datasets of images, texts, videos and sounds, be them fully or partially reproduced. Captions such as “entirely made by AI”, “made using AI-generated material” should become the standard to ensure the proper functioning of AIs themselves.

5) The distinction between “copyrighted material” and “public domain" is no longer adequate to identify what can and cannot be used for the datasets. Learning datasets contain personal sensitive data, protected by the privacy laws, but not by copyright. We can find examples of material released when it would not have been possible to foresee its use in a dataset to train an AI model. Any data used in training model shall be curated and authorized by its legitimate owner and willingly inserted in the dataset by its author with full knowledge of it. AI companies shall produce internally original materials for the training or license external material following terms and contracts previously established with the authors and rightful owners of said material.

Each signature matters: sign our manifesto and let our voices be heard.

For more information please visit www.egair.eu

Support now
Signatures: 9,720Next Goal: 10,000
Support now
Share this petition in person or use the QR code for your own material.Download QR Code