by Milagros Miceli and Adio Dinika
3 min read
Behind artificial intelligence’s gleaming promise of innovation lies an uncomfortable truth: its success depends on an invisible workforce doing exhausting, often traumatic work. These data workers – predominantly in the Global South – annotate, moderate and process the vast data streams that power AI systems. They are the uncredited architects of artificial intelligence, their labour rendered invisible by design.
Recognising the need to confront this erasure, we launched the Data Workers’ Inquiry (DWI) to better understand the realities of this hidden workforce. This pioneering initiative, funded by the Distributed AI Research Institute (DAIR), the Weizenbaum Institute and TU Berlin, seeks to amplify the voices of data workers and recentre their epistemic authority. For us, it was imperative that the workers were not simply research participants or interviewees; they were community researchers co-researching with us. From shaping the inquiries to determining what forms their output would take, they took ownership and leadership over the project, deciding what mattered most to them and how it should be presented.
Our team, led by Dr Milagros Miceli, alongside Dr Adio Dinika, Krystal Kauffmann, Camilla Salim Wagner and Laurenz Sachenbacher, collaborated with 17 co-researchers from diverse geographical locations – Kenya, Venezuela, Colombia, Nigeria, Syria, Lebanon, Brazil and Germany – and a range of employment contexts. The community researchers, including content moderators and data annotators working for outsourcing companies and platforms, produced a remarkable variety of outputs: vlogs, documentaries, animations, essays, podcasts, comic strips and more. Through their creative and personal contributions, the project provides a rare and intimate glimpse into the lives of those who sustain the AI industry, exposing the profound human cost of this labour.
Their findings are both harrowing and essential. In Nairobi, data workers admit turning to dangerous drugs as a desperate measure to cope with the relentless psychological trauma of reviewing graphic violence daily, compounded by the absence of adequate mental health support. In Venezuela, Alexis Chávez shares the indignity of being paid in gift cards, forcing him into convoluted exchanges to convert them into usable currency. In her powerful report, Fasica Berhane Gebrekidan, a content moderator from Tigray, Ethiopia, recounts the heart-wrenching burden of reviewing material tied to the war ravaging her own community. She describes the agony of being exposed to graphic videos and images of atrocities, often recognising places and people from her home region. This proximity to the conflict magnifies the emotional toll, as she is forced to repeatedly view and categorise images of violence from her home region, leaving her with a haunting sense of complicity – an involuntary participant in documenting atrocities she cannot prevent. Another particularly poignant project is the Zine by Botlhokwa Ranta, which focuses on the challenges faced by mothers working in content moderation, underscoring the gendered dimensions of this labour and its impact on families and communities.
These accounts lay bare the systemic exploitation ingrained in the AI industry, revealing a web of vulnerabilities that outsourcing firms specifically exploit. Refugee workers, already marginalised by their precarious legal status and limited employment options, are prime targets for firms that offer minimal compensation, no benefits and zero job security. In Kenya, for instance, the firms targeted the recruitment of people living in the slums. These firms capitalise on desperation, knowing that these workers have few alternatives. The targeting of vulnerable people also extends to countries like Germany, where migrants often face the additional burden of contracts tied to their residency permits, leaving them at the mercy of employers who can dictate exploitative terms without fear of reprisal. The situation of migrant data workers in Germany is graphically described in Sakine Mohamadi Bozorg’s essay and the podcast produced by Meta content moderators at the outsourcing firm Telus Digital in Essen. For these workers, the stakes are high – not just their livelihoods, but their legal right to stay in the country depends on keeping their jobs, no matter how unjust the conditions.
In Lebanon, Roukaya al-Hammada’s report reveals a wider pattern of instability. Short-term contracts, designed to provide only subsistence-level income, create a precarious existence where planning for the future becomes impossible. This economic fragility traps workers in cycles of uncertainty, forcing them to live paycheck to paycheck while enduring relentless workloads. The exploitative practices are further compounded by the global fragmentation of the workforce, where workers are scattered across geographies to prevent unionisation or collective action. These accounts underscore how the AI industry, while championed as the frontier of innovation, relies on labour practices that entrench systemic inequities and perpetuate a global divide.
One particularly insidious practice is the rampant use of non-disclosure agreements (NDAs) to silence workers about the nature of their jobs. These NDAs go beyond the standard restrictions of discussing company information with the press or the public – they extend into workers’ private lives, cutting off essential avenues of support. For instance, one worker shared how her NDA prohibited her from revealing the specifics of her role even to her psychotherapist. Unable to discuss the graphic content she reviewed daily, she was forced to say, “I work for a call centre” – a vague explanation that rendered her counselling ineffective. In the context of data work, NDAs deepen the isolation of already vulnerable workers, creating an environment where exploitation thrives in the dark. By forbidding transparency, these agreements not only hide the true cost of AI labour, but also perpetuate the very conditions that lead to worker burnout, trauma and disempowerment.
The Data Workers’ Inquiry (DWI) unpicks this fragmentation and disempowerment. Through participatory action research, we foster solidarity among workers, creating bridges that connect their stories and experiences across borders. Workers who were once isolated and silenced are beginning to find common ground, sharing strategies and building alliances. What emerged from this process were not victim narratives, but bold collective testimonies from agents of change – workers reclaiming their agency and demanding accountability from an industry that has long thrived on their invisibility. In addition, encouragingly, workers are beginning to organise and communicate, challenging the industry’s grip on their labour. In Africa, data workers have formed the African Content Moderators Union, creating a collective platform to advocate for their rights and push back against exploitative practices. These efforts are not limited to one region. Transnational solidarity is growing, with organisations such as Turkopticon, which supports workers on platforms such as Amazon Mechanical Turk, engaging with the African Content Moderators Union to share insights, resources and strategies. This cross-border collaboration signals a growing movement that transcends geographic and cultural divides, laying the foundation for a global network of resistance.
Through these efforts, the narrative is shifting. Workers are no longer passive recipients of injustice, but active participants in shaping their futures. Their stories and collective actions remind us that systemic change is not just possible; it is already underway, driven by the courage and determination of those who refuse to be erased. Imagine an AI industry where the well-being of its workforce is a priority, where labour rights are upheld, and where innovation is driven by principles of justice. This is the future that the Data Workers’ Inquiry envisions, and that the workers’ blueprints call upon us to create.
Dr. Milagros Miceli is a sociologist and computer scientist investigating how ground-truth data for machine learning is produced. Her research focuses on labor conditions and power dynamics in data work. Since 2018, Milagros has continuously engaged with communities of data workers globally.
Dr. Adio Dinika is a research fellow at the Distributed AI Research Institute (DAIR). His work focuses on the intersection of digital labor, AI governance, and platform economies, with a particular emphasis on labor exploitation in the Global South. He co-leads the Data Workers Inquiry, a research initiative examining the working conditions of data workers in the AI supply chain.
Technology, Employment and Wellbeing is an FES blog that offers original insights on the ways new technologies impact the world of work. The blog focuses on bringing different views from tech practitioners, academic researchers, trade union representatives and policy makers.
Cours Saint Michel 30e 1040 Brussels Belgium
+32 2 329 30 32
futureofwork(at)fes.de
Meet the team
Follow us on LinkedIn and X
Subscribe to receive our Newsletter
Watch our short videos and recorded events Youtube
This site uses third-party website tracking technologies to provide and continually improve our services, and to display advertisements according to users' interests. I agree and may revoke or change my consent at any time with effect for the future.
These technologies are required to activate the core functionality of the website.
This is an self hosted web analytics platform.
Data Purposes
This list represents the purposes of the data collection and processing.
Technologies Used
Data Collected
This list represents all (personal) data that is collected by or through the use of this service.
Legal Basis
In the following the required legal basis for the processing of data is listed.
Retention Period
The retention period is the time span the collected data is saved for the processing purposes. The data needs to be deleted as soon as it is no longer needed for the stated processing purposes.
The data will be deleted as soon as they are no longer needed for the processing purposes.
These technologies enable us to analyse the use of the website in order to measure and improve performance.
This is a video player service.
Processing Company
Google Ireland Limited
Google Building Gordon House, 4 Barrow St, Dublin, D04 E5W5, Ireland
Location of Processing
European Union
Data Recipients
Data Protection Officer of Processing Company
Below you can find the email address of the data protection officer of the processing company.
https://support.google.com/policies/contact/general_privacy_form
Transfer to Third Countries
This service may forward the collected data to a different country. Please note that this service might transfer the data to a country without the required data protection standards. If the data is transferred to the USA, there is a risk that your data can be processed by US authorities, for control and surveillance measures, possibly without legal remedies. Below you can find a list of countries to which the data is being transferred. For more information regarding safeguards please refer to the website provider’s privacy policy or contact the website provider directly.
Worldwide
Click here to read the privacy policy of the data processor
https://policies.google.com/privacy?hl=en
Click here to opt out from this processor across all domains
https://safety.google/privacy/privacy-controls/
Click here to read the cookie policy of the data processor
https://policies.google.com/technologies/cookies?hl=en
Storage Information
Below you can see the longest potential duration for storage on a device, as set when using the cookie method of storage and if there are any other methods used.
This service uses different means of storing information on a user’s device as listed below.
This cookie stores your preferences and other information, in particular preferred language, how many search results you wish to be shown on your page, and whether or not you wish to have Google’s SafeSearch filter turned on.
This cookie measures your bandwidth to determine whether you get the new player interface or the old.
This cookie increments the views counter on the YouTube video.
This is set on pages with embedded YouTube video.
This is a service for displaying video content.
Vimeo LLC
555 West 18th Street, New York, New York 10011, United States of America
United States of America
Privacy(at)vimeo.com
https://vimeo.com/privacy
https://vimeo.com/cookie_policy
This cookie is used in conjunction with a video player. If the visitor is interrupted while viewing video content, the cookie remembers where to start the video when the visitor reloads the video.
An indicator of if the visitor has ever logged in.
Registers a unique ID that is used by Vimeo.
Saves the user's preferences when playing embedded videos from Vimeo.
Set after a user's first upload.
This is an integrated map service.
Gordon House, 4 Barrow St, Dublin 4, Ireland
https://support.google.com/policies/troubleshooter/7575787?hl=en
United States of America,Singapore,Taiwan,Chile
http://www.google.com/intl/de/policies/privacy/