The nature of the work is traumatic. Workers must label data that describes, the investigation found, child sexual abuse, bestiality, murder, suicide, torture, self harm, and incest. In order to teach programs like ChatGPT how not to indulge in such topics, it must first identify them. Scores of marginalized workers in Kenya, for example, are charged with this thankless task while also being paid wages that range from $1.32 to $2 (Williams, Miceli & Gebru, 2022). It should also be noted that Meta (previously Facebook) has hired content moderators via the same firm, and tasked them with labeling disturbing videos under heavy surveillance. As a result, they reportedly suffer from a slew of mental health issues (including PTSD, anxiety, and depression) (Perrigo, 2023).