ClientUSCToolsQualitative Coding, Data Structuring, NVivo, Python-adjacent annotation workflowsTimeline2017–2020TagsAI Ethics, Moral Psychology, NLP Training, Civic Reasoning, Cross-Disciplinary Research, Responsible AI, Dataset Curation, Research for Impact, Annotated Language Models
Artificial intelligence systems increasingly shape how we access information, interact socially, and even understand justice. But for these systems to behave ethically, they must be trained on data that reflects real human values. At USC’s Lab of Social and Moral Cognition, our challenge was to understand how morality shows up in everyday language—on platforms like Twitter or news media—and how we could transform that raw data into annotated corpora that train more value-aligned AI systems. We explored how moral concepts like fairness, harm, authority, and loyalty manifest in civic discourse and whether those patterns could be reliably detected and modeled.
This work deepened my understanding of how to bridge human insight with the technical demands of AI systems. I learned to annotate and structure large qualitative datasets using a custom schema—balancing rigor and empathy—to train NLP models that could detect real-world social signals, such as approximate crime rates in underreported areas. It also showed me how moral concepts like fairness and harm can be translated into computational patterns, expanding what AI can detect and understand.
But I also learned how complex and subjective human morality really is—like how some saw a simple “Happy birthday” as a moral comment, while others viewed it as an obligatory social norm. Grappling with that ambiguity taught me that responsible AI demands both technical precision and deep social and cultural awareness. Ultimately, this experience reinforced my conviction that behavioral scientists are critical to developing AI that is not just technically advanced, but also socially meaningful and ethically grounded.