Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
π Abstract
Abstract: Recent advances in vision-language models have enabled rich semantic
understanding across modalities. However, these encoding methods lack the
ability to interpret or reason about the moral dimensions of content-a crucial
aspect of human cognition. In this paper, we address this gap by introducing
MoralCLIP, a novel embedding representation method that extends multimodal
learning with explicit moral grounding based on Moral Foundations Theory (MFT).
Our approach integrates visual and textual moral cues into a unified embedding
space, enabling cross-modal moral alignment. MoralCLIP is grounded on the
multi-label dataset Social-Moral Image Database to identify co-occurring moral
foundations in visual content. For MoralCLIP training, we design a moral data
augmentation strategy to scale our annotated dataset to 15,000 image-text pairs
labeled with MFT-aligned dimensions. Our results demonstrate that explicit
moral supervision improves both unimodal and multimodal understanding of moral
content, establishing a foundation for morally-aware AI systems capable of
recognizing and aligning with human moral values.
Authors (3)
Ana Carolina Condez
Diogo Tavares
JoΓ£o MagalhΓ£es
Key Contributions
Introduces MoralCLIP, a novel method to extend vision-language models with explicit moral grounding based on Moral Foundations Theory (MFT). It enables cross-modal moral alignment by integrating visual and textual moral cues, addressing the lack of moral reasoning capabilities in current multimodal AI.
Business Value
Enables the development of AI systems that can understand and reason about the ethical and moral implications of content, leading to safer and more responsible AI applications in areas like content moderation and recommendation systems.