Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 95% Match Research Paper AI Ethics Researchers,AI Safety Researchers,Developers of responsible AI systems,Social Scientists 5 days ago

MoralCLIP: Contrastive Alignment of Vision-and-Language Representations with Moral Foundations Theory

ai-safety β€Ί alignment
πŸ“„ Abstract

Abstract: Recent advances in vision-language models have enabled rich semantic understanding across modalities. However, these encoding methods lack the ability to interpret or reason about the moral dimensions of content-a crucial aspect of human cognition. In this paper, we address this gap by introducing MoralCLIP, a novel embedding representation method that extends multimodal learning with explicit moral grounding based on Moral Foundations Theory (MFT). Our approach integrates visual and textual moral cues into a unified embedding space, enabling cross-modal moral alignment. MoralCLIP is grounded on the multi-label dataset Social-Moral Image Database to identify co-occurring moral foundations in visual content. For MoralCLIP training, we design a moral data augmentation strategy to scale our annotated dataset to 15,000 image-text pairs labeled with MFT-aligned dimensions. Our results demonstrate that explicit moral supervision improves both unimodal and multimodal understanding of moral content, establishing a foundation for morally-aware AI systems capable of recognizing and aligning with human moral values.
Authors (3)
Ana Carolina Condez
Diogo Tavares
JoΓ£o MagalhΓ£es
Submitted
June 6, 2025
arXiv Category
cs.CV
arXiv PDF

Key Contributions

Introduces MoralCLIP, a novel method to extend vision-language models with explicit moral grounding based on Moral Foundations Theory (MFT). It enables cross-modal moral alignment by integrating visual and textual moral cues, addressing the lack of moral reasoning capabilities in current multimodal AI.

Business Value

Enables the development of AI systems that can understand and reason about the ethical and moral implications of content, leading to safer and more responsible AI applications in areas like content moderation and recommendation systems.