Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Methodology/System Paper AI safety researchers,Developers of autonomous agents,AI ethicists 4 weeks ago

Contextual Integrity in LLMs via Reasoning and Reinforcement Learning

ai-safety › alignment
📄 Abstract

Abstract: As the era of autonomous agents making decisions on behalf of users unfolds, ensuring contextual integrity (CI) -- what is the appropriate information to share while carrying out a certain task -- becomes a central question to the field. We posit that CI demands a form of reasoning where the agent needs to reason about the context in which it is operating. To test this, we first prompt LLMs to reason explicitly about CI when deciding what information to disclose. We then extend this approach by developing a reinforcement learning (RL) framework that further instills in models the reasoning necessary to achieve CI. Using a synthetic, automatically created, dataset of only $\sim700$ examples but with diverse contexts and information disclosure norms, we show that our method substantially reduces inappropriate information disclosure while maintaining task performance across multiple model sizes and families. Importantly, improvements transfer from this synthetic dataset to established CI benchmarks such as PrivacyLens that has human annotations and evaluates privacy leakage of AI assistants in actions and tool calls.

Key Contributions

Addresses contextual integrity (CI) for autonomous agents by developing a framework that instills reasoning about appropriate information disclosure. It combines explicit prompting with a reinforcement learning approach, significantly reducing inappropriate disclosures while maintaining task performance across various LLMs.

Business Value

Crucial for building trustworthy AI systems, especially those acting on behalf of users, by preventing privacy breaches and ensuring responsible information handling.