Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 92% Match Research Paper LLM Researchers,Machine Learning Engineers,NLP Practitioners 4 weeks ago

Submodular Context Partitioning and Compression for In-Context Learning-short paper

large-language-models › model-architecture
📄 Abstract

Abstract: In-context learning (ICL) enables efficient few-shot learning in large language models (LLMs) without training, but suffers from the quadratic input complexity of transformers, limiting the maximum number of exemplars. While various efficient ICL approaches partition the context into blocks to process (e.g., ensembling, compression, cross-attention), they often ignore the information redundancy or under-representation caused by different partition strategies, leading to suboptimal performance. To tackle this problem, we propose Sub-CP, a block-aware context selection framework that leverages submodular objectives to control block diversity. Sub-CP supports a flexible spectrum of selection strategies, allowing each block to range from globally diverse to locally coherent. This allows fine-grained control over semantic structure while enabling precomputation. Extensive experiments across diverse tasks on multiple datasets show that Sub-CP consistently improves performance across model scales.

Key Contributions

Proposes Sub-CP, a block-aware context selection framework that uses submodular objectives to control block diversity for in-context learning. This approach mitigates information redundancy and under-representation caused by partitioning strategies, leading to improved performance across diverse tasks.

Business Value

Enables more effective and efficient use of LLMs in few-shot learning scenarios, reducing computational costs and improving accuracy. This is valuable for applications requiring rapid adaptation to new tasks with limited data.