Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: In-context learning (ICL) enables efficient few-shot learning in large
language models (LLMs) without training, but suffers from the quadratic input
complexity of transformers, limiting the maximum number of exemplars. While
various efficient ICL approaches partition the context into blocks to process
(e.g., ensembling, compression, cross-attention), they often ignore the
information redundancy or under-representation caused by different partition
strategies, leading to suboptimal performance. To tackle this problem, we
propose Sub-CP, a block-aware context selection framework that leverages
submodular objectives to control block diversity. Sub-CP supports a flexible
spectrum of selection strategies, allowing each block to range from globally
diverse to locally coherent. This allows fine-grained control over semantic
structure while enabling precomputation. Extensive experiments across diverse
tasks on multiple datasets show that Sub-CP consistently improves performance
across model scales.
Key Contributions
Proposes Sub-CP, a block-aware context selection framework that uses submodular objectives to control block diversity for in-context learning. This approach mitigates information redundancy and under-representation caused by partitioning strategies, leading to improved performance across diverse tasks.
Business Value
Enables more effective and efficient use of LLMs in few-shot learning scenarios, reducing computational costs and improving accuracy. This is valuable for applications requiring rapid adaptation to new tasks with limited data.