Skip to yearly menu bar Skip to main content


Virtual presentation / poster accept

Compositional Task Representations for Large Language Models

NAN SHAO · Zefan Cai · Hanwei Xu · Chonghua Liao · Yanan Zheng · Zhilin Yang

Keywords: [ Deep Learning and representational learning ]


Abstract:

Large language models have shown a remarkable cross-task generalization ability. Most prior work assumed that prompts effectively extract knowledge from language models to facilitate generalization to new tasks. This perspective led to numerous studies on improving prompts. In contrast, we introduce a new perspective, compositional generalization, that views each task as a composition of latent codes and generalizes to test tasks by a new composition of seen codes. To this end, we propose a novel prompt-free approach, Compositional Task Representations (CTR), that employs multi-task training to learn a discrete, compositional codebook. Empirically, our CTR substantially outperforms prompt-based methods in zero-label learning on average. According to our analysis, some of the learned CTR codes are interpretable to human and demonstrate a certain degree of controllability.

Chat is not available.