On Conditional and Compositional Language Model Differentiable Prompting
On Conditional and Compositional Language Model Differentiable Prompting
Jonathan Pilault, Can Liu, Mohit Bansal, Markus Dreyer
Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence
Main Track. Pages 4136-4144.
https://doi.org/10.24963/ijcai.2023/460
Prompts have been shown to be an effective method to adapt a frozen Pretrained Language Model (PLM) to perform well on downstream tasks. Prompts can be represented by a human-engineered word sequence or by a learned continuous embedding.
In this work, we investigate conditional and compositional differentiable prompting.
We propose a new model, Prompt Production System (ProPS), which learns to transform task instructions or input metadata, into continuous prompts that elicit task-specific outputs from the PLM.
Our model uses a modular network structure based on our neural formulation of Production Systems, which allows the model to learn discrete rules -- neural functions that learn to specialize in transforming particular prompt input patterns, making it suitable for compositional transfer learning and few-shot learning.
We present extensive empirical and theoretical analysis and show that ProPS consistently surpasses other PLM adaptation techniques, and often improves upon fully fine-tuned models, on compositional generalization tasks, controllable summarization and multilingual translation, while needing fewer trainable parameters.
Keywords:
Machine Learning: ML: Multi-task and transfer learning
Machine Learning: ML: Neuro-symbolic methods