论文标题
使用冷冻语言模型临床及时学习
Clinical Prompt Learning with Frozen Language Models
论文作者
论文摘要
及时学习是自然语言处理(NLP)领域的新范式,它在许多自然语言任务上表现出令人印象深刻的表现,并以完整,很少且零射的火车评估设置为常见的基准测试文本数据集。最近,甚至已经观察到,迅速学习的大型但冷冻的预训练的语言模型(PLM)的表现要素较小但微调的模型。但是,与许多最近的NLP趋势一样,即使是最大的PLM(例如GPT-3)的性能在专业领域(例如医学文本)上也不能很好地表现,并且实现最新状态的共同做法(SOTA)的结果仍然包括预训练和在下游任务上对PLM进行微调。在通常在非GPU环境中保存数据的临床环境中,依赖微调大型PLM是有问题的,而培训专业领域模型的更多资源有效方法至关重要。我们研究了迅速学习对临床意义的决策任务的生存能力,并与更传统的微调方法直接进行了比较。结果部分与及时的学习文献一致,及时学习能够匹配或改进传统的微调,具有较少的训练参数,并且需要更少的培训数据。我们认为,迅速学习因此提供了适用于临床环境的较低的计算资源成本,这可以作为尺寸PLM的微调替代品的替代方法。可以在:https://github.com/ntaylorox/public_clinical_prompt上找到复制实验的补充代码。
Prompt learning is a new paradigm in the Natural Language Processing (NLP) field which has shown impressive performance on a number of natural language tasks with common benchmarking text datasets in full, few-shot, and zero-shot train-evaluation setups. Recently, it has even been observed that large but frozen pre-trained language models (PLMs) with prompt learning outperform smaller but fine-tuned models. However, as with many recent NLP trends, the performance of even the largest PLMs such as GPT-3 do not perform well on specialized domains (e.g. medical text), and the common practice to achieve State of the Art (SoTA) results still consists of pre-training and fine-tuning the PLMs on downstream tasks. The reliance on fine-tuning large PLMs is problematic in clinical settings where data is often held in non-GPU environments, and more resource efficient methods of training specialized domain models is crucial. We investigated the viability of prompt learning on clinically meaningful decision tasks and directly compared with more traditional fine-tuning methods. Results are partially in line with the prompt learning literature, with prompt learning able to match or improve on traditional fine-tuning with substantially fewer trainable parameters and requiring less training data. We argue that prompt learning therefore provides lower computational resource costs applicable to clinical settings, that can serve as an alternative to fine-tuning ever increasing in size PLMs. Complementary code to reproduce experiments presented in this work can be found at: https://github.com/NtaylorOX/Public_Clinical_Prompt.