论文标题
在基于迅速的多任务学习中衡量社会偏见
On Measuring Social Biases in Prompt-Based Multi-Task Learning
论文作者
论文摘要
大型语言模型经过培训的NLP任务的混合物,这些任务被使用提示转换为文本对文本格式,可以将其推广到新颖的语言形式并处理新任务。迅速工程中的大量工作试图了解输入形式和提示实现卓越性能的影响。我们考虑一种替代措施,并询问输入的编码方式是否影响产出的社会偏见。在本文中,我们研究了T0,这是一种大规模的多任务文本到文本语言模型,该语言模型训练了基于及时的学习。我们考虑了两种不同形式的语义上等效输入:问题 - 答案格式和前提 - 假设格式。我们对以前的烧烤使用现有的偏差基准,并用手工编写的假设创建了自然语言推断BBNLI的第一个偏差基准,同时还将每个基准转换为另一种形式。两种基准的结果表明,在两个不同的表述中,与前提 - 假设形式相比,在训练中可以看出的两个不同的输入,T0显着作用更有偏见,这在训练中可以看出,这与其培训示例不同。代码和数据以https://github.com/feyzaakyurek/bbnli发布。
Large language models trained on a mixture of NLP tasks that are converted into a text-to-text format using prompts, can generalize into novel forms of language and handle novel tasks. A large body of work within prompt engineering attempts to understand the effects of input forms and prompts in achieving superior performance. We consider an alternative measure and inquire whether the way in which an input is encoded affects social biases promoted in outputs. In this paper, we study T0, a large-scale multi-task text-to-text language model trained using prompt-based learning. We consider two different forms of semantically equivalent inputs: question-answer format and premise-hypothesis format. We use an existing bias benchmark for the former BBQ and create the first bias benchmark in natural language inference BBNLI with hand-written hypotheses while also converting each benchmark into the other form. The results on two benchmarks suggest that given two different formulations of essentially the same input, T0 conspicuously acts more biased in question answering form, which is seen during training, compared to premise-hypothesis form which is unlike its training examples. Code and data are released under https://github.com/feyzaakyurek/bbnli.