论文标题
零拍摄实体和推文表征,具有设计的条件提示和上下文
Zero-shot Entity and Tweet Characterization with Designed Conditional Prompts and Contexts
论文作者
论文摘要
从过去的十年开始,在线新闻和社交媒体是事实上的媒体,可以在全球传播信息。但是,意图的内容和目的的偏见不受监管,管理偏见是内容消费者的责任。在这方面,了解新闻来源对特定实体的立场和偏见变得很重要。为了解决这个问题,我们使用了验证的语言模型,这些语言模型已证明在没有特定特定任务培训或几次训练的情况下取得了良好的成绩。在这项工作中,我们将表征指定实体和推文描述为开放式文本分类和开放式事实探测问题的问题。我们评估生成预审计的变形金刚2(GPT-2)的零光语言模型能力,以主观的人类心理学和逻辑的条件条件和逻辑条件的前提和上下文来表征实体和推文。首先,我们在足够大的新闻语料库上微调GPT-2模型,并通过前缀启动来评估语料库中流行实体的主观表征。其次,我们用一些流行的主题标签中的推文语料库微调了GPT-2,并通过使用前缀,问题和上下文提要提示来启动语言模型来评估表征推文。实体表征结果在衡量标准和人类评估之间是正面的。
Online news and social media have been the de facto mediums to disseminate information globally from the beginning of the last decade. However, bias in content and purpose of intentions are not regulated, and managing bias is the responsibility of content consumers. In this regard, understanding the stances and biases of news sources towards specific entities becomes important. To address this problem, we use pretrained language models, which have been shown to bring about good results with no task-specific training or few-shot training. In this work, we approach the problem of characterizing Named Entities and Tweets as an open-ended text classification and open-ended fact probing problem.We evaluate the zero-shot language model capabilities of Generative Pretrained Transformer 2 (GPT-2) to characterize Entities and Tweets subjectively with human psychology-inspired and logical conditional prefixes and contexts. First, we fine-tune the GPT-2 model on a sufficiently large news corpus and evaluate subjective characterization of popular entities in the corpus by priming with prefixes. Second, we fine-tune GPT-2 with a Tweets corpus from a few popular hashtags and evaluate characterizing tweets by priming the language model with prefixes, questions, and contextual synopsis prompts. Entity characterization results were positive across measures and human evaluation.