Time-Aware Language Models as Temporal Knowledge Bases

Author:

Dhingra Bhuwan1,Cole Jeremy R.2,Eisenschlos Julian Martin3,Gillick Daniel4,Eisenstein Jacob5,Cohen William W.6

Affiliation:

1. Google Research. bdhingra@google.com

2. Google Research. jrcole@google.com

3. Google Research. eisenjulian@google.com

4. Google Research. dgillick@google.com

5. Google Research. jeisenstein@google.com

6. Google Research. wcohen@google.com

Abstract

Abstract Many facts come with an expiration date, from the name of the President to the basketball team Lebron James plays for. However, most language models (LMs) are trained on snapshots of data collected at a specific moment in time. This can limit their utility, especially in the closed-book setting where the pretraining corpus must contain the facts the model should memorize. We introduce a diagnostic dataset aimed at probing LMs for factual knowledge that changes over time and highlight problems with LMs at either end of the spectrum—those trained on specific slices of temporal data, as well as those trained on a wide range of temporal data. To mitigate these problems, we propose a simple technique for jointly modeling text with its timestamp. This improves memorization of seen facts from the training time period, as well as calibration on predictions about unseen facts from future time periods. We also show that models trained with temporal context can be efficiently “refreshed” as new data arrives, without the need for retraining from scratch.

Publisher

MIT Press - Journals

Subject

Artificial Intelligence,Computer Science Applications,Linguistics and Language,Human-Computer Interaction,Communication

Reference48 articles.

1. Towards a human-like open-domain chatbot;Adiwardana;CoRR,2020

2. Dynamic word embeddings;Bamler,2017

3. Editing factual knowledge in language models;Cao,2021

4. Learning dynamic author representations with temporal language models;Delasalles,2019

5. BERT: Pre-training of deep bidirectional transformers for language understanding;Devlin,2019

Cited by 23 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Unmasking large language models by means of OpenAI GPT-4 and Google AI: A deep instruction-based analysis;Intelligent Systems with Applications;2024-09

2. Do large language models have a legal duty to tell the truth?;Royal Society Open Science;2024-08

3. Time-CoT for Enhancing Time Reasoning Factual Question Answering in Large Language Models;2024 International Joint Conference on Neural Networks (IJCNN);2024-06-30

4. Continual Learning for Temporal-Sensitive Question Answering;2024 International Joint Conference on Neural Networks (IJCNN);2024-06-30

5. TIQ: A Benchmark for Temporal Question Answering with Implicit Time Constraints;Companion Proceedings of the ACM Web Conference 2024;2024-05-13

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3