Prompt Tuning or Fine-Tuning: Investigating Relational Knowledge in Pre-Trained Language Models.

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

Externe Organisationen

  • Technische Universität Braunschweig
  • Vrije Universiteit Amsterdam
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des SammelwerksConference on Automated Knowledge Base Construction
Seitenumfang15
PublikationsstatusVeröffentlicht - 2021
Extern publiziertJa

Abstract

Extracting relational knowledge from large pre-trained language models by a cloze-style sentence serving as a query has shown promising results. In particular, language models can be queried similar to knowledge graphs. The performance of the relational fact extraction task depends significantly on the query sentence, also known under the term prompt. Tuning these prompts has shown to increase the precision on standard language models by a maximum of around 12% points. However, usually large amounts of data in the form of existing knowledge graph facts and large text corpora are needed to train the required additional model. In this work, we propose using a completely different approach: Instead of spending resources on training an additional model, we simply perform an adaptive fine-tuning of the pre-trained language model on the standard fill-mask task using a small training dataset of existing facts from a knowledge graph. We investigate the differences between complex prompting techniques and adaptive fine-tuning in an extensive evaluation. Remarkably, adaptive fine-tuning outperforms all baselines, even by using significantly fewer training facts. Additionally, we analyze the transfer learning capabilities of this adapted language model by training on a restricted set of relations to show that even fewer training relations are needed to achieve high knowledge extraction quality.

Zitieren

Prompt Tuning or Fine-Tuning: Investigating Relational Knowledge in Pre-Trained Language Models. / Fichtel, Leandra; Kalo, Jan-Christoph; Balke, Wolf-Tilo.
Conference on Automated Knowledge Base Construction. 2021.

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Fichtel, L., Kalo, J.-C., & Balke, W.-T. (2021). Prompt Tuning or Fine-Tuning: Investigating Relational Knowledge in Pre-Trained Language Models. In Conference on Automated Knowledge Base Construction https://doi.org/10.24432/C5RC75
Fichtel L, Kalo JC, Balke WT. Prompt Tuning or Fine-Tuning: Investigating Relational Knowledge in Pre-Trained Language Models. in Conference on Automated Knowledge Base Construction. 2021 doi: 10.24432/C5RC75
Fichtel, Leandra ; Kalo, Jan-Christoph ; Balke, Wolf-Tilo. / Prompt Tuning or Fine-Tuning : Investigating Relational Knowledge in Pre-Trained Language Models. Conference on Automated Knowledge Base Construction. 2021.
Download
@inproceedings{3196ec75bf914320b77d0fdd96fd3ceb,
title = "Prompt Tuning or Fine-Tuning: Investigating Relational Knowledge in Pre-Trained Language Models.",
abstract = "Extracting relational knowledge from large pre-trained language models by a cloze-style sentence serving as a query has shown promising results. In particular, language models can be queried similar to knowledge graphs. The performance of the relational fact extraction task depends significantly on the query sentence, also known under the term prompt. Tuning these prompts has shown to increase the precision on standard language models by a maximum of around 12% points. However, usually large amounts of data in the form of existing knowledge graph facts and large text corpora are needed to train the required additional model. In this work, we propose using a completely different approach: Instead of spending resources on training an additional model, we simply perform an adaptive fine-tuning of the pre-trained language model on the standard fill-mask task using a small training dataset of existing facts from a knowledge graph. We investigate the differences between complex prompting techniques and adaptive fine-tuning in an extensive evaluation. Remarkably, adaptive fine-tuning outperforms all baselines, even by using significantly fewer training facts. Additionally, we analyze the transfer learning capabilities of this adapted language model by training on a restricted set of relations to show that even fewer training relations are needed to achieve high knowledge extraction quality.",
author = "Leandra Fichtel and Jan-Christoph Kalo and Wolf-Tilo Balke",
year = "2021",
doi = "10.24432/C5RC75",
language = "English",
booktitle = "Conference on Automated Knowledge Base Construction",

}

Download

TY - GEN

T1 - Prompt Tuning or Fine-Tuning

T2 - Investigating Relational Knowledge in Pre-Trained Language Models.

AU - Fichtel, Leandra

AU - Kalo, Jan-Christoph

AU - Balke, Wolf-Tilo

PY - 2021

Y1 - 2021

N2 - Extracting relational knowledge from large pre-trained language models by a cloze-style sentence serving as a query has shown promising results. In particular, language models can be queried similar to knowledge graphs. The performance of the relational fact extraction task depends significantly on the query sentence, also known under the term prompt. Tuning these prompts has shown to increase the precision on standard language models by a maximum of around 12% points. However, usually large amounts of data in the form of existing knowledge graph facts and large text corpora are needed to train the required additional model. In this work, we propose using a completely different approach: Instead of spending resources on training an additional model, we simply perform an adaptive fine-tuning of the pre-trained language model on the standard fill-mask task using a small training dataset of existing facts from a knowledge graph. We investigate the differences between complex prompting techniques and adaptive fine-tuning in an extensive evaluation. Remarkably, adaptive fine-tuning outperforms all baselines, even by using significantly fewer training facts. Additionally, we analyze the transfer learning capabilities of this adapted language model by training on a restricted set of relations to show that even fewer training relations are needed to achieve high knowledge extraction quality.

AB - Extracting relational knowledge from large pre-trained language models by a cloze-style sentence serving as a query has shown promising results. In particular, language models can be queried similar to knowledge graphs. The performance of the relational fact extraction task depends significantly on the query sentence, also known under the term prompt. Tuning these prompts has shown to increase the precision on standard language models by a maximum of around 12% points. However, usually large amounts of data in the form of existing knowledge graph facts and large text corpora are needed to train the required additional model. In this work, we propose using a completely different approach: Instead of spending resources on training an additional model, we simply perform an adaptive fine-tuning of the pre-trained language model on the standard fill-mask task using a small training dataset of existing facts from a knowledge graph. We investigate the differences between complex prompting techniques and adaptive fine-tuning in an extensive evaluation. Remarkably, adaptive fine-tuning outperforms all baselines, even by using significantly fewer training facts. Additionally, we analyze the transfer learning capabilities of this adapted language model by training on a restricted set of relations to show that even fewer training relations are needed to achieve high knowledge extraction quality.

U2 - 10.24432/C5RC75

DO - 10.24432/C5RC75

M3 - Conference contribution

BT - Conference on Automated Knowledge Base Construction

ER -

Von denselben Autoren