Diagnosing what is in a Language Model: On the Pitfalls of Probes and Prompts

Thumbnail

Event details

Date 11.11.2022
Hour 15:0017:00
Speaker Deniz Bayazit
Location
Category Conferences - Seminars
EDIC candidacy exam
Exam president: Prof. Boi Faltings
Thesis advisor: Prof. Antoine Bosselut
Co-examiner: Prof. Martin Jaggi

Abstract
Probing and prompting results show that pre-trained language models can encode linguistic and factual properties. These approaches make assumptions, both at the probing and prompting level, that fundamentally affect the conclusions on the abilities of language models.
In this proposal, to verify this hypothesis, we first investigate how probes can memorize linguistic tasks through supervision. Then, we examine how different prompting schemes may overfit the prediction distribution to a standard dataset's golden answer distribution. Finally, we review a propositional logic augmentation to language model prompting that can infer robust and consistent answers.

Following these works on mitigating the pitfalls of probing and prompting, we propose developing behavioral diagnostic tools that can more robustly provide insights into the encoded properties of language models.

Background papers
Designing and Interpreting Probes with Control Tasks
https://aclanthology.org/D19-1275/ 
EMNLP 2019

Knowledgeable or Educated Guess? Revisiting Language Models as Knowledge Bases
https://aclanthology.org/2021.acl-long.146/ 
ACL 2021

Maieutic Prompting: Logically Consistent Reasoning with Recursive Explanations
https://arxiv.org/abs/2205.11822 
EMNLP 2022

 

Practical information

  • General public
  • Free

Tags

EDIC candidacy exam

Share