Publication Type

Conference Proceeding Article

Version

publishedVersion

Publication Date

3-2022

Abstract

Recent studies have shown that strong Natural Language Understanding (NLU) models are prone to relying on annotation biases of the datasets as a shortcut, which goes against the underlying mechanisms of the task of interest. To reduce such biases, several recent works introduce debiasing methods to regularize the training process of targeted NLU models. In this paper, we provide a new perspective with causal inference to fnd out the bias. On the one hand, we show that there is an unobserved confounder for the natural language utterances and their respective classes, leading to spurious correlations from training data. To remove such confounder, the backdoor adjustment with causal intervention is utilized to fnd the true causal effect, which makes the training process fundamentally different from the traditional likelihood estimation. On the other hand, in inference process, we formulate the bias as the direct causal effect and remove it by pursuing the indirect causal effect with counterfactual reasoning. We conduct experiments on large-scale natural language inference and fact verifcation benchmarks, evaluating on bias sensitive datasets that are specifcally designed to assess the robustness of models against known biases in the training data. Experimental results show that our proposed debiasing framework outperforms previous stateof-the-art debiasing methods while maintaining the original in-distribution performance.

Discipline

Artificial Intelligence and Robotics | Graphics and Human Computer Interfaces

Research Areas

Data Science and Engineering

Publication

Proceedings of the 36th AAAI Conference on Artificial Intelligence, Virtual Conference, February 22 - March 1

Volume

36

First Page

11376

Last Page

11384

ISBN

9781577358763

Identifier

10.1609/aaai.v36i10.21389

Publisher

AAAI

City or Country

Virtual Conference

Additional URL

http://doi.org/10.1609/aaai.v36i10.21389

Share

COinS