File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Perturbed Masking: Parameter-free Probing for Analyzing and Interpreting BERT

TitlePerturbed Masking: Parameter-free Probing for Analyzing and Interpreting BERT
Authors
Issue Date2020
PublisherAssociation for Computational Linguistics.
Citation
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), Virtual Conference, 5-10 July 2020, p. 4166-4176 How to Cite?
AbstractBy introducing a small set of additional parameters, a probe learns to solve specific linguistic tasks (e.g., dependency parsing) in a supervised manner using feature representations (e.g., contextualized embeddings). The effectiveness of such probing tasks is taken as evidence that the pre-trained model encodes linguistic knowledge. However, this approach of evaluating a language model is undermined by the uncertainty of the amount of knowledge that is learned by the probe itself. Complementary to those works, we propose a parameter-free probing technique for analyzing pre-trained language models (e.g., BERT). Our method does not require direct supervision from the probing tasks, nor do we introduce additional parameters to the probing process. Our experiments on BERT show that syntactic trees recovered from BERT using our method are significantly better than linguistically-uninformed baselines. We further feed the empirically induced dependency structures into a downstream sentiment classification task and find its improvement compatible with or even superior to a human-designed dependency schema.
Persistent Identifierhttp://hdl.handle.net/10722/289172
ISBN

 

DC FieldValueLanguage
dc.contributor.authorWu, Z-
dc.contributor.authorChen, Y-
dc.contributor.authorKao, CM-
dc.contributor.authorLiu, FUN-
dc.date.accessioned2020-10-22T08:08:51Z-
dc.date.available2020-10-22T08:08:51Z-
dc.date.issued2020-
dc.identifier.citationProceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), Virtual Conference, 5-10 July 2020, p. 4166-4176-
dc.identifier.isbn9781952148255-
dc.identifier.urihttp://hdl.handle.net/10722/289172-
dc.description.abstractBy introducing a small set of additional parameters, a probe learns to solve specific linguistic tasks (e.g., dependency parsing) in a supervised manner using feature representations (e.g., contextualized embeddings). The effectiveness of such probing tasks is taken as evidence that the pre-trained model encodes linguistic knowledge. However, this approach of evaluating a language model is undermined by the uncertainty of the amount of knowledge that is learned by the probe itself. Complementary to those works, we propose a parameter-free probing technique for analyzing pre-trained language models (e.g., BERT). Our method does not require direct supervision from the probing tasks, nor do we introduce additional parameters to the probing process. Our experiments on BERT show that syntactic trees recovered from BERT using our method are significantly better than linguistically-uninformed baselines. We further feed the empirically induced dependency structures into a downstream sentiment classification task and find its improvement compatible with or even superior to a human-designed dependency schema.-
dc.languageeng-
dc.publisherAssociation for Computational Linguistics.-
dc.relation.ispartofProceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL)-
dc.titlePerturbed Masking: Parameter-free Probing for Analyzing and Interpreting BERT-
dc.typeConference_Paper-
dc.identifier.emailKao, CM: kao@cs.hku.hk-
dc.identifier.authorityKao, CM=rp00123-
dc.description.naturelink_to_OA_fulltext-
dc.identifier.doi10.18653/v1/2020.acl-main.383-
dc.identifier.hkuros316283-
dc.identifier.spage4166-
dc.identifier.epage4176-
dc.publisher.placeStroudsburg, PA, USA-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats