Mostrar registro simples

dc.contributor.advisorBecker, Karinpt_BR
dc.contributor.authorSáenz, Carlos Abel Córdovapt_BR
dc.date.accessioned2022-08-20T04:55:52Zpt_BR
dc.date.issued2022pt_BR
dc.identifier.urihttp://hdl.handle.net/10183/247549pt_BR
dc.description.abstractBERT produces state-of-the-art solutions for many natural language processing tasks at the cost of interpretability. As works discuss the value of BERT’s attention weights to this purpose, we contribute with an attention-based interpretability framework to identify the most influential words for stance classification using BERT-based models. Unlike related work, we develop a broader level of interpretability focused on the overall model behavior instead of single instances. We aggregate tokens’ attentions into words’ attention weights that are more meaningful and can be semantically related to the domain. We propose attention metrics to assess words’ influence in the correct classification of stances. We use three case studies related to COVID-19 to assess the proposed framework in a broad experimental setting encompassing six datasets and four BERT pre-trained models for Portuguese and English languages, resulting in sixteen stance classification models. Through establishing five different research questions, we obtained valuable insights on the usefulness of attention weights to interpret stance classification that allowed us to generalize our findings. Our results are independent of a particular pre-trained BERT model and comparable to those obtained using an alternative baseline method. High attention scores improve the probability of finding words that positively impact the model performance and influence the correct classification (up to 82% of identified influential words contribute to correct predictions). The influential words represent the domain and can be used to identify how the model leverages the arguments expressed to predict a stance.en
dc.format.mimetypeapplication/pdfpt_BR
dc.language.isoporpt_BR
dc.rightsOpen Accessen
dc.subjectInterpretabilityen
dc.subjectProcessamento de linguagem naturalpt_BR
dc.subjectCOVID-19pt_BR
dc.subjectBERTen
dc.subjectAttentionen
dc.subjectStance classificationen
dc.titleUnderstanding stance classification of BERT models : an attention-based mechanismpt_BR
dc.typeDissertaçãopt_BR
dc.identifier.nrb001147927pt_BR
dc.degree.grantorUniversidade Federal do Rio Grande do Sulpt_BR
dc.degree.departmentInstituto de Informáticapt_BR
dc.degree.programPrograma de Pós-Graduação em Computaçãopt_BR
dc.degree.localPorto Alegre, BR-RSpt_BR
dc.degree.date2022pt_BR
dc.degree.levelmestradopt_BR


Thumbnail
   

Este item está licenciado na Creative Commons License

Mostrar registro simples