Michael Wiegand,Josef Ruppenhofer,Elisabeth Eder
Correct Metadata for
Abstract
Abusive language detection is an emerging field in natural language processing which has received a large amount of attention recently. Still the success of automatic detection is limited. Particularly, the detection of implicitly abusive language, i.e. abusive language that is not conveyed by abusive words (e.g. dumbass or scum), is not working well. In this position paper, we explain why existing datasets make learning implicit abuse difficult and what needs to be changed in the design of such datasets. Arguing for a divide-and-conquer strategy, we present a list of subtypes of implicitly abusive language and formulate research tasks and questions for future research.
- Anthology ID:
- 2021.naacl-main.48
- Volume:
- Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
- Month:
- June
- Year:
- 2021
- Address:
- Online
- Editors:
- Kristina Toutanova,Anna Rumshisky,Luke Zettlemoyer,Dilek Hakkani-Tur,Iz Beltagy,Steven Bethard,Ryan Cotterell,Tanmoy Chakraborty,Yichao Zhou
- Venue:
- NAACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 576–587
- Language:
- URL:
- https://aclanthology.org/2021.naacl-main.48/
- DOI:
- 10.18653/v1/2021.naacl-main.48
- Bibkey:
- Cite (ACL):
- Michael Wiegand, Josef Ruppenhofer, and Elisabeth Eder. 2021. Implicitly Abusive Language – What does it actually look like and why are we not getting there?. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 576–587, Online. Association for Computational Linguistics.
- Cite (Informal):
- Implicitly Abusive Language – What does it actually look like and why are we not getting there? (Wiegand et al., NAACL 2021)
- Copy Citation:
- PDF:
- https://aclanthology.org/2021.naacl-main.48.pdf
- Video:
- https://aclanthology.org/2021.naacl-main.48.mp4
PDFCiteSearchVideoFix data
Export citation
- BibTeX
- MODS XML
- Endnote
- Preformatted
@inproceedings{wiegand-etal-2021-implicitly-abusive, title = "Implicitly Abusive Language {--} What does it actually look like and why are we not getting there?", author = "Wiegand, Michael and Ruppenhofer, Josef and Eder, Elisabeth", editor = "Toutanova, Kristina and Rumshisky, Anna and Zettlemoyer, Luke and Hakkani-Tur, Dilek and Beltagy, Iz and Bethard, Steven and Cotterell, Ryan and Chakraborty, Tanmoy and Zhou, Yichao", booktitle = "Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", month = jun, year = "2021", address = "Online", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2021.naacl-main.48/", doi = "10.18653/v1/2021.naacl-main.48", pages = "576--587", abstract = "Abusive language detection is an emerging field in natural language processing which has received a large amount of attention recently. Still the success of automatic detection is limited. Particularly, the detection of implicitly abusive language, i.e. abusive language that is not conveyed by abusive words (e.g. dumbass or scum), is not working well. In this position paper, we explain why existing datasets make learning implicit abuse difficult and what needs to be changed in the design of such datasets. Arguing for a divide-and-conquer strategy, we present a list of subtypes of implicitly abusive language and formulate research tasks and questions for future research."}
Download as File
<?xml version="1.0" encoding="UTF-8"?><modsCollection xmlns="http://www.loc.gov/mods/v3"><mods ID="wiegand-etal-2021-implicitly-abusive"> <titleInfo> <title>Implicitly Abusive Language – What does it actually look like and why are we not getting there?</title> </titleInfo> <name type="personal"> <namePart type="given">Michael</namePart> <namePart type="family">Wiegand</namePart> <role> <roleTerm authority="marcrelator" type="text">author</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Josef</namePart> <namePart type="family">Ruppenhofer</namePart> <role> <roleTerm authority="marcrelator" type="text">author</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Elisabeth</namePart> <namePart type="family">Eder</namePart> <role> <roleTerm authority="marcrelator" type="text">author</roleTerm> </role> </name> <originInfo> <dateIssued>2021-06</dateIssued> </originInfo> <typeOfResource>text</typeOfResource> <relatedItem type="host"> <titleInfo> <title>Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title> </titleInfo> <name type="personal"> <namePart type="given">Kristina</namePart> <namePart type="family">Toutanova</namePart> <role> <roleTerm authority="marcrelator" type="text">editor</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Anna</namePart> <namePart type="family">Rumshisky</namePart> <role> <roleTerm authority="marcrelator" type="text">editor</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Luke</namePart> <namePart type="family">Zettlemoyer</namePart> <role> <roleTerm authority="marcrelator" type="text">editor</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Dilek</namePart> <namePart type="family">Hakkani-Tur</namePart> <role> <roleTerm authority="marcrelator" type="text">editor</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Iz</namePart> <namePart type="family">Beltagy</namePart> <role> <roleTerm authority="marcrelator" type="text">editor</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Steven</namePart> <namePart type="family">Bethard</namePart> <role> <roleTerm authority="marcrelator" type="text">editor</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Ryan</namePart> <namePart type="family">Cotterell</namePart> <role> <roleTerm authority="marcrelator" type="text">editor</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Tanmoy</namePart> <namePart type="family">Chakraborty</namePart> <role> <roleTerm authority="marcrelator" type="text">editor</roleTerm> </role> </name> <name type="personal"> <namePart type="given">Yichao</namePart> <namePart type="family">Zhou</namePart> <role> <roleTerm authority="marcrelator" type="text">editor</roleTerm> </role> </name> <originInfo> <publisher>Association for Computational Linguistics</publisher> <place> <placeTerm type="text">Online</placeTerm> </place> </originInfo> <genre authority="marcgt">conference publication</genre> </relatedItem> <abstract>Abusive language detection is an emerging field in natural language processing which has received a large amount of attention recently. Still the success of automatic detection is limited. Particularly, the detection of implicitly abusive language, i.e. abusive language that is not conveyed by abusive words (e.g. dumbass or scum), is not working well. In this position paper, we explain why existing datasets make learning implicit abuse difficult and what needs to be changed in the design of such datasets. Arguing for a divide-and-conquer strategy, we present a list of subtypes of implicitly abusive language and formulate research tasks and questions for future research.</abstract> <identifier type="citekey">wiegand-etal-2021-implicitly-abusive</identifier> <identifier type="doi">10.18653/v1/2021.naacl-main.48</identifier> <location> <url>https://aclanthology.org/2021.naacl-main.48/</url> </location> <part> <date>2021-06</date> <extent unit="page"> <start>576</start> <end>587</end> </extent> </part></mods></modsCollection>
Download as File
%0 Conference Proceedings%T Implicitly Abusive Language – What does it actually look like and why are we not getting there?%A Wiegand, Michael%A Ruppenhofer, Josef%A Eder, Elisabeth%Y Toutanova, Kristina%Y Rumshisky, Anna%Y Zettlemoyer, Luke%Y Hakkani-Tur, Dilek%Y Beltagy, Iz%Y Bethard, Steven%Y Cotterell, Ryan%Y Chakraborty, Tanmoy%Y Zhou, Yichao%S Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies%D 2021%8 June%I Association for Computational Linguistics%C Online%F wiegand-etal-2021-implicitly-abusive%X Abusive language detection is an emerging field in natural language processing which has received a large amount of attention recently. Still the success of automatic detection is limited. Particularly, the detection of implicitly abusive language, i.e. abusive language that is not conveyed by abusive words (e.g. dumbass or scum), is not working well. In this position paper, we explain why existing datasets make learning implicit abuse difficult and what needs to be changed in the design of such datasets. Arguing for a divide-and-conquer strategy, we present a list of subtypes of implicitly abusive language and formulate research tasks and questions for future research.%R 10.18653/v1/2021.naacl-main.48%U https://aclanthology.org/2021.naacl-main.48/%U https://doi.org/10.18653/v1/2021.naacl-main.48%P 576-587
Download as File
Markdown (Informal)
[Implicitly Abusive Language – What does it actually look like and why are we not getting there?](https://aclanthology.org/2021.naacl-main.48/) (Wiegand et al., NAACL 2021)
- Implicitly Abusive Language – What does it actually look like and why are we not getting there? (Wiegand et al., NAACL 2021)
ACL
- Michael Wiegand, Josef Ruppenhofer, and Elisabeth Eder. 2021. Implicitly Abusive Language – What does it actually look like and why are we not getting there?. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 576–587, Online. Association for Computational Linguistics.