... | ... | @@ -8,11 +8,11 @@ We are happy that the team agrees to tell us more about their approach and exper |
|
|
|
|
|
**We cordially invite to a JULAIN Talk**
|
|
|
|
|
|
**How to spot toxic, engaging, and fact-claiming comments in German social media - A field report from the GermEval competition**
|
|
|
Tobias Bornheim, Niklas Grieger, Stephan Bialonski
|
|
|
**How to spot toxic, engaging, and fact-claiming comments in German social media - A field report from the GermEval competition**<br>
|
|
|
Tobias Bornheim, Niklas Grieger, Stephan Bialonski <br>
|
|
|
Data Science, FB Medizintechnik und Technomathematik, FH Aachen
|
|
|
|
|
|
**Monday, 20 September 2021, 11am**
|
|
|
**Monday, 20 September 2021, 11am**<br>
|
|
|
**Link to the VC:** https://webconf.fz-juelich.de/b/wen-mym-pj7
|
|
|
|
|
|
**Abstract:** The availability of language representations learned by large pretrained neural network models (such as BERT and ELECTRA) has led to improvements in many downstream Natural Language Processing tasks in recent years. Pretrained models usually differ in pretraining objectives, architectures, and datasets they are trained on which can affect downstream performance. In this talk, we introduce our approach which was ranked 1st place in the GermEval 2021 competition for identifying toxic and fact-claiming comments. We created ensembles of BERT and ELECTRA models and investigated whether and how classification performance depends on the number of ensemble members, their composition, and whether exploiting label-label correlation were helpful to improve classification performance.
|
... | ... | |