Show simple item record

dc.contributor.authorImran, Ali Shariq
dc.contributor.authorHodnefjeld, Henrik
dc.contributor.authorKastrati, Zenun
dc.contributor.authorFatima, Noureen
dc.contributor.authorDaudpota, Sher Muhammad
dc.contributor.authorWani, Mudasir Ahmad
dc.date.accessioned2023-10-11T10:25:58Z
dc.date.available2023-10-11T10:25:58Z
dc.date.created2023-07-07T12:29:42Z
dc.date.issued2023
dc.identifier.citationIEEE Access. 2023, 11 55664-55676.en_US
dc.identifier.issn2169-3536
dc.identifier.urihttps://hdl.handle.net/11250/3095761
dc.description.abstractIn the field of text classification, researchers have repeatedly shown the value of transformer-based models such as Bidirectional Encoder Representation from Transformers (BERT) and its variants. Nonetheless, these models are expensive in terms of memory and computational power but have not been utilized to classify long documents of several domains. In addition, transformer models are also often pre-trained on generalized languages, making them less effective in language-specific domains, such as legal documents. In the natural language processing (NLP) domain, there is a growing interest in creating newer models that can handle more complex input sequences and domain-specific languages. Keeping the power of NLP in mind, this study proposes a legal documentation classifier that classifies the legal document by using the sliding window approach to increase the maximum sequence length of the model. We used the ECHR (European Court of Human Rights) publicly available dataset which to a large extent is imbalanced. Therefore, to balance the dataset we have scrapped the case articles from the web and extracted the data. Then, we employed conventional machine learning techniques such as SVM, DT, NB, AdaBoost, and transformer-based neural networks models including BERT, Legal-BERT, RoBERTa, BigBird, ELECTRA, and XLNet for the classification task. The experimental findings show that RoBERTa outperformed all the mentioned BERT versions by obtaining precision, recall, and F1-score of 89.1%, 86.2%, and 86.7%, respectively. While from conventional machine learning techniques, AdaBoost outclasses SVM, DT, and NB by achieving scores of 81.9%, 81.5%, and 81.7% for precision, recall, and F1-score, respectively.en_US
dc.language.isoengen_US
dc.publisherIEEE, Institute of Electrical and Electronics Engineersen_US
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 Internasjonal*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/deed.no*
dc.titleClassifying European Court of Human Rights Cases Using Transformer-Based Techniquesen_US
dc.title.alternativeClassifying European Court of Human Rights Cases Using Transformer-Based Techniquesen_US
dc.typePeer revieweden_US
dc.typeJournal articleen_US
dc.description.versionpublishedVersionen_US
dc.source.pagenumber55664-55676en_US
dc.source.volume11en_US
dc.source.journalIEEE Accessen_US
dc.identifier.doi10.1109/ACCESS.2023.3279034
dc.identifier.cristin2161399
cristin.ispublishedtrue
cristin.fulltextoriginal
cristin.qualitycode1


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

Attribution-NonCommercial-NoDerivatives 4.0 Internasjonal
Except where otherwise noted, this item's license is described as Attribution-NonCommercial-NoDerivatives 4.0 Internasjonal