Vis enkel innførsel

dc.contributor.authorShaikh, Sarang
dc.contributor.authorDaudpota, Sher Muhammad
dc.contributor.authorImran, Ali Shariq
dc.contributor.authorKastrati, Zenun
dc.date.accessioned2021-03-08T09:50:29Z
dc.date.available2021-03-08T09:50:29Z
dc.date.created2021-01-19T12:16:16Z
dc.date.issued2021
dc.identifier.citationApplied Sciences. 2021, 11 (2), .en_US
dc.identifier.issn2076-3417
dc.identifier.urihttps://hdl.handle.net/11250/2732089
dc.description.abstractData imbalance is a frequently occurring problem in classification tasks where the number of samples in one category exceeds the amount in others. Quite often, the minority class data is of great importance representing concepts of interest and is often challenging to obtain in real-life scenarios and applications. Imagine a customers’ dataset for bank loans-majority of the instances belong to non-defaulter class, only a small number of customers would be labeled as defaulters, however, the performance accuracy is more important on defaulters labels than non-defaulter in such highly imbalance datasets. Lack of enough data samples across all the class labels results in data imbalance causing poor classification performance while training the model. Synthetic data generation and oversampling techniques such as SMOTE, AdaSyn can address this issue for statistical data, yet such methods suffer from overfitting and substantial noise. While such techniques have proved useful for synthetic numerical and image data generation using GANs, the effectiveness of approaches proposed for textual data, which can retain grammatical structure, context, and semantic information, has yet to be evaluated. In this paper, we address this issue by assessing text sequence generation algorithms coupled with grammatical validation on domain-specific highly imbalanced datasets for text classification. We exploit recently proposed GPT-2 and LSTM-based text generation models to introduce balance in highly imbalanced text datasets. The experiments presented in this paper on three highly imbalanced datasets from different domains show that the performance of same deep neural network models improve up to 17% when datasets are balanced using generated text.en_US
dc.language.isoengen_US
dc.rightsNavngivelse 4.0 Internasjonal*
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/deed.no*
dc.titleTowards Improved Classification Accuracy on Highly Imbalanced Text Dataset Using Deep Neural Language Modelsen_US
dc.typePeer revieweden_US
dc.typeJournal articleen_US
dc.description.versionpublishedVersionen_US
dc.source.volume11en_US
dc.source.journalApplied Sciencesen_US
dc.source.issue2en_US
dc.identifier.doi10.3390/app11020869
dc.identifier.cristin1874228
dc.description.localcodeThis is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.en_US
cristin.ispublishedtrue
cristin.fulltextoriginal
cristin.qualitycode1


Tilhørende fil(er)

Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel

Navngivelse 4.0 Internasjonal
Med mindre annet er angitt, så er denne innførselen lisensiert som Navngivelse 4.0 Internasjonal