Multi-label sentiment analysis on 100 languages with dynamic weighting for label imbalance

buir.contributor.authorYılmaz , Selim Fırat
buir.contributor.authorKaynak , Ergün Batuhan
buir.contributor.authorKoç , Aykut
buir.contributor.authorDibeklioğlu, Hamdi
buir.contributor.authorKozat , Süleyman Serdar
buir.contributor.orcidYılmaz, Selim Fırat|0000-0002-0486-7731
buir.contributor.orcidKaynak, Ergün Batuhan|0000-0002-3249-3343
buir.contributor.orcidKoç, Aykut|0000-0002-6348-2663
buir.contributor.orcidDibeklioğlu, Hamdi|0000-0003-0851-7808
buir.contributor.orcidKozat, Süleyman Serdar|0000-0002-6488-3848
dc.citation.epage343en_US
dc.citation.issueNumber1
dc.citation.spage331
dc.citation.volumeNumber34
dc.contributor.authorYılmaz, Selim Fırat
dc.contributor.authorKaynak, Ergün Batuhan
dc.contributor.authorKoç, Aykut
dc.contributor.authorDibeklioğlu, Hamdi
dc.contributor.authorKozat, Süleyman Serdar
dc.date.accessioned2024-03-19T05:20:23Z
dc.date.available2024-03-19T05:20:23Z
dc.date.issued2023-01-01
dc.departmentDepartment of Electrical and Electronics Engineering
dc.departmentDepartment of Computer Engineering
dc.departmentNational Magnetic Resonance Research Center (UMRAM)
dc.description.abstractWe investigate cross-lingual sentiment analysis, which has attracted significant attention due to its applications in various areas including market research, politics, and social sciences. In particular, we introduce a sentiment analysis framework in multi-label setting as it obeys Plutchik’s wheel of emotions. We introduce a novel dynamic weighting method that balances the contribution from each class during training, unlike previous static weighting methods that assign non-changing weights based on their class frequency. Moreover, we adapt the focal loss that favors harder instances from single-label object recognition literature to our multi-label setting. Furthermore, we derive a method to choose optimal class-specific thresholds that maximize the macro-f1 score in linear time complexity. Through an extensive set of experiments, we show that our method obtains the state-of-the-art performance in seven of nine metrics in three different languages using a single model compared with the common baselines and the best performing methods in the SemEval competition. We publicly share our code for our model, which can perform sentiment analysis in 100 languages, to facilitate further research.
dc.description.provenanceMade available in DSpace on 2024-03-19T05:20:23Z (GMT). No. of bitstreams: 1 Multi-label_sentiment_analysis_on_100_languages_with_dynamic_weighting_for_label_imbalance.pdf: 1518693 bytes, checksum: 183cfd89fab38fd144a502cd9390a229 (MD5) Previous issue date: 2023-01-01en
dc.identifier.doi10.1109/TNNLS.2021.3094304en_US
dc.identifier.eissn2162-2388en_US
dc.identifier.issn2162-237Xen_US
dc.identifier.urihttps://hdl.handle.net/11693/114920en_US
dc.language.isoEnglishen_US
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_US
dc.relation.isversionofhttps://dx.doi.org/10.1109/TNNLS.2021.3094304
dc.source.titleIEEE Transactions on Neural Networks and Learning Systems
dc.subjectCross-lingual
dc.subjectLabel imbalance
dc.subjectMacro-f1 maximization
dc.subjectMulti-label
dc.subjectNatural language processing (NLP)
dc.subjectSentiment analysis
dc.subjectSocial media
dc.titleMulti-label sentiment analysis on 100 languages with dynamic weighting for label imbalance
dc.typeArticle

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Multi-label_sentiment_analysis_on_100_languages_with_dynamic_weighting_for_label_imbalance.pdf
Size:
1.45 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
2.01 KB
Format:
Item-specific license agreed upon to submission
Description: