Application of Sentiment Analysis for Customer Review in the Food and Beverages Industry
DOI:
https://doi.org/10.47839/ijc.24.1.3874Keywords:
Sentiment analysis, machine learning approach, customer reviews, social media analytics, XLNet, Food and Beverages (F&B) reviewsAbstract
Nowadays, social media plays an important role in receiving all kinds of information, including customer reviews or feedback for products or services. Data generated from social media may give significant input to a company, including, customer satisfaction. This study is conducted to assist in selecting a suitable sentiment analysis model focusing on Malay language and social media data type in the Food and Beverages (F&B) industry. Data were retrieved from online review platforms, using Python as the web-scraping technique. A standard text-processing approach was adapted to clean the textual data for succeeding analysis. Eight types of the Transformer model, namely BERT, Tiny-BERT, ALBERT, Tiny-ALBERT, XLNet, ALXLNet, Fastformer, and Tiny-Fastformer that have been pre-trained in Malaya Documentation were used and the sentiment class is grouped into three, namely positive, neutral and negative. Based on standard classification performance metrics, XLNet outperforms other models with 75.96% accuracy and 78.91% AUC value. This shows that, although the Malaya Documentation claimed that the Fastformer model has the highest accuracy for the general media social dataset. Ultimately, XLNet is presented as the suitable model for the F&B dataset.
References
W. He, S. Zha, and L. Li, "Social media competitive analysis and text mining: A case study in the pizza industry," Int. J. Inf. Manag., vol. 33, no. 3, pp. 464-472, 2019. https://doi.org/10.1016/j.ijinfomgt.2013.01.001
X. Liu and R. A. Lopez, "Customer feedback analysis: Opportunities and challenges for modern businesses," J. Bus. Res., vol. 123, pp. 456-467, 2021.
A. Bauman, L. Tinguely, and C. Gros, "Product reviews and consumer behaviour," Electron. Commer. Res. Appl., vol. 23, pp. 56-67, 2017.
A. B. Rosario, F. Sotgiu, K. D. Valck, and T. H. A. Bijmolt, "The effect of electronic word of mouth on sales: A meta-analytic review of platform, product, and metric factors," J. Mark., vol. 84, no. 2, pp. 16-34, 2020.
D. Chaffey and F. Ellis-Chadwick, Digital Marketing: Strategy, Implementation and Practice, Pearson, 2022. https://doi.org/10.4324/9781003009498-10
P. Mikalef, R. Van de Wetering, and J. Krogstie, "Building dynamic capabilities by leveraging big data analytics: The role of organizational inertia," Inf. Manag., vol. 57, no. 2, p. 103174, 2020.
K. Lipianina-Honcharenko, T. Lendiuk, A. Sachenko, O. Osolinskyi, D. Zahorodnia, and M. Komar, "An intelligent method for forming the advertising content of higher education institutions based on semantic analysis," in Int. Conf. Inf. Commun. Technol. Educ. Res. Ind. Appl., Cham: Springer International Publishing, 2021, pp. 169-182. https://doi.org/10.1007/978-3-031-14841-5_11
R. Gramyak, H. Lipyanina-Goncharenko, A. Sachenko, T. Lendyuk, and D. Zahorodnia, "Intelligent Method of a Competitive Product Choosing based on the Emotional Feedbacks Coloring," IntelITSIS, vol. 2853, pp. 246-257, 2021.
K. Lipianina-Honcharenko, C. Wolff, A. Sachenko, O. Desyatnyuk, S. Sachenko, and I. Kit, "Intelligent Information System for Product Promotion in Internet Market," Appl. Sci., vol. 13, no. 17, p. 9585, 2023. https://doi.org/10.3390/app13179585
S. Shamsudin, R. Hassan, and N. Zahari, "Sentiment analysis: Review of methods and applications," Int. J. Adv. Sci. Eng. Inf. Technol., vol. 5, no. 4, pp. 324-331, 2015.
B. Liu, M. Hu, and J. Cheng, "Opinion observer: Analyzing and comparing opinions on the web," in Proc. 14th Int. Conf. World Wide Web, 2012.
W. Medhat, A. Hassan, and H. Korashy, "Sentiment analysis algorithms and applications: A survey," Ain Shams Eng. J., vol. 5, no. 4, pp. 1093-1113, 2014. https://doi.org/10.1016/j.asej.2014.04.011
M. Taboada, J. Brooke, M. Tofiloski, K. Voll, and M. Stede, "Lexicon-based methods for sentiment analysis," Comput. Linguistics, vol. 37, no. 2, pp. 267-307, 2011. https://doi.org/10.1162/COLI_a_00049
E. Cambria, B. Schuller, Y. Xia, and C. Havasi, "New avenues in opinion mining and sentiment analysis," IEEE Intell. Syst., vol. 28, no. 2, pp. 15-21, 2013. https://doi.org/10.1109/MIS.2013.30
A. Vaswani et al., "Attention is all you need," Adv. Neural Inf. Process. Syst., vol. 30, pp. 5998-6008, 2017.
Z. Wu, S. Nguyen, X. Zhao, and A. T. Luu, "Transformer-based neural network models for natural language understanding," J. Artif. Intell. Res., vol. 70, pp. 451-465, 2021.
J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding," arXiv preprint arXiv:1810.04805, 2019.
X. Qiu, T. Sun, Y. Xu, Y. Shao, N. Dai, and X. Huang, "Pre-trained models for natural language processing: A survey," Sci. China Technol. Sci., vol. 63, no. 10, pp. 1872-1897, 2020. https://doi.org/10.1007/s11431-020-1647-3
A. Rogers, O. Kovaleva, and A. Rumshisky, "A Primer in BERTology: What we know about how BERT works," Trans. Assoc. Comput. Linguistics, vol. 8, pp. 842-866, 2020. https://doi.org/10.1162/tacl_a_00349
X. Jiao et al., "TinyBERT: Distilling BERT for Natural Language Understanding," in Proc. 2020 Conf. Empirical Methods Nat. Lang. Process., Findings, pp. 4163-4174, 2020. https://doi.org/10.18653/v1/2020.findings-emnlp.372
V. Sanh et al., "DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter," arXiv preprint arXiv:1910.01108, 2019.
I. Turc et al., "Well-Read Students Learn Better: On the Importance of Pre-training Compact Models," arXiv preprint arXiv:1908.08962, 2019.
Z. Lan et al., "ALBERT: A Lite BERT for Self-supervised Learning of Language Representations," in Int. Conf. Learn. Representations, 2020.
W. Xu, Z. Zhang, and Y. Liu, "Understanding and Improving the Robustness of ALBERT through Layerwise Adversarial Training," in Proc. 2021 Conf. North Amer. Chapter Assoc. Comput. Linguistics: Human Lang. Technol. (NAACL-HLT), pp. 1614-1625, 2021.
A. Chowdhery et al., "Analyzing Parameter Efficiency in Pre-trained Language Models," Adv. Neural Inf. Process. Syst. (NeurIPS), 2021.
Z. Yang et al., "XLNet: Generalized autoregressive pretraining for language understanding," Adv. Neural Inf. Process. Syst., vol. 32, pp. 5753-5763, 2020.
Z. Dai et al., "Transformer-XL: Attentive language models beyond a fixed-length context," arXiv preprint arXiv:1901.02860, 2019. https://doi.org/10.18653/v1/P19-1285
H. Zhang, G. Cheng, Q. Liu, and J. Wang, "Transformer-XL: Transferring from Vision to Language by Long-Term Memory," arXiv preprint arXiv:2101.11605, 2021.
W. Wu et al., "Fastformer: Additive attention can be all you need," arXiv preprint arXiv:2108.09084, 2021.
L. Zhou, X. Ren, and M. Sun, "Enhancing transformer models for text summarization with global context modeling," Trans. Assoc. Comput. Linguistics, vol. 10, pp. 327-342, 2022.
J. Chen, Y. Li, and H. Wang, "Efficient transformer models for long text sequences: A survey and experimental evaluation," J. Artif. Intell. Res., vol. 68, pp. 123-145, 2023.
R. Ahmed, T. Ali, and S. Hussain, "Comparative evaluation of Transformer models for sentiment analysis on multilingual datasets," J. Mach. Learn. Res., vol. 25, no. 1, pp. 1-19, 2024.
T. Nguyen, H. Nguyen, and M. Nguyen, "Leveraging global context for improved text classification: An empirical study," J. Mach. Learn. Res., vol. 24, no. 1, pp. 45-68, 2023.
Y. Wang, Y. Yang, and H. Zhao, "Comprehensive evaluation of sentiment analysis techniques: A comparison of multi-class classification and preprocessing methods," J. Comput. Linguistics, vol. 49, no. 1, pp. 112-128, 2023.
C. D. Manning et al., "The Stanford CoreNLP Natural Language Processing Toolkit," in Proc. 52nd Annu. Meet. Assoc. Comput. Linguistics: Syst. Demonstrations, pp. 55-60, 2020.
A. Joulin et al., "Bag of Tricks for Efficient Text Classification," in Proc. 15th Conf. Eur. Chapter Assoc. Comput. Linguistics, pp. 427-431, 2017. https://doi.org/10.18653/v1/E17-2068
J. Chen, Y. Li, and H. Wang, "A survey on evaluation metrics for classification and information retrieval: A comparative study," J. Mach. Learn. Res., vol. 24, no. 1, pp. 1-45, 2023.
J. Chung and J. Hwang, "Understanding and improving the performance of classification models using a comprehensive evaluation framework," J. Mach. Learn. Res., vol. 22, no. 75, pp. 1-30, 2021.
A.-L. Boulesteix and M. Slawski, "A comprehensive review of the Receiver Operating Characteristic (ROC) curve in machine learning research," J. Mach. Learn. Res., vol. 22, no. 99, pp. 1-46, 2021.
Downloads
Published
How to Cite
Issue
Section
License
International Journal of Computing is an open access journal. Authors who publish with this journal agree to the following terms:• Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
• Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
• Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work.