
Int J Performability Eng ›› 2026, Vol. 22 ›› Issue (5): 253-262.doi: 10.23940/ijpe.26.05.p3.253262
Previous Articles Next Articles
Ritika Bidlan* and Sonal Chawla
Submitted on
;
Revised on
;
Accepted on
Contact:
* E-mail address: ritika_dcsa@pu.ac.in
Ritika Bidlan and Sonal Chawla. Emotion-Driven Music Recommender System: A Novel Deep Learning Approach for Enhanced User Experience [J]. Int J Performability Eng, 2026, 22(5): 253-262.
Add to citation manager EndNote|Reference Manager|ProCite|BibTeX|RefWorks
| [1] Pandeya Y.R., andLee J., 2021. Deep learning-based late fusion of multimodal information for emotion classification of music video. [2] Lucia-Mulas M.J., Revuelta-Sanz P., Ruiz-Mezcua B., andGonzalez-Carrasco I., 2023. Automatic music emotion classification model for movie soundtrack subtitling based on neuroscientific premises: MJ lucia-mulas et al. [3] Gonçalves A.R., Fernandes C., Pasion R., Ferreira-Santos F., Barbosa F., andMarques-Teixeira J., 2018. Effects of age on the identification of emotions in facial expressions: A meta-analysis. [4] Meena G., Mohbey K.K., Indian A., Khan M.Z., andKumar S., 2024. Identifying emotions from facial expressions using a deep convolutional neural network-based approach. [5] Xu C.,2018. A novel recommendation method based on social network using matrix factorization technique. [6] Deshmukh P., andKale G., 2018. A survey of music recommendation system. [7] Krumhansl C.L.,2002. Music: A link between cognition and emotion. [8] Yang Y.H., andChen H.H., 2012. Machine recognition of music emotion: A review. [9] Mahsal Khan S., Abdul Hamid N., andMohd Rashid S., 2019. Music and its familiarity affection on audience decision making. [10] Balakrishnan A., andDixit K., 2014. Deepplaylist: using recurrent neural networks to predict song similarity.Stanfort University, pp. 1-7. [11] Park T., andJeong O.R., 2015. Social network based music recommendation system. [12] Polignano M., Narducci F., de Gemmis M., andSemeraro G., 2021. Towards emotion-aware recommender systems: an affective coherence model based on emotion-driven behaviors. [13] Cowen A.S., andKeltner D., 2017. Self-report captures 27 distinct categories of emotion bridged by continuous gradients. [14] Gunes H., Schuller B., Pantic M., andCowie R., 2011. Emotion representation, analysis and synthesis in continuous space: A survey. In2011 IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp. 827-834. [15] Mary Little Flower T., Jaya T., andChristopher Ezhil Singh S., 2024. Data augmentation using a 1D-CNN model with MFCC/MFMC features for speech emotion recognition. [16] Mishra S.P., Warule P., andDeb S., 2024. Speech emotion recognition using mfcc-based entropy feature. [17] Wani T.M., Gunawan T.S., Qadri S.A.A., Kartiwi M., andAmbikairajah E., 2021. A comprehensive review of speech emotion recognition systems. [18] Yu Z., Zhao M., Wu Y., Liu P., andChen H., 2020. Research on automatic music recommendation algorithm based on facial micro-expression recognition. In2020 39th Chinese Control Conference (CCC), pp. 7257-7263. [19] Bhuiyan A.,2021. [20] Sarkar R., Choudhury S., Dutta S., Roy A., andSaha S.K., 2020. Recognition of emotion in music based on deep convolutional neural network. [21] Yang P.T., Kuang S.M., Wu C.C., andHsu J.L., 2020. Predicting music emotion by using convolutional neural network. InInternational Conference on Human-Computer Interaction, pp. 266-275. [22] Delbouys R., Hennequin R., Piccoli F., Royo-Letelier J., andMoussallam M., 2018. Music mood detection based on audio and lyrics with deep neural net. [23] Dong Y., Yang X., Zhao X., andLi J., 2019. Bidirectional convolutional recurrent sparse network (BCRSN): an efficient model for music emotion recognition. [24] Coutinho E., andCangelosi A., 2011. Musical emotions: predicting second-by-second subjective feelings of emotion from low-level psychoacoustic features and physiological measurements. [25] Barthet M., Fazekas G., andSandler M., 2013. Music emotion recognition: from content-to context-based models. InInternational Symposium on Computer Music Modeling and Retrieval, pp. 228-252. [26] Qian Y., Zhang Y., Ma X., Yu H., andPeng L., 2019. EARS: emotion-aware recommender system based on hybrid information fusion. [27] Lu C.C., andTseng V.S., 2009. A novel method for personalized music recommendation. [28] Han B.J., Rho S., Jun S., andHwang E., 2010. Music emotion classification and context-based music recommendation. [29] Wang X., andWang Y., 2014. Improving content-based and hybrid music recommendation using deep learning. InProceedings of the 22nd ACM International Conference on Multimedia, pp. 627-636. [30] Wang S., Xu C., Ding A.S., andTang Z., 2021. A novel emotion-aware hybrid music recommendation method using deep neural network. [31] Kumble A., Medatati S., andBhatt A., 2023. Emotion-based music recommendation. InInternational Conference on Advanced Computational and Communication Paradigms, pp. 147-155. [32] Deldjoo Y., Schedl M., andKnees P., 2024. Content-driven music recommendation: evolution, state of the art, and challenges. [33] Younis E.M., Mohsen S., Houssein E.H., andIbrahim O.A.S., 2024. Correction: machine learning for human emotion recognition: a comprehensive review. [34] Benzirar A., Hamidi M., andBouami M.F., 2025. Conception of speech emotion recognition methods: A review. [35] Kumari H.M.L.S., Kumari H.M.N.S., andNawarathne U.M.M.P.K., 2024. Speech emotion recognition with hybrid CNN-LSTM and transformers models: evaluating the hybrid model using grad-CAM. [36] Jin Z., andZai W., 2025. Audiovisual emotion recognition based on bi-layer LSTM and multi-head attention mechanism on RAVDESS dataset: Z. Jin, W [37] Ning J., andZhang W., 2025. Speech-based emotion recognition using a hybrid RNN-CNN network. [38] Mares A., Diaz-Arango G., Perez-Jacome-Friscione J., Vazquez-Leal H., Hernandez-Martinez L., Huerta-Chua J., Jaramillo-Alvarado A.F., andDominguez-Chavez A., 2025. Advancing spanish speech emotion recognition: A comprehensive benchmark of pre-trained models. [39] Livingstone S.R., andRusso F.A., 2018. The ryerson audio-visual database of emotional speech and song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in north american english. [40] Hema C., andMarquez F.P.G., 2023. Emotional speech recognition using cnn and deep learning techniques. [41] Ammanavar Y., Patil S., Bidargaddi A.P., Deshanur P., Rathod R., andSM M., 2024. Speech emotion recognition on RAVDESS dataset using deep learning. In2024 5th International Conference for Emerging Technology (INCET), pp. 1-6. [42] Rochlani Y.R., andRaut A.B., 2024. Machine learning approach for detection of speech emotions for RAVDESS audio dataset. In2024 Fourth International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT), pp. 1-7. [43] Liau C.S., andHong K.S., 2023. Multilingual speech emotion recognition using deep learning approach. InInternational Visual Informatics Conference, pp. 526-540. [44] Dal Ri F.A., Ciardi F.C., andConci N., 2023. Speech emotion recognition and deep learning: an extensive validation using convolutional neural networks. [45] Bhangale K., andKothandaraman M., 2023. Speech emotion recognition based on multiple acoustic features and deep convolutional neural network. [46] Rao A.S., Reddy A.P., Vulpala P., Rani K.S., andHemalatha P., 2023. Deep learning structure for emotion prediction using MFCC from native languages. [47] Singh J., Saheer L.B., andFaust O., 2023. Speech emotion recognition using attention model. |
| [1] | Shivani Agarwal, Pancham Singh, Ashish Kumar, Aditya Pratap Singh, and Sakshi Pandey. A Multimodal Deep Learning Framework for Detecting Violence-Encouraging Content on Social Media [J]. Int J Performability Eng, 2026, 22(4): 218-226. |
| [2] | Bhawana Sharma and Komal Saxena. Federated Learning for Heterogeneous Multimodal Emotion Recognition on Edge Devices [J]. Int J Performability Eng, 2026, 22(3): 149-157. |
| [3] | Somen Kumar Roy, Jyothi Pillai, Ani Thomas and Gopal Behera. LGAT: Lightweight Graph Attention Model for Real-Time Session-Based Recommendation System [J]. Int J Performability Eng, 2026, 22(3): 158-166. |
| [4] | Sharma Ji and Abhishek Mishra. A Hybrid Deep Learning-Based IoT System Security Framework for 5G-Enabled Smart Cities [J]. Int J Performability Eng, 2025, 21(7): 361-371. |
| [5] | Somen Roy, Jyothi Pillai, and Ani Thomas. Enhancing Real-Time Session-Based Recommendation System using Light Graph Convolutional Network [J]. Int J Performability Eng, 2025, 21(7): 392-400. |
| [6] | Latika Pinjarkar, Devanshu Sawarkar, Pratham Agrawal, Devansh Motghare, and Nidhi Bansal. Multi Object Image Captioning via CNNs and Transformer Model [J]. Int J Performability Eng, 2025, 21(5): 259-268. |
| [7] | Pancham Singh, Updesh Kumar Jaiswal, Eshank Jain, Nikhil Kumar, and Vimlesh Mishra. A Novel Methodology Utilizing Modern CCTV Cameras and Software as a Service Model for Crime Detection and Prediction [J]. Int J Performability Eng, 2025, 21(2): 112-121. |
| [8] | Shiv Shanker Singh and Ajitesh Kumar. Explainable Adaptive Fusion and Multi-Hop Attention (EAF-MA): an Interpretable Framework for Robust Visual Question Answering [J]. Int J Performability Eng, 2025, 21(12): 686-696. |
| [9] | Kamaljit Singh Saini and Sumit Chaudhary. HydraBoost++: An Optimized Deep Fusion Network for Multi-Class Intruder Detection in IoT Network Security [J]. Int J Performability Eng, 2025, 21(10): 593-604. |
| [10] | Seema Kalonia and Amrita Upadhyay. Comparative Analysis of Machine Learning Model and PSO Optimized CNN-RNN for Software Fault Prediction [J]. Int J Performability Eng, 2025, 21(1): 48-55. |
| [11] | Manu Banga. Enhancing Software Fault Prediction using Machine Learning [J]. Int J Performability Eng, 2024, 20(9): 529-540. |
| [12] | Aditya Dayal Tyagi, and Krishna Asawa. Influence Maximization in Social Network using Community Detection and Node Modularity [J]. Int J Performability Eng, 2024, 20(9): 552-562. |
| [13] | Ekta Singh, and Parma Nand. Efficient Multi-Class Facial Emotion Recognition using YOLOv9: A Deep Learning Approach for Real-Time Applications [J]. Int J Performability Eng, 2024, 20(9): 581-590. |
| [14] | Manpreet Singh, Gauri Jindal, Akshita Oberoi, and Rohan Dhangar. Improving Crime Detection Through Geo-MDA: A Hybrid Linear Regression Approach in Data Mining [J]. Int J Performability Eng, 2024, 20(8): 469-477. |
| [15] | Nilesh Shelke, Deepali Sale, Sagar Shinde, Atul Kathole, and Rachna Somkunwar. A Comprehensive Framework for Facial Emotion Detection using Deep Learning [J]. Int J Performability Eng, 2024, 20(8): 487-497. |
|
||