Team 5A- Automated Sentiment Based Music Genration_sping
Team 5A- Automated Sentiment Based Music Genration_sping
Speech Content
Swathi Gowroju1, P.srinivas Rao2 ,A.yashwanth Kumar3, B.Sai Siddartha4, V.Sneharika5, K.Sriram6
1
Associate Professor, Dept of CSE (AI&ML)
Sreyas Institute of Engineering and Technology,
Telangana,India.
[email protected]
2
Associate Professor,Dept of CSE (AI&ML)
Sreyas Institute of Engineering and Technology,
Telangana, India
[email protected]
3
Dept of CSE (AI&ML)
Sreyas Institute of Engineering and Technology,
Telangana, India
[email protected]
4
Dept of CSE (AI&ML)
Sreyas Institute of Engineering and Technology,
Telangana, India
[email protected]
5
Dept of CSE (AI&ML)
Sreyas Institute of Engineering and Technology,
Telangana, India
[email protected]
6
Dept of CSE (AI&ML)
Sreyas Institute of Engineering and Technology,
Telangana, India
[email protected]
I. Introduction
Especially in digital content making, one of the greatest increases in demand in
recent times has been that of personalized and emotionally engaging multimedia
experiences. With increasing use of podcasts, audiobooks, storytelling, and virtual
assistants, there arises a need for further enhancement of spoken content by
integrating music that conveys the intended emotional tones. Background music
can enhance emotional quality, set a specific mood, and draw listeners deeper into
an experience. However, manually composing or selecting the right music for
spoken content can often be quite tedious and difficult, requiring a good
understanding of the emotional undertone within the speech while knowing the
expanse of music that supports that emotion.
The difficulties are due to a fact that speech emotions are often nuanced and
varied, making it difficult to locate or generate music that follows the mood in a
particular conversation or narrative. Music selection through human intuition or
pre-set playlists may lead to mismatches where the listening experience is less
than favourable. This gap between the emotional tone of speech and the music to
accompany it has made it clear that a simple automation of assessing speech
content and generation of complimentary background music, in real-time, is quite
needed.
Music: a lively music Balances clarity and Speech and music are
file loudness while mixed together in this
combining the two audio file, with the music
audio sources. matching the sentiment.
Speech and Music in a Produces a merged file Final Product: A voice
Mixed Audio File that may be played recording with uplifting
back. music playing in the
background.
Ⅴ. CONCLUSION
The proposed method was implemented successfully using Python and relevant
libraries. For the proposed method, the background music data set from
Kaggle.com was used. Using the natural language processing technique, speech
was converted to text, from which the user sentiment was identified for generating
the background music. Based on sentiment, specific background music would be
selected and mixed with the original speech to create a new speech with
background music. In existing methods, background music is manually added to
original speech. The proposed methodology provides an advanced and efficient
solution that allows for automated sentiment-based generation of music for speech
content. The methodology here proposed outperforms those aspired by the
existing state-of-the-art technologies in music generation.
References
[1] Ferreira, Lucas N. and E. James Whitehead. “Learning to Generate Music With Sentiment.”
International Society for Music Information Retrieval Conference (2021).
[2] Jean-Pierre Briot, Gaëtan Hadjeres, and François Pa-chet. Deep learning techniques for music
generation-a survey. ArXiv preprint arXiv:1709.01620, 2017.
[3] Gaëtan Hadjeres, François Pachet, and Frank Nielsen. Deepbach: a steerable model for bach
chorales generation. In Proceedings of the 34th International Conference on Machine Learning-
Volume 70, pages 1362–1371. JMLR. org, 2017.
[4] Ben Krause, Iain Murray, Steve Renals, and Liang Lu. Multiplicative LSTM for sequence
modelling. ICLR Workshop track, 2017.
[5] Sageev Oore, Ian Simon, Sander Dieleman, and Doug Eck. Learning to create piano
performances. In NIPS 2017 Workshop on Machine Learning for Creativity and Design, 2017.
[6] Ruining He and Julian McAuley. Ups and downs: Modeling the visual evolution of fashion trends
with one-class collaborative filtering. In Proceedings of the 25th International Conference on
World Wide Web, WWW ’16, pages 507–517, Republic and Canton of Geneva, Switzerland,
2016. International World Wide Web Conferences Steering Committee.
[7] Sixian Chen, John Bowers, and Abigail Durrant. ’ambient walk’: A mobile application for
mindful walking with sonification of biophysical data. In Proceedings of the 2015 British HCI
Conference, British HCI ’15,pages 315–315, New York, NY, USA, 2015. ACM.
[8] Hannah Davis and Saif M Mohammad. Generating music from literature. Proceedings of the 3rd
Workshop on Computational Linguistics for Literature (CLfL),pages 1–10, 2014.
[9] Eduardo R Miranda, Wendy L Magee, John J Wilson, Joel Eaton, and Ramaswamy Palaniappan.
Brain computer music interfacing (bcmi): from basic research to the real world of special needs.
Music & Medicine, 3(3):134–140, 2011.
[10] Kristine Monteith, Tony R Martinez, and Dan Ventura. Automatic generation of music for
inducing emotive response. In International Conference on Computational Creativity, pages 140–
149, 2010.
[11] Briot, Jean-Pierre, and François Pachet. "Deep learning for music generation: challenges and
directions." Neural Computing and Applications 32, no. 4 (2020): 981-993.
[12] Conklin, Darrell. "Music generation from statistical models." In Proceedings of the AISB 2003
Symposium on Artificial Intelligence and Creativity in the Arts and Sciences, pp. 30-35. 2003.
[13] Ji, Shulei, Xinyu Yang, and Jing Luo. "A survey on deep learning for symbolic music generation:
Representations, algorithms, evaluations, and challenges." ACM Computing Surveys 56, no. 1
(2023): 1-39.
[14] Van Der Merwe, Andries, and Walter Schulze. "Music generation with markov models." IEEE
multimedia 18, no. 3 (2010): 78-85.
[15] Mangal, Sanidhya, Rahul Modak, and Poorva Joshi. "LSTM based music generation system."
arXiv preprint arXiv:1908.01080 (2019).
[16] Gowroju, Swathi, Sandeep Kumar, Aarti, and Anshu Ghimire. "Deep Neural Network for
Accurate Age Group Prediction through Pupil Using the Optimized UNet Model." Mathematical
Problems in Engineering 2022 (2022): 1-24.
[17] Swathi, A., Aarti, and Sandeep Kumar. "A smart application to detect pupil for small dataset with
low illumination." Innovations in Systems and Software Engineering 17 (2021): 29-43.
[18] Gowroju, Swathi, Aarti, and Sandeep Kumar. "Review on secure traditional and machine
learning algorithms for age prediction using IRIS image." Multimedia Tools and Applications 81,
no. 24 (2022): 35503-35531.
[19] Swathi Gowroju, “A novel implementation of fast phrase search for encrypted cloud storage”
(IJSREM-2019), volume-3-issue-09. ISSN: 2590-1892
[20] Swathi, A., and Shilpa Rani. "Intelligent fatigue detection by using ACS and by avoiding false
alarms of fatigue detection." In Innovations in Computer Science and Engineering: Proceedings
of the Sixth ICICSE 2018, pp. 225-233. Springer Singapore, 2019.
[21] Gowroju, Swathi, and Sandeep Kumar. "Robust deep learning technique: U-Net architecture for
pupil segmentation." In 2020 11th IEEE Annual Information Technology, Electronics and Mobile
Communication Conference (IEMCON), pp. 0609-0613. IEEE, 2020.
[22] Swathi, A., Aarti, V. Swathi, Y. Sirisha, M. Rishitha, S. Tejaswi, L. Shashank Reddy, and M.
Sujith Reddy. "A Reliable Novel Approach of Bio-Image Processing—Age and Gender
Prediction." In Proceedings of Fourth International Conference on Computer and
Communication Technologies: IC3T 2022, pp. 329-336. Singapore: Springer Nature Singapore,
2023.
[23] Swathi, A., and Shilpa Rani. "Intelligent fatigue detection by using ACS and by avoiding false
alarms of fatigue detection." In Innovations in Computer Science and Engineering: Proceedings
of the Sixth ICICSE 2018, pp. 225-233. Springer Singapore, 2019.