Emotion Based Music Recommendation System
Emotion Based Music Recommendation System
Emotion Based Music Recommendation System
ISSN No:-2456-2165
Abstract:- The internet and mobile technology have recommends songs to user according to mood of the user
developed quickly and made it possible for us to freely which was earlier classified into different emotions. In this
access various music resources. While the music industry way user can get recommendation of songs based on his
might lean more toward certain genres of music. But current mood and will change dynamically based on current
there is no particular way by which we can understand mood.
what exactly user wants to listen based on current mood
or emotion. Music is a great way to express emotions and II. RELATED WORK
moods. For example, people like to listen to happy songs
when they are feeling good, a soothing song can help us Emotions are a basic part of human nature. Human’s
to relax when we're feeling stressed or exhausted and emotional states and present moods can be quickly seen
people tend to listen some sort of sad songs when they through their facial expressions. Basic emotions (happy, sad,
angry, excited, surprised, disgusted, fear, and neutral) were
are feeling down. So in this project, we are going to
taken into consideration when developing this system. In
develop a system which will capture the real time
emotion of user by conversating with user or by other this research, face identification was accomplished using a
means and based on that emotion related songs will be convolutional neural network.
recommended. We are going to categorize songs into the Recommended manual playlist segregation and song
groups based on the categories like Happy, Sad, Neutral annotation based on the user's emotional state as a time-
etc. Then according to the captured emotion from the consuming and labor-intensive task. There have been many
user, the songs related to that emotion will be algorithms suggested to automate this process.
recommended. In this way, user can listen the songs
according to the mood. However, the currently used algorithms are slow, use
extra hardware, raise the system's overall cost, and have
Keywords:- Recommendation System, Emotion Recognition, substantially the overall cost of the device and have much
Interactive UI, Mood based classifier. poorer precision.
I. INTRODUCTION It aims to increase the accuracy of the system design.
People frequently use their facial expressions to contrasting a face expression recognition module of the
convey their feelings. It has long been recognized that music system with a dataset that is both user-dependent and user-
may change a person's disposition. A user's mind can be impartial serves to validate it.
gradually calmed down and an overall nice effect can be According to AyushGuidel [1] et al, one may quickly
produced by capturing and recognizing the emotion being determine a person's emotional condition by observing their
uttered by the person and playing appropriate tunes facial expressions. Basic emotions (happy, sad, angry,
matching the one's mood.[7] The goal of the project is to excited, surprised, disgusted, fear, and neutral) were taken
record a person's facial expressions as they exhibit emotion. into account when developing this system. In this research,
The web camera interface for computer systems is used by a face detection was implemented using a convolutional
music player to record human emotion. The software takes a neural network. On the whole, people refer to music as a
user's image and uses image segmentation and image "language of feelings."
processing techniques to extract information from a target
person's face in an effort to determine what emotion they are Sadhvika [4] et al. recommended manual playlist
attempting to convey. The idea attempts to uplift the user's segregation and song annotation based on the user's current
mood by playing music that fits their needs while emotional state as a time-consuming and labor-intensive
simultaneously taking their photograph. Facial expression task. There have been many algorithms suggested to
recognition has been the most effective method of automate this process. However, the currently used
expression analysis known to humanity since ancient times. algorithms are slow, use extra hardware (such EEG
Facial expressions are the best way that people may deduce structures and sensors), raise the system's overall cost, and
or evaluate the emotion, sentiment, or thoughts that another have substantially lower accuracy.
person is attempting to express. Mood modification may
occasionally be useful in overcoming depressive and sad The report suggested by Ramya Ramanathan [3]
circumstances. Many health dangers can be prevented with described an intelligent music player that used emotion
the use of expression analysis, and actions can be done to recognition. A fundamental aspect of human nature is
improve a user's mood as well. emotion. They have the most significant role in life. Human
emotions are intended to be shared and understood by
Our project focuses classification of songs based on others. The local music library of the user is initially sorted
acoustic features. Then system determines the mood of the in accordance with the feelings the album arouses. The
user through its facial expressions and eventually system lyrics of the music are frequently taken into consideration to
[5 ]Given diagram gives the clear understanding of training set, whereas 3,589 examples make up the
how the system is going to work. This includes sequential public test set.
steps in which every step has its own significance. First
system classifies the songs on its acoustic feature like b) Music Mood Dataset
loudness, tempo, energy etc. and the mood of the music is A labeled dataset with the size of 686 is used.
determined. In this way all the songs in the dataset are
classified into different moods and based on the input It has 19 columns including name, album, artist,
received from the user system recommends the songs to the id, release_date, popularity, length, danceability,
user. acousticness, energy, instrumentalness, liveness,
valence, loudness, speechness, tempo, key, time
System mainly divided into 4 components. It includes signature, and mood.
Real Time Capture, Face Recognition, Emotion Detection,
Music Recommendation. While recommending a song The moods are in four categories: Calm, Happy,
system goes throughout these steps. In real time capture Energetic, Sad.The features of Length, Danceability,
system captures the face of the user and in the subsequent Acousticness, Energy, Instrumentalness, Liveness,
step CNN algorithm evaluates the features of that image. Valence, Loudness, Speechiness and Tempo are used
After detecting the mood or emotion of the image system for training the model because they have more
recommends the appropriate song to the user. influence to classify the tracks.
IV. METHODOLOGY
A. Dataset Description:
a) Image Dataset:
The FER2013 dataset was utilized to develop the
model. 48x48 pixel grayscale portraits of faces make
up the data. Anger, Disgust, Fear, Happy, Sad,
Surprise, and Neutral are the seven emotions that are
assigned to each image. 28,709 samples make up the Fig. 2: Music Mood Dataset
The image's class will either be binary or multi-class to When emotion is detected through facial expressions, we
distinguish between different types of clothing or to identify assign the appropriate category to that mood, for example if
digits. The learned properties in a neural network cannot be detected mood is happy then system will recommend songs
understood because neural networks are like a "black box." from happy category those were already categorized.
Consequently, the CNN model basically returns the findings In this way, song is recommended to the user.
after receiving an input image. The model that was learned
by weights using CNN is loaded to detect emotions. When a V. FUTURE SCOPE
user takes a real-time image, the image is submitted to a
CNN model that has already been trained. The CNN model This system is completely operational, however there
then predicts the emotion and adds a label to the image. is still opportunity for advancement. A lot of changes may
be made to the programme to enhance user experience
C. Music Recommendation: generally and provide better outcomes.
Our dataset contains various features that too are also
considered as acoustic features of that particular song, every Currently existing systems face difficulties while
feature has its magnitude which defines the intensity of the giving manual inputs but our system performs well when
that feature. Magnitude may range in different scales. In there is difficulty in giving manual inputs as we are
general, less Magnitude Or negative value of the Magnitude dynamically determining the mood of a person from its
represent that least contribution of that feature to overall expressions. In addition to this, this system not only used as
mood of the song, and obviously if Magnitude has larger a feature but also as a independent system. If further
value then that feature dominantly contribute to the overall technical advancements done and if it improved the
mood of that song.In this way, there are 4-5 main features efficiency of the algorithm used then we can also
amongst 18-19 that contribute more.[8]We get desired mood recommend the songs on crucial and trivial expressions. [6]
for the song after training the model based on that dataset. In Current system mainly focuses on doing the
this way, songs are classified into different categories like recommendation on content based features that is based on
happy, sad, neutral and energetic. its acoustic parameters. But in future hybrid model can be