Professional Documents
Culture Documents
Music Mood
Chang Bae Moon,1 HyunSoo Kim,1 Dong Won Lee,2
Byeong Man Kim1*
1
Department of Computer Software Engineering, Kumoh National Institute of Technology, Gumi, Gyeongbuk, South Korea
INTRODUCTION
Fig. 1.
Mood models. (a) Russell Model. (b) Hevner Model. (c) Thayers Two-Dimensional Model.
RELATED STUDIES
expressing different emotions. Different hues were associated with different emotions; furthermore, dark colors
were associated with minor tonality and light colors with
major tonality. Odbert et al.9 surveyed the mood of music
using Hevners model and asked subjects what color they
associated with the music. Through analyzing the relationship between the mood of music and the colors suggested by that music, They demonstrated that subjects
who disagree on the mood of a piece of music tend to
also disagree on the colors associated with that music.
These results were very similar to those obtained by studies in which subjects were asked to name the color best
fitting certain mood words. Palmer et al.26 provided
experimental evidence that music and colors were mediated by emotional association in two cultures, US and
Mexico, by showing that there were strong correlations
between the emotional associations of the music and
those of the colors chosen to go with the music. They
showed that faster music in the major mode made participants choose more saturated lighter colors and yellower
colors whereas slower. They also showed that minor
music produced the opposite pattern, that is, desaturated,
darker, and bluer colors were chosen for minor music.
Some researchers6,1114 have investigated the relationship between color and mood. Manav6 defined this relationship using adjective mood words. In this study,
subjects were provided with 41 colors and were asked to
select which of 30 provided adjectives (e.g., vivid, boring, cold, warm, exciting, fearful, mysterious, peaceful,
and relaxing) best matched them. Manav further analyzed responses for 10 colors and looked at associations
between subjects education level, age, and gender.
Based on these findings, he recommended certain colors
be used for bedrooms, bathrooms, and childrens room.
Valdez and Mehrabian12 investigated the relationship
between mood and color using the PleasureArousal
Dominance (PAD) emotion model, providing PAD-value
prediction equations with the parameters of hue, saturation, and brightness. Valdez and Mehrabian12 demonstrated strong and reliable relationships between
emotional reactions and each of brightness and saturation, but only a weak relationship with hue. Ou et al.14
demonstrated that color preference can be determined
using three color-emotion scales (cleandirty, tense
relaxed, and heavylight), with cleandirty being the
predominant scale. They also showed that color preference can be determined by the three color-appearance
attributes of hue, lightness, and chroma; the most disliked color was found to be at the hue angle of 105
with the chroma value of 31.
Previously,27 we investigated the relationship between
mood and color. However, the findings of our prior study
were not useful for informing our lighting system, as the
relationship focused on users individual musical preferences. As our mood-lighting system is intended to be
installed in public spaces such as parks, plazas, and
squares, we needed a model that mapped mood to color
without considering individual users preferences.
203
Fig. 2.
ments were extracted from each music piece, and that the
moods of these segments may differ. Distributions of 101
pieces of music can be summarized and classified into six
types. The first type is a piece of music which has one
predominant mood (Fig. 3a); the second type is for a
piece in which all moods are similar (Fig. 3b); the third
type contains a wide distribution of moods covering two
quadrants (Fig. 3c); the fourth type contains different
moods, all of which are in one quadrant (Fig. 3d); the
fifth type contains different moods with a wide range of
distribution (Fig. 3e); and the sixth type has high frequencies of several moods (Fig. 3f).
In Fig. 3, different colors denote different segments of
the same piece of music; there were an average of three
segments per piece. This analysis supports our assumption
that individual music pieces contain several moods, making it necessary to change the lighting color when playing
music.
Mapping Music Segments to Moods
To map a music segment to a mood automatically, it is
necessary to learn the relationship between its musical
features and its mood. In this article, features of music
segments were extracted using MIRtoolbox.28 These features can be divided into five categories: dynamics,
rhythm, timbre, pitch, and tonality. Dynamics includes
root mean square (RMS) energy, and rhythm includes
fluctuation summary, tempo, and attack times. Timbre
includes zero cross rate, brightness, and roll-off, as well
as spectral centroid, spread, skewness, kurtosis, entropy,
and flatness. Pitch refers to pitch and chromagram, and
tonality includes the clarity, mode, and harmonic change
detection function. We used the 391-dimensional vector
produced by the mirfeatures function of MIRtoolbox,
which calculates statistics such as the mean, standard
deviation, and slope, as well as the frequency, amplitude,
and entropy of periods, instead of the complete features
themselves.
If all 391 features of the feature vector are used, mapping performance may decrease due to the effect of noise.
For this reason, we chose some noise features using the
well-known dimensional reduction method, R2 reduction.29 In our experiments, we used the 50 features with
COLOR research and application
Fig. 3. Mood distribution of music. (a) Case 1: Music 37. (b) Case 2: Music 13. (c) Case 3: Music 95. (d) Case 4: Music
74 (e) Case 5: Music 1. (f) Case 6: Music 28.
the largest r2 values [see Eq. (1)] (Table I). The performances of these selected features are given in the section
12
n
n n
P
P
P
B
C
n
xi yi 2
xi
yi
B
C
i51
i51
i51
2 B
C
r 5Bs
C
n n 2s
n n
2
n
@
A
P
P 2P
P 2
P 2
n
xi 2
xi
yi 2
yi
yi
n
i51
i51
i51
i51
(1)
i51
TONAL
Subcategory
Sub-sub category
RMS energy
Attack slope
Attack time
Tempo
Spectrum
Spectral irregularity
Spectral kurtosis
Zero crossingrate
Brightness
Entropy of spectrum
Rolloff (85%)
Rolloff (95%)
Spectral centroid
Spectral flatness
Spectral flux
Roughness
Spectral skewness
Mel-frequency cepstral
coefficients (MFCC)
Delta MFCC (DMFCC)
Delta Delta MFCC (DDMFCC)
Centroid of chromagram
Chromagram
Harmonic change detection Function
Keyclarity
Mode
Peak chromagram
SD
Period entropy
Mean, period entropy
SD
Mean, SD
Mean, SD
Mean, SD
Mean, SD
Mean
Mean
Mean
Mean
Mean
Mean
Mean
Mean, Period Amp
SD
Mean (1, 8, 9, 10), Period Freq
(9, 11, 12, 13)SD (6, 7, 8, 9, 10, 11, 12, 13)
Period Amp (2), Std (4)
Std (4)
Mean
Mean
Mean, Period Amp
Mean, Std
Period Amp
Peak Pos Mean
206
nodes or the number of iterations of learning. So, we conducted experiments varying the number of hidden nodes
from 2 to 25 and employing two different iterations
(3000 times and 4000 times) in the learning stage of the
neural network. As shown in the previous section, for
performance reasons, we did not use 391 features, but
used only the top 50 features after sorting the features
with the regression coefficient calculated by Eq. (1). The
experimental results are shown in Fig. 5. With 3000
learning iterations, we obtained the best performance
(66.46%) using 17 hidden nodes (Fig. 5b). With 4000
learning iterations, we obtained the best performance
(65.24%) using 7 hidden nodes (Fig. 5a). From the
results, we could conclude that the number of hidden
nodes and the number of iterations of learning did not
contribute to the performance improvement of our model
impressively. The performance we got is not suitable for
commercial use, but we believe it confirms the promise
Fig. 5.
Fig. 6.
Mood
Color
1
2
3
4
5
6
7
8
9
10
11
Excited
Happy
Pleased
Angry
Annoying
Nervous
Sad
Bored
Sleepy
Calm
Peaceful
255
255
255
204
255
153
102
204
255
153
204
51
153
255
0
0
0
102
204
255
255
255
51
153
51
0
0
0
0
255
204
153
153
0
0
42
0
0
0
42
170
42
85
63
204
102
204
255
255
255
255
51
51
102
102
255
255
255
204
255
253
102
255
255
255
255
12
Relaxed
Bright red
Bright red
Bright yellow
Red
Red
Dark red
Dark yellow
Bright blue
Bright yellow
Bright green
Bright
light green
Bright green
153
255
153
85
102
255
207
Proposed
Manav6
Angry, annoying
Excited, happy
Nervous
Bored
Pleased, sleepy
Barbiere et al.7
Bresin et al.8
Odbert et al.9
Happy
Happy
Anger, jealousy
Exciting
Hygiene, pure
Calm, peaceful,
modern, relaxing
Hygiene, pure
Simple, classic, plain
Happy
Happy
Love, fear
Tender
Happy
Happy
Happiness, pride
Playful
Hygiene, pure
Happy
Happy
Sad
Calm, relaxed
Peaceful
Leisurely
Sad
Sad
Hygiene, pure, plain,
simple
Curiosity
Contentment, shame
Gay
Warm, romantic,
enjoying, cheerful,
striking
Purple
Violet
Solemn
Sadness, love,
tenderness, disgust
The configuration of the emotional lighting system representing the mood of music is shown in Fig. 9. A PC
Fig. 7. Color distribution according to AV model area. (a) Color Distribution of Group 1. (b) Color Distribution of Group 2.
(c) Color Distribution of Group 3. (d) Color Distribution of Group 4.
208
Fig. 8. Brightness and saturation distribution for groups of mood words. (a) Group 1 brightness and saturation distribution. (b) Group 2 brightness and saturation distribution. (c) Group 3 brightness and saturation distribution. (d) Group 4
brightness and saturation distribution.
Fig. 9. Mood lighting system reflecting the mood of music. (a) Mood collection system configuration. (b) Mood lighting
system configuration. (c) Internal configuration of system.
209
Fig. 10.
For example, if the mood changes from Angry to Annoying (Fig. 12a), HSV color vectors (Fig. 12b) are generated
to display the colors located along the connecting dotted
line, with the time interval calculated by considering the
music playing time and the number of colors to be displayed during that interval.
Music Playing Module
Fig. 12. Example of creating a lighting scenario. (a) Music player. (b) Music playlist. (c) Status of lighting devices.
(d) Communication status.
210
211
LED lighting device. To do this, we analyzed the associations between mood words and colors using Moon
et al.s27 mood and color datasets. We observed the color
distribution associated with each mood and set a representative color for each mood word.
In this article, we mapped music to color indirectly,
first determining the mood of a music segment and then
selecting a color that matched that mood. To determine
the mood of a music segment, we took a similar approach
with prior approaches based on the Thayer model. However, the data used in the article differed from their dataset in that music segments were automatically generated
through a structural analysis, not manually. Furthermore,
our mood data set was much larger and was collected
from Korea. Finally, prior studies have treated musical
pieces as unchanging in mood, while our model separated
music into short segments that could have different
moods.
We obtained our own data on the correlations between
mood and color based on a large sample of Korean participants instead of using the correlations previously
determined by researchers such as Manav,6 Barbiere
et al.,7 Bresin,8 Odbert et al.,9 Spence,10 Wright and
Rainwater,11 Valdez and Mehrabian,12 DAndrade and
Egan,13 and Ou et al,14 whose results differed greatly
from ours. This discrepancy may reflect cultural differences in samples of different nationalities; however, our
findings should not be generalized to the Korean population as a whole, as our sample was mostly university
students.
The prototype proposed in this article lays the foundation for a commercial system; however, before such a
system can be developed, mood classification performance must be improved. The model also needs to be
expanded to cover 12 distinct moods, as opposed to 4
mood groups, as was done in the current study. We
assumed that the algorithm successfully extracted the
mood from the music, and a color was chosen to match
that mood. So, in the near future, we need to show the
evidence that participants agree that the color matches the
music. Furthermore, we need to get a more continuous
map from color to music through mood by allowing
music to represent different moods at once and colors to
represent the multiple moods as well.
1. Russell JA. A circumplex model of affect. J Personality Social Psychol
1980;39(6):11611178.
2. Hevner K. Experimental studies of the elements of expression in music.
Am J Psychol 1936;48(2):246268.
3. Thayer RE. The Biopsychology of Mood and Arousal. New York:
Oxford University Press; 1989.
4. Katayose H, Imai M, Inokuchi S. Sentiment extraction in music. International Conference on Pattern Recognition, 1988. p 10831087.
5. Lee JI, Yeo D-G, Kim BM, Lee H-Y. Automatic Music Mood Detection through Musical Structure Analysis. International Conference on
Computer Science and its Application CSA, jeju, Korea, 2009. p 510
515.
212