Escolar Documentos
Profissional Documentos
Cultura Documentos
A neuro-computational study of
laughter
Alonso MF., Loste P., Navarro J., del Moral R., Marijuán PC.
Abstract
The revival of laughter research during last two decades [Provine, 2000]
has been very fertile concerning specialized achievements in the
neuroimaging, neurophysiological, sound analysis, physiological
(respiratory & phonatory), ethological, evolutionary, social and health
aspects related to laughter. However, the conceptual counterpart of
putting together the most relevant strands of thought in order to gain
more advanced synthetic views or even to establish a new core
hypothesis has not been developed sufficiently. The preliminary idea of
this paper is to establish a coherent link between the evolutionary roots
of laughter and the origins of language, aligned with the “social brain”
hypothesis [Allman, 1999; Dunbar, 2004].
The social brain hypothesis posits that, in primate societies, selection has
favored larger brains and more complex cognitive capabilities as a mean
to cope with the challenges of social life [Silk, 2007; Dunbar, 2004].
Social networks in primates seem to be very different from those found
in most other mammals: they are cognitive, memory-loaded, based on
bonded relationships of a kind found only in pairbonds of other taxa
[Dunbar and Shultz, 2007)]. Frequent pair-wise grooming in between
individuals, however, imposes a strict time limitation regarding group
size: depending on diet, 20% of time is the upper ecological limit that
grooming can reach. This factor necessarily restricts the size of grooming
networks and, thus, of natural groups in primate societies (composed, at
most, of a few dozen individuals).
So, how could human societies have organized their “grooming” within
increasingly larger natural groups, of around 100 or 150 individuals? As
Dunbar (1998, 2004) has argued, human language was the evolutionary
solution. “Languaging” was co-opted as a virtual system for social
grooming-massaging, plus other specifically human adaptations for
group cohesion: laughter, crying, gaze-facial expressions, music, dance...
2
It is by following this line of thought, that the enigmatic presence of
laughter along the human life cycle may be further clarified.
Far from being a stereotyped signal, laughter becomes one of the most
variable acoustic expressions of humans, comparable to language except
for the combinatory richness of the latter. Typical laughter is composed
of separate elements or “calls” or “syllables”, plosives, over which a
vibrato of some fundamental frequency Fo is superimposed [Rothganger
et al. 1997]. A typical laughter episode may last around one second (or
slightly less) and will contain around 5 plosives (more in general, in
between 2 and 8). An important distinction to make is between
“vocalized” and “unvocalized” laughter; even though the former induces
significantly more emotional responses in listeners, the latter appears
3
consistently in many laughter records, comprising a large variety of
sounds (snort-like, grunt-like, giggles, chuckles, etc.).
All of the previous elements could form part of the inbuilt cues to
laugher identity, which have been proposed to play an important role in
listener emotional responses [Baworowski and Owren, 2001]. In
particular, the pitch or tone curve described by Fo, together with the
distribution of plosives, would show consistent differences between
laugh forms associated with emotional states of positive and negative
valence [Devillers and Vidrascu, 2007]. The main trend is that the energy
and duration becomes higher for “positive” than for “negative” laugh,
and vice versa for the relative presence of unvoiced frames, more
frequent in ironic and hostile laughs than in joyful ones. Notwithstanding
that, there is not much consensus established yet –neither significant
hypothesis to put to test– on how the interrelationship between plosives,
tones, melodies and other variables of laughter may be systematically
involved in encoding and distinguishing the underlying emotional states
[Bea and Marijuán, 2003; Bachorowski and Owren, 2008; Marijuán and
Navarro, 2010].
The sentic forms hypothesis, framed by M. Clynes in the 70’s, could help
in the exploration of new directions for such open questions. If laughter
6
contains inner “melodies” or pitch patterns of emotional character, how
could they be structured? Following the sentic paradigm developed
around tactile emotional communication by means of exchange of
pressure gradients, there appears a set of universal dynamic forms that
faithfully express the emotional interactions of the subjects (see Figure 1)
[Clynes, 1979; Winter, 1999].
3. Experimental Methodology
8
The plosive automatic detector was implemented in Matlab. For this
purpose, we created a program that reads every sound archive and
computes the mean energy values of concatenated signal frames, with a
frame length of 21 ms and a shift of 5 ms. Due to the wide frequency
distribution of unvoiced laughs (snort, grunt, etc), a spectral-domain
method isn't adequate. Then the program finds the lowest energy points
using a valley detection algorithm. To confirm if these points are
accurately plosive limits, it should be at least 10% of maximum energy
between two valleys, and duration of 100 ms to discard spurious
vocalizations. The number of plosives in an episode varied significantly,
for example, one bout can contain in between one and eight plosives.
However, to perform straightforward comparisons between laughter
types, it is better to build vectors of fixed length. The features extracted
in our study for every plosive were:
• Total time duration in seconds. Breathing is the primal driving
force of laughter, and marks a very definite pattern different from
other expressions. In that regard, the evolution of burst durations
gives an overview of the temporal structure of a laughter episode.
• Fundamental frequency mean. It can be defined as the inverse of
the smallest period of the vocal fold signal, in the interval being
analyzed (every 10 ms). However, since pitch has a short-time
variation, we compute its mean. The algorithm is based on the
normalized cross-correlation function. Emotional expression is
dominated by effects created in the glottis and laryngeal muscles.
We consider the fundamental frequency important because vocal
articulators are less accurate and slower in nonverbal expressions,
and therefore vocal chords have a key role in laughter
characterization.
• Standard deviation of the aforementioned fundamental frequency. It
was computed over successive short time frames of the signal, in
order to measure range and variability of an instantaneous measure.
• First three formants. A formant is defined as spectral peaks of the
spectrum of the voice, which give information about the acoustic
resonances of the vocal tract position at every plosive. We use a
method based on averaging different linear predictive code
spectrograms, and detecting the first three maximum values. Vocal
articulators act as filters of a resonant tube, changing the
fundamental frequency. When communicating different emotions
or intentions these articulators are a key mechanism for changing
emotional information.
9
• Average power or energy per sample. The amplitude of sound
affects our perception of fundamental frequency, which is called
pitch. The pitch is an interpretation of the listener, and one of the
parameters that influence the subjective interpretation is the
intensity or sound energy.
• Shannon’s entropy. This feature measures the information
contained in a message as opposed to the portion of the message
that is determined.
• Jitter, a cycle-to-cycle measure of the variations of fundamental
frequency, as a percentage. It is known that emotions are expressed
in many cases as unintentional acts, such as quiver or tremble. Jitter
is defined as vocal chords disturbances that occur during the glottal
phonation of the vowel and affect the pitch.
• Shimmer, defined as the average absolute difference between the
amplitudes of consecutive periods, expressed as a percentage. It is
equivalent to jitter in signal amplitude, that is, rapid variations in
loudness.
• Harmonic to Noise Ratio (HNR), described as the energy ratio of
fundamental frequency harmonics to noise components. HNR is
found to be a more sensitive index of vocal function than other
features.
• Percentage of voiced signal over unvoiced signal, that is, the time
vocal cords vibration spends over a plosive. This measure shows
how much value has the harmonic expression from the vocal cords,
as opposed to sound generated in other parts of the vocal tract.
This is an ongoing work and the results will be published in a next paper.
A library of natural laughs has been compiled, so that an artificial neural
network should be able to distinguish and categorize the concerned
10
sound structures, gauging the soundness and conceptual possibilities of
the sentic forms hypothesis. We were able to find patterns in acoustical
features between the different laugh-types according to the expressed
emotion. As it may be expected, the differences would be in the features
corresponding to the vibration cycle in the pharynx, and they translate in
different acoustical patterns that approximately coincide with the sentic
forms proposed by Manfred Clynes, but in reverse order.
7. References
- Allman J.M. 1999. Evolving Brains. Scientific American Library, New York.
- Bachorowski JA. & Owren M.J. 2008 . Sounds of Emotion 1 Vocal Expressions of
Emotion. Guilford, New York.
- Bachorowski JA, Smoski M and Owren MJ. 2001. The acoustic features of human
laughter. J. Acoust. Soc. Am. 110, 1581–1597.
- Bachorowski JA, Owren MJ. 2001. Not all laughs are alike: voiced but not unvoiced
laughter readily elicits positive affect. Psychol Sci. 12(3):252-7.
- Banse R, Scherer KR. 1996. Acoustic profiles in vocal emotion expression. J Pers Soc
Psychol. 70(3):614-36.
- Bea JA. & Marijuán PC. 2003. Informal patterns of laughter. Entropy 5: 205-213.
www.mdpi.org/entropy
- Bendor D. & Wang X. 2005. The neuronal representation of pitch in primate auditory
cortex. Nature. 436(7054):1161-5.
- Buzsáki, G. 2006. Rhythms of the Brain. Oxford University Press, New York.
- Collins KP. & Marijuán PC. 1997. El Cerebro Dual: Un Acercamiento Interdisciplinar a
la Naturaleza del Conocimiento Humano y Biológico. Editorial Hacer, Barcelona.
- Clynes M. 1979. The Source of Laughter and Essentic Form: Is Evolution Dis-Covery?
Somatics 2 (3): 3-9.
- Clynes M. 1988. Generalized Emotion: How it May be Produced, and Sentic Cycle
Therapy. In: Manfred Clynes & Jaak Panksepp. Editor. Emotions and Psychopathology.
- D'Ausilio A. 2007. The role of the mirror system in mapping complex sounds into
actions. J Neurosci. 27(22): 5847-8.
- Dehaene, S. 2009. Reading in the Brain. New York, NY, USA: Penguin.
- Devillers L. & Vidrascu L. 2007. Positive and Negative emotional states behind the
laughs in spontaneous spoken dialogs. Interdisciplinary Workshop on The Phonetics of
Laughter. Saarbrücken, Germany, 4-5 August 2007.
- Di Pellegrino G, Fadiga L, Fogassi L, Gallese V, Rizzolatti G. 1992. Understanding
motor events: a neurophysiological study. Exp Brain Res 91(1):176-80.
- Dunbar R. & Shultz S. 2007. Evolution in the Social Brain. Science 317: 1344-47.
- Dunbar R. 1998. The Social Brain Hypothesis. Evolutionary Anthropology, 178-90.
- Dunbar R. 2004. The Human Story: A New History of Mankind’s Evolution. Faber &
Faber Ltd., London.
- Dupont S, Dubuisson T, Mills JA, Moinet A, Siebert X, Tardieu D, Urbain J. 2009.
"LaughterCycle", QPSR of the numediart research program, volume 2, nbr. 2.
12
- Edelman G.M. & Tononi G. 2000. A Universe of Consciousness: Basic Books, NY.
- Fishman YI, Reser DH, Arezzo JC, Steinschneider M. 1998. Pitch vs. spectral
encoding of harmonic complex tones in primary auditory cortex of the awake monkey.
Brain Res. 786(1-2):18-30.
- Goel V. & Dolan R.J. 2001. The functional anatomy of humor: segregating cognitive
and affective components. Nat Neurosci. Mar; 4 (3): 237-8.
- Hill PR, Hartley DE, Glasberg BR, Moore BC, Moore DR. 2004. Auditory processing
efficiency and temporal resolution in children and adults. J Speech Lang Hear Res.
47(5):1022-9
- Marijuán PC & Navarro J. 2010. “The Bonds of Laughter: A Multidisciplinary Inquiry
into the Information Processes of Human Laughter”. Cite as: arXiv:1010.5602v1 [q-
bio.NC]
- Patterson RD, Uppenkamp S, Johnsrude IS, Griffiths TD. 2002. The processing of
temporal pitch and melody information in auditory cortex. Neuron. 36(4):767-76.
- Penagos H, Melcher JR, Oxenham AJ. 2004. A neural representation of pitch salience
in nonprimary human auditory cortex revealed with functional magnetic resonance
imaging. J Neurosci. 24(30):6810-5.
- Provine RR. 2000. Laughter: A scientific investigation. Faber & Faber, London.
- Provine RR. & Emmorey K. 2006. Laughter Among Deaf Signers. J. Deaf. Stud. Deaf.
Educ. 11 (4): 403-9.
- Ramachandra V. 2009. On whether mirror neurons play a significant role in processing
affective prosody. Percept Mot Skills. 108(1):30-6.
- Rothgänger H. 1997. Comparison on standarized fundamental frequency of crying and
lalling, 6th. International Workshop on Infant Cry Research. Bowness.
- Schwarz DW, Tomlinson RW. 1990. Spectral response patterns of auditory cortex
neurons to harmonic complex tones in alert monkey (Macaca mulatta). J Neurophysiol.
64(1):282-98.
- Silk JB. 2007. Social Components of Fitness in Primate Groups. Science 317: 1347-51.
- Smoski M. 2004. The Development of Antiphonal Laugther Between Friends and
Strangers. PhD dissertation Vanderbilt University. Psychology. August, 2004. Nashville.
- Smoski, M. & Bachorowski, JA. 2003. Antiphonal laughter between friends and
strangers. Cognition & Emotion, 17, 327-40.
- Sporns, O. 2011. The non-random brain: Efficiency, economy, and complex dynamics.
Front. Comput. Neurosci., 5, 1–13.
- Szameitat DP, Kreifelts B, Alter K, Szameitat AJ, Sterr A, Grodd W, Wildgruber D.
2010. It is not always tickling: distinct cerebral responses during perception of different
laughter types. Neuroimage 53(4):1264-71.
- Szameitat DP, Darwin CJ, Wildgruber D, Alter K, Szameitat AJ. 2011. Acoustic
correlates of emotional dimensions in laughter: arousal, dominance, and valence. Cogn
Emot. 25(4):599-611.
- Trouvain J. "Segmenting Phonetic Units in Laughter", proceedings of the 15th
International congress of phonetic sciences, Barcelona 2003.
- Winter D. 1999. How to touch: How the Geometry of Pressure in Your Touch-Predicts
the Harmonics of Emotion in Music-& Love. http://soulinvitation.com/touch/touch.html
13