0% found this document useful (0 votes)
14 views10 pages

Fassnidge 2017

Uploaded by

lucas.c.jofilsan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
14 views10 pages

Fassnidge 2017

Uploaded by

lucas.c.jofilsan
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

Consciousness and Cognition 49 (2017) 15–24

Contents lists available at ScienceDirect

Consciousness and Cognition


journal homepage: www.elsevier.com/locate/concog

A deafening flash! Visual interference of auditory signal


detection
Christopher Fassnidge a, Claudia Cecconi Marcotti b, Elliot Freeman a,⇑
a
Cognitive Neuroscience Research Unit, Department of Psychology, City, University of London, London, UK
b
Neuroscience Masters Programme, King’s College London, UK

a r t i c l e i n f o a b s t r a c t

Article history: In some people, visual stimulation evokes auditory sensations. How prevalent and how
Received 6 October 2016 perceptually real is this? 22% of our neurotypical adult participants responded ‘Yes’ when
Accepted 9 December 2016 asked whether they heard faint sounds accompanying flash stimuli, and showed signifi-
cantly better ability to discriminate visual ‘Morse-code’ sequences. This benefit might arise
from an ability to recode visual signals as sounds, thus taking advantage of superior
Keywords: temporal acuity of audition. In support of this, those who showed better visual relative
Synaesthesia
to auditory sequence discrimination also had poorer auditory detection in the presence
Perception
Audiovisual integration
of uninformative visual flashes, though this was independent of awareness of visually-
Individual differences evoked sounds. Thus a visually-evoked auditory representation may occur subliminally
Psychophysics and disrupt detection of real auditory signals. The frequent natural correlation between
visual and auditory stimuli might explain the surprising prevalence of this phenomenon.
Overall, our results suggest that learned correspondences between strongly correlated
modalities may provide a precursor for some synaesthetic abilities.
Ó 2016 Elsevier Inc. All rights reserved.

1. Introduction

In synaesthesia an inducing stimulus consistently and involuntarily evokes a consciously experienced concurrent sensa-
tion in a different sensory dimension or modality. Though there are many variants of synaesthesia, one feature they have in
common is that they are rare: the most frequent types (e.g. grapheme and colour, or sound and shape) are found in only
about 2–4% of the population (Simner et al., 2006; Ward, 2013). The low prevalence of cases is consistent with suggestions
that synaesthesia represents an aberrant genotype (Brang, Williams, & Ramachandran, 2012; Tomson et al., 2011), which
may result in unusual patterns of neural cross-wiring or cross-activation between adjacent cortical regions
(Ramachandran & Hubbard, 2001; Rouw & Scholte, 2007). However, many researchers have argued that some forms of
synaesthesia might be grounded on normal mechanisms involved in forming and reinforcing associations between different
modalities and sensory dimensions (Brang et al., 2012; Cohen, 2013; Cohen Kadosh, Henik, Catena, Walsh, & Fuentes, 2009;
Cytowic, 2003; Grossenbacher & Lovelace, 2001; Ramachandran & Hubbard, 2001; Ward, Huckstep, & Tsakanikos, 2006). On
this latter view, the rarity of synaesthesia might be explained by the observation that the kinds of exotic associations that
typify synaesthesia are very rarely found in nature. For example, grapheme-colour synaesthesia might be rare because con-
sistent correspondences between letters and colours are themselves rare and thus do not typically reinforce strong associ-
ations, even though repeated exposure to consistent letter-colour pairings (found in fridge-magnets, educational materials,

⇑ Corresponding author at: City, University of London, Northampton Square, London EC1V 0HB, UK.
E-mail address: [email protected] (E. Freeman).

http://dx.doi.org/10.1016/j.concog.2016.12.009
1053-8100/Ó 2016 Elsevier Inc. All rights reserved.
16 C. Fassnidge et al. / Consciousness and Cognition 49 (2017) 15–24

or experimental stimuli) might shape and reinforce grapheme-colour associations in individuals who are susceptible to them
(Bor, Rothen, Schwartzman, Clayton, & Seth, 2014; Witthoft, Winawer, & Eagleman, 2015).
Such reasoning leads to the hypothesis that we might find synaesthetic associations more frequently when they occur
between stimulus dimensions that are naturally correlated, so that their associations are regularly and consistently rein-
forced. For example, visual events naturally correlate very frequently with sounds, whenever two objects collide or a person
speaks and their lips move. An opportunity to test this association frequency hypothesis is presented by a past report that
transient visual stimulation such as flashes or moving dots can induce conscious concurrent auditory sensations in some
individuals (Saenz & Koch, 2008). Saenz and Koch (2008) devised an elegant objective test of this phenomenon, where par-
ticipants had to compare two paired ‘Morse code’ type sequences, presented both either as sounds or flashes. A small sample
of participants self-identified as ‘hearing-motion’ synaesthetes showed relatively high performance for both auditory and
visual stimuli, while control participants found the task significantly harder in the visual modality. This superior visual per-
formance in synaesthetes was explained on the assumption that they were benefiting from additional temporal information
provided by recoding the visual stimulation into the auditory modality (Glenberg, Mann, Altman, Forman, & Procise, 1989;
Guttman, Gilroy, & Blake, 2006). Apart from other rare reports of individuals who hear distinct musical sounds associated
with visual colours (Baron-Cohen, Burt, Smith-Laittan, Harrison, & Bolton, 1996; Goller, Otten, & Ward, 2009) there has been
no other published research on this visual-to-auditory direction of association to our knowledge. Two unresolved questions
are raised, which we consider here: how prevalent is this phenomenon, and is it perceptually real?
Prevalence cannot be assessed from Saenz and Koch’s (2008) original study, as the few participants who were identified
as synaesthetes were not randomly sampled but self-selected. Our first goal was therefore to make a tentative estimate of
the prevalence of visually-evoked auditory sensations in a random sample, using a combination of subjective questioning
and objective tests based on Saenz and Koch’s (2008) paradigm. Our association frequency hypothesis predicts relatively
high frequency of reports of visually-evoked sensations, in neurotypical individuals, along with task performance resembling
that of Saenz & Koch’s synaesthetes.
Though intended as an objective diagnostic of ‘hearing-motion’ synaesthesia, Saenz and Koch’s (2008) sequence discrim-
ination paradigm only provides an indirect test of the effects of visually-evoked auditory sensations on the visual modality,
rather than on hearing. Our second goal was therefore to probe the effects of visual stimulation on actual auditory signal
detection, and measure the correlation of such effects with performance on Saenz & Koch’s sequence discrimination para-
digm. We measured sensitivity for detecting a white-noise burst, in the presence of an irrelevant and temporally non-
predictive visual event (a high-contrast drifting radial grating), compared to no visual stimulus (following Lovelace, Stein,
& Wallace, 2003). We predicted that participants who benefitted from auditory-recoding of visual stimuli in the sequence
discrimination task might be affected more by irrelevant flashes in the auditory signal detection task, because they would
be experiencing concurrent visually-evoked auditory sensations. We also measured and controlled for individual biases in
the dominance of the auditory modality relative to visual (Colavita, 1974; Koppen, Levitan, & Spence, 2009), in case this
influenced our other objective measures.

2. Methods

2.1. Participants

A total of 40 naïve participants with normal hearing (by self-report) took part in the research and were paid for their par-
ticipation. Two participants who did not have time to complete the debriefing questions were excluded from analysis, and a
further participant was excluded because of chance performance in both auditory and visual sequence discrimination tasks.
The final sample had 24 females, aged 19–36 (mean 24.24, standard deviation 4.68). One participant had an absolute audi-
tory threshold of more than 3 standard deviations higher than the sample mean, but was included as performance on the
main tasks was in the normal range and exclusion made no difference to the pattern of results. A subsample of 24 partici-
pants (18 female, mean age 23.8, SD 4.27) also performed an additional ‘Colavita’ test (see below). All participants had nor-
mal or corrected vision and reported normal hearing. All procedures were carried out with informed consent and were
approved by the local Psychology ethics committee.

2.2. Apparatus and stimuli

The experimental procedure was conducted using an Apple Mac Mini connected to a 1700 Sony HMD-A420 cathode ray
tube (CRT) display. Auditory stimuli were presented through two Labtec PC speakers both positioned next to each other
directly in front of and below the centre of the monitor. Video mode was 800  600 pixels with a 120 Hz refresh rate and
viewing distance was approximately 57 cm (controlled using a chin rest). A small white fixation point marked the centre
of the display. Subject responses were collected using the arrow keys on a standard computer keyboard. Experimental pro-
cedures and stimuli were programmed using Psychtoolbox for Matlab.
Stimuli for each main task closely followed the methods described in the originating studies (Lovelace et al., 2003; Saenz
& Koch, 2008). For the Sequence Discrimination task (Fig. 1a), visual stimuli consisted of white circular discs of luminance
81 cdm2, presented centrally on a black background. Disk diameter was 3 degrees of visual angle. Auditory stimuli were
C. Fassnidge et al. / Consciousness and Cognition 49 (2017) 15–24 17

a Sequence discrimination

75ms 100ms 300ms

Interval 1
Same
Interval 2

Interval 1
Diff
Interval 2

b Auditory detection

+ Y< >N + Y< >N + Y< >N

Y N Y
500ms 500ms

150ms 350ms

c Sensory dominance

A< >V
+ + + +

A A V V
1300-1700ms

50ms
Fig. 1. Illustrative stimuli and display sequences for (a) sequence discrimination task, showing part of a typical trial presented either in the visual or
auditory modality; sample sequences are shown below for typical ‘same’ versus ‘different’ trials; (b) auditory signal detection task: two audiovisual trials
are illustrated with the auditory target present then absent, then an auditory-only trial; (c) sensory dominance (Colavita) task: unimodal auditory, bimodal,
and unimodal visual trials are illustrated.
18 C. Fassnidge et al. / Consciousness and Cognition 49 (2017) 15–24

sine wave tones of maximum 91 dBA sound pressure level, frequency 360 Hz. ‘Short’ and ‘Long’ events were presented for
periods of either 75 ms or 300 ms respectively, during which stimulation amplitude immediately decayed linearly from
maximum to zero amplitude.
For the Auditory Detection task (Fig. 1b), auditory stimuli were white noise bursts of 300 ms duration with amplitude
modulated by a Tukey window with 150 ms rise and fall time. This was presented on a continuous white noise background
of 45 dBA sound pressure level. The visual stimulus was a grey-level radial grating, of maximum luminance 72 cdm2, diam-
eter 4 deg, and frequency of four cycles per revolution, with the grating phase incremented at two cycles per second giving
the appearance of a rotating windmill. This animated stimulus was chosen with the aim of amplifying any auditory sensa-
tions induced by visual motion. Interval marker digits and response prompts were displayed in white 18 pt Helvetica. In the
visual dominance task (Fig. 1c), visual stimuli were circular white discs of luminance 48 cdm2 and diameter 4.5 deg, pre-
sented centrally on a black background. Auditory stimuli were white noise bursts of 60 dBA sound pressure level. All stimuli
were presented for 50 ms.

2.3. Design

In the sequence discrimination task the independent variable was the modality of the stimuli (either visual or auditory)
and the dependent variable was same/different discrimination accuracy. In the auditory detection task the independent vari-
able was also the modality of the stimuli (either auditory or audiovisual) and the dependent variable was auditory target
detection accuracy. Results from each paradigm were coded as the proportion of correct detects or correct ‘different’
responses, versus false alarms or incorrect ‘different’ responses, for the purpose of analysis based on signal detection theory
(Green & Swets, 1966), which allows perceptual detection or discriminability of signals to be measured independently of any
cognitive or attentional factors which might bias the decision criteria. The order of these two tasks was counterbalanced
between subjects. We also included a measure of visual dominance over audition (Colavita, 1974; Koppen et al., 2009), with
three randomly interleaved conditions: auditory only, visual only and audiovisual; the dependent measure was the accuracy
for identifying the modality of the stimulus.

2.4. Procedure

In the Sequence Discrimination task, the procedure closely followed Saenz and Koch (2008). On each trial two successive
rhythmic sequences of stimuli were presented. In half of the trials the events were all visual, and in the other half all audi-
tory. The modality of each trial was randomized between trials. Within each sequence, constituent stimuli (events) could be
either short (75 ms) or long (300 ms) with a total of eight events per sequence. Sequences were randomly generated and
consisted of a minimum of four and a maximum of five transitions (i.e. a short event followed by a long event, or a long event
followed by a short). There was an inter-event interval of 100 ms, and an interval of 500 ms between the first and second
sequence. On half of the trials, the two sequences were identical, and on the other half they differed. In ‘different’ trials,
the first two events and the last event were always identical between pairs, while the order of the remaining events was
randomly permuted. Immediately following the second sequence, participants were required to indicate whether they
thought the two sequences were same or different by pressing either the left or right arrow key on a PC keyboard,
respectively. No error feedback was given. The response initiated the next trial. After the final trial in each block of 20 trials,
participants could take a break and the next block would begin when they pressed the spacebar. There was a total of 100
trials per session.
Before beginning the Auditory Detection task, we used a two-alternative forced choice staircase procedure to find audi-
tory detection thresholds in the absence of any visual stimuli. Participants were required to detect a white noise burst in the
presence of on-going background white noise. Each block began with a central fixation dot, which was followed after a key-
press by a central digit ‘‘1” for 500 ms, followed by a blank stimulation interval of 500 ms. This was immediately followed by
a ‘‘2” for another 500 ms, then another blank stimulation interval. This was terminated by a visual prompt to make a
response. The target stimulus was a 300 ms burst of white noise, which could be presented, after a delay of 150 ms, in either
one of the two stimulation intervals. This procedure ensured that there was no visual stimulation present during the audi-
tory stimulation, while providing clear visual temporal markers bracketing each stimulation interval (similar to Lovelace
et al., 2003). The participant had to indicate whether the target was in either the first or second interval, using left or right
arrow keys respectively. Participants heard a single click if their response was correct, and two clicks if incorrect. On each
trial, target intensity was chosen depending on responses to the previous trials, using the Quest algorithm (Watson &
Pelli, 1983). Each block contained 40 trials, and there were 6 blocks in total. After each block the algorithm calculated the
auditory thresholds, which were then averaged to achieve a final threshold estimate. This was used to set the amplitude
of the target for the main auditory detection task, which then remained fixed.
The main auditory detection task consisted of single-interval trials in which a target was either present or absent. Partic-
ipants had to indicate whether the target was present or absent, using left or right arrow keys respectively, and received
error feedback. The target timing and stimulus characteristics were the same as described before. Prior to each trial a central
fixation dot was presented for 500 ms, which then disappeared at the onset of the 500 ms stimulation interval. The end of the
stimulus interval was marked by the appearance of a response prompt, instructing participants to press the left or right
arrow to indicate whether they thought the target was present or absent, respectively. On half of the trials the interval
C. Fassnidge et al. / Consciousness and Cognition 49 (2017) 15–24 19

contained no visual stimuli (auditory-only trial), while the other half contained a white rotating radial grating which was
presented throughout the 500 ms interval (audiovisual trials). Because it filled the entire stimulation interval, the visual
stimulus provided no more temporally predictive information than the fixation offsets and onsets that already flanked
the stimulation interval. Participants were told that the visual stimulus could be ignored because it provided no information.
Auditory and audiovisual trials were blocked and the order of blocks was randomly permuted for each participant. Each
block contained 40 trials and there were 6 blocks in total (three for each condition).
After participants had completed both tasks they were asked a short series of questions:

 What strategy were you using when you saw flashes in the visual sequencing experiment?
 Were you aware of using the flashes as if they were sounds, e.g. ‘‘flash, flash-flash” = ‘‘beep, beep-beep”
 Did you actually hear faint sounds when you saw flashes?
 In everyday life, are you ever aware of hearing sounds when you see flashing lights or movement, e.g. shop displays, car
indicators, or people walking?
 Do you ever experience colours associated with letters, or with music, or tastes or smells associated with sounds?
 Have you ever been diagnosed as a synaesthete or do you suspect you might be one?

In a later session we administered the test of sensory dominance (Colavita, 1974) to 24 of our original participants who
were still available. We presented participants with 10 blocks of 100 randomised trials, where 20% of trials were bimodal and
the remaining 80% were divided equally between unimodal visual and auditory trials. Each block consisted of a stream of
stimuli, each presented for 50 ms interspersed by random intervals between 1300 and 1700 ms. Participants were required
to respond as quickly as possible with the left arrow key for auditory trials, the right arrow key for visual trials and both keys
simultaneously for bimodal trials.

3. Results

3.1. Subjective results

Subjective data were collected by asking a series of questions at the end of the experimental tasks. When initially asked to
talk about strategies used in the sequence discrimination task, without prompting 22 out of 37 participants (59%) reported
deliberately converting visual sequences to internal sounds (i.e. they ‘mentally replayed’ them as auditory sequences), while
17 (46%) said they tried counting and remembering the temporal positions of long vs. short flashes. Some reported attempt-
ing both strategies. When asked ‘In the sequence discrimination task were you aware of using the flashes as if they were sounds’,
27 answered ‘yes’ (73%, Table 1). When then asked ‘In the sequence discrimination task did you actually hear faint sounds when
you saw flashes?’, eight participants answered ‘yes’ (22%). Responses were conservatively interpreted, so if participants were
unsure about whether or not they heard any auditory sensation on seeing the flashes, this was coded as a negative response.
We then asked whether they were ‘ever aware of hearing sounds when you see flashing lights or movement, e.g. shop displays, car
indicators, or people walking?’, to which four answered in the affirmative, though only one of these had answered ‘yes’ to the
previous question. None said they had been previously diagnosed as synaesthetes, though three did report that they
sometimes experienced pairings between different senses.

3.2. Objective group statistics

Performance was on average significantly higher for auditory versus visual sequence discrimination [mean d0 (and stan-
dard error): Visual 1.74 (0.13); Auditory 3.14 (0.20); t(36) = 7.87, p < 0.00001, Cohen’s D = 1.36], and likewise for both of the
subgroups defined by responses to the question ‘Did you actually hear faint sounds when you saw flashes?’ [No-Hear: t(28)
= 6.83, p < 0.00001, Cohen’s D = 1.3; Yes-Hear: t(7) = 4.01, p < 0.005, Cohen’s D = 1.5]. Participants were also significantly less
cautious about making ‘different’ responses to visual sequences (Mean criterion 1.67, SE 0.078) compared to auditory (Mean
2.25, SE 0.12) [t(36) = 5.62, p < 0.00001, Cohen’s D = 0.96].
For between group comparisons we analysed sequence discrimination in a mixed ANOVA, with auditory/visual task and
Yes/No-Hear groups as the two factors, after first confirming that there were no significant violations of homogeneity of vari-
ances [Levene’s F(1, 35) = 0.76, p = 0.39]. As found in Saenz and Koch (2008), performance was significantly better in the
auditory modality [Visual Mean d0 (SE): 1.74 (3.14); Auditory: 3.14 (1.22); F(1, 35) = 41.8, p < 0.00001], however performance
overall was significantly better for participants who answered ‘Yes’ to the question [Yes-Hear Mean d0 (SE): 3.02 (0.29);

Table 1
Results for debrief questionnaire.

Use flashes Hear flashes Ever hear flashes? Experience pairings?


Sum 27 8 4 3
% (N = 37) 73.0% 21.6% 10.8% 8.1
20 C. Fassnidge et al. / Consciousness and Cognition 49 (2017) 15–24

Heard faint sounds? Sequence discrimination


a b
5
5

4 No
4
Seq. discr. d'

Yes

3
3

d'
2
2

1 1

0 0
V A V A

Auditory detection
3
c d 1.5
2.5
1
Aud. det. d' AV-A

2
Visual benefit

0.5
d'

1.5

0
1

-0.5
0.5

r(37) = -0.42, p< 0.0075


-1
0
AV A -5 -4 -3 -2 -1 0 1
Seq. discr. d' V-A
Visual better re: auditory
Fig. 2. Experimental results. (a) Mean d0 sensitivity for discriminating sequences of visual (V) and auditory (A) sequences, grouped by responses to the
question ‘Did you hear faint sounds when you saw flashes?’ (Yes: Yellow; No: Blue). Brackets indicate statistically significant differences [p < 0.05, see
Results]; (b) sequence discriminability d0 for individual participants; same colour scheme as above. (c) d0 for auditory signal detection, in the context of an
irrelevant visual stimulus (AV) or alone (A); (d) scatterplot of the benefit of an irrelevant visual flash on d0 for auditory signal detection (AV-A on y-axis),
against visual relative to auditory sequence discriminability (V-A, on x-axis). Orange symbols represent participants who answered positive to the ‘heard
faint sounds’ question. (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article.)

No-Hear: 2.27 (0.15); F(1, 35) = 5.17, p = 0.029] (Fig. 2a). The interaction was not significant [F(1, 35) = 0.90, ns]. Bonferroni-
corrected comparisons showed that only the visual task benefited significantly in participants confirming awareness of audi-
tory sensations compared to those who did not [Visual: t(35) = 2.43, p = 0.02; Auditory t(35) = 1.64, p = 0.11] (see bar chart of
means in Fig. 2a). Because of possible concerns about unequal sample sizes, we performed an alternative bootstrapping anal-
ysis in which we compared performance in the eight ‘Yes’ participants with 1000 samples of eight participants randomly
selected with replacement from the ‘No’ group. This showed a similar pattern, where only performance in the visual ‘Yes’
group condition was outside the 95% confidence interval of the bootstrapped ‘No’ distribution [visual z = 2.79, p = 0.005;
auditory z = 1.92, p = 0.055]. Comparison of visual-minus-auditory performance across groups (equivalent to the interaction
term in the above ANOVA) was however still not significant using this method. An ANOVA analysis for the question of
whether participants were ‘aware of using the flashes as if they were sounds’ showed no significant main effect of group
[F(1, 35) = 0.48, ns]. There was a significant interaction with task modality [F(1, 35) = 5.96, p = 0.02], where those who
C. Fassnidge et al. / Consciousness and Cognition 49 (2017) 15–24 21

answered ‘Yes’ had higher auditory d0 s than the others [Yes-Use Mean d0 (SE): 3.32 (0.23); No-Use: 2.63 (0.38)], however post
hoc comparisons showed no significant differences.
For the Auditory Detection task, there were also wide individual differences in auditory detection d0 , and in the effects of
the additional visual stimulation (Fig. 2c). There were no significant differences in d0 or criterion between Auditory or
Audiovisual conditions overall nor for each participant group, nor in the difference in performance between conditions.

3.3. Correlation between tasks

According to Saenz and Koch (2008), their ‘hearing-motion’ synaesthetes differed less across modalities in their sequence
discrimination ability compared to control subjects, as if they were using similar resources to perform the task in different
modalities. We hypothesised that participants showing this pattern of behaviour would have lower auditory detection sen-
sitivity in the presence of visual stimulation, compared to without. To assess this, we first subtracted the auditory d0 from the
visual d0 for sequence discrimination (SEQd0 V-A). Higher values (less negative) indicate more similar performance for both
modalities. Our second measure of the visual influence on auditory detection was computed by subtracting d0 in the
auditory-only condition from d0 in the audiovisual condition (DETd0 AV-A). Negative values would be consistent with an effect
of visual stimulation interfering with auditory detection. There was a significant negative correlation between scores on
these two measures [Pearson’s r(35) = 0.43, p < 0.0037]: the closer visual sequence discriminability was to auditory, the
more visual stimulation impaired auditory detection, or the less it benefited (Fig. 2d). This result is consistent with the
notion that internal auditory noise evoked by the visual stimulus can actually interfere with an externally-originating signal.
We also found a significant positive correlation between DETd0 AV-A and auditory sequence discrimination [r(35) = 0.40,
p = 0.01], but not visual sequence discrimination [r(35) = 0.03, p = 0.85; all three above analyses used Bonferroni-adjusted
alpha]. There was no significant correlation of sequence discrimination with the absolute threshold for auditory signal detec-
tion measured prior to the main task, that could account for individual differences in performance. Visual stimulation may
thus interfere with auditory detection more in individuals whose auditory modality is less adept at sequence discrimination.
A similar analysis using criterion scores instead of d0 for the detection task revealed no significant correlations with the
sequence discrimination conditions.
The above negative correlation between d0 difference scores (Fig. 2b) might arise if each task’s absolute d0 measurements
were actually positively correlated, but each subject to ceiling versus floor effects respectively. However there was no
significant positive correlation between absolute measures (averaged across visual and auditory conditions for each task)
[r(35) = 0.18, p < 0.29].

3.4. Sensory dominance task

One explanation for the negative correlation above could be that people who are naturally biased to attend more to visual
events can benefit when such events are relevant (resulting in better visual sequence discrimination) but are more distracted
when they are irrelevant (resulting in poorer auditory detection in the context of visual stimulation). To assess this, we
subsequently obtained a measure of visual dominance over audition following Colavita (1974) from 24 of our original
participants, including 6 who had answered ‘Yes’ to the question about hearing the flashes. This subsample had similar
descriptive statistics and showed a very similar pattern of significant and non-significant results as reported above. We
successfully replicated the ‘Colavita Effect’, finding that when participants were presented with an audiovisual stimulus, they
tended on average to erroneously report seeing only a visual stimulus about twice as frequently than reporting only an audi-
tory stimulus [Visual error rate: M = 10.8%, SE = 1.9; Auditory: Mean 5.1%, SE 0.11; t(23) = 3.46, p = 0.002, Cohen’s D = 0.75].
Such visual bias did not correlate significantly with our difference measures of detection or sequence discrimination
[DETd0 AV-A: r(22) = 0.34, p = 0.10; SEQd0 V-A: r(22) = 0.05, p = 0.81]. In a multiple regression analysis, we further confirmed
that the critical relationship (i.e. shown in Fig. 2d) between sequence discrimination and auditory signal detection
(SEQd0 V-A vs. DETd0 AV-A) was still reliable in this sub-sample after controlling for sensory dominance [t(21) = 2.46,
p = 0.022, R2 = 0.31], which itself was not significantly predictive [t(21) = 1.78, p = 0.091].

4. Discussion

The present combination of subjective and objective results provides the first estimate of the prevalence of visual-induced
auditory sensations, and a test of the perceptual reality of such sensations. 22% of our normal participants reported hearing
sounds accompanying visual flashes. These participants performed significantly better in a visual sequence discrimination
task, supporting previous suggestions that the ability to recode visual events as sounds may influence visual discrimination
of rhythmic sequences (Guttman et al., 2006; Saenz & Koch, 2008). Furthermore, our study is the first to suggest that such
auditory-recoding of visual events not only indirectly affects visual performance but directly affects auditory signal detec-
tion. This kind of visually-evoked auditory response (V-EAR) is thus apparently not only quite common but it can impact
on detection of real sounds.
In Saenz and Koch’s (2008) study, ‘hearing-motion’ synaesthetes who reported routinely hearing visually-evoked sounds
were almost as good at discriminating visual flash sequences as they were discriminating auditory tone sequences, while in
22 C. Fassnidge et al. / Consciousness and Cognition 49 (2017) 15–24

non-synaesthetes sequence discrimination was significantly poorer for flashes than for tones. It was inferred from this that
the synaesthetes were uniquely able to recode the flashes as sounds and thus benefit from the better auditory temporal acu-
ity of the auditory modality (Glenberg et al., 1989; Guttman et al., 2006). However, this ability to recode flashes into the
auditory modality might not be restricted to synaesthetes. Testing normal participants, Guttman et al. (2006) found that
visual sequence discrimination was disrupted more by irrelevant auditory rhythms than by irrelevant visual rhythms, sug-
gesting that the visual sequences were being processed using the same auditory resources as the auditory distractors.
Indeed, Guttman et al. (2006) informally reported that they themselves experienced ‘‘a natural tendency to hear the temporal
sequencing of these [visual] changes” (p. 2). However the above studies only obtained an indirect measure of the effects of such
auditory recoding on visual performance. We hypothesised that if there is such a natural tendency for visual flashes to evoke
an auditory response, then these should impact on the detection of real sounds. Our results confirmed this: participants who
showed similar visual and auditory sequence discriminability that resembled Saenz & Koch’s synaesthetes tended to have
poorer auditory detection sensitivity when accompanied by irrelevant visual flashes. This association was found despite
the difference between task and stimuli, which were chosen to closely replicate the methods of the originating studies.
The influence of the visual modality on audition is presumably involuntary rather than a result of a deliberate strategy, as
it did not benefit detection performance. It might occur due to masking of the auditory signal by increased internal noise,
though this does not easily explain why auditory detection benefited from visual stimulation in some participants. Alterna-
tively poorer sensitivity might reflect an attempt in some participants to actively discount internally-originating sensations
from externally-originating signals (Sinke et al., 2014), while others integrate or ignore them; conversely it has been pro-
posed that people with poorer audiovisual integration may develop compensatory synaesthetic tendencies (Sinke et al.,
2014). Interestingly, we found that greater visual interference in the auditory detection task was significantly associated
with poorer auditory sequence discrimination. This supports the idea that an ability to recode visual information as sounds
might compensate for a poorer ability to discriminate pure auditory sequences. This interpretation has some neurophysio-
logical support from evidence of increased visual evoked potentials in temporal auditory areas in hearing-impaired subjects
(Campbell & Sharma, 2014).
In addition to the association with objective auditory detection performance, our subjective data supports the assumption
that performance on the sequence discrimination task reflects an ability to recode flashes as sounds. On debriefing, 22% of
participants answered ‘Yes’ to the surprise question of whether they had been aware of faint sounds accompanying the
flashes in the sequence discrimination task. In agreement with Saenz and Koch (2008), these ‘Yes-Hear’ participants per-
formed significantly better on the visual sequence discrimination tasks compared to those who did not report such sounds.
Given that our participants reported no established history of synaesthesia, the proportion of ‘Yes-Hear’ participants is sur-
prisingly high. This proportion might have been inflated if participants had answered that they actually ‘heard’ flashes when
instead they had just experienced a vivid form of imagery. However only ‘Yes-Hear’ responses to this question predicted sig-
nificantly better performance on the visual sequence discrimination task relative to ‘No-Hear’ responders, while the many
more participants who admitted using a strategy of actively imagining the flashes as sounds did not perform significantly
differently to others on the visual measure. This suggests that the phenomenon of interest is not merely based on imagery,
but on auditory sensations.
Though the subjective data predict some aspects of objective performance, the objective signs of visually-evoked auditory
sensations could also occur independently of awareness. For example, some ‘No-Hear’ participants showed high visual
sequence discriminability; furthermore, the effect of flashes on auditory detection was also not dependent on reporting
awareness of hearing flashes. Two methodological differences from the original study might account for the predominance
of such patterns here. Firstly, we sampled randomly while the original study presumably contrasted self-selected partici-
pants who did versus did not report synaesthesia for hearing visual events. It is therefore possible that Saenz & Koch’s control
group might have underrepresented any normally-occurring tendencies to hear flashes. Conversely, our variant of their para-
digm, with slightly longer events and shorter retention delays between the two comparison intervals, may have been more
sensitive to latent tendencies found in normal participants. A similar dissociation between subjective and objective mea-
sures has been found in studies that have attempted to induce synaesthesia-like associations by training and have found
robust objective effects on information processing but weak evidence of subjective concurrents (reviewed in Deroy &
Spence, 2013). In the present case, reporting a subjective experience may depend on first explicitly noticing the auditory
sensations, which might be easily ignored due to the highly predictable natural association of auditory and visual events
(see below).
Given this apparent partial dissociation of subjective and objective measures, it might be wondered whether the objective
correspondence between sequence discrimination performance and auditory detection is related to other factors jointly
affecting both tasks that are not directly associated with the subjective effects of auditory sensations evoked by visual
events. In particular, a bias towards processing stimuli in the visual modality might have benefitted visual sequence discrim-
ination relative to auditory, while causing visual stimuli to distract more from auditory signal detection. However, our mea-
sure of the Colavita effect (Colavita, 1974; Koppen et al., 2009) helps to discount this possibility. Though we successfully
replicated the Colavita effect, this did not predict performance in either task, thus there is no evidence that the present
results can be accounted for by general differences in bias towards the visual modality.
The above indications of prevalence are consistent with our original hypothesis that some forms of synaesthesia might
occur more frequently if the inducing and concurrent sensations are more frequently associated in nature. This might be
expected on the basis of evidence that some synaesthetic associations can be reinforced by exposure to consistent sensory
C. Fassnidge et al. / Consciousness and Cognition 49 (2017) 15–24 23

pairings (Bor et al., 2014; Witthoft et al., 2015). While artificial correspondences such as consistent grapheme/colour asso-
ciations are extremely rare in nature, visual and auditory events are highly predictive of each other in nature: whenever we
see two objects colliding, or a person’s lips moving to speak, this is usually accompanied by a sudden onset of sound.
Consequently, if an individual has any predisposition towards synaesthetic perceptions, they should be more likely to have
formed audiovisual associations than other less frequently occurring kinds.
The very predictability of visual-to-auditory correspondences might also counterintuitively explain why there have been
so few reports of visually-evoked auditory concurrents (Baron-Cohen et al., 1996; Goller et al., 2009; Saenz & Koch, 2008).
When a faint auditory sensation accompanies a visual event we might easily ignore it because it would not seem surprising
or unusual. We might even assume it was real, and think we are actually hearing the footfalls of somebody walking ahead of
us. In contrast, experiencing a visually-evoked taste instead would be more surprising and thus harder to discount or ignore.
Similar reasoning might explain why rare shape-taste correspondences are classified as synaesthetic, but not taste-smell
associations (e.g. the ‘sweet’ smell of chocolate (Stevenson & Tomiczek, 2007; van Campen, 2008) which naturally belong
together and are thus highly predictable. This perspective of ignored or discounted concurrents predicts that awareness
of visually-evoked sounds might be promoted in some individuals if attention is drawn to them under verifiably silent con-
ditions. There have been cases where individuals have become aware of latent synaesthesias in adulthood through introspec-
tive attention (Tyler, 2005). In tentative support of this, some of our lab assistants and author (CF) informally reported that
they started to hear flash-evoked sounds after repeated exposure to the visual sequence discrimination paradigm.
In terms of brain mechanisms, it is currently debated to what extent synaesthesias reflect genotypically unusual cross-
wiring between sensory areas (Brang et al., 2012; Tomson et al., 2011), versus physiological disinhibition of normally-
occurring connections (Cohen Kadosh & Walsh, 2006; Cohen Kadosh et al., 2009; Grossenbacher & Lovelace, 2001;
Neufeld et al., 2012). This debate might be informed by whether the behavioural measures have a discontinuous or contin-
uous distribution (Cohen Kadosh, 2013; Deroy & Spence, 2015; Martino & Marks, 2001; Simner, 2012). Although we did not
test participants with a history of synaesthesia, we can distinguish a subgroup that reports visually-evoked sounds; however
such subjective awareness seems to be the only dichotomous variable, for even though on average ‘Yes-Hear’ participants
were significantly better at visual discrimination, the distribution of our objective measures appears smooth rather than dis-
continuous. This apparent behavioural continuity seems more likely to reflect normal mechanisms linking auditory and
visual neural representations, either directly or via subcortical or multimodal areas (Ward et al., 2006), rather than unusual
patterns of neural cross-wiring present only in a few individuals. Indeed, neurons responding to visual stimulation have been
found in auditory cortex which may receive stimulation from higher areas or from subcortical regions via interconnections
found in normal anatomy (Ghazanfar & Schroeder, 2006; Schroeder & Foxe, 2005). Such connections may function to
improve spatial localisation of sound or amplify the auditory response to visually identifiable sources (Schroeder & Foxe,
2005). Given this normal connectivity, the additional emergence of a conscious visually-evoked auditory concurrent might
thus be more readily explained by individual physiological variations, which might influence whether existing connections,
or their interactions with higher areas, are inhibited or unmasked (Cohen Kadosh & Walsh, 2006; Cohen Kadosh et al., 2009;
Grossenbacher & Lovelace, 2001; Neufeld et al., 2012), to greater or lesser degree.
This study has provided the first evidence that a surprisingly high proportion of randomly-sampled neurotypical individ-
uals can sometimes report auditory sensations evoked by visual flashes. These internal sounds seem to be perceptually real
enough to interfere with the detection of externally-generated sounds, as well as benefiting visual sequence discrimination,
even if they not always subjectively reportable. The greater prevalence of this phenomenon compared to canonical synaes-
thesias might derive from the strength of statistical correspondence between visual events and the sounds they cause. The
continuous distribution of performance across individuals on our objective measures, and the partial independence of these
from subjective measures seems consistent with the role of normal rather than abnormal neural connections between visual
and auditory representations, the disinhibition of which may occasionally result in emergence of conscious visually-evoked
auditory concurrents.

References

Baron-Cohen, S., Burt, L., Smith-Laittan, F., Harrison, J., & Bolton, P. (1996). Synaesthesia: Prevalence and familiality. Perception, 25(9), 1073–1079. http://dx.
doi.org/10.1068/p251073.
Bor, D., Rothen, N., Schwartzman, D. J., Clayton, S., & Seth, A. K. (2014). Adults can be trained to acquire synesthetic experiences. Scientific Reports, 4, 7089.
http://dx.doi.org/10.1038/srep07089.
Brang, D., Williams, L. E., & Ramachandran, V. S. (2012). Grapheme-color synesthetes show enhanced crossmodal processing between auditory and visual
modalities. Cortex; A Journal Devoted to the Study of the Nervous System and Behavior, 48(5), 630–637. http://dx.doi.org/10.1016/j.cortex.2011.06.008.
Campbell, J., & Sharma, A. (2014). Cross-modal re-organization in adults with early stage hearing loss. PLoS ONE, 9(2), e90594. http://dx.doi.org/10.1371/
journal.pone.0090594.
Cohen, J. (2013). Synesthetic perception as continuous with ordinary perception, or: We’re all synesthetes now. In Sensory blending: New essays on
synaesthesia inproceedings (pp. 1–32). Oxford University Press.
Cohen Kadosh, R. (2013). Using transcranial electrical stimulation to enhance cognitive functions in the typical and atypical brain. Translational
Neuroscience, 4(1), 20–33. http://dx.doi.org/10.2478/s13380-013-0104-7.
Cohen Kadosh, R., Henik, A., Catena, A., Walsh, V., & Fuentes, L. J. (2009). Induced cross-modal synaesthetic experience without abnormal neuronal
connections. Psychological Science, 20(2), 258–265. http://dx.doi.org/10.1111/j.1467-9280.2009.02286.x.
Cohen Kadosh, R., & Walsh, V. (2006). Cognitive neuroscience: Rewired or crosswired brains? Current Biology, 16(22), R962–3. http://dx.doi.org/10.1016/j.
cub.2006.10.017.
Colavita, F. B. (1974). Human sensory dominance. Perception & Psychophysics, 16(2), 409–412. http://dx.doi.org/10.3758/BF03203962.
Cytowic, R. E. (2003). The man who tasted shapes. Imprint Academic. book.
24 C. Fassnidge et al. / Consciousness and Cognition 49 (2017) 15–24

Deroy, O., & Spence, C. (2013). Training, hypnosis, and drugs: Artificial synaesthesia, or artificial paradises? Frontiers in Psychology, 4(October), 1–15. http://
dx.doi.org/10.3389/fpsyg.2013.00660.
Deroy, O., & Spence, C. (2015). Lessons of synaesthesia for consciousness: Learning from the exception, rather than the general. Neuropsychologia, 1–9.
http://dx.doi.org/10.1016/j.neuropsychologia.2015.08.005.
Ghazanfar, A. A., & Schroeder, C. E. (2006). Is neocortex essentially multisensory? Trends in Cognitive Sciences, 10(6), 278–285. http://dx.doi.org/10.1016/j.
tics.2006.04.008.
Glenberg, A. M., Mann, S., Altman, L., Forman, T., & Procise, S. (1989). Modality effects in the coding and reproduction of rhythms. Memory & Cognition, 17(4),
373–383. http://dx.doi.org/10.3758/BF03202611.
Goller, A. I., Otten, L. J., & Ward, J. (2009). Seeing sounds and hearing colors: An event-related potential study of auditory-visual synesthesia. Journal of
Cognitive Neuroscience, 21(10), 1869–1881. http://dx.doi.org/10.1162/jocn.2009.21134.
Green, D. M., & Swets, J. A. (1966). Signal detection theory and psychophysics. New York: Wiley.
Grossenbacher, P. G., & Lovelace, C. T. (2001). Mechanisms of synesthesia: Cognitive and physiological constraints. Trends in Cognitive Sciences, 5(1), 36–41.
http://dx.doi.org/10.1016/S1364-6613(00)01571-0.
Guttman, S. E., Gilroy, L. A., & Blake, R. (2006). Hearing what the eyes see. Psychological Science, 16(3), 228–235. http://dx.doi.org/10.1111/j.0956-
7976.2005.00808.x.Hearin.
Koppen, C., Levitan, C. A., & Spence, C. (2009). A signal detection study of the Colavita visual dominance effect. Experimental Brain Research, 196(3), 353–360.
http://dx.doi.org/10.1007/s00221-009-1853-y.
Lovelace, C. T., Stein, B. E., & Wallace, M. T. (2003). An irrelevant light enhances auditory detection in humans: A psychophysical analysis of multisensory
integration in stimulus detection. Cognitive Brain Research, 17(2), 447–453. Retrieved from <http://www.sciencedirect.com/science/article/pii/
S0926641003001605>.
Martino, G., & Marks, L. E. (2001). Synesthesia: Strong and weak. Current Directions in Psychological Science, 10(2), 61–65. http://dx.doi.org/10.1111/1467-
8721.00116.
Neufeld, J., Sinke, C., Zedler, M., Dillo, W., Emrich, H. M., Bleich, S., & Szycik, G. R. (2012). Disinhibited feedback as a cause of synesthesia: Evidence from a
functional connectivity study on auditory-visual synesthetes. Neuropsychologia, 50(7), 1471–1477. http://dx.doi.org/10.1016/j.neuropsychologia.
2012.02.032.
Ramachandran, S., & Hubbard, E. M. (2001). Perception, thought and language. Journal of Consciousness Studies, 8(12), 3–34.
Rouw, R., & Scholte, H. S. (2007). Increased structural connectivity in grapheme-color synesthesia. Nature Neuroscience, 10(6), 792–797. http://dx.doi.org/
10.1038/nn1906.
Saenz, M., & Koch, C. (2008). The sound of change: Visually-induced auditory synesthesia. Current Biology, 18(15), 650–651. Retrieved from <http://
www.sciencedirect.com/science/article/pii/S0960982208007343>.
Schroeder, C. E., & Foxe, J. (2005). Multisensory contributions to low-level, ‘‘unisensory” processing. Current Opinion in Neurobiology, 15(4), 454–458. http://
dx.doi.org/10.1016/j.conb.2005.06.008.
Simner, J. (2012). Defining synaesthesia. British Journal of Psychology. Retrieved from <http://onlinelibrary.wiley.com/doi/10.1348/000712610X528305/full>.
Simner, J., Mulvenna, C., Sagiv, N., Tsakanikos, E., Witherby, S. A., Fraser, C., ... Ward, J. (2006). Synaesthesia: The prevalence of atypical cross-modal
experiences. Perception, 35(8), 1024–1033.
Sinke, C., Neufeld, J., Zedler, M., Emrich, H. M., Bleich, S., Münte, T. F., & Szycik, G. R. (2014). Reduced audiovisual integration in synesthesia – Evidence from
bimodal speech perception. Journal of Neuropsychology, 8(1), 94–106. http://dx.doi.org/10.1111/jnp.12006.
Stevenson, R. J., & Tomiczek, C. (2007). Olfactory-induced synesthesias: A review and model. Psychological Bulletin, 133(2), 294–309. http://dx.doi.org/
10.1037/0033-2909.133.2.294.
Tomson, S. N., Avidan, N., Lee, K., Sarma, A. K., Tushe, R., Milewicz, D. M., & Ellipsis Eagleman, D. M. (2011). The genetics of colored sequence synesthesia:
Suggestive evidence of linkage to 16q and genetic heterogeneity for the condition. Behavioural Brain Research, 223(1), 48–52. http://dx.doi.org/10.1016/j.
bbr.2011.03.071.
Tyler, C. W. (2005). Varieties of synesthetic experience. Synesthesia: Perspectives from Cognitive Neuroscience, 34–44.
van Campen, C. (2008). The hidden sense: Synesthesia in art and science. Cambridge, MA: MIT Press.
Ward, J. (2013). Synesthesia. Annual Review of Psychology, 64, 49–75. http://dx.doi.org/10.1146/annurev-psych-113011-143840.
Ward, J., Huckstep, B., & Tsakanikos, E. (2006). Sound-colour synaesthesia: To what extent does it use cross-modal mechanisms common to us all? Cortex, 42
(2), 264–280. http://dx.doi.org/10.1016/S0010-9452(08)70352-6.
Watson, A. B., & Pelli, D. G. (1983). QUEST: A Bayesian adaptive psychometric method. Perception & Psychophysics, 33(2), 113–120.
Witthoft, N., Winawer, J., & Eagleman, D. M. (2015). Prevalence of learned grapheme-color pairings in a large online sample of synesthetes. PLoS ONE, 10(3),
e0118996. http://dx.doi.org/10.1371/journal.pone.0118996.

You might also like