Abstract
Previous studies evidenced transfer effects from professional music training to novel word learning. However, it is unclear whether such an advantage is driven by cascading, bottom–up effects from better auditory perception to semantic processing or by top–down influences from cognitive functions on perception. Moreover, the long-term effects of novel word learning remain an open issue. To address these questions, we used a word learning design, with four different sets of novel words, and we neutralized the potential perceptive and associative learning advantages in musicians. Under such conditions, we did not observe any advantage in musicians on the day of learning (Day 1 [D1]), at neither a behavioral nor an electrophysiological level; this suggests that the previously reported advantages in musicians are likely to be related to bottom–up processes. Nevertheless, 1 month later (Day 30 [D30]) and for all types of novel words, the error increase from D1 to D30 was lower in musicians compared to nonmusicians. In addition, for the set of words that were perceptually difficult to discriminate, only musicians showed typical N400 effects over parietal sites on D30. These results demonstrate that music training improved long-term memory and that transfer effects from music training to word learning (i.e., semantic levels of speech processing) benefit from reinforced (long-term) memory functions. Finally, these findings highlight the positive impact of music training on the acquisition of foreign languages.
INTRODUCTION
Playing a musical instrument at a professional level is a multidimensional task that requires acute auditory perception, focused attention, the ability to maintain auditory information in short-term memory (STM) and long-term memory, and specific motor abilities. Typically, professional musicians start playing their instrument at a very young age and train for years, many hours per week. Accumulating evidence indicates that such an intensive training improves auditory perception and attention (Bidelman, Krishnan, & Gandour, 2011; Strait, Kraus, Parbery-Clark, & Ashley, 2010; Micheyl, Delhommeau, Perrot, & Oxenham, 2006; Tervaniemi, Castaneda, Knoll, & Uther, 2006; Fujioka, Trainor, Ross, Kakigi, & Pantev, 2004; Kishon-Rabin, Amir, Vexler, & Zaltz, 2001; Koelsch, Schröger, & Tervaniemi, 1999; Trainor, Desjardins, & Rockel, 1999; Besson & Faïta, 1995; Spiegel & Watson, 1984) and strongly promotes brain plasticity (Münte, Altenmüller, & Jäncke, 2002). Most importantly, for this study, music-training-related advantages have been shown to transfer to the language domain as well (for a review, see Asaridou & McQueen, 2013; Besson, Chobert, & Marie, 2011; Kraus & Chandrasekaran, 2010), having effects at the segmental (Linnavalli, Putkinen, Lipsanen, Huotilainen, & Tervaniemi, 2018; Intartaglia, White-Schwoch, Kraus, & Schön, 2017; Gordon, Fehd, & McCandliss, 2015; Bidelman, Weiss, Moreno, & Alain, 2014; Elmer, Meyer, & Jäncke, 2012; Chobert, Marie, François, Schön, & Besson, 2011), suprasegmental (Torppa, Faulkner, Laasonen, Lipsanen, & Sammler, 2020; Marie, Delogu, Lampis, Belardinelli, & Besson, 2011; Wong & Perrachione, 2007), and even syntactic (Fitzroy & Sanders, 2013) level of speech processing. The evidence also suggests that long-term music training improves second-language acquisition (children's listening comprehension, phonological and word knowledge of English: Yang, Ma, Gong, Hu, & Yao, 2014), auditory attention (Strait, Slater, O'Connell, & Kraus, 2015) and speech-in-noise perception (Slater et al., 2015), visual attention (Wang, Ossher, & Reuter-Lorenz, 2015), working and verbal memory (Roden, Grube, Bongard, & Kreutz, 2014; George & Coch, 2011), executive functions (Habibi, Damasio, Ilari, Elliott Sachs, & Damasio, 2018; Jaschke, Honing, & Scherder, 2018; Zuk, Benjamin, Kenyon, & Gaab, 2014; for reviews, see Benz, Sellaro, Hommel, & Colzato, 2016; Moreno & Farzan, 2015; Moreno & Bidelman, 2014), and general intelligence (Schellenberg, 2004, 2011). However, it has also been argued that genetic factors are possibly mediating the relationships between music training and intelligence (Swaminathan, Schellenberg, & Khalil, 2017; Mosing, Madison, Pedersen, & Ullén, 2016) and that motivation may be more important than music training itself so that children who find music rewarding are more likely to continue music training than children who do not (Schellenberg, Corrigall, Dys, & Malti, 2015).
Recently, by comparing musicians and nonmusicians in three independent samples (children, young adults, and older adults), Dittinger and colleagues (Dittinger, Scherer, Jäncke, Besson, & Elmer, 2019; Dittinger, Chobert, Ziegler, & Besson, 2017; Dittinger et al., 2016) showed that musical expertise influences the semantic level of speech processing. More specifically, the authors focused on novel word learning (a multidimensional task requiring both perceptive and cognitive functions) and conducted a series of experiments that comprised phonological categorization tasks, a picture-word learning phase, and a test phase. The test phase included a matching task (i.e., does the word match or not with the previously learned picture–word association?) and a semantic task (i.e., is the word semantically related or not to a novel picture?). First, both behavioral measures and ERPs indicated that the musicians outperformed the nonmusicians during the phonological categorization tasks based on aspiration and pitch, two nonnative phonetic contrasts for the French-speaking participants (Dittinger, D'Imperio, & Besson, 2018; Dittinger et al., 2016). Second, only the musicians showed a significant N400 increase in amplitude over fronto-central regions, from the first to second half of the picture–word learning phase (François, Cunillera, Garcia, Laine, & Rodriguez-Fornells, 2017; Batterink & Neville, 2011; McLaughlin, Osterhout, & Kim, 2004). Third, in both tasks of the test phase, musicians were characterized by significantly larger N400 amplitude to words that were unexpected compared to words that were expected, based on the previously learned picture–word associations. These N400 effects (i.e., the difference between unexpected minus expected words) were largest over centro-parietal sites, a scalp distribution that is comparable to the N400 effect observed for known words (Kutas, Van Petten, & Besson, 1988). Finally, these results at the electrophysiological level were accompanied by a higher level of performance in the semantic task for the musicians' group, compared to the nonmusicians' group. Importantly, the series of results described above for young adults (Dittinger et al., 2016) were similar in children (Dittinger et al., 2017) but less clear-cut in older adults (Dittinger et al., 2019).
Two main interpretations have been proposed to explain transfer effects from music training to novel word learning (Dittinger et al., 2016). First, according to the cascading interpretation, enhanced auditory perception drive transfer effects in musicians (Mankel & Bidelman, 2018) by facilitating the bottom–up processing stages involved in novel word learning (speech perception, building phonological, lexical, and semantic representations; Besson, Barbaroux, & Dittinger, 2017). Second, according to the multidimensional, top–down interpretation musicians do show not only enhanced speech perception but also improved attention (Strait et al., 2015), working memory and/or STM (Roden et al., 2014; Schulze, Dowling, & Tillmann, 2012; George & Coch, 2011), and executive functions (Habibi et al., 2018; Moreno & Bidelman, 2014; Zuk et al., 2014), and these different functions may, in turn, all contribute to novel word learning. The aim of the present series of experiments was to disentangle these two interpretations and better understand why musicians have an advantage in novel word learning.
To this end, we performed a series of four experiments (see Figure 1). In Experiment 1 (E1), we used a similar design to Dittinger et al. (2016) that aimed at learning word–picture associations, using the same stimuli (monosyllabic Thai words) but with one important modification aimed at reducing the influence of auditory perception and associative learning. That is, rather than performing only one learning phase (to learn picture–word associations) followed by only one matching task (to test for associative learning), participants performed a variable number of learning–matching cycles, until each participant reached a level of 83% correct responses in the matching task. This ensured that all participants had learned the associations equally well before performing the semantic task. Similarly, in Experiment 2 (E2), the same learning–matching task cycles procedure was used, only with Finnish disyllabic words, which, by contrast to the monosyllabic Thai words, were easy to discriminate for all participants. This presumably further attenuated the potential influence of enhanced auditory speech perception in musicians (Strait et al., 2010, 2015; Bidelman et al., 2014). Consequently, if differences in novel word learning between musicians and nonmusicians are still to be found in E1 and E2, these would likely reflect the top–down influences of higher cognitive functions (i.e., attention; better integration of novel words into preexisting semantic networks and/or better memory).
Experimental design. (A) Participants performed four independent novel word learning experiments on 2 days (D1 and D2). Levels of performance in all experiments were retested after 30 days. (B) Each experiment was composed of learning–matching task cycles and a semantic task. The cycles included word learning phases during which each word was paired with its respective picture and matching tasks during which words were presented with one of the pictures, either matching or mismatching the previously learned associations. Once the participant reached 83% correct responses, she or he entered the semantic task where words were presented with novel pictures that were either semantically related or unrelated to the novel words.
Experimental design. (A) Participants performed four independent novel word learning experiments on 2 days (D1 and D2). Levels of performance in all experiments were retested after 30 days. (B) Each experiment was composed of learning–matching task cycles and a semantic task. The cycles included word learning phases during which each word was paired with its respective picture and matching tasks during which words were presented with one of the pictures, either matching or mismatching the previously learned associations. Once the participant reached 83% correct responses, she or he entered the semantic task where words were presented with novel pictures that were either semantically related or unrelated to the novel words.
Experiments 3 and 4 (E3 and E4) were designed as control experiments accounting for musicians' potential advantage in cross-modal integration of visual picture and auditory word associations (Bidelman, 2016; Paraskevopoulos, Kuchenbuch, Herholz, & Pantev, 2012). In E3, disyllabic Finnish words were associated with nonobjects (i.e., objects with no semantic content), so that participants had to learn an association between two unknown items (i.e., neither the pictures nor the words had an established semantic meaning). By contrast, in E4, participants had to learn an association between two known items. That is, known French words were associated with known but noncorresponding pictures (e.g., the word “fromage” [cheese] was presented together with the picture of a hat). If professional musicians perform better than nonmusicians in E3 and E4, this would suggest that facilitation in cross-modal integration contributes to explaining their advantage over the nonmusicians in novel word learning.
After having established whether (1) semantic integration is faster and/or more efficient in musicians even without the advantage of better auditory perception (E1 and E2) and (2) better cross-modal integration in musicians contributes to novel word learning (E3 and E4), the third aim was to investigate the long-term effects of novel word learning. Participants performed the semantic tasks from E1, E2, E3, and E4 again, after a period of 1 month. The aim was to determine whether musicians would remember the previously learned picture–word associations better than nonmusicians. A positive finding would be taken to reflect enhanced attention top–down processes in musicians compared to nonmusicians and would extend previous results showing enhanced working memory and STM in musicians (Schulze et al., 2012; George & Coch, 2011).
METHODS
Participants
Thirty-three participants contributed data, of which 17 were professional musicians (MUS, eight women) and 16 were nonmusician control participants without formal music training (NM, eight women), but who were involved in a regular leisure activity (e.g., sports, dance, theater). Of these, two MUS and one NM were considered outliers based on behavioral performance (±2 SDs away from the mean; see Statistical Analyses section), resulting in equal samples of 15 analyzed data sets for each group. The experiments lasted for altogether 9 hr, and all participants were invited to the laboratory on 3 different days (i.e., two sessions on consecutive days and the third session about 30 days later). The two groups did not significantly differ in age (MUS: mean age = 25.7 years, age range = 19–36 years, SD = 1.5; NM: mean age = 26.0 years, age range = 20–35 years, SD = 1.5), F(1, 28) = 0.02, p = .90. All participants were native French speakers, had comparable education levels (university degree), and reported no past or current auditory or neurological deficits. The MUS group practiced their instruments for an average of 18.3 years (range = 11–29 years, SD = 5.1) and included five violinists, three clarinetists, two pianists, two flautists, one oboist, one trombonist, and one trumpeter. None of the participants was bilingual, but all spoke English as a second language. All participants had a basic knowledge of a third language (mainly Spanish, Italian, or German). The study was conducted in accordance with the Declaration of Helsinki concerning human testing. All participants gave their informed written consent before enrolling in the experiment and received monetary compensation for participating.
Screening Measures
Cognitive Ability
Standardized psychometric tests were used to examine verbal STM and working memory (forward and reverse Digit Span, The Wechsler Adult Intelligence Scale [WAIS-III; Wechsler, 1997]), auditory attention (Associated Responses, adapted from the Neuropsychological [NEPSY-II] child battery; Korkman, Kirk, & Kemp, 2007), visual attention (D2-R; Brickenkamp, Schmidt-Atzert, & Liepmann, 2015), lexical and semantic fluency (Verbal Fluency; Cardebat, Doyon, Puel, Goulet, & Joanette, 1990), and nonverbal general intelligence (Matrices, WAIS-III; Wechsler, 1997).
Musical Aptitude
Participants performed two musicality tests (adapted from the Montreal Battery of Evaluation of Amusia (MBEA) battery; Peretz, Champod, & Hyde, 2003) consisting of judging whether pairs of piano melodies were the same or different. One test was based on melodic information, and the other test was based on rhythmic information.
Experimental Stimuli
Auditory Stimuli
For E1, the stimuli were the same as those used by Dittinger et al. (2016). They comprised 12 natural Thai monosyllabic words: /ba1/, /pa1/, /pha1/, /ba:1/, /pa:1/, /pha:1/, /ba0/, /pa0/, /pha0/, /ba:0/, /pa:0/, and /pha:0/.1 These words varied in vowel duration, with short (261 msec, on average) and long (531 msec, on average) vowels, in fundamental frequency, with low-tone (F0 = 175 Hz, on average) and mid-tone (F0 = 218 Hz, on average) vowels, and in voice onset time (/b/ = −144 msec vs. /p/ = 3 msec vs. /ph/ = 77 msec). Stimuli were recorded by a female Thai–French bilingual, this way ensuring that they were produced naturally. For E2 and E3, 48 natural Finnish disyllabic words were selected: tyyny, lintu, parsa, kirves, jänis, huntu, pullo, ruoho, sanka, farkut, ryhmä, hanhi, norsu, teltta, marsu, hylje, lehmä, laskin, molli, ketju, jalka, myrsky, sänky, huilu, nuoli, lettu, maila, sakset, tähti, harja, kangas, rumpu, juusto, noppa, lisko, sähkö, sydän, varvas, mökki, patja, hillo, kongi, huulet, pyssy, mylly, raastin, järvi, and huivi. These words included nonnative phonetic features for French speakers (i.e., geminate stops, e.g., teltta; initials pronounced as “h,” e.g., hanhi; nonnative vowels for French speakers, e.g., sähkö) and were presented in the two experiments (i.e., two lists of 24 words counterbalanced across E2 and E3 and across participants). To ensure clear perceptual differences, each list contained words without the same initial syllables. Finally, for E4, 24 natural French disyllabic words without the same initial syllable were chosen: fléchette, valise, echelle, bougie, balance, collier, girafe, orteil, gilet, abeille, passoire, mouton, soleil, fenêtre, râteau, cactus, pinceau, tortue, maison, asperge, camion, poupée, croissant, and bocal. A female Finnish–French bilingual recorded the words of E2, E3, and E4, again ensuring that all words were produced naturally. Furthermore, for each word in each experiment, four versions were digitally recorded to reproduce natural speech variability; these different versions of the same word were randomly assigned between word lists and participants. Sound pressure level was normalized across all words to a mean level of 70 dB by using the Praat software (Boersma & Weenink, 2011).
Visual Stimuli
For the learning phases in all four experiments, black and white line drawings were selected. In E1, E2, and E4, these drawings represented familiar objects that were chosen from the Snodgrass and Vanderwart (1980) pictures' set.2 To ensure compatibility with the auditory stimuli, pictures with monosyllabic French names were selected in E1; and pictures with disyllabic French names, in E2 and E4. On the basis of the French normative measures for the Snodgrass and Vanderwart pictures (Alario & Ferrand, 1999), the pictures were controlled for name agreement, image agreement, familiarity, image complexity, age of acquisition, and word frequency. For E1, the same pictures as in Dittinger et al. (2016) were used, whereas for E2 and E4, the remaining pictures were counterbalanced across participants. In E3, line drawings corresponded to the best-rated nonobjects (i.e., do not resemble a real object) created by Kroll and Potter (1984).3 For all four experiments, the same pictures as in the learning phase were then presented in the matching task. For the semantic tasks of E1, E2, and E4, new pictures that the participants had not seen before in the learning–matching task cycles and that were semantically related or unrelated to the meaning of the newly learned words were selected from the Internet. For the semantic task of E3, new visually related or unrelated pictures were created manually in Office PowerPoint (for examples, see Figure 1B). All pictures were pretested with university students (n = 15; age range = 18–30 years), who were asked to rate the semantic or visual relatedness between new and old pictures on a scale from 1 to 5 (1 = not related at all and 5 = strongly related). When the semantic or visual association between the new picture and the old picture was rated as 4 or 5, the pair was later used as a related pair. When the new–old picture association was rated as 1 or 2, the pair was later used as an unrelated pair.
Experimental Tasks
Participants were tested individually in a quiet and shielded (i.e., Faraday cage) experimental room, where they sat in a comfortable chair at about 1 m from a CRT computer screen. Auditory stimuli were presented through HiFi headphones (Sennheiser, HD590) at 70-dB sound pressure level. Visual and auditory stimuli presentation, as well as the collection of behavioral data, was controlled via the Presentation software (NeuroBehavioral Systems, Version 11.0). Four independent experiments were performed, two on D1 (always E1 and E4, with E1 first) and two on Day 24 (always E2 and E3, with E2 first), and each experiment comprised Learning–Matching task cycles, followed by the semantic task (see Figure 1A and B). Moreover, to test for long-term memory effects, participants performed again the semantic tasks of all four experiments (without learning–matching task cycles) around 30 days (range: 24–41 days) after D1.
Learning–Matching Task Cycles
One learning–matching task cycle consisted of one learning block (i.e., one presentation of each picture–word association), followed by one matching task (i.e., one match and one mismatch condition for each picture–word association). The number of cycles varied across participants based on the percentage of correct responses in the matching task. That is, participants performed learning–matching task cycles until they reached at least 83% correct responses twice in a row (i.e., until participants correctly answered 10 of 12 picture–word associations twice in a row). Then, the semantic task was presented (described below).
Word learning phase.
Participants were asked to learn the meaning of the novel words using picture–word associations. In E1, 12 picture–word associations had to be learned, based on known pictures and natural Thai monosyllabic words previously used by Dittinger et al. (2016). In E2, E3, and E4, 24 picture–word associations had to be learned, with known pictures and natural Finnish disyllabic words in E2, nonobjects and natural Finnish disyllabic words in E3, and known pictures and known French disyllabic words in E4 (see Figure 1B for the full list). For instance, in E1, a drawing of a flower was followed by the auditory presentation of the word /pa1/ and, thus, /pa1/ was the word for flower in our “foreign” language. Similarly, in E2, a drawing of a bird was followed by the word “Lintu”; therefore, “Lintu” was the word for bird. In E3, participants were asked to learn the visual appearance (given by the picture of the nonobject) of the novel word “Rumpu.” Finally, In E4, participants learned that “fromage” (meaning cheese in French) was the word for hat. The number of associations to be learned in each experiment (12 in E1 and 24 in E2, E3, and E4) was determined based on pilot data showing that associations were more difficult to learn in E1 than in the other three experiments. In one block of learning, each picture–word association was presented once, resulting in 12 trials for E1 and 24 trials for E2, E3, and E4. The picture was presented first and followed after 750 msec by one of the words. The total trial duration was 2000 msec. Different lists were built, so that across participants, different pictures were associated with different words. No behavioral response was required from the participants during this word learning phase. The duration of one learning block was about half a minute for E1 and 1 min for each of E2, E3, and E4.
Matching task.
One of the pictures was presented, followed after 750 msec by an auditory word that either matched or mismatched the association previously learned (in the word learning phase). After having heard the word, participants were asked to press one of two response keys as quickly and accurately as possible, to indicate whether the presented association was correct or incorrect. After 2750 msec, a row of “Xs” appeared on the screen for 1000 msec, and participants were asked to blink during this period, to minimize eye movements during the next trial. The total trial duration was 3750 msec. For instance, in E1, the drawing of a flower followed by /pa1/ (i.e., flower) was a match and the drawing of a key followed by /pa1/ was a mismatch (see Figure 1B for more examples in E2, E3, and E4). The response hands for indicating matching/mismatching responses were counterbalanced across participants. During one block of the matching task, each word was presented twice, once in a match condition and once in a mismatch condition. Thus, 24 trials for E1 and 48 trials for E2, E3, and E4 were presented in one block, whereas the order was pseudorandomized such that no more than four successive “matches” or “mismatches” were presented. The duration of one matching block was 1.5 min in E1 and 3 min in E2, E3, and E4.
Semantic Task
In E1, E2, and E4, new pictures were presented that were semantically related or unrelated to the words used previously in each of these experiments. In E3, new pictures were visually similar or not to the pictures used previously to learn the picture–word associations. In all four experiments, pictures were presented first followed after 1500 msec by the auditory word. Similar to the matching task, after having heard the word, participants were asked to press one of two response keys as quickly and accurately as possible, to indicate whether the picture and the words were related or unrelated. After 3500 msec, a row of “Xs” appeared on the screen for 1000 msec, and participants were asked to blink during this period. The total trial duration was 4500 msec, and the response hands were counterbalanced across participants. For instance, whereas the picture of soil was semantically related to the previously learned word /pa1/ (i.e., “flower”) in E1, the picture of a lock was semantically unrelated to /pa1/ (see Figure 1B for more examples in E2, E3, and E4). In all experiments, the semantic task started with a short practice block containing four trials, to familiarize participants with the task. After that, 144 trials grouped in two blocks were presented in E1, and 288 trials grouped in four blocks were presented in each of E2, E3, and E4. In total and for each experiment, every previously learned word was presented 12 times, but none of the new pictures was repeated, so that on each trial, the word was associated with a different related or unrelated picture. Half of the picture–word pairs were semantically or visually related, and half were semantically or visually unrelated. The trial order was pseudorandomized such that no more than four successive “related” or “unrelated” pairs were presented. The duration of one block was 5.4 min.
Long-Term Memory Session
To test for long-term memory of the novel words, participants came back to the laboratory about 1 month after the first experimental session, to perform the semantic tasks from all four experiments again (as described above). The order of experiments was counterbalanced across participants for this session.
EEG Data Acquisition and Analysis
The EEG was continuously recorded at a sampling rate of 512 Hz with a band-pass filter of 0–102.4 Hz by using a Biosemi amplifier system (BioSemi Active 2) with 32 active Ag/Cl electrodes (Biosemi Pintype), located at standard positions according to the International 10–20 System (Jasper, 1958). The EOG was recorded from flat-type active electrodes placed 1 cm to the left and right of the external canthi and from an electrode placed beneath the right eye. Two additional electrodes were placed on the left and right mastoids. Electrode impedance was kept below 5 kΩ. EEG data were analyzed using the Brain Vision Analyzer software (Version 1.05.0005; Brain Products, Gmbh). All data were rereferenced offline to the averaged left and right mastoids and filtered with a bandpass filter from 0.1 to 40 Hz. An independent component analysis and an inverse independent component analysis were used to identify and remove components associated with vertical and horizontal ocular movements. Data were segmented into 1200-msec epochs, time-locked to word onset, and included a 200-msec baseline. DC-detrend and removal of artifacts above a gradient criterion of 10 μV/msec or a max–min criterion of 100 μV over the entire epoch were applied. Only trials in which participants gave a correct response were included in the averages, and the number of accepted epochs varied between around 60% (E1) and 80% (E2, E3, and E4). Averages were computed for each participant and for each condition in each experiment, and these individual averages were then averaged into the grand averages across all participants.
Statistical Analyses
ANOVAs were computed using the Statistica software (Version 12.0, StatSoft Inc.). ANOVAs always included Group (MUS vs. NM) as a between-participant factor and specific within-participant factors for each task. Regarding the behavioral data registered on D1, univariate ANOVAs (only including the Group factor) were computed for all experiments on the number of learning–matching task cycles. For the matching and semantic tasks, ANOVAs were computed on error rates (ERRs) and on RTs and included Condition (Matching task: match vs. mismatch and Semantic task: related vs. unrelated) as a within-participant factor. For the matching task, Cycles (first vs. final cycles5) was included as an additional within-participant factor. ANOVAs were first computed by considering each experiment separately and second by including Experiment (E1 vs. E2 vs. E3 vs. E4) as an additional within-participant factor.
Regarding long-term memory for novel words, ANOVAS for D30 always included Group as a between-participant factor and were computed on the difference in ERRs and in RTs between D30 and D1, for both related and unrelated words. As for D1, ANOVAs were first computed separately for each experiment and, second, by including Experiment (E1 vs. E2 vs. E3 vs. E4) as an additional within-participant factor.
Regarding the ERP analyses, for each experiment and on the basis of previous results and visual inspection of the ERP waveforms, the N400 component was analyzed in the semantic task by computing the mean amplitude in the 400- to 550-msec time window. Only correct responses were considered in these analyses. ANOVAs always included Group (MUS vs. NM) as a between-participant factor and Condition (related vs. unrelated) as a within-participant factor, together with Laterality (left: F3, C3, P3; midline: Fz, Cz, Pz; right: F4, C4, P4) and Anterior/Posterior (frontal: F3, Fz, F4; central: C3, Cz, C4; parietal: P3, Pz, P4). As for the behavioral analyses, ANOVAs were first computed for each experiment separately and, second, by including Experiment (E1 vs. E2 vs. E3 vs. E4) as an additional within-participant factor. Post hoc Tukey tests (reducing the probability of Type I errors) were used to determine the origin of significant main effects and interactions. Statistical significance was defined at the .05 alpha level, and results are reported together with the partial eta-squared effect sizes (ηp2).
RESULTS
Psychometric Measures
Results of univariate ANOVAs showed no significant between-group differences regarding the nonverbal general intelligence (F < 1), auditory or visual attention (F(1, 28) = 1.55, p = .22, ηp2 = .052, and F(1, 28) = 1.96, p = .17, ηp2 = .065, respectively), or lexical or semantic fluency (both Fs < 1). By contrast, MUS showed better working memory and STM abilities than NM, F(1, 28) = 6.90, p = .01, ηp2 = .197, and F(1, 28) = 12.83, p = .001, ηp2 = .314, respectively.
Musicality Task
Results of a 2 × 2 ANOVA (2 Groups [MUS vs. NM] × 2 Tasks [Melodic vs. Rhythmic]) showed that MUS made fewer errors (8.1%, SD = 1.6) than NM (15.2%, SD = 1.6; main effect of Group: F(1, 28) = 9.32, p = .005, ηp2 = .249), and all participants performed better on the rhythmic (9.3%, SD = 1.1) than on the melodic (14.1%, SD = 1.8) task (main effect of Task: F(1, 28) = 6.56, p = .02, ηp2 = .189). No Group × Task interaction was observed.
Learning–Matching Task Cycles: Behavioral Results
Number of Cycles
Results of univariate ANOVAs (including Group: MUS vs. NM) for each experiment showed that the number of cycles necessary to reach the threshold did not significantly differ for MUS and for NM in any of the four experiments (main effects of Group: E1, E2, E3, E4: all Fs < 1). Moreover, the between-experiments comparison using two-way ANOVAs (2 Groups × 4 Experiments) showed that participants needed more cycles in E1 (9.0 cycles, SD = 0.5) and fewer cycles in E4 (2.4 cycles, SD = 0.1), compared to E2 (3.6 cycles, SD = 0.2; Tukey, p < .001 and p = .02, respectively) and E3 (3.8 cycles, SD = 0.2; Tukey, p < .001 and p = .005, respectively), which did not differ from each other (number of cycles: E1 > E2 = E3 > E4; main effect of Experiment: F(3, 84) = 105.08, p < .001, ηp2 = .789).
Matching Task
Results of separate ANOVAs (2 Groups [MUS vs. NM] × 2 Cycles [First vs. Final] × 2 Conditions [Match vs. Mismatch]; see Table 1A for F values, p values, and effect sizes [ηp2] and Table 2 for mean ERR and RT values) for each experiment showed that ERRs and RTs did not significantly differ between MUS and NM. Moreover, in all four experiments, participants made fewer errors and responded faster in the final compared to the first cycles.
. | A. Matching Task . | B. Semantic Task D1 . | C. Semantic Task D30 vs. D1 . | ||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
ERRs . | RTs . | ERRs . | RTs . | Error Increase . | RT Decrease . | ||||||||||||||
F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | ||
Main effect of Group | E1 | 0.42 | .52 | .014 | 0.00 | .99 | .000 | 0.32 | .58 | .011 | 0.07 | .79 | .000 | 0.82 | .37 | .028 | 0.10 | .75 | .003 |
E2 | 0.69 | .41 | .024 | 0.51 | .48 | .017 | 1.14 | .29 | .039 | 0.27 | .61 | .009 | 7.50 | .01 | .211 | 2.90 | .10 | .093 | |
E3 | 0.02 | .88 | .000 | 1.10 | .31 | .037 | 0.02 | .90 | .000 | 0.01 | .94 | .000 | 4.93 | .04 | .149 | 1.63 | .21 | .005 | |
E4 | 0.39 | .54 | .013 | 1.45 | .24 | 0.49 | 1.55 | .22 | .052 | 3.47 | .07 | .002 | 5.56 | .03 | .165 | 7.16 | .01 | .203 | |
Main effect of Condition | E1 | 5.71 | .02 | .169 | 4.21 | .05 | .130 | 4.76 | .04 | .145 | 49.46 | .001 | .638 | 1.18 | .29 | .040 | 15.93 | .001 | .362 |
E2 | 38.45 | .001 | .579 | 0.40 | .53 | .014 | 68.34 | .001 | .709 | 67.26 | .001 | .706 | 5.86 | .02 | .173 | 23.12 | .001 | .452 | |
E3 | 41.87 | .001 | .599 | 0.18 | .67 | .006 | 75.01 | .001 | .728 | 73.89 | .001 | .725 | 6.94 | .01 | .198 | 21.17 | .001 | .430 | |
E4 | 25.65 | .001 | .478 | 2.90 | .10 | .093 | 66.03 | .001 | .702 | 41.80 | .001 | .598 | 22.57 | .001 | .446 | 8.94 | .006 | .242 | |
Group × Condition | E1 | 0.24 | .63 | .008 | 0.00 | .98 | .000 | 0.01 | .96 | .000 | 0.03 | .86 | .001 | 9.16 | .005 | .246 | 0.43 | .52 | .015 |
E2 | 0.02 | .88 | .000 | 1.02 | .32 | .035 | 0.80 | .38 | .027 | 0.25 | .62 | .008 | 3.59 | .07 | .113 | 0.40 | .53 | .014 | |
E3 | 0.01 | .93 | .000 | 2.83 | .10 | .091 | 0.02 | .88 | .000 | 0.83 | .37 | .028 | 2.39 | .13 | .094 | 1.02 | .32 | .035 | |
E4 | 1.30 | .26 | .044 | 0.04 | .84 | .001 | 0.00 | .99 | .000 | 1.51 | .23 | .051 | .27 | .60 | .009 | 0.94 | .34 | .032 | |
Main Effect of First/Final | E1 | 365.63 | .001 | .928 | 139.85 | .001 | .833 | ||||||||||||
E2 | 337.20 | .001 | .923 | 159.91 | .001 | .850 | |||||||||||||
E3 | 302.00 | .001 | .915 | 137.80 | .001 | .831 | |||||||||||||
E4 | 54.46 | .001 | .660 | 97.81 | .001 | .777 | |||||||||||||
Group × First/Final | E1 | 1.81 | .19 | .060 | 2.25 | .14 | .074 | ||||||||||||
E2 | 2.03 | .17 | .067 | 0.03 | .87 | .001 | |||||||||||||
E3 | 1.11 | .30 | .038 | 0.14 | .71 | .004 | |||||||||||||
E4 | 0.69 | .41 | .024 | 1.42 | .24 | .048 |
. | A. Matching Task . | B. Semantic Task D1 . | C. Semantic Task D30 vs. D1 . | ||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
ERRs . | RTs . | ERRs . | RTs . | Error Increase . | RT Decrease . | ||||||||||||||
F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | F(1, 28) . | p . | ηp2 . | ||
Main effect of Group | E1 | 0.42 | .52 | .014 | 0.00 | .99 | .000 | 0.32 | .58 | .011 | 0.07 | .79 | .000 | 0.82 | .37 | .028 | 0.10 | .75 | .003 |
E2 | 0.69 | .41 | .024 | 0.51 | .48 | .017 | 1.14 | .29 | .039 | 0.27 | .61 | .009 | 7.50 | .01 | .211 | 2.90 | .10 | .093 | |
E3 | 0.02 | .88 | .000 | 1.10 | .31 | .037 | 0.02 | .90 | .000 | 0.01 | .94 | .000 | 4.93 | .04 | .149 | 1.63 | .21 | .005 | |
E4 | 0.39 | .54 | .013 | 1.45 | .24 | 0.49 | 1.55 | .22 | .052 | 3.47 | .07 | .002 | 5.56 | .03 | .165 | 7.16 | .01 | .203 | |
Main effect of Condition | E1 | 5.71 | .02 | .169 | 4.21 | .05 | .130 | 4.76 | .04 | .145 | 49.46 | .001 | .638 | 1.18 | .29 | .040 | 15.93 | .001 | .362 |
E2 | 38.45 | .001 | .579 | 0.40 | .53 | .014 | 68.34 | .001 | .709 | 67.26 | .001 | .706 | 5.86 | .02 | .173 | 23.12 | .001 | .452 | |
E3 | 41.87 | .001 | .599 | 0.18 | .67 | .006 | 75.01 | .001 | .728 | 73.89 | .001 | .725 | 6.94 | .01 | .198 | 21.17 | .001 | .430 | |
E4 | 25.65 | .001 | .478 | 2.90 | .10 | .093 | 66.03 | .001 | .702 | 41.80 | .001 | .598 | 22.57 | .001 | .446 | 8.94 | .006 | .242 | |
Group × Condition | E1 | 0.24 | .63 | .008 | 0.00 | .98 | .000 | 0.01 | .96 | .000 | 0.03 | .86 | .001 | 9.16 | .005 | .246 | 0.43 | .52 | .015 |
E2 | 0.02 | .88 | .000 | 1.02 | .32 | .035 | 0.80 | .38 | .027 | 0.25 | .62 | .008 | 3.59 | .07 | .113 | 0.40 | .53 | .014 | |
E3 | 0.01 | .93 | .000 | 2.83 | .10 | .091 | 0.02 | .88 | .000 | 0.83 | .37 | .028 | 2.39 | .13 | .094 | 1.02 | .32 | .035 | |
E4 | 1.30 | .26 | .044 | 0.04 | .84 | .001 | 0.00 | .99 | .000 | 1.51 | .23 | .051 | .27 | .60 | .009 | 0.94 | .34 | .032 | |
Main Effect of First/Final | E1 | 365.63 | .001 | .928 | 139.85 | .001 | .833 | ||||||||||||
E2 | 337.20 | .001 | .923 | 159.91 | .001 | .850 | |||||||||||||
E3 | 302.00 | .001 | .915 | 137.80 | .001 | .831 | |||||||||||||
E4 | 54.46 | .001 | .660 | 97.81 | .001 | .777 | |||||||||||||
Group × First/Final | E1 | 1.81 | .19 | .060 | 2.25 | .14 | .074 | ||||||||||||
E2 | 2.03 | .17 | .067 | 0.03 | .87 | .001 | |||||||||||||
E3 | 1.11 | .30 | .038 | 0.14 | .71 | .004 | |||||||||||||
E4 | 0.69 | .41 | .024 | 1.42 | .24 | .048 |
F values, p values, and effect sizes (ηp2) are indicated, and significant p values are in bold.
. | E1 . | E2 . | E3 . | E4 . | |
---|---|---|---|---|---|
ERRs (%) | MUS | 27.0 (1.2) | 21.7 (1.5) | 22.1 (1.4) | 9.2 (1.7) |
NM | 28.1 (1.2) | 19.9 (1.5) | 22.4 (1.4) | 7.7 (1.6) | |
First cycles | 44.8 (2.1) | 34.0 (2.4) | 36.3 (2.3) | 15.2 (2.9) | |
Final cycles | 10.3 (1.1) | 7.6 (1.0) | 8.3 (1.1) | 1.6 (0.7) | |
Match overall | 24.7 (2.1) | 27.0 (2.1) | 29.1 (1.9) | 11.2 (2.2) | |
First cycles | 43.1 (2.4) | 44.0 (2.3) | 48.0 (2.3) | 20.4 (2.8) | |
Final cycles | 6.3 (1.2) | 10.0 (1.1) | 10.1 (1.1) | 2.0 (0.6) | |
Mismatch overall | 30.4 (2.0) | 14.6 (2.0) | 15.5 (2.2) | 5.7 (1.4) | |
First cycles | 46.5 (1.2) | 24.0 (2.5) | 24.6 (2.6) | 10.1 (1.7) | |
Final cycles | 14.3 (1.6) | 5.1 (0.7) | 6.4 (0.9) | 1.3 (0.5) | |
RTs (msec) | MUS | 1212 (42) | 1130 (42) | 1130 (45) | 906 (36) |
NM | 1212 (42) | 1172 (42) | 1197 (45) | 844 (36) | |
First cycles | 1423 (60) | 1297 (48) | 1310 (52) | 981 (44) | |
Final cycles | 1002 (34) | 1006 (43) | 1017 (46) | 770 (34) | |
Match overall | 1189 (48) | 1147 (40) | 1166 (47) | 862 (35) | |
First cycles | 1398 (50) | 1296 (32) | 1313 (40) | 973 (29) | |
Final cycles | 980 (26) | 998 (33) | 1019 (32) | 751 (26) | |
Mismatch overall | 1235 (41) | 1156 (47) | 1161 (46) | 889 (40) | |
First cycles | 1448 (43) | 1298 (38) | 1308 (35) | 988 (36) | |
Final cycles | 1023 (28) | 1014 (33) | 1015 (34) | 790 (25) |
. | E1 . | E2 . | E3 . | E4 . | |
---|---|---|---|---|---|
ERRs (%) | MUS | 27.0 (1.2) | 21.7 (1.5) | 22.1 (1.4) | 9.2 (1.7) |
NM | 28.1 (1.2) | 19.9 (1.5) | 22.4 (1.4) | 7.7 (1.6) | |
First cycles | 44.8 (2.1) | 34.0 (2.4) | 36.3 (2.3) | 15.2 (2.9) | |
Final cycles | 10.3 (1.1) | 7.6 (1.0) | 8.3 (1.1) | 1.6 (0.7) | |
Match overall | 24.7 (2.1) | 27.0 (2.1) | 29.1 (1.9) | 11.2 (2.2) | |
First cycles | 43.1 (2.4) | 44.0 (2.3) | 48.0 (2.3) | 20.4 (2.8) | |
Final cycles | 6.3 (1.2) | 10.0 (1.1) | 10.1 (1.1) | 2.0 (0.6) | |
Mismatch overall | 30.4 (2.0) | 14.6 (2.0) | 15.5 (2.2) | 5.7 (1.4) | |
First cycles | 46.5 (1.2) | 24.0 (2.5) | 24.6 (2.6) | 10.1 (1.7) | |
Final cycles | 14.3 (1.6) | 5.1 (0.7) | 6.4 (0.9) | 1.3 (0.5) | |
RTs (msec) | MUS | 1212 (42) | 1130 (42) | 1130 (45) | 906 (36) |
NM | 1212 (42) | 1172 (42) | 1197 (45) | 844 (36) | |
First cycles | 1423 (60) | 1297 (48) | 1310 (52) | 981 (44) | |
Final cycles | 1002 (34) | 1006 (43) | 1017 (46) | 770 (34) | |
Match overall | 1189 (48) | 1147 (40) | 1166 (47) | 862 (35) | |
First cycles | 1398 (50) | 1296 (32) | 1313 (40) | 973 (29) | |
Final cycles | 980 (26) | 998 (33) | 1019 (32) | 751 (26) | |
Mismatch overall | 1235 (41) | 1156 (47) | 1161 (46) | 889 (40) | |
First cycles | 1448 (43) | 1298 (38) | 1308 (35) | 988 (36) | |
Final cycles | 1023 (28) | 1014 (33) | 1015 (34) | 790 (25) |
Regarding the comparison between experiments (2 Groups [MUS vs. NM] × 4 Experiments [E1 vs. E2 vs. E3 vs. E4] × 2 Conditions [Match vs. Mismatch] × 2 Cycles [First vs. Final]) in the first cycles, ERRs were highest and RTs were slowest in E1, intermediate in E2 and E3, and lowest and fastest in E4 (matching task: E1 > E2 = E3 > E4; Tukey, ERRs and RTs: all ps < .001). By contrast, in the final cycles, ERRs and RTs were not significantly different in E1, E2, and E3 but were still lower and faster in E4 (E1 = E2 = E3 > E4; Tukey, ERRs: all ps < .01; and RTs: all ps < .001; Cycle × Experiment interactions: ERRs: F(3, 84) = 29.21, p < .001, ηp2 = .510; and RTs: F(3, 84) = 10.87, p < .001, ηp2 = .279; level of performance: E4 > E2 = E3 = E1).
Semantic Task, D1: Behavioral Results
The results of separate ANOVAs (2 Groups [MUS vs. NM] × 2 Conditions [Related vs. Unrelated]; see Table 1B for F values, p values, and effect sizes [ηp2]; Table 3 for mean ERR and RT values; and Figure 2A) for each experiment showed that ERRs and RTs did not significantly differ between MUS and NM. Moreover, participants made more errors and responded faster to related than unrelated words in all four experiments.
. | E1 . | E2 . | E3 . | E4 . | |
---|---|---|---|---|---|
ERRs: D1 | MUS | 33.5 (1.5) | 17.9 (1.8) | 24.9 (2.0) | 17.2 (2.3) |
NM | 34.7 (1.5) | 15.1 (1.9) | 24.5 (2.0) | 13.1 (2.3) | |
Related | 37.7 (1.7) | 25.2 (2.1) | 36.6 (2.5) | 23.0 (2.4) | |
Unrelated | 30.4 (2.2) | 7.8 (1.0) | 12.8 (1.4) | 7.3 (1.1) | |
ERRs: D30 | MUS | 33.8 (2.3) | 26.0 (2.7) | 33.3 (2.2) | 27.9 (3.2) |
NM | 37.1 (2.3) | 32.1 (2.7) | 39.5 (2.2) | 33.6 (3.2) | |
Related | 41.1 (2.6) | 42.1 (3.1) | 53.0 (2.4) | 47.2 (3.4) | |
Unrelated | 29.8 (2.7) | 16.0 (2.5) | 19.8 (2.9) | 14.3 (2.7) | |
Error increase: D1–D30 | MUS | 0.3 (1.7) | 8.1 (2.3) | 8.4 (2.1) | 10.7 (2.9) |
NM | 2.5 (1.7) | 17.0 (2.3) | 14.9 (2.1) | 20.4 (2.9) | |
Related | 3.4 (2.3) | 16.9 (2.9) | 16.4 (2.7) | 24.2 (3.1) | |
Unrelated | −0.6 (2.1) | 8.2 (1.9) | 7.0 (1.9) | 7.0 (2.4) | |
RTs: D1 | MUS | 1308 (45) | 1352 (53) | 1297 (47) | 1339 (50) |
NM | 1325 (46) | 1314 (52) | 1292 (46) | 1207 (51) | |
Related | 1242 (31) | 1271 (36) | 1246 (33) | 1232 (34) | |
Unrelated | 1392 (36) | 1395 (40) | 1342 (33) | 1315 (39) | |
RTs: D30 | MUS | 1156 (54) | 1269 (54) | 1222 (56) | 1254 (59) |
NM | 1153 (54) | 1325 (54) | 1277 (56) | 1258 (59) | |
Related | 1121 (39) | 1292 (41) | 1243 (41) | 1249 (45) | |
Unrelated | 1189 (39) | 1302 (38) | 1255 (39) | 1263 (42) | |
RT decrease: D1–D30 | MUS | −152 (43) | −83 (40) | −75 (33) | −85 (36) |
NM | −172 (44) | 12 (40) | −16 (33) | 51 (36) | |
Related | −121 (34) | 21 (30) | −3 (28) | 18 (27) | |
Unrelated | −203 (30) | −93 (30) | −88 (23) | −52 (29) |
. | E1 . | E2 . | E3 . | E4 . | |
---|---|---|---|---|---|
ERRs: D1 | MUS | 33.5 (1.5) | 17.9 (1.8) | 24.9 (2.0) | 17.2 (2.3) |
NM | 34.7 (1.5) | 15.1 (1.9) | 24.5 (2.0) | 13.1 (2.3) | |
Related | 37.7 (1.7) | 25.2 (2.1) | 36.6 (2.5) | 23.0 (2.4) | |
Unrelated | 30.4 (2.2) | 7.8 (1.0) | 12.8 (1.4) | 7.3 (1.1) | |
ERRs: D30 | MUS | 33.8 (2.3) | 26.0 (2.7) | 33.3 (2.2) | 27.9 (3.2) |
NM | 37.1 (2.3) | 32.1 (2.7) | 39.5 (2.2) | 33.6 (3.2) | |
Related | 41.1 (2.6) | 42.1 (3.1) | 53.0 (2.4) | 47.2 (3.4) | |
Unrelated | 29.8 (2.7) | 16.0 (2.5) | 19.8 (2.9) | 14.3 (2.7) | |
Error increase: D1–D30 | MUS | 0.3 (1.7) | 8.1 (2.3) | 8.4 (2.1) | 10.7 (2.9) |
NM | 2.5 (1.7) | 17.0 (2.3) | 14.9 (2.1) | 20.4 (2.9) | |
Related | 3.4 (2.3) | 16.9 (2.9) | 16.4 (2.7) | 24.2 (3.1) | |
Unrelated | −0.6 (2.1) | 8.2 (1.9) | 7.0 (1.9) | 7.0 (2.4) | |
RTs: D1 | MUS | 1308 (45) | 1352 (53) | 1297 (47) | 1339 (50) |
NM | 1325 (46) | 1314 (52) | 1292 (46) | 1207 (51) | |
Related | 1242 (31) | 1271 (36) | 1246 (33) | 1232 (34) | |
Unrelated | 1392 (36) | 1395 (40) | 1342 (33) | 1315 (39) | |
RTs: D30 | MUS | 1156 (54) | 1269 (54) | 1222 (56) | 1254 (59) |
NM | 1153 (54) | 1325 (54) | 1277 (56) | 1258 (59) | |
Related | 1121 (39) | 1292 (41) | 1243 (41) | 1249 (45) | |
Unrelated | 1189 (39) | 1302 (38) | 1255 (39) | 1263 (42) | |
RT decrease: D1–D30 | MUS | −152 (43) | −83 (40) | −75 (33) | −85 (36) |
NM | −172 (44) | 12 (40) | −16 (33) | 51 (36) | |
Related | −121 (34) | 21 (30) | −3 (28) | 18 (27) | |
Unrelated | −203 (30) | −93 (30) | −88 (23) | −52 (29) |
(A) ERRs (top) and RTs (bottom) in the semantic tasks on D1 are shown separately for the four experiments. Results for related (Rel; full bars) and unrelated (Unrel; empty bars) words are illustrated for musicians (MUS; red) and nonmusicians (NM; black). (B) Error increase (top) and RT decrease (bottom) from D1 to D30 in the semantic tasks are shown. The level of significance is represented by asterisks with *p < .05, **p < .01, and ***p < .001.
(A) ERRs (top) and RTs (bottom) in the semantic tasks on D1 are shown separately for the four experiments. Results for related (Rel; full bars) and unrelated (Unrel; empty bars) words are illustrated for musicians (MUS; red) and nonmusicians (NM; black). (B) Error increase (top) and RT decrease (bottom) from D1 to D30 in the semantic tasks are shown. The level of significance is represented by asterisks with *p < .05, **p < .01, and ***p < .001.
Regarding the comparison between experiments (i.e., 2 Groups [MUS vs. NM] × 4 Experiments [E1 vs. E2 vs. E3 vs. E4] × 2 Conditions [Related vs. Unrelated]), participants made most errors in E1, intermediate errors in E3, and fewest errors in E2 and E4 (ERRs: E1 > E3 > E2 = E4; Tukey, all ps < .001; main effect of Experiment: F(3, 84) = 80.36, p < .001, ηp2 = .741). RTs were similar in all four experiments for related words (Tukey, all ps > .10) but faster in E3 and E4 compared to E1 and E2 for unrelated words (Tukey, all ps < .001; Experiment × Condition: F(3, 84) = 5.24, p = .002, ηp2 = .157).
Semantic Task, D1: ERP Results
Turning to the electrophysiological data (see Figure 3 for MUS and Figure 4 for NM), results of the separate ANOVAs for each experiment (2 Groups [MUS vs. NM] × 2 Conditions [Related vs. Unrelated] × 3 Laterality [Left vs. Central vs. Right] × 3 Anterior/Posterior [Frontal vs. Central vs. Parietal]; see Table 4 for μV values and Figures 3 and 4) showed that, in E1, the Group × Anterior/Posterior and Group × Condition × Laterality interactions were significant, F(2, 56) = 3.40, p = .04, ηp2 = .108, and F(2, 56) = 3.23, p = .05, ηp2 = .103, respectively. Separate ANOVAs revealed larger N400 to related than unrelated words over frontal sites in MUS (related: −2.62 μV, SD = 1.43; unrelated: −1.90 μV, SD = 1.40; Tukey, p = .02; Condition × Anterior/Posterior: F(2, 28) = 5.19, p = .01, ηp2 = .270). By contrast, in NM, N400 was larger to unrelated than related words over midline and right hemisphere (unrelated: −0.03 μV, SD = 1.37, and −0.78 μV, SD = 1.16, respectively; related: 0.53 μV, SD = 1.45, and −0.01 μV, SD = 1.20, respectively; Tukey, p = .03 and p = .002, respectively; Condition × Laterality: F(2, 28) = 3.32, p = .05, ηp2 = .191; see Figure 4, “E1”). In E2, E3, and E4, neither the main effect of Group nor any interactions involving the Group factor were significant. Finally, in E1, E2, and E3, the N400 was larger to unrelated than related words over parietal sites and to related than unrelated words over frontal sites in E2 and E3 (Condition × Anterior/Posterior interactions, E1: F(2, 56) = 3.27, p = .05, ηp2 = .104; E2: F(2, 56) = 44.93, p < .001, ηp2 = .616; and E3: F(2, 56) = 53.09, p < .001, ηp2 = .654; Tukey, E1: p = .01, E2 and E3: all ps < .001). In E4, the N400 was larger to related than unrelated words over frontal and central sites (Tukey, both ps < .001; Condition × Anterior/Posterior: F(2, 56) = 21.62, p < .001, ηp2 = .435; see Figure 7, “D1”).
Semantic tasks of the four experiments (E1, E2, E3, and E4) on D1 are shown for musicians. ERPs recorded at frontal (Fz), central (Cz), and parietal (Pz) sites are overlapped for semantically related (solid-red lines) and unrelated (dashed-red lines) words. In this and subsequent figures, time in milliseconds is in abscissa and the amplitude of the effects in microvolt is in ordinate. Time 0 corresponds to word onset, and negativity is plotted upward. Latency windows for statistical analyses are indicated with gray dotted lines, and the level of significance is represented by asterisks with *p < .05, **p < .01, and ***p < .001. Topographic voltage distribution maps of the unrelated minus related differences are illustrated for N400 components. Voltage values are scaled from −1.5 to +1.0 μV.
Semantic tasks of the four experiments (E1, E2, E3, and E4) on D1 are shown for musicians. ERPs recorded at frontal (Fz), central (Cz), and parietal (Pz) sites are overlapped for semantically related (solid-red lines) and unrelated (dashed-red lines) words. In this and subsequent figures, time in milliseconds is in abscissa and the amplitude of the effects in microvolt is in ordinate. Time 0 corresponds to word onset, and negativity is plotted upward. Latency windows for statistical analyses are indicated with gray dotted lines, and the level of significance is represented by asterisks with *p < .05, **p < .01, and ***p < .001. Topographic voltage distribution maps of the unrelated minus related differences are illustrated for N400 components. Voltage values are scaled from −1.5 to +1.0 μV.
Semantic tasks of the four experiments (E1, E2, E3, and E4) on D1 are shown for nonmusicians. ERPs recorded at frontal (Fz), central (Cz), and parietal (Pz) sites are overlapped for semantically related (solid-black lines) and unrelated (dashed-black lines) words. For E1, an additional overlap of related and unrelated words is shown over the right hemisphere (average of F4, C4, and P4).
Semantic tasks of the four experiments (E1, E2, E3, and E4) on D1 are shown for nonmusicians. ERPs recorded at frontal (Fz), central (Cz), and parietal (Pz) sites are overlapped for semantically related (solid-black lines) and unrelated (dashed-black lines) words. For E1, an additional overlap of related and unrelated words is shown over the right hemisphere (average of F4, C4, and P4).
. | E1 . | E2 . | E3 . | E4 . | |
---|---|---|---|---|---|
D1 | Related | −0.24 (1.50) | −2.17 (1.28) | −1.93 (1.08) | −2.33 (1.56) |
Frontal | −1.76 (0.97) | −3.38 (0.82) | −2.66 (0.75) | −3.46 (0.97) | |
Central | −0.46 (0.93) | −2.58 (0.81) | −2.35 (0.66) | −2.80 (0.97) | |
Parietal | 1.50 (0.83) | −0.55 (0.74) | −0.78 (0.60) | −0.74 (0.89) | |
Unrelated | −0.32 (1.55) | −2.00 (1.16) | −2.11 (1.09) | −1.80 (1.48) | |
Frontal | −1.61 (0.94) | −2.37 (0.67) | −1.93 (0.70) | −2.37 (0.90) | |
Central | −0.49 (0.99) | −2.46 (0.73) | −2.58 (0.68) | −2.33 (0.94) | |
Parietal | 1.15 (0.90) | −1.17 (0.71) | −1.83 (0.58) | −0.70 (0.83) | |
D30 | Related | 0.69 (1.50) | −0.79 (1.26) | −0.68 (1.30) | −0.62 (1.24) |
Frontal | −0.52 (0.96) | −1.55 (0.83) | −1.43 (0.86) | −1.43 (0.82) | |
Central | 0.34 (0.93) | −1.16 (0.81) | −1.10 (0.76) | −0.96 (0.78) | |
Parietal | 2.26 (0.87) | 0.34 (0.71) | 0.51 (0.75) | 0.52 (0.73) | |
Unrelated | 0.57 (1.44) | −1.03 (1.25) | −1.68 (1.09) | −1.33 (1.24) | |
Frontal | −0.35 (0.85) | −1.49 (0.76) | −1.82 (0.76) | −1.93 (0.87) | |
Central | 0.33 (0.88) | −1.38 (0.79) | −2.06 (0.66) | −1.79 (0.78) | |
Parietal | 1.73 (0.89) | −0.22 (0.73) | −1.15 (0.61) | −0.29 (0.67) |
. | E1 . | E2 . | E3 . | E4 . | |
---|---|---|---|---|---|
D1 | Related | −0.24 (1.50) | −2.17 (1.28) | −1.93 (1.08) | −2.33 (1.56) |
Frontal | −1.76 (0.97) | −3.38 (0.82) | −2.66 (0.75) | −3.46 (0.97) | |
Central | −0.46 (0.93) | −2.58 (0.81) | −2.35 (0.66) | −2.80 (0.97) | |
Parietal | 1.50 (0.83) | −0.55 (0.74) | −0.78 (0.60) | −0.74 (0.89) | |
Unrelated | −0.32 (1.55) | −2.00 (1.16) | −2.11 (1.09) | −1.80 (1.48) | |
Frontal | −1.61 (0.94) | −2.37 (0.67) | −1.93 (0.70) | −2.37 (0.90) | |
Central | −0.49 (0.99) | −2.46 (0.73) | −2.58 (0.68) | −2.33 (0.94) | |
Parietal | 1.15 (0.90) | −1.17 (0.71) | −1.83 (0.58) | −0.70 (0.83) | |
D30 | Related | 0.69 (1.50) | −0.79 (1.26) | −0.68 (1.30) | −0.62 (1.24) |
Frontal | −0.52 (0.96) | −1.55 (0.83) | −1.43 (0.86) | −1.43 (0.82) | |
Central | 0.34 (0.93) | −1.16 (0.81) | −1.10 (0.76) | −0.96 (0.78) | |
Parietal | 2.26 (0.87) | 0.34 (0.71) | 0.51 (0.75) | 0.52 (0.73) | |
Unrelated | 0.57 (1.44) | −1.03 (1.25) | −1.68 (1.09) | −1.33 (1.24) | |
Frontal | −0.35 (0.85) | −1.49 (0.76) | −1.82 (0.76) | −1.93 (0.87) | |
Central | 0.33 (0.88) | −1.38 (0.79) | −2.06 (0.66) | −1.79 (0.78) | |
Parietal | 1.73 (0.89) | −0.22 (0.73) | −1.15 (0.61) | −0.29 (0.67) |
Regarding the comparison between experiments (2 Groups [MUS vs. NM] × 4 Experiments [E1 vs. E2 vs. E3 vs. E4] × 2 Conditions [Related vs. Unrelated] × 3 Laterality [Left vs. Central vs. Right] × 3 Anterior/Posterior [Frontal vs. Central vs. Parietal]), N400 amplitude was smaller in E1 than in E2, E3, and E4 (Tukey, all ps < .001; main effect of Experiment: F(3, 84) = 16.91, p < .001, ηp2 = .376).
Semantic Task: Change in Behavioral Performance from D1 to D30
The results of separate ANOVAs for each experiment (2 Groups [MUS vs. NM] × 2 Conditions [Related vs. Unrelated]; see Table 1C for F values, p values, and effect sizes [ηp2]; Table 3 for mean ERR and RT values; and Figure 2B) revealed that, in E2, E3, and E4, the error increase from D1 to D30 was smaller in MUS than in NM (main effects of Group, E2: F(1, 28) = 7.50, p = .01, ηp2 = .211; E3: F(1, 28) = 4.93, p < .04, ηp2 = .149; and E4: F(1, 28) = 5.56, p < .03, ηp2 = .165). In E1, the group effect was significant for unrelated words only (Group × Condition interaction: F(1, 28) = 9.16, p = .005, ηp2 = .246, with a significant main effect of Group for unrelated words, F(1, 28) = 8.98, p = .006, ηp2 = .242, but not for related words, F(1, 28) = 3.84, p = .06, ηp2 = .120). RT decreases from D1 to D30 were larger in MUS than in NM only in E4 (main effect of Group: F(1, 28) = 7.16, p < .01, ηp2 = .203).
Regarding the comparison between experiments (2 Groups [MUS vs. NM] × 4 Experiments [E1 vs. E2 vs. E3 vs. E4] × 2 Conditions [Related vs. Unrelated]), error increases were smaller and RT decreases were larger in E1 compared to the other three experiments (Tukey, all ps < .001). Moreover, error increases were smaller in E3 compared to E4 (Tukey, p = .05; main effects of Experiment: ERRs: F(3, 84) = 33.90, p < .001, ηp2 = .547; and RTs: F(3, 84) = 9.51, p < .001, ηp2 = .253).
Semantic Task, D30: ERP Results
The results of separate ANOVAs for each experiment (2 Groups [MUS vs. NM] × 2 Conditions [Related vs. Unrelated] × 3 Laterality [Left vs. Central vs. Right] × 3 Anterior/Posterior [Frontal vs. Central vs. Parietal]; see Table 4 for μV values, Figure 5 for MUS, and Figure 6 for NM) revealed that the Group × Condition × Anterior/Posterior interaction was significant in E1, F(2, 56) = 3.27, p = .05, ηp2 = .104. Separate ANOVAs showed larger N400 to unrelated than related words over parietal sites in MUS (unrelated: 1.56 μV, SD = 1.37; related: 2.56 μV, SD = 1.45; Tukey, p < .001; Condition × Anterior/Posterior: F(2, 28) = 9.86, p < .001, ηp2 = .413), but not in NM (main effect of Condition and Condition × Anterior/Posterior interaction: both Fs < 1). In E2, E3, and E4, neither the main effect of Group nor any interaction involving the Group factor was significant. Finally, in E1 and E2, results showed larger N400 to unrelated than related words over parietal sites (Tukey, p = .009 and p = .004, respectively; Condition × Anterior/Posterior interactions: F(2, 56) = 6.13, p = .004, ηp2 = .179, and F(2, 56) = 4.50, p = .02, ηp2 = .138, respectively) and over all scalp sites in E3 and E4 (main effect of Condition: F(1, 28) = 13.53, p < .001, ηp2 = .325, and F(1, 28) = 7.54, p = .01, ηp2 = .212, respectively; see Figure 7).
Semantic tasks of the four experiments (E1, E2, E3, and E4) on D30 are shown for musicians. ERPs recorded at frontal (Fz), central (Cz), and parietal (Pz) sites are overlapped for semantically related (solid-red lines) and unrelated (dashed-red lines) words.
Semantic tasks of the four experiments (E1, E2, E3, and E4) on D30 are shown for nonmusicians. ERPs recorded at frontal (Fz), central (Cz), and parietal (Pz) sites are overlapped for semantically related (solid-black lines) and unrelated (dashed-black lines) words.
Semantic tasks of the four experiments (E1, E2, E3, and E4) on D1 (top row) and D30 (bottom row) are shown for all participants (averaged across musicians and nonmusicians). ERPs recorded at frontal (Fz) and parietal (Pz) sites are overlapped for semantically related (solid-blue lines) and unrelated (dashed-blue lines) words.
Semantic tasks of the four experiments (E1, E2, E3, and E4) on D1 (top row) and D30 (bottom row) are shown for all participants (averaged across musicians and nonmusicians). ERPs recorded at frontal (Fz) and parietal (Pz) sites are overlapped for semantically related (solid-blue lines) and unrelated (dashed-blue lines) words.
Regarding the comparison between experiments (2 Groups [MUS vs. NM] × 4 Experiments [E1 vs. E2 vs. E3 vs. E4] × 2 Conditions [Related vs. Unrelated] × 3 Laterality [Left vs. Central vs. Right] × 3 Anterior/Posterior [Frontal vs. Central vs. Parietal]), the N400s were smaller in E1 than in E2, E3, and E4 (Tukey, all ps < .001; main effect of Experiment: F(3, 84) = 12.50, p < .001, ηp2 = .308).
DISCUSSION
On the basis of previous results showing better novel word learning in adult musicians than in nonmusicians (Dittinger et al., 2016), we conducted a series of experiments in which we reduced the influence of auditory perception and associative learning on semantic processing. Of interest was to determine whether, under such conditions, musicians would still outperform nonmusicians in the semantic task, whether audiovisual integration contributed to the musician's advantage in novel word learning, and whether musicians would remember the newly learned words better than nonmusicians 1 month later. Overall, the results of both behavioral and electrophysiological data did not bring evidence for a musician's advantage in semantic processing on D1, thereby showing that, when between-group differences in auditory perception are neutralized, musicians no longer outperform nonmusicians in the semantic task. Interestingly, however, on D30, musicians showed better long-term memory for the novel words. Finally, the results did not reveal better audiovisual integration in musicians than in nonmusicians. These findings are discussed in detail below.
D1: Learning–Matching Task Cycles
To neutralize the influence of auditory perception and associative learning, participants performed a variable number of short learning–matching task cycles until each participant reached a level of 83% correct responses in the matching task. Successful learning in all four experiments was reflected by lower ERRs and faster RTs in the final compared to the first cycles of the matching tasks. Participants needed five cycles on average to reach the learning threshold (i.e., five repetitions of each picture–word association in the learning phases and five repetitions in the matching tasks). Such fast mapping (Carey, 1978) is not surprising in view of the word learning literature evidencing that novel word encoding can be successful only with few repetitions (Borovsky, Elman, & Kutas, 2012; Batterink & Neville, 2011; Dobel et al., 2010; Mestres-Missé, Rodriguez-Fornells, & Münte, 2007; Perfetti, Wlotko, & Hart, 2005).
Interestingly, the number of cycles varied as a function of the experiment, with significantly more cycles in E1 (n = 9), intermediate number in E2 and E3 (n = 3.6 and n = 3.8, respectively), and fewer cycles in E4 (n = 2.4; E1 > E2 = E3 > E4). Because the same experimental design was used in the four experiments, these differences likely result from the different levels of familiarity and discriminability of the stimuli. Whereas in E4, the novel words were familiar French words, in E1, the novel words were unfamiliar monosyllabic Thai words (including aspiration, tonal, and duration contrasts) that were difficult to discriminate for French participants, and in E2 and E3, the stimuli were disyllabic Finnish words that were unfamiliar but still easy to discriminate. Thus, participants needed more cycles in E1, even if the number of to-be-learned words was half the number of words in the other three Experiments (i.e., 12 words compared to 24 words in E2, E3, and E4).6 Note that results in E1, showing that 18 repetitions of each picture–word association (nine cycles, 18 repetitions) were needed on average to reach the threshold, are comparable to the results reported by Dittinger et al. (2016), in which 20 repetitions of each picture–word association were used to reach a similar level of performance in the matching task. Taken together, these results point to the importance of the stimulus material in driving novel word learning.
The interpretation of E1 being the most difficult and E4 being the easiest experiment is supported by the results in the first cycles of the matching tasks showing the highest ERRs and slowest RTs in E1 as well as the lowest ERRs and fastest RTs in E4 (E1 > E2 = E3 > E4). Moreover, whereas E4 remained the easiest task in the final cycles, the level of performance in E1 was similar to that in E2 and E3 (no significant differences on ERRs and on RTs; E1 = E2 = E3 > E4), thereby showing that, although the difficult Thai words in E1 needed more repetitions, they were comparably-well encoded in the final cycles, relative to the (easier) Finnish words.
Most importantly for the specific goal of this study, ERRs and RTs in the matching task were not significantly different for musicians and nonmusicians in any of the four experiments (see Figure 2A). This is evidence that the implemented manipulation was successful and that potential group differences based on auditory perception, and perceptual learning-related processes (as shown, for instance, by the higher level of performance of musicians compared to nonmusicians in both the rhythmic and melodic tasks) were neutralized before participants performed the semantic task. Nevertheless, the possibility remains that musicians and nonmusicians used different strategies to reach similar outcomes.
D1: Semantic Task
When the perceptual differences between musicians and nonmusicians were neutralized, the results showed no between-group difference at a behavioral level and only in E1 at the electrophysiological level. However, as the E1 results were not expected based on previous results (Dittinger et al., 2016, 2017), they would need to be replicated before being considered further.
Semantic Priming Effects
Typical semantic priming effects (Meyer & Schvaneveldt, 1971) were found in both groups of participants and in all four experiments, with shorter RTs for related than for unrelated trials (see Figure 2A). This is taken as evidence that novel words were already integrated into semantic networks, so that both musicians and nonmusicians were able to rapidly generalize the novel words' meanings to novel concepts. Moreover, this finding also clearly points to a relationship between word discriminability at the phonological level and word learning at the semantic level. However, in all four experiments, and as reported in the group of young adults tested by Dittinger et al. (2016), participants made more errors to related than unrelated words, thereby producing a speed–accuracy trade-off. This possibly reflects a response bias toward rejection (i.e., considering the word as unrelated to the picture), when the task generates a high degree of response uncertainty (Gigerenzer, 1991).
The interpretation that the novel words were integrated into semantic networks and produced semantic priming effects is supported by the electrophysiological data, showing larger N400 components to the unrelated than to the related words over parietal sites in E1, E2, and E3 and larger N400s to the related than to the unrelated words over frontal sites in E2, E3, and E4 (see Figure 7). In line with previous literature (François et al., 2017; Dittinger et al., 2016; Borovsky et al., 2012; Cunillera et al., 2009; De Diego Balaguer, Toro, Rodriguez-Fornells, & Bachoud-Lévi, 2007; Mestres-Missé et al., 2007), the parietal N400 is taken as evidence that the meaning of the novel words has been integrated into semantic networks, whereas the presence of a frontal reversed N400 is taken to reflect manipulation of novel information in working memory (Hagoort, 2014) and/or ongoing consolidation of episodic memory traces (Mestres-Missé et al., 2007).
Interestingly in E4, results showed larger N400 to related than unrelated words (inversed N400 effects) over fronto-central sites, together with no typical N400 effect over parietal sites. In this experiment (E4), the novel word learning required participants to replace an existing association stored in long-term memory (e.g., the picture of a hat associated with the word hat, “chapeau”) by a new association (e.g., expect the word “fromage,” cheese, after the picture of a hat; see Figure 1). The fronto-central inversed N400 effect is in line with the interpretations proposed above. However, the absence of a parietal N400 effect possibly indicates that the new association (“hat” [picture] – “cheese” [word]) has not yet overridden the old association (“hat” [picture] – “hat” [word]) in long-term memory, even if E4 seemed the easiest of the four experiments performed on D1 (as reflected by fastest learning and lowest ERRs in the matching and semantic tasks).
Finally, we turn to the second question that this series of experiments was designed to answer. On the basis of previous results showing better audiovisual binding (Bidelman, 2016) and higher detection rate of audiovisual violations of a preestablished rule (i.e., “the higher the pitch of the tone, the higher the position of the circle”; Paraskevopoulos et al., 2012) in musicians than in nonmusicians, we asked whether musical expertise would have an influence on the audiovisual integration when picture–word associations did or did not contain semantic information (control experiments E3 and E4). Results revealed that the musicians did not show improved audiovisual integration, by comparison to the nonmusicians, when the musicians' advantage on auditory perception and associative learning has been neutralized. Thus, coming to an answer, these results suggest that the more efficient audiovisual integration and better novel word learning through picture–word associations reported in musicians than in nonmusicians in previous experiments likely relied on better auditory perception, that is, on bottom–up rather than on top–down process.
D30: Semantic Task
Participants were retested 1 month later, to investigate the long-term memory effects of novel word learning and whether long-term memory for novel words would differ between musicians and nonmusicians. Results showed that musicians outperformed nonmusicians in all four experiments on D30 (see Figure 2A). This was reflected by smaller error increases from D1 to D30 for musicians than for nonmusicians, for both related and unrelated words in all experiments, except in E1, where this effect was only significant for unrelated words. RT decreases from D1 to D30 were also larger in musicians than in nonmusicians in E4. In the ERPs, results in E1 showed between-group differences with larger N400 components to unrelated than related words over parietal sites in musicians, but not in nonmusicians (compare Figures 5 and 6). Thus, in addition to better behavioral performance in all four experiments, the between-group differences on semantic processing on D30 were also reflected by the N400 component, but only when the words were unfamiliar and difficult to discriminate (i.e., Thai words in E1; see discussion above).
At least three interpretations possibly account for better long-term memory in musicians. First, although here there was no evidence showing that musicians encoded the words better at a behavioral level (i.e., no between-group differences in the number of learning cycles, no between-group difference in the matching tasks and in the semantic tasks on D1), results of psychometric tests revealed better working memory and STM abilities in musicians that in nonmusicians, which possibly contributed to form stronger memory traces of the novel words, for later recall (Wojcik, 2013; Morris, Bransford, & Franks, 1977). Thus, words remembered on D30 were possibly more strongly encoded and better integrated into the semantic networks, already on D1. Second, the musicians' higher level of performance on D30 was accompanied by significant N400 effects only in E1. This suggests that the superior auditory skills of musicians may still play a role under difficult encoding conditions of Thai words that called upon preexisting representations of duration and pitch. Thus, in this case, the stronger memory traces of the novel words for musicians than for non-musicians may emerge from better auditory processing, and better auditory processing may also foster recognition of the Thai stimuli on D30. Finally, it is also possible that long-term consolidation processes are more efficient in musicians than in nonmusicians. Previous results have clearly highlighted the importance of consolidation periods in novel word learning (Bakker, Takashima, van Hell, Janzen, & McQueen, 2015; Takashima, Bakker, van Hell, Janzen, & McQueen, 2014; Dumay & Gaskell, 2007). These different interpretations can be disentangled in further experiments by sorting ERPs recorded during novel word encoding (i.e., learning phase on D1) as a function of whether the words were subsequently remembered on D30 or not (i.e., examining differences at encoding based on subsequent memory, the so-called Dm effect; Paller, Kutas, & Mayes, 1987).
Semantic Priming Effects
The N400 was larger to unrelated than related words across all scalp sites in E3 and E4 and only over parietal sites in E1 and E2 (see Figure 7). These findings suggest that, under ecologically valid conditions for novel word learning and semantic processing (i.e., in E1 and E2, in which novel Thai or Finnish words are attached to known concepts [i.e., meaningful pictures], as it is the case in foreign language learning), N400 effects are localized over parietal scalp sites, as previously reported in the literature (Kutas & Federmeier, 2011; Kutas & Hillyard, 1980). By contrast, results of the control conditions for associative learning (i.e., E3: no semantic information in the novel Finnish words and in the novel nonmeaningful pictures; E4: violations of already well-established semantic associations between French words and meaningful but unrelated pictures) showed largely distributed effects over the scalp, possibly reflecting the call into play of more general processes than the specific semantic processes involved in E1 and E2.
Between-Experiment Comparison
As discussed above, although E1 was very difficult for all participants, the error increase from D1 to D30 was lower, the RT decrease was larger, and the N400 amplitude was smaller than in the other three experiments (see Figures 2B and 7). These results may be explained by the higher number of repetitions in the learning–matching task cycles in E1 than in the other experiments (i.e., nine cycles in E1 vs. 2.4 in E4, 3.6 in E2, and 3.8 in E3). This larger number of repetitions possibly favored deeper encoding and stronger traces in long-term memory. Such an interpretation would also account for the results in E4: Although participants made the fewest errors in E4 compared to the other experiments on D1, the error increase in E4 was higher than in E3 on D30, suggesting that fast learning, with the lowest number of repetitions on D1, was not sufficient to establish strong memory traces. This is an interesting issue for future experiments that may test for the influence of the number of repetitions on long-term memory for words, for example, by testing whether participants who learn most slowly (i.e., with the highest number of learning–matching task cycles) remember best after 30 days.
Conclusion
By using an experimental design similar to Dittinger et al. (2016), and by comparing four sets of novel words in different experiments, results showed three main findings. First, enhanced auditory perception seem to largely contribute to explain the musicians' advantage in novel word learning. If such advantages are neutralized by testing participants when they have reached the same level of performance in the matching task, results showed no between-group differences in the semantic task. Second, results on D30 clearly showed that fast mapping between a novel word and a concept (represented by a picture) through few repetitions is sufficient to establish a memory trace of the novel word meaning that can be accessed 1 month later. Third and most importantly, results revealed better long-term memory for novel words in musicians than in nonmusicians in all four experiments, a result that, to the best of our knowledge, has not been reported before. These results extend to long-term memory, transfer effects previously described from music expertise to working memory and STM (Kraus, Strait, & Parbery-Clark, 2012; Schulze et al., 2012; George & Coch, 2011; Ho, Cheung, & Chan, 2003; for a meta-analysis, see Talamini, Altoè, Carretti, & Grassi, 2017).
Turning to limitations, the present experiment is based on a correlational design and results do not demonstrate causality. Longitudinal experiments are needed to further test for a causal link between musical expertise and long-term memory. Another limitation is that, because of the complexity of the experimental design involving four experiments and several dependent variables, we only analyzed the N400 and no other ERP components that may have revealed complementary information.
Importantly, whereas results on D1 suggest that the musicians' advantage in novel word learning was mainly driven by auditory perception, results of D30 nevertheless point to differences in long-term memory. Taken together, these findings suggest that enhanced auditory perception, reported in previous experiments (Strait et al., 2010; Micheyl et al., 2006; Kishon-Rabin et al., 2001; Spiegel & Watson, 1984), combine with better long-term memory, as found here, to optimize novel word learning in musicians. Such interactions of perceptive and memory functions, as well as transfer effects from musical expertise to long-term memory and novel word learning, add evidence in favor of domain-general networks in the brain underlying perception, cognition, and language processing (for a review on the hotly debated issue of domain-specific vs. domain-general networks in the brain, see Besson, Dittinger, & Barbaroux, 2018). Moreover, because STM and long-term memory are crucial cognitive functions central to language processing, these results highlight the potential impact of music training starting at a young age.
Funding Information
The present work was carried out within the Laboratory of Cognitive Neuroscience (LNC) and was funded by the French government, through the French National Agency for Research (ANR), for the Labex BLRI (ANR-11-LABX-0036) and the program “Investissements d'Avenir” for the Institute for Language and Communication in the Brain (ANR-11-IDEX-0001-02). E. D. was supported by a doctoral fellowship from the BLRI (http://dx.doi.org/10.13039/501100001665), and B. K. benefited from a grant from the A*MIDEX program (ANR-11-IDEX-0001-02).
Diversity in Citation Practices
A retrospective analysis of the citations in every article published in this journal from 2010 to 2020 has revealed a persistent pattern of gender imbalance: Although the proportions of authorship teams (categorized by estimated gender identification of first author/last author) publishing in the Journal of Cognitive Neuroscience (JoCN) during this period were M(an)/M = .408, W(oman)/M = .335, M/W = .108, and W/W = .149, the comparable proportions for the articles that these authorship teams cited were M/M = .579, W/M = .243, M/W = .102, and W/W = .076 (Fulvio et al., JoCN, 33:1, pp. 3–7). Consequently, JoCN encourages all authors to consider gender balance explicitly when selecting which articles to cite and gives them the opportunity to report their article's gender citation balance.
Acknowledgments
We thank all the participants, as well as Chotiga Pattamadilok, Aino Niklas-Salminen, and Mathilde Niklas-Salminen, for registering the auditory stimuli. Moreover, we thank Ken Paller and James McQueen for helpful discussions in the building of the task design. The present work was carried out within The Laboratoire de Neurosciences Cognitives (LNC) Labex Brain and Language Language Institute (ANR-11-LABX-0036) and within the Institute for Language and Communication in the Brain and has been supported by the French government, through the French National Agency for Research (ANR), under the program “Investissements d'Avenir” (ANR-11-IDEX-0001-02). E. D. was supported by a doctoral fellowship from the BLRI (http://dx.doi.org/10.13039/501100001665), and B. K. benefited from a grant from the A*MIDEX program (ANR-11-IDEX-0001-02) funded by the French Government under the program “Investissements d'Avenir.”
Reprint requests should be sent to Mireille Besson, Université Publique, Laboratory of Cognitive Neuroscience, Centre Saint-Charles, Pole 3C, Case C, 3 Place Victor Hugo, 13331 Marseille Cedex 3, France, or via e-mail: mireille.besson@univ-amu.fr.
Notes
Following phonetic transcription in Thai, 1 refers to low-tone; 0, to mid-tone; ph, to aspirated voicing; and the colon, to long vowel duration.
Pictures were based on the standardized set of 260 pictures built by Snodgrass and Vanderwart (1980) but retraced in Office PowerPoint to ensure sufficient resolution and quality.
To simplify, we refer to both Day 1 and Day 2 as “Day 1” (i.e., initial learning session) throughout the paper.
“First cycles” refers to the trials of the matching tasks of the first two learning–matching task cycles, and “Final cycles” refers to the trials of the matching tasks of the last two learning–matching task cycles.
The number of words to be learned in each experiment was decided based on results of pilot studies.