Skip to main content
Have a personal or library account? Click to login
MGPHot: A Dataset of Musicological Annotations for Popular Music (1958–2022) Cover

MGPHot: A Dataset of Musicological Annotations for Popular Music (1958–2022)

Open Access
|May 2025

Full Article

1 Introduction

Music annotation datasets have become essential in advancing research in music information retrieval (MIR), machine learning, and artificial intelligence (Bertin‑Mahieux et al., 2011; Chen et al., 2018; Fonseca et al., 2017). These datasets provide the foundation for a variety of tasks, such as music recommendation, genre classification, emotion detection, and computational musicology. Arguably, the most comprehensive effort in annotating and analyzing music is the Music Genome Project,® a pioneering initiative that catalogs music based on detailed musical attributes, referred to as ‘genes’. With millions of songs analyzed, it is the largest dataset of this type, featuring more than 20 years of musicological analysis by a team of trained musicologists. It has played a significant role in the early advancements of music streaming and personalized music recommendation (Glaser et al., 2006; Landau, 2022). This proprietary dataset has been utilized in various publications contributing to MIR research on topics such as multimodal representation learning (Ferraro et al., 2023), music retrieval and recommendation (Oramas et al., 2024), artist similarity (Korzeniowski et al., 2022), natural language understanding (Oramas et al., 2021), modeling of music genres (Prockup et al., 2015a) and rhythm characteristics (Prockup et al., 2015b), and audio classification (McCallum et al., 2022; Pons et al., 2018).

The MGPHot dataset1 proposed in this paper is derived from it. Unlike conventional datasets, which typically rely on basic metadata and tags like artist names, genres, or moods, MGPHot offers a rich, high‑dimensional set of human‑curated annotations. These annotations go beyond binary labels, offering scores that reflect the intensity of specific musical attributes. This detailed data allow for deeper insights into musical structure, lyrical content, and technical characteristics, enabling more nuanced approaches to tasks such as music classification and recommendation, as well as facilitating unique musicological studies. In Table 1, we compare MGPHot with some of the most commonly used public datasets in MIR.

Table 1

Some public music datasets used in MIR, with song counts, attributes, categories, and curation details.

Dataset#Songs#Attributes#CategoriesCurated by expertsAttribute scoresChart data
GZTAN (Tzanetakis and Cook, 2002)1,000101NoNoNo
Ballroom (Gouyon et al., 2006)69881NoNoNo
MagnaTagATune (Law et al., 2009)5,4051881NoNoNo
MGPHot21,320587YesYesYes
MTG‑Jamendo (Bogdanov et al., 2019)55,7011953NoNoNo
FMA (Defferrard et al., 2017a)106,5741631NoNoNo
MuMu (Oramas et al., 2017)147,2952501NoNoNo
MSD500 (Won et al., 2021)158,3235007NoNoNo
MSD‑last.fm (Bertin‑Mahieux et al., 2011)505,216522,3661NoNoNo
Audioset (Gemmeke et al., 2017)2,084,3205277NoNoNo

The MGPHot dataset includes 58 musical attribute annotations, grouped into seven categories: rhythm, instrumentation, sonority, harmony, compositional focus, vocals, and lyrics, covering 21,320 songs that appeared on the Billboard Hot 100 charts for at least one week between 1958 and 2022. The Billboard Hot 100 charts compile data from record sales, radio airplay, and streaming numbers. These charts allow researchers to identify the most popular songs from specific time periods and quantify the zeitgeist of musical taste from that era. However, these charts are focused on the US market, resulting in a bias towards the tastes and music consumption habits of that population. The Billboard Hot 100 charts have been extensively employed in MIR research, including chart prediction based on audio (Zangerle et al., 2019) or social media activity (Zangerle et al., 2016) and the study of the evolution of popular music (Mauch et al., 2015). Additionally, this and other charts from Billboard have been used for cultural and musicological studies, such as analyzing lyrical tendencies of popular songs (Pettijohn and Sacco Jr, 2009), changes in self‑promotion language (McAuslan and Waung, 2018), and race and gender studies (Lafrance et al., 2011; Watson, 2019; Wells,1991).

In this work, we describe the composition, annotation methodology, and potential applications of the MGPHot dataset in music technology research. Additionally, we illustrate its intrinsic value through a musicological analysis of the evolution of popular music.

Prior studies have explored the evolution of popular music through historical and cultural perspectives (Adorno and Simpson, 1941; Adorno and Rabinbach, 1975; Chapple and Garofalo, 1977; Garofalo and Waksman, 2016; Shuker, 2016), as well as through MIR techniques applied to audio (Askin and Mauskapf, 2017; Interiano et al., 2018; Lambert et al., 2020; Mauch et al., 2015; O’Toole and Horvát, 2023; Serrà et al., 2012), symbolic notation (Nakamura, 2023), and text (Brand et al., 2019; Oramas et al., 2016, 2018). Despite advancements in computational methods, the study of musical evolution using features extracted from audio or symbolic data faces notable limitations. For instance, MIDI data often suffer from issues related to coverage and quality, as a significant portion of available data is of low fidelity or incomplete, thereby limiting its reliability for detailed analyses. Conversely, audio‑based features are constrained by the performance of predictive supervised models, which are inherently dependent on the quality of training data and may struggle with accuracy in complex scenarios. Features derived from unsupervised models, while useful, often lack semantic specificity, making them less interpretable or relevant for nuanced studies. In contrast, human‑labeled data offer high‑quality annotations with exceptional semantic specificity, especially when curated by domain experts. However, the manual nature of this approach typically results in limited coverage, thereby restricting its scalability for large datasets or broad temporal analyses. The release of MGPHot addresses this gap by providing a large‑scale, systematically annotated dataset of high quality.

To illustrate the quality and potential of the MGPHot dataset annotations, we conduct a study on the evolution of musical attributes of popular music over the past 65 years, focusing on when and how changes occurred. Specifically, we identify three major music revolutions in 1964, 1983, and 2016, along with two minor revolutions in 1991 and 2007. Each of these revolutions is characterized by the evolution of distinct musical attributes, offering insights into the dynamic nature of popular music.

This study advances our understanding of the evolution of popular music from a musicological perspective, providing both a robust dataset and a framework for future research. Moreover, by making this dataset publicly available, we aim to foster innovation in music research and support the development of algorithms that leverage detailed musical features for MIR applications.

2 Dataset Description

2.1 On the Music Genome Project®

The Music Genome Project® is arguably the most comprehensive musicological analysis of music. Since the year 2000, a team of trained music analysts has carefully annotated millions of individual songs across all major genres through a detailed, structured, and quality‑controlled process.

The skills required to be part of this team of music analysts include the ability to transcribe harmonic progressions and identify melodic intervals by ear, and a mastery of Music Theory equivalent to a four‑year college degree. Music analysts also possess broad genre knowledge and a detailed understanding of musical cultural movements.

A typical music analysis involves an average of 15 minutes of careful listening while the music analyst annotates hundreds of predefined attributes per song. These attributes include details of harmony, melody, rhythm, instrumentation, vocal performance, genre and stylistic influence, lyrical subjects and sentiments, compositional focus, overall sonority, and production techniques.

Music analysts use custom software to make annotations efficient. Of the hundreds of attributes, not all need to be considered for every song. For example, instrumental music will exclude many dozens of attributes related to vocal qualities and lyrics. When instruments are absent, there is no need to consider multiple attributes used to describe the activity and tonal qualities of instruments. Due to this variability, the annotation time varies considerably. A New Age solo piano piece might take only a few minutes, while a heavily orchestrated and structurally complex Bollywood song might take 30 minutes.

Many annotations describing the presence of attributes make use of the ‘incidental to dominant’ scaling concept. For example, in the case of an instrument, if the score is zero, the instrument is absent. Low scores, on the ‘incidental’ side of the scale, indicate a minor role for that instrument. For example, the instrument is only present in part of the song or is barely noticeable in the mix. High scores, on the ‘dominant’ side of the scale, indicate a prominent or lead role which dominates the experience of the song. In most popular music, it is common for instruments to receive a middling score, indicating they are performing an expected supporting role for a dominant lead vocal.

Several quality control measures are used to ensure accuracy and alignment among analysts. Although the labor‑intensive annotation process makes it impractical to redundantly analyze every song, each analyst is subject to expert review of a percentage of their annotated songs by dedicated Quality Assurance analysts, and feedback is delivered on a monthly cadence. In addition, analysts meet regularly in genre‑aligned teams to discuss evolutions in music and how to maintain alignment as new popular music trends emerge.

2.2 The MGPHot dataset

MGPHot is derived from the Music Genome Project.® The song selection for MGPHot is based on a dataset collected from the Billboard Hot 100 chart data ranging from August of 1958 until September of 2022.2 The purpose of the Billboard Hot 100 is to provide the music industry a measure of a song’s popularity in the United States. As music consumption has evolved, so has the methodology of the Hot 100 chart. For example, in 2005, Billboard started incorporating digital song sales and in 2013, digital streams were considered, while in 2017 Billboard stopped considering the sales of physical singles. Users of this dataset should be aware that the methodology for compiling the Billboard Hot 100 has shifted over time, but its core objective of reflecting music popularity has remained consistent (Trust, 2023). In addition, the chart may reflect certain biases inherent to the structure and evolution of the music industry.

The artist names and track titles of the Billboard charts were mapped to an internal music catalog using an internal search API and some filtering heuristics. A common issue in this process is the presence of variations in how collaborators are represented, either in the track title or the artist name. To address this, we compared the results provided by the search API with the names listed in the charts, pre‑processing the strings and utilizing n‑grams and string similarity measures to evaluate the accuracy of the mappings. After applying these filtering techniques, we could successfully map a total of 21,320 tracks.

To evaluate the quality of the mapping, we conducted a manual verification of 300 randomly selected track mappings. The process revealed an accuracy rate of 94%. Among the 18 identified non‑exact mappings, only one represented an actual error, where the track was incorrectly mapped to the wrong artist. The remaining 17 were variations of the original song, including three live versions, nine remixes, one karaoke version, and four versions featuring additional artists. Considering only a single actual error, the effective accuracy rate increases to 99%.

The year assigned to each track in the dataset is its first year of appearance in the charts. In Figure 1, we show the number of tracks per year in the original Billboard Hot 100 dataset and those successfully mapped that make up the final dataset. Note that the coverage of mapped tracks increases with the years. Although the coverage is partial, especially for the early years of the charts, it is a large enough proportion (always higher than 50%).

Figure 1

Number of tracks per year in the Billboard Hot 100 charts and the subset successfully mapped. The X‑axis represents the chart year, and the Y‑axis indicates the number of tracks.

Song attributes were selected for full coverage of the dataset. The final list of attributes is grouped into seven different categories: rhythm, compositional focus, harmony, instrumentation, sonority, vocals, and lyrics. The complete list of attributes and their descriptions is reported in Table 2. Attributes are manually annotated on a 0‑to‑5 scale, which we translated to a 0‑to‑1 scale. Therefore, an attribute with a score towards 1 will imply high dominance of that attribute in the song, and a score towards 0 its absence. However, there is an exception in the case of the Major/Minor Key Tonality gene, where a score towards 0 indicates a predominance of a minor harmony and towards 1, a predominance of a major harmony.

Table 2

List of MGPHot attributes.

CategoryNameDescription
VocalsVocal RegisterDescribes the vocal range of lead vocal performance on a scale from low to high.
Vocal Timbre Thin to FullExpresses the timbre from thin and wispy to full and resonant.
Vocal BreathinessIndicates breathiness in the vocal delivery, characterized by airiness in the voice.
Vocal SmoothnessIndicates smoothness, reflecting the absence of roughness or raspiness.
Vocal GrittinessReflects the presence of roughness or raspiness in vocal delivery.
Vocal NasalityMeasures nasality, the pinched or ‘plugged‑up’ quality in vocal delivery.
Vocal AccompanimentIndicates the importance of non‑lead vocal accompaniment in a track.
HarmonyMinor/Major Key TonalityIndicates whether the tonality is minor, major, or ambiguous.
Harmonic SophisticationCaptures the complexity of harmony, from simple to complex chromatic notes.
RhythmTempoDescribes the song’s tempo and how other factors affect the perceived speed.
Cut Time FeelReflects the presence of a ‘cut time’ feel, where the rhythm is felt in half‑time.
Triple MeterIndicates the presence of a triple meter, such as 3/4 time.
Compound MeterIndicates the presence of compound meter, combining triple and duple rhythms.
Odd MeterReflects the presence of odd meters, such as 5 or 7 beats per measure.
Swing FeelMeasures swing feel, where the first 8th note is longer than the second.
Shuffle FeelSimilar to swing feel, but with more pronounced articulation of each note.
Syncopation Low to HighIndicates syncopation, where rhythm emphasizes offbeats or anticipations.
BackbeatMeasures the dominance of a backbeat rhythm, with emphasis on beats 2 and 4.
DanceabilityRates how suitable the song is for dancing, from low to high.
InstrumentationDrum SetIndicates the presence and dominance of a drum set in the song.
Drum AggressivenessReflects the aggressiveness of the drum set performance.
Synthetic DrumsIndicates the presence of synthetic drums, often programmed.
PercussionReflects the dominance of percussion in the song, excluding drums.
Electric GuitarIndicates the presence and dominance of electric guitar(s).
Electric Guitar DistortionMeasures the degree of guitar distortion, from clean to ‘dirty’.
Acoustic GuitarIndicates the presence of acoustic guitar(s).
String EnsembleReflects the presence and dominance of a string ensemble in the song.
Horn EnsembleIndicates the presence of a horn ensemble, from small to large.
PianoIndicates the presence of a piano in the song.
OrganReflects the presence of an organ in the instrumentation.
RhodesIndicates the presence of a Fender Rhodes or other electric piano.
SynthesizerReflects the presence of synthesizers in the instrumentation.
Synth TimbreDescribes synthesizer timbres, from ambient to robotic or industrial.
Bass GuitarReflects the presence and dominance of a bass guitar.
Reed InstrumentReflects the presence of reed instruments like saxophones or clarinets.
LyricsAngry LyricsMeasures the presence and dominance of angry lyrics in the song.
Sad LyricsMeasures the presence and dominance of sad lyrics.
Happy/Joyful LyricsReflects the presence of happy or joyful lyrics.
Humorous LyricsIndicates the presence of humorous or funny lyrics.
Love/Romance LyricsMeasures the presence of romantic or love‑themed lyrics.
Social/Political LyricsIndicates the presence of lyrics about social or political issues.
Abstract LyricsMeasures the presence of abstract or whimsical lyrics.
Explicit LyricsMeasures the explicitness of lyrics, from clean to very explicit.
SonorityLive RecordingIndicates whether the song was recorded live or in a studio.
Audio ProductionMeasures the quality of the audio production, from poor to excellent.
Aural IntensityMeasures the song’s overall loudness or softness.
Acoustic SonorityIndicates the presence of acoustic instruments or voices.
Electric SonorityMeasures the presence of electric instruments.
Synthetic SonorityReflects the presence of synthetic instruments like synthesizers.
CompositionFocus on Lead VocalReflects the importance of lead vocals to the overall track.
Focus on LyricsMeasures the importance of lyrics in the overall track.
Focus on MelodyIndicates the importance of melody in the track.
Focus on Vocal AccompanimentReflects the importance of backing vocals in the track.
Focus on Rhythmic GrooveIndicates how important the rhythmic groove is to the track.
Focus on Musical ArrangementsReflects the importance of the arrangement and orchestration.
Focus on FormMeasures the importance of the song’s form or structure.
Focus on RiffsReflects the importance of instrumental riffs in the track.
Focus on PerformanceMeasures the importance of instrumental performance in the track.

The 58 attributes fall in a spectrum from subjective to objective. For example, measuring the beats per minute, or the presence of a certain rhythmic meter can be done with objective precision, while the level of ‘danceability,’ or the compositional importance of melody, will always be colored by each individual’s taste and experience with music. Although it is impossible to reduce a work of art into a set of purely objective measures, we have worked hard to codify each attribute through extensive training, rigorous QA processes, and detailed documentation including many musical examples, ensuring the best possible alignment between annotators. For instance, when considering the ‘explicitness’ of lyrics, we have developed and systematically applied an internal standard that evaluates the presence of certain explicit words and thematic content, grounded in a contemporary Western context.

3 Applications of MGPHot in MIR

The release of the MGPHot dataset opens up a wide range of potential applications in MIR and in musicology, offering a rich, high‑dimensional set of annotations that go beyond traditional metadata. With its detailed coverage of 58 musical attributes across 21,320 tracks from the Billboard Hot 100 charts, MGPHot can significantly enhance various MIR tasks by providing deeper insights into the structure, composition, and characteristics of popular music. Below, we outline some— non‑exhaustive—applications where MGPHot can make a substantial impact.

3.1 Auto‑Tagging

One of the most immediate applications of the MGPHot dataset is the development of automated tagging systems. Traditional auto‑tagging models are typically trained to classify songs into broad categories such as genre or mood, utilizing datasets like the Ballroom dataset (Gouyon et al., 2006), The Million Song Dataset (Bertin‑Mahieux et al., 2011), the FMA dataset (Defferrard et al., 2017b), or the MTG‑Jamendo dataset (Bogdanov et al., 2019). However, MGPHot offers a more comprehensive set of musical attributes, including rhythm, instrumentation, harmony, vocals, compositional focus, and lyrics, some of which are unique and not found in other datasets. Additionally, MGPHot provides detailed scores that quantify the strength of each attribute. These characteristics make MGPHot a valuable resource for developing and evaluating more sophisticated auto‑tagging models, enabling the automatic generation of tags that more accurately reflect human‑curated annotations.

3.2 Music recommendation

MGPHot provides an opportunity to improve music recommendation algorithms by incorporating multidimensional representations of songs. Many existing recommendation systems rely on collaborative filtering and/or audio features, which may not capture the full complexity of a track’s musical attributes. By integrating the 58 musicological attributes from MGPHot, recommendation systems can offer more personalized and contextually relevant suggestions. For instance, a user’s listening preferences can be matched with specific rhythmic patterns, compositional structures, or lyrical themes, leading to more accurate and satisfying recommendations.

3.3 Chart prediction

Given that the MGPHot dataset focuses on songs that appeared in the Billboard Hot 100 between 1958 and 2022, it offers a unique resource for chart prediction models. These models aim to predict whether a song will become a hit based on its musical features. With MGPHot’s extensive annotations, chart prediction can move beyond basic audio features and consider deeper attributes such as rhythmic complexity, lyrical themes, and instrumental choices. This allows for a more refined analysis of what makes a song popular, providing researchers with better tools for understanding and forecasting musical success.

3.4 Musicological and cultural studies

The detailed time‑based coverage of MGPHot makes it a useful resource for studying the temporal evolution of popular music and conducting broader musicological and cultural analyses. Researchers can use the dataset to explore how specific musical attributes, such as rhythm or harmony, have shifted over the years and correlate these changes with cultural and societal trends. In addition to temporal analysis, MGPHot rich annotations may be useful for musicological and cultural studies. Researchers can investigate how specific musical genres or attributes have driven the development of new styles, while also exploring the cultural impact of different musical trends. The dataset allows for the examination of how shifts in musical preferences reflect broader societal changes, such as through the prominence of certain lyrical themes or compositional styles in relation to historical events. Overall, we hope this dataset serves as a valuable resource for musicological studies and aids in exploring the evolution and cultural significance of popular music.

4 Case Study: The Evolution of Popular Music

To demonstrate the potential of the MGPHot dataset, we conduct an in‑depth case study focusing on the evolution of popular music from 1958 to 2022. By leveraging the 58 musical attributes in MGPHot, we aim to identify key stylistic shifts and revolutions in popular music and explore how various musical elements, such as rhythm, harmony, and instrumentation, have evolved over time. This case study not only reaffirms findings from previous research but also provides unique new insights into the underlying musical attributes driving these changes.

4.1 Visualizing trend curves

In our quest to comprehend the evolution of popular music, we aim to compare the songs of each year with those of all other years. Each song within our dataset is represented as a 58‑dimensional vector of musical attributes, ranging from 0 to 1, forming what we refer to as the MGPHot track embedding. By leveraging these embeddings, we compute the centroid of all tracks from a given year, thereby generating an embedding that captures the essence of that particular year, referred to as the MGPHot year embedding.

To analyze the evolution of each of the musical attributes present in the MGPHot year embedding, we calculate a smoothed version of the values for each dimension across the years by applying 1‑D convolutions with a Hanning window of 10, following the methodology in (Oramas et al., 2018), and then visualize the trend curves of these musical attributes grouped by category (see Figure 2). Vertical lines in the plots indicate the years of faster change in the dataset (or revolutions), as explained in Section 4.2.

Figure 2

Trend curve of all the attributes by category across the years.

This visualization provides a comprehensive overview of the attributes in the dataset and highlights the evolution of musical characteristics over time. Several insights are immediately apparent from these plots. Notable trends include a shift towards minor harmonies, a transition in sonority from acoustic to electric and later to synthetic sounds, a reduction in the compositional emphasis on riffs and form, an increase in breathiness and nasality in vocals, and a growing prevalence of explicit lyrics. Additionally, we can observe sudden changes at specific points in time, such as the rise of synthetic sonorities, synthesizers, and synthetic drums in the early 1980s, the surge in explicit lyrics in the mid‑2010s, and the end of the prominence of swing feel in the mid‑1960s. While we highlight only a few insights, a deeper analysis of these plots can yield many more detailed conclusions. To facilitate this, we provide an Appendix as supplementary material with all the plots of individual attributes, including the pre‑smoothed mean values and the standard deviation of each year.

4.2 Identifying music revolutions

As observed in the previous plots, some musical changes happen gradually, while others occur more abruptly. These rapid shifts contribute to the fragmented perception of music history, dividing it into distinct eras. To pinpoint the years where these discontinuities occur, we applied the methodology used in Mauch et al. (2015), which analyzed music evolution using descriptors extracted from audio. Our goal was to determine if using musicological annotations would lead us to similar conclusions, while also allowing for a deeper exploration of the specific attributes driving these changes.

Following the aforementioned methodology, we first measure the similarity of the music within years, calculating the pairwise cosine distance among all MGPHot year embeddings, and visualizing the results through a heatmap (see Figure 3). This visualization enables us to visually identify distinct regions—or eras—of popular music. Then, to better identify the inflection points, we employ the Foote novelty metric (Foote, 2000) with a kernel size of 10 years. We selected this kernel size to be consistent with the window size used to compute the 1‑D convolutions of the trend curves. This metric aims to measure local changes by correlating a small checkerboard‑like kernel along the main diagonal of the similarity matrix. This results in a novelty function that unveils peaks at specific years where the kernel meets a transition between two contrasting blocks (see Figure 4).

Figure 3

Similarity matrix of MGPHot year embeddings. Darker means more similar, lighter means less similar.

Figure 4

Foote novelty metric for the 58 attributes.

Applying this function, we identify three major peaks in years 1964, 1983, and 2016, and two minor peaks in 1991 and 2007. This result agrees with the prior study on the evolution of music between 1960 and 2010 based on audio descriptors (Mauch et al., 2015), which identified three music revolutions in 1964, 1983, and 1991. Our results not only confirm these three revolutions but also reveal two additional ones in 2007 and 2016.

4.3 Breakdown by category

To understand the factors contributing to the identified peaks, we conduct a similar analysis on various subsets of dimensions within the MGPHot year embeddings. Since each dimension corresponds to a distinct musical attribute, and these attributes are categorized accordingly, we generate a year embedding for each category. This approach yields distinct year embeddings representing rhythm, instrumentation, sonority, compositional focus, harmony, vocals, and lyrics attributes. Subsequently, we iterate the process of computing pairwise cosine distance between all year embeddings within each category, followed by plotting the Foote novelty metric (refer to Figure 5).

Figure 5

Foote novelty metric for the attributes in each category.

These plots reveal when revolutions occurred for different musical characteristics, allowing us to identify the key drivers behind each revolution. For instance, rapid changes in rhythm occurred in 1964, while instrumentation and sonority shifted significantly in 1983. In contrast, compositional focus, harmony, vocals, and lyrics saw their most substantial changes in 2016. The peaks observed in these isolated categories align with the revolutionary years identified in the previous section, enabling us to associate specific musical changes with those major turning points.

To further explore the relationships between musical attributes across different categories, we complement the novelty metric with a Mantel test (Mantel and Valand, 1970). This test, commonly used in ecology to measure the linear correlation and statistical significance between two proximity matrices, is applied here to compare the distance matrices across different musical categories. Instead of comparing species, we compare the distances between years, using the distance matrices calculated for each category. A heatmap is then generated to illustrate the Pearson correlations among these categorical distance matrices (see Figure 6). The statistical significance of the results is determined via a two‑tailed test with 10,000 permutations, yielding a p‑value < 0.05 for all comparisons.

Figure 6

Pearson correlation between categorical distances according to the Mantel test, having all p < 0.05.

The analysis reveals strong correlations between lyrics, composition, and vocals, as well as between instrumentation and sonority, which is consistent with the way these attributes changed together during the identified musical revolutions. Rhythm, on the other hand, shows the weakest correlations, with only moderate associations to sonority and instrumentation.

4.4 A closer look into revolutions

Through our analysis, we identified key moments in history when rapid changes occurred in music. In this section, we propose several hypotheses to explain these musical revolutions, drawing on the information available in the dataset. Note, however, that these hypotheses would require further investigation, e.g., through interdisciplinary studies and causal modeling to validate the underlying causes of these transformative periods.

4.4.1 First revolution in 1964: Rhythm

The revolution of 1964 corresponds to the emergence of a new era of rock music. As we see in Figure 2b, the rhythmic changes involved a significant decline in the use of swung 8th notes and compound meters (e.g. the 12/8 meter so common in earlier soul and pop music), in favor of a straight 8th note, 4/4 rhythmic feel that would set the template for decades of rock music to come. Prime examples of this trend can be found on the 1964 Hot 100, including songs like ‘You Really Got Me’ by The Kinks, ‘Twist And Shout’ by The Beatles, and ‘Louie Louie’ by The Kingsmen, all of which feature a straight 8th rhythmic feel in 4/4 meter.

4.4.2 Second revolution in 1983: Sonority and instrumentation

The music revolution of 1983 centered around a tipping point for synth‑driven pop music. Although the synthesizer had been fairly common in pop and rock music throughout the 1970s, it was not until the early 80s that it became a dominant sound on the charts, displacing the guitar‑heavy sound of 1970s classic rock. Songs such as ‘Sweet Dreams’ by the Eurythmics, ‘Electric Avenue’ by Eddy Grant, and ‘She Blinded Me With Science’ by Thomas Dolby are just a few examples that demonstrate this shift towards synthetic sonorities and a more dominant use of synthesizers.

We observe in Figures 5b and 5c important peaks around 1983, implying major shifts in sonority and instrumentation. Looking more closely at the sonority attributes present in Figure 2e, we see a clear increase in synthetic sonority that reaches its peak slope around this year, and a corresponding decrease of acoustic and electric sonorities. Regarding instrumentation, we observe in Figures 2g and 2i the maximum slope in the increase of synthetic drums and synthesizers, as well as a shift in the types of synth timbres, moving away from more ambient sounds towards more aggressive, techno‑inspired sounds. This trend began some years earlier and ended a couple of years after this peak of change. We can observe in Figure 2h just the opposite for acoustic guitars and in Figure 2i for string ensembles and piano. From these observations, we can infer that peaks of the Foote novelty metric are associated with steeper slopes in the curves showing the evolution of musicological features.

4.4.3 Third revolution in 2016: Vocals, harmony, and lyrics

We observe in Figures 5d–5g that this revolution is mainly driven by changes in harmony, vocals, and lyrics, as well as compositional focus. We see in Figure 2a that the composition focus of songs was shifting away from instrumental riffs and elaborated song forms, beginning in 2007. This tendency continued in 2016, accompanied by a loss of focus on melody. At the same time, vocal timbres became more nasal, while smooth and gritty vocal timbres became less common. Looking at harmony, we see an acceleration of an existing trend away from major key tonalities towards minor keys (see Figure 2d). By 2020, for the first time in history, minor keys became more prevalent on the charts than major keys, with the highest peak in the novelty curve (see Figure 5e) occurring in 2016 during this revolution. Regarding lyrics, we see in Figure 2e a tremendous increase of explicit language and subject matter (see Figure 2c). In terms of sonority, we see a renewed emphasis on synthetic sonorities.

All of the enumerated characteristics observed in the data coincide with the rise of modern hip‑hop as the dominant genre. Our analysis suggests that at this point, the influence of hip‑hop became one of the most dominant forces in modern music, achieving mainstream status from its once underground level of cultural influence. The year 2016 saw the rapper Drake holding eight songs in the Billboard Top 100 charts.

As highlighted in prior research (Oliveira et al., 2020), there has been a substantial uptick in artist collaborations from diverse genres in the last 10 years, contributing to the widespread dissemination of hip‑hop’s stylistic attributes and sonic characteristics to other genres such as pop or even rock.

4.4.4 Minor revolutions of 1991 and 2007

The factors behind the revolution we observe in 1991 are a little murkier compared to the revolutions of 1964 and 1983. We observe minor peaks for instrumentation, vocals, compositional focus, and lyrics in the novelty plots (see Figure 5). We observe in Figures 2h and 2i a shift in instrumentation, in favor of acoustic guitars over electric guitars or synthetizers. Compositional interest in form and melody decreased, while vocal accompaniment became more of a focus (see Figure 2a).

The revolution around 2007 seems to be related to changes in compositional focus, small changes in vocals and lyrics (see Figure 5), the rise of happy lyrics (see Figure 2c), and a short‑term reversal in the longer term trend away from major harmonies (see Figure 2d).

5 Conclusions

With this paper, we release MGPHot, a dataset of music annotations of unprecedented quality, derived from the Music Genome Project.® It covers songs that have appeared at least once in the Billboard Hot 100 charts from 1958 until 2022. We hope that this release represents a significant milestone for MIR research. MGPHot offers a rich, high‑dimensional set of human‑curated annotations, which provide deep insights into musical structure, lyrical content, and technical characteristics, enabling more nuanced approaches to tasks such as music classification and recommendation, as well as facilitating unique musicological studies.

To illustrate the usefulness of MGPHot, we conducted a data‑driven musicological analysis of popular music evolution from 1958 to 2022. We corroborated previous research identifying distinct musical eras, showing significant shifts in musical trends. And we provided unique insights into the specific music characteristics that underwent changes during each stylistic revolution.

We believe the MGPHot dataset holds immense value for further research, and this paper represents only a fraction of its potential utility. Moving forward, we envision several avenues for future work, such as, e.g., genre‑specific analysis, interdisciplinary studies on cultural influences, and exploring patterns for predicting future hits.

Acknowledgments

A huge thank you to all the Music Analysts who have contributed to this massive undertaking over the years.

Competing Interests

The authors have no competing interests to declare.

Notes

Additional File

The additional file for this article can be found using the links below:

Supplementary Appendix
DOI: https://doi.org/10.5334/tismir.236 | Journal eISSN: 2514-3298
Language: English
Submitted on: Nov 6, 2024
Accepted on: Mar 22, 2025
Published on: May 28, 2025
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2025 Sergio Oramas, Fabien Gouyon, Steve Hogan, Camilo Landau, Andreas Ehmann, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.