PapersFlow Research Brief
Music Technology and Sound Studies
Research Guide
What is Music Technology and Sound Studies?
Music Technology and Sound Studies is a field that develops and applies interactive evolutionary music systems and instruments, combining human evaluation with evolutionary computation optimization, encompassing music generation, digital musical instruments, gesture recognition, machine learning, sound synthesis, and the intersection of art and technology in musical performance.
The field includes 148,265 works focused on topics such as interactive evolutionary computation, music generation, human-computer interaction, digital musical instruments, gesture recognition, machine learning, sound synthesis, musical performance, artificial intelligence, and acoustic ecology. "WaveNet: A Generative Model for Raw Audio" by van den Oord et al. (2016) introduced a deep neural network for generating raw audio waveforms, achieving 3565 citations. "librosa: Audio and Music Signal Analysis in Python" by McFee et al. (2015) provides Python implementations for music information retrieval functions, with 2755 citations.
Topic Hierarchy
Research Sub-Topics
Interactive Evolutionary Computation in Music
This sub-topic examines evolutionary algorithms optimized through human feedback for music generation and composition. Researchers study population-based search methods integrated with user evaluations to evolve musical structures.
Digital Musical Instruments
This area focuses on the design, implementation, and evaluation of novel digital interfaces for musical expression. Researchers investigate mapping strategies, real-time performance, and user interaction paradigms.
Gesture Recognition for Musical Interfaces
Researchers develop computer vision and sensor-based systems to interpret performer gestures for controlling sound. Studies cover feature extraction, machine learning classification, and real-time responsiveness in live settings.
Sound Synthesis Techniques
This sub-topic explores algorithmic methods for generating audio waveforms, including physical modeling, granular synthesis, and neural approaches. Researchers analyze perceptual quality, computational efficiency, and timbral control.
Music Information Retrieval
Focuses on algorithms for audio analysis, including genre classification, onset detection, and audio feature extraction using tools like librosa. Researchers develop machine learning models for content-based music search and recommendation.
Why It Matters
Music Technology and Sound Studies enables practical tools for audio processing and generation used in research and industry. Mirelo, a Berlin startup, raised $41 million in seed funding to generate sound effects for videos using AI. AudioShake secured $14 million in Series A funding to enhance sound usability in media applications. "librosa: Audio and Music Signal Analysis in Python" by McFee et al. (2015) supports music information retrieval in Python, cited 2755 times for signal processing tasks. "WaveNet: A Generative Model for Raw Audio" by van den Oord et al. (2016) generates raw audio waveforms, applied in speech synthesis and music production with 3565 citations.
Reading Guide
Where to Start
"librosa: Audio and Music Signal Analysis in Python" by McFee et al. (2015) is the starting point for beginners, as it offers practical Python tools for audio and music signal processing essential for hands-on analysis in music information retrieval.
Key Papers Explained
"librosa: Audio and Music Signal Analysis in Python" by McFee et al. (2015) provides feature extraction foundations that support classification methods in "Musical genre classification of audio signals" by Tzanetakis and Cook (2002). "WaveNet: A Generative Model for Raw Audio" by van den Oord et al. (2016) builds on signal processing by generating raw waveforms, extending analysis to synthesis. openSMILE by Eyben et al. (2010) complements these with unified feature extraction from speech and music domains.
Paper Timeline
Most-cited paper highlighted in red. Papers ordered chronologically.
Advanced Directions
Recent preprints highlight IRCAM's work on sound analysis/synthesis, physical models, and computer-aided composition. "Science and Technology of Music and Sound: The IRCAM Roadmap" outlines links between signal and symbolic music levels. Centers like the Center for Computer Research in Music and Acoustics offer seminars on computational models of sound perception and audio signal processing.
Papers at a Glance
| # | Paper | Year | Venue | Citations | Open Access |
|---|---|---|---|---|---|
| 1 | Abstract Harmonic Analysis. | 1966 | American Mathematical ... | 3.8K | ✕ |
| 2 | WaveNet: A Generative Model for Raw Audio | 2016 | arXiv (Cornell Univers... | 3.6K | ✓ |
| 3 | <i>The Theory of Sound</i> | 1957 | Physics Today | 3.5K | ✕ |
| 4 | librosa: Audio and Music Signal Analysis in Python | 2015 | Proceedings of the Pyt... | 2.8K | ✓ |
| 5 | Musical genre classification of audio signals | 2002 | IEEE Transactions on S... | 2.7K | ✕ |
| 6 | Opensmile | 2010 | — | 2.5K | ✓ |
| 7 | A simple model of feedback oscillator noise spectrum | 1966 | Proceedings of the IEEE | 2.4K | ✕ |
| 8 | Acoustics: An Introduction to Its Physical Principles and Appl... | 1984 | Journal of vibration a... | 2.3K | ✓ |
| 9 | Emotion and Meaning in Music | 1961 | — | 2.0K | ✕ |
| 10 | The Journal of the Acoustical Society of America | 1939 | The Journal of the Aco... | 1.6K | ✕ |
In the News
Musicians-turned-AI founders at Mirelo land $41M to make ...
# Musicians-turned-AI founders at Mirelo land $41M to make sound the next frontier of generative media bySofia Chesnokova December 15, 2025 **2 minute read Image credits: Mirelo AI LinkedIn Total...
Berlin startup Mirelo raises $41m in seed funding for AI- ...
Mirelo , a Berlin-based artificial intelligence company that automatically generates sound effects for videos, has secured $41 million in seed funding.
AudioShake Raises $14M to Make Sound More Usable ...
SAN FRANCISCO,Oct. 1, 2025/PRNewswire/ -- San Francisco startup AudioShake announced today it has raised $14 million in Series A funding. The round was led by Shine Capital, with participation from...
GRAMMY MUSEUM® GRANT PROGRAM AWARDS ...
Generously funded by the Recording Academy, the GRAMMY Museum Grant Program provides funding annually to organizations and individuals to support efforts that advance the archiving and preservatio...
Press Releases
Jan 14, 2025 National Endowment for the Arts Supports the Arts with Nearly $36.8 Million in Funding Nationwide
Code & Tools
Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokeniz...
We introduce ACE-Step, a novel open-source foundation model for music generation that overcomes key limitations of existing approaches and achieves...
Amphion (/æmˈfaɪən/) is a toolkit for Audio, Music, and Speech Generation. Its purpose is to support reproducible research and help junior research...
Tone.js is a Web Audio framework for creating interactive music in the browser. The architecture of Tone.js aims to be familiar to both musicians a...
Back To Top ↥ ## About Python library for audio and music analysis librosa.org/ ### Topics audio python music dsp scipy librosa ### Resources
Recent Preprints
Science and Technology of Music and Sound: The IRCAM Roadmap
levels of musical information. The addressed subjects include sound analysis/synthesis, physical models, sound spatialisation, computer-aided composition, and interdisciplinary transversal themes c...
IRCAM
IRCAM is an internationally recognized research center dedicated to creating new technologies for music. The institute offers a unique experimental environment where composers strive to enlarge the...
Center for Computer Research in Music and Acoustics
** Music 256A **Music, Computing, and Design I: Software Paradigms for Computer Music ** Music 319 **Research Seminar on Computational Models of Sound Perception ** Music 320 **Introduction to Audi...
Digital Technology and the Study of Music
technology in the study of music. Firstly as a tool, secondly as an instrument and lastly as a medium for thinking. As our societies become increasingly embroiled in digital media for representatio...
What you listen to makes a difference: The impact of music on ...
Sounds constantly surround us, serving as sensory cues that help humans interpret the world and navigate the flood of stimuli they encounter. Research has shown that sounds and music can influence ...
Latest Developments
Recent developments in music technology and sound studies research as of February 2026 include advancements in AI-driven music creation, with Spotify partnering with major labels to develop generative AI tools (billboard.com), and ongoing exploration of AI's role in co-creativity, instrument design, and performance practices (frontiersin.org, frontiersin.org). Additionally, key trends include the integration of brain–computer interfaces, quantum computing, and ethical considerations in AI-enhanced music (namm.org, imusician.pro).
Sources
Frequently Asked Questions
What is WaveNet?
WaveNet is a deep neural network for generating raw audio waveforms introduced by van den Oord et al. (2016). The model is fully probabilistic and autoregressive, conditioning each audio sample on all previous ones. It generates high-fidelity audio for applications like speech and music synthesis.
How does librosa support music analysis?
librosa is a Python package for audio and music signal processing by McFee et al. (2015). It implements functions for music information retrieval, including feature extraction. Version 0.4.0 provides tools like chroma and spectral analysis.
What methods are used for musical genre classification?
Musical genre classification of audio signals by Tzanetakis and Cook (2002) characterizes genres by instrumentation, rhythmic structure, and harmonic content. The approach uses categorical labels created by humans. It applies signal processing techniques to audio features.
What features does openSMILE extract?
openSMILE by Eyben et al. (2010) extracts audio low-level descriptors like CHROMA, CENS, loudness, Mel-frequency cepstral coefficients, and perceptual linear prediction. It unites algorithms from speech processing and music information retrieval. The toolkit supports feature extraction for analysis tasks.
What is the focus of IRCAM?
IRCAM is a research center for creating new technologies for music, as described in recent preprints. It addresses sound analysis/synthesis, physical models, sound spatialisation, and computer-aided composition. The institute provides an experimental environment for composers.
Open Research Questions
- ? How can evolutionary computation optimize interactive music systems while incorporating real-time human feedback?
- ? What architectures improve autoregressive generation of raw audio waveforms beyond WaveNet?
- ? How do gesture recognition techniques enhance control of digital musical instruments?
- ? Which machine learning models best integrate sound synthesis with musical performance?
- ? How does acoustic ecology inform AI-driven music generation?
Recent Trends
Mirelo raised $41 million in December 2025 for AI-generated sound effects in videos.
AudioShake secured $14 million in October 2025 to improve sound usability.
Preprints from IRCAM (January 2026) emphasize new technologies for music, including sound spatialisation.
The field maintains 148,265 works with tools like Audiocraft and Amphion advancing audio generation.
Research Music Technology and Sound Studies with AI
PapersFlow provides specialized AI tools for Computer Science researchers. Here are the most relevant for this topic:
AI Literature Review
Automate paper discovery and synthesis across 474M+ papers
Code & Data Discovery
Find datasets, code repositories, and computational tools
Deep Research Reports
Multi-source evidence synthesis with counter-evidence
AI Academic Writing
Write research papers with AI assistance and LaTeX support
See how researchers in Computer Science & AI use PapersFlow
Field-specific workflows, example queries, and use cases.
Start Researching Music Technology and Sound Studies with AI
Search 474M+ papers, run AI-powered literature reviews, and write with integrated citations — all in one workspace.
See how PapersFlow works for Computer Science researchers