Interfacing Sounds: Hierarchical Audio-Content Morphologies for Creative Re-purposing in earGram 2.0
Gilberto Bernardes, and Gilberto Bernardes
Proceedings of the International Conference on New Interfaces for Musical Expression
- Year: 2020
- Location: Birmingham, UK
- Pages: 537–542
- DOI: 10.5281/zenodo.4813176 (Link to paper)
- PDF link
- Presentation Video
Abstract:
Audio content-based processing has become a pervasive methodology for techno-fluent musicians. System architectures typically create thumbnail audio descriptions, based on signal processing methods, to visualize, retrieve and transform musical audio efficiently. Towards enhanced usability of these descriptor-based frameworks for the music community, the paper advances a minimal content-based audio description scheme, rooted on primary musical notation attributes at the threefold sound object, meso and macro hierarchies. Multiple perceptually-guided viewpoints from rhythmic, harmonic, timbral and dynamic attributes define a discrete and finite alphabet with minimal formal and subjective assumptions using unsupervised and user-guided methods. The Factor Oracle automaton is then adopted to model and visualize temporal morphology. The generative musical applications enabled by the descriptor-based framework at multiple structural hierarchies are discussed.
Citation:
Gilberto Bernardes, and Gilberto Bernardes. 2020. Interfacing Sounds: Hierarchical Audio-Content Morphologies for Creative Re-purposing in earGram 2.0. Proceedings of the International Conference on New Interfaces for Musical Expression. DOI: 10.5281/zenodo.4813176BibTeX Entry:
@inproceedings{NIME20_103, abstract = {Audio content-based processing has become a pervasive methodology for techno-fluent musicians. System architectures typically create thumbnail audio descriptions, based on signal processing methods, to visualize, retrieve and transform musical audio efficiently. Towards enhanced usability of these descriptor-based frameworks for the music community, the paper advances a minimal content-based audio description scheme, rooted on primary musical notation attributes at the threefold sound object, meso and macro hierarchies. Multiple perceptually-guided viewpoints from rhythmic, harmonic, timbral and dynamic attributes define a discrete and finite alphabet with minimal formal and subjective assumptions using unsupervised and user-guided methods. The Factor Oracle automaton is then adopted to model and visualize temporal morphology. The generative musical applications enabled by the descriptor-based framework at multiple structural hierarchies are discussed.}, address = {Birmingham, UK}, author = {Bernardes, Gilberto and Bernardes, Gilberto}, booktitle = {Proceedings of the International Conference on New Interfaces for Musical Expression}, doi = {10.5281/zenodo.4813176}, editor = {Romain Michon and Franziska Schroeder}, issn = {2220-4806}, month = {July}, pages = {537--542}, presentation-video = {https://youtu.be/zEg9Cpir8zA}, publisher = {Birmingham City University}, title = {Interfacing Sounds: Hierarchical Audio-Content Morphologies for Creative Re-purposing in earGram 2.0}, url = {https://www.nime.org/proceedings/2020/nime2020_paper103.pdf}, year = {2020} }