Skip to main content
Have a personal or library account? Click to login
Music Information Retrieval and Contemporary Classical Music: A Successful Failure Cover

Music Information Retrieval and Contemporary Classical Music: A Successful Failure

Open Access
|Sep 2020

References

  1. Andén, J., & Mallat, S. (2014). Deep scattering spectrum. IEEE Transactions on Signal Processing, 62(16), 41144128. DOI: 10.1109/TSP.2014.2326991
  2. Andersen, K., & Knees, P. (2016). Conversations with expert users in music retrieval and research challenges for creative MIR. In Proceedings of the International Society for Music Information Retrieval Conference (ISMIR), pages 122128.
  3. Assayag, G., Rueda, C., Laurson, M., Agon, C., & Delerue, O. (1999). Computer-assisted composition at IRCAM: From PatchWork to OpenMusic. Computer Music Journal, 23(3), 5972. DOI: 10.1162/014892699559896
  4. Boulanger, R. C., editor. (2000). The Csound Book: Perspectives in Software Synthesis, Sound Design, Signal Processing, and Programming. MIT Press.
  5. Briot, J.-P., Hadjeres, G., & Pachet, F. (2019). Deep Learning Techniques for Music Generation. Computational Synthesis and Creative Systems Series. Springer Verlag. DOI: 10.1007/978-3-319-70163-9
  6. Bruna, J., & Mallat, S. (2013). Invariant scattering convolution networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8), 18721886. DOI: 10.1109/TPAMI.2012.230
  7. Burred, J. J., Cella, C.-E., Peeters, G., Roebel, A., & Schwarz, D. (2008). Using the SDIF sound description interchange format for audio features. In Proceedings of the International Conference on Music Information Retrieval (ISMIR), pages 427432.
  8. Cardoso, A., Veale, T., & Wiggins, G. A. (2009). Converging on the divergent: The history (and future) of the International Joint Workshops in Computational Creativity. AI Magazine, 30(3), 1522. DOI: 10.1609/aimag.v30i3.2252
  9. Carpentier, G., Tardieu, D., Assayag, G., & Saint-James, E. (2007). An evolutionary approach to computer-aided orchestration. In M. Giacobini, editors, Applications of Evolutionary Computing: EvoWorkshops 2007, volume 4448, pages 488497. Springer. DOI: 10.1007/978-3-540-71805-5_54
  10. Carpentier, G., Tardieu, D., Harvey, J., Assayag, G., & Saint-James, E. (2010). Predicting timbre features of instrument sound combinations: Application to automatic orchestration. Journal of New Music Research, 39(1), 4761. DOI: 10.1080/09298210903581566
  11. Cella, C.-E. (2011a). On symbolic representations of music. PhD thesis, University of Bologna.
  12. Cella, C.-E. (2011b). Sound-types: A new framework for symbolic sound analysis and synthesis. In Proceedings of the International Computer Music Conference (ICMC), pages 179184.
  13. Cella, C.-E. (2017). Machine listening intelligence. In Proceedings of the International Workshop on Deep Learning for Music, pages 5055.
  14. Cella, C.-E., & Burred, J. J. (2013). Advanced sound hybridizations by means of the theory of soundtypes. In Proceedings of the International Computer Music Conference (ICMC), pages 3946.
  15. Cella, C.-E., & Esling, P. (2018). Open-source modular toolbox for computer-aided orchestration. In Proceedings of Timbre 2018: Timbre is a Many-Splendored Thing, pages 9394.
  16. Choi, K., Fazekas, G., Sandler, M., & Cho, K. (2017). Convolutional recurrent neural networks for music classification. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 23922396. DOI: 10.1109/ICASSP.2017.7952585
  17. Crayencour, H.-C., & Cella, C.-E. (2019). Learning, probability and logic: Toward a unified approach for content-based music information retrieval. Frontiers in Digital Humanities, 6(6). DOI: 10.3389/fdigh.2019.00006
  18. Davis, S. B., & Mermelstein, P. (1980). Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences. IEEE Transactions on Acoustics, Speech, and Signal Processing, 28(4), 357366. DOI: 10.1109/TASSP.1980.1163420
  19. Deserno, S. (2015). Algorithmic composition: An overview of the field, inspired by a criticism of its methods. Seminar topics in computer music, RWTH Aachen University.
  20. Dieleman, S., & Schrauwen, B. (2014). End-to-end learning for music audio. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 69646968. DOI: 10.1109/ICASSP.2014.6854950
  21. Donin, N., & Feneyrou, L., editors. (2017). Théories de la composition musicale au xxe siècle. Symétrie.
  22. Fernández, J. D., & Vico, F. (2013). AI methods in algorithmic composition: A comprehensive survey. Journal of Artificial Intelligence Research, 48(1), 513582. DOI: 10.1613/jair.3908
  23. Gabrielli, L., Cella, C.-E., Vesperini, F., Droghini, D., Principi, E., & Squartini, S. (2018). Deep learning for timbre modification and transfer: An evaluation study. In Proceedings of the Audio Engineering Society (AES) Convention 144.
  24. Ghisi, D. (2017). Music Across Music: Towards a Corpus-Based, Interactive Computer-Aided Composition. PhD thesis, IRCAM.
  25. Gillick, J., Cella, C.-E., & Bamman, D. (2019). Estimating unobserved audio features for target-based orchestration. In Proceedings of the International Society for Music Information Retrieval Conference (ISMIR), pages 192199.
  26. He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770778. DOI: 10.1109/CVPR.2016.90
  27. Hirn, M., Mallat, S., & Poilvert, N. (2017). Wavelet scattering regression of quantum chemical energies. Journal of Multiscale Modeling Simulation, 15(2), 827863. DOI: 10.1137/16M1075454
  28. Humphrey, E. J., Bello, J. P., & LeCun, Y. (2015). Feature learning and deep architectures: New directions for music informatics. Journal of Intelligent Information Systems, 41(3), 461481. DOI: 10.1007/s10844-013-0248-5
  29. Humphrey, E. J., Turnbull, D., & Collins, T. (2013). A brief review of creative MIR. In International Society for Music Information Retrieval Conference (ISMIR), Late-Breaking News and Demos.
  30. Klien, V., Grill, T., & Flexer, A. (2012). On automated annotation of acousmatic music. Journal of New Music Research, 41(2), 153173. DOI: 10.1080/09298215.2011.618226
  31. Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. In Proceedings of the International Conference on Neural Information Processing Systems (NIPS), pages 10971105.
  32. Lacoste, A., & Eck, D. (2005). Onset detection with artificial neural networks. In Music Information Retrieval Evaluation eXchange (MIREX), pages 10971105.
  33. LeCun, Y., Bengio, Y., & Hinton, G. (2015). Deep learning. Nature, 521, 436444. DOI: 10.1038/nature14539
  34. Lipp, C. (1996). Real-time interactive digital signal processing: A view of computer music. Computer Music Journal, 20(4), 2124. DOI: 10.2307/3680412
  35. Lostanlen, V., & Cella, C.-E. (2016). Deep convolutional networks on the pitch spiral for music instrument recognition. In Proceedings of the International Society for Music Information Retrieval Conference (ISMIR), pages 612618.
  36. Mallat, S. (2012). Group invariant scattering. Communications on Pure and Applied Mathematics, 65(10), 13311398. DOI: 10.1002/cpa.21413
  37. Mallat, S. (2016). Understanding deep convolutional networks. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 374(2065).
  38. Maresz, Y. (2013). On computer-assisted orchestration. Contemporary Music Review, 32(1), 99109. DOI: 10.1098/rsta.2015.0203
  39. McAdams, S. (1999). Perspectives on the contribution of timbre to musical structure. Computer Music Journal, 23(3), 85102. DOI: 10.1080/07494467.2013.774515
  40. McAdams, S., & Giordano, B. L. (2016). The perception of musical timbre. In S. Hallam, I. Cross & M. Thaut, editors, The Oxford Handbook of Music Psychology (2nd ed.). Oxford University Press. DOI: 10.1162/014892699559797
  41. Mehri, S., Kumar, K., Gulrajani, I., Kumar, R., Jain, S., Sotelo, J., Courville, A., & Bengio, Y. (2015). SampleRNN: An unconditional end-to-end neural audio generation model. In Proceedings of the International Conference on Learning Representations (ICLR). DOI: 10.1093/oxfordhb/9780198722946.013.12
  42. Mermelstein, P. (1976). Distance measures for speech recognition, psychological and instrumental. Pattern Recognition and Artificial Intelligence, 116, 374388.
  43. Müller, M. (2015). Fundamentals of Music Processing: Audio, Analysis, Algorithms, Applications. Springer.
  44. Nilsson, N. J. (2009). The Quest for Artificial Intelligence. Cambridge University Press. DOI: 10.1017/CBO9780511819346
  45. Papadopoulos, H., & Peeters, G. (2007). Large-scale study of chord estimation algorithms based on chroma representation and HMM. In Proceedings of the IEEE International Workshop on Content-Based Multimedia Indexing (CBMI), pages 5360. DOI: 10.1109/CBMI.2007.385392
  46. Papadopoulos, H., & Peeters, G. (2011). Joint estimation of chords and downbeats. IEEE Transactions on Audio, Speech, and Language Processing, 19(1), 138152. DOI: 10.1109/TASL.2010.2045236
  47. Peeters, G. (2004). A large set of audio features for sound description (similarity and classification) in the CUIDADO project. Technical report, IRCAM.
  48. Sak, H., Senior, A., & Beaufays, F. (2014). Long short-term memory recurrent neural network architectures for large scale acoustic modeling. In Proceedings of the Annual Conference of the International Speech Communication Association (INTERSPEECH), pages 338342.
  49. Schlüter, J., & Böck, S. (2014). Improved musical onset detection with convolutional neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 69796983. DOI: 10.1109/ICASSP.2014.6854953
  50. Sifre, L., & Mallat, S. (2013). Rotation, scaling and deformation invariant scattering for texture discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 12331240. DOI: 10.1109/CVPR.2013.163
  51. Smalley, D. (1997). Spectromorphology: Explaining sound-shapes. Organised Sound, 2(2), 107126. DOI: 10.1017/S1355771897009059
  52. Srinivas, S., Sarvadevabhatla, R. K., Mopuri, K. R., Prabhu, N., Kruthiventi, S. S. S., & Babu, R. V. (2016). A taxonomy of deep convolutional neural nets for computer vision. Frontiers in Robotics and AI, 2, 36. DOI: 10.3389/frobt.2015.00036
  53. van den Oord, A., Dieleman, S., Zen, H., Simonyan, K., Vinyals, O., Graves, A., Kalchbrenner, N., Senior, A., & Kavukcuoglu, K. (2016). WaveNet: A generative model for raw audio. arXiv preprint arXiv:1609.03499.
  54. Vinet, H. (2003). The representation level of music information. In Proceedings of the International Symposium on Computer Music Modeling and Retrieval (CMMR), pages 193209. DOI: 10.1007/978-3-540-39900-1_17
  55. Vinet, H. (2008). Science and technology of music and sound: The IRCAM roadmap. Journal of New Music Research, 36(3), 207226. DOI: 10.1080/09298210701859313
  56. Wiggins, G. A. (2009). Semantic gap?? Schemantic schmap!! Methodological considerations in the scientific study of music. In Proceedings of the IEEE International Symposium on Multimedia, pages 477482. DOI: 10.1109/ISM.2009.36
  57. Wiggins, G. A., Pearce, M. T., & Müllensiefen, D. (2009). Computational modelling of music cognition and musical creativity. In R. T. Dean, editor. The Oxford Handbook of Computer Music, pages 383420. Oxford University Press.
  58. Ycart, A., & Benetos, E. (2017). A study on LSTM networks for polyphonic music sequence modelling. In Proceedings of the International Society for Music Information Retrieval Conference (ISMIR), pages 421427.
DOI: https://doi.org/10.5334/tismir.55 | Journal eISSN: 2514-3298
Language: English
Submitted on: Feb 29, 2020
Accepted on: Jul 2, 2020
Published on: Sep 1, 2020
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2020 Carmine-Emanuele Cella, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.