You are currently viewing a new version of our website. To view the old version click .
Arts
  • Article
  • Open Access

6 September 2019

Artificial Intelligence and Music: Open Questions of Copyright Law and Engineering Praxis

,
,
,
and
1
Division of Speech, Music and Hearing, School of Electrical Engineering and Computer Science, Royal Institute of Technology (KTH), 100 44 Stockholm, Sweden
2
Joint Research Centre, European Commission, 1050 Brussels, Belgium
3
Department of Performing Arts, Kingston University, Kingston KT2 7LB, UK
4
Music Technology Group, Universitat Pompeu Fabra, 08002 Barcelona, Spain
This article belongs to the Special Issue Music and the Machine: Contemporary Music Production

Abstract

The application of artificial intelligence (AI) to music stretches back many decades, and presents numerous unique opportunities for a variety of uses, such as the recommendation of recorded music from massive commercial archives, or the (semi-)automated creation of music. Due to unparalleled access to music data and effective learning algorithms running on high-powered computational hardware, AI is now producing surprising outcomes in a domain fully entrenched in human creativity—not to mention a revenue source around the globe. These developments call for a close inspection of what is occurring, and consideration of how it is changing and can change our relationship with music for better and for worse. This article looks at AI applied to music from two perspectives: copyright law and engineering praxis. It grounds its discussion in the development and use of a specific application of AI in music creation, which raises further and unanticipated questions. Most of the questions collected in this article are open as their answers are not yet clear at this time, but they are nonetheless important to consider as AI technologies develop and are applied more widely to music, not to mention other domains centred on human creativity.

1. Introduction

Artificial intelligence (AI) is a well-established discipline of computer science focused on making computers perform tasks that would normally require human intelligence (Russell and Norvig 1995). Due to the convergence of massive data availability, computational resources and novel deep-learning-based architectures, the machine learning (ML) sub-field of AI has experienced major breakthroughs over the past decade (Goodfellow et al. 2018). Such great progress has been made that AI technology is now a major driver of global business and investment.1 AI is being deployed in varied practical scenarios, from machine translation to medical diagnosis, detecting fraudulent credit card use, transcribing speech, summarizing sporting events and financial reports, targeted advertising on social media, and autonomous vehicles. The combination of AI and robotics is also widely transforming manufacturing (Economics 2019). Real-world applications of AI have generated controversy as well, such as tracking people using biometric indicators (e.g., face, speech, gait), prediction of criminal recidivism, judicial sentencing recommendations, drone warfare, predicting protected characteristics (e.g., sexuality, pregnancy), and creating and distributing propaganda (e.g., deep fakes).
AI has also impacted music, from its creation to its distribution. Significant among these are music streaming services using AI for music recommendation and information retrieval, e.g., Spotify2 and Shazam;3 and in product suggestions for online retailers, such as Amazon.4 The application of AI to music creation has appeared in academic halls during the past 60 years, e.g., Hiller and Isaacson (1959); Biles (1999); Cope (1991); Dannenberg et al. (1997); Ebcioğlu (1988); Fernández and Vico (2013); Lewis (2000); Pachet (2003); Sánchez-Quintana et al. (2013); Sturm et al. (2016), but AI is now being used visibly in popular forms of music. Recent examples include: Taryn Southern’s 2017 album, “I AM AI”, which features music generated by a commercially developed music AI system; the 2018 album “Hello World”, billed as “the first music album composed by AI + artists”;5 Holly Herndon’s 2019 album, “Proto”;6 and dozens of albums created by the “first-ever algorithm to sign [a] major label deal”.7 Several companies have also been founded recently to capitalize on advancements of AI applied to music content creation, particularly for production music, i.e., music to accompany film, radio and other media. Examples of these companies include Aiva8 and Amper9 for creating soundtracks for advertisements; and Melodrive10 for automatically creating music in video games. Some companies are also devoting resources to creating software for artists exploiting AI technology. Examples include LANDR for mastering music,11 IBM12 and Sony13 for music composition, and Google’s Project Magenta14 for sound and music synthesis.
The combination of large quantities of recorded music and accompanying data, effective algorithms, and high-power computational hardware is now producing progress audibly moving so fast that alarms are sounding.15 What does AI mean for music? As in so many areas of labour where AI and technology leads to more efficient production lines and increased profit but human redundancy and deskilling, can the same happen in music? Some say that AI will free people of mundane tasks, but this can be “tantamount to liberation from the ability to make a living” (Drott 2019). How will AI help and harm the various participants contributing to and benefiting from music, e.g., composers, musicians, educators, listeners, and organisations? Should consumers be informed about the involvement of AI in the music they listen to, much the same way ingredients of food products are communicated? And how should this information be presented, and to what level of detail?
A tangible example of many challenges and open questions of music AI is given by the folkrnn project (Sturm et al. 2016).16 This project has built and trained several music AI models on data produced from tens of thousands of transcriptions of folk music available online.17 The resulting models are able to generate novel transcriptions that exhibit many characteristics similar to the original data. An online implementation18 offers users the functionality of generating transcriptions with the click of a button. Figure 1 shows the parameters available to the user, and a notated transcription generated by the AI. Users can submit the transcription to a growing collection of such tunes at another website.19 Material generated by folkrnn models has been used in variety of ways,20 including an album situated within the style of Irish traditional music (Sturm and Ben-Tal 2018).21
Figure 1. The menu of http://folkrnn.org (left) gives the user parameters with which to direct the music generation process of a selected AI model. Here, the user has selected the model “thesession.org (w/ :| |:)”, a random meter and mode, a temperature of 1, and a random number seed of 793470. Clicking “Compose” begins the generation process, which creates the transcription shown notated at right. Figure created by authors.
The experience of developing and using folkrnn models brings focus to the numerous questions identified above, and also motivates other questions. Who owns the rights to the material generated by folkrnn models, and to any music derived from it? What implications are there that folkrnn models are trained on a collective source of music transcriptions, some of which may be derived from copyright-protected work? How might folkrnn—and tools like it— impact music creators and music communities, traditional or otherwise? What kinds of misuses of folkrnn are possible, and how can its developers guard against these? How is the value of music affected by AI system that can generate and synthesize millions of “folk music” recordings without the direct input of human musicians? Should a person who uses material generated by folkrnn models in the composition process specify where it came from, i.e., that it was not entirely their own creation? Should music created in such ways, with the “input of AI”, be considered inferior to music created without?
These questions motivate taking a closer look at what is actually occurring in the application of AI to music from two different and complementary perspectives: copyright law and engineering praxis. In Section 2, we look specifically at copyright law, and how it might address some of the questions surrounding material generated by music AI. The legal perspective is important because the use of AI in music creation is posing novel challenges to intellectual property law, moral rights, and the protection of creative human production. In Section 3, we look at engineering praxis in AI and ML, and some of the questions arising from their research and development methodologies. This perspective shifts the focus in engineering from developing music AI, and the incentives of doing so, to surveying the impact of such technology and its development, intended or not, and the implicit assumptions made in performing such work. Section 4 returns to folkrnn to discuss aspects of these two perspectives in a more grounded way. The conclusion looks to the future of AI in music, and speculates what it may mean for copyright law and engineering praxis. Overall, this article contributes new perspectives to the interdisciplinary domain of music AI development, which is currently heavily focused on how to create such technology. There are no easy answers to many of the questions raised in this article, but they must continue to be identified and formalized to assist their future discussion.

3. Engineering Praxis Perspective

Since engineers bring new technologies into the world, it can be argued that they share the responsibility for the resulting outcomes, positive and negative, intended and unintended. The integration of AI into societal and political spheres can appear as fostering objectivity, but the data-centric nature of ML can unintentionally perpetuate discrimination by encoding existing social biases learned from real data used by an engineer for training (Barocas and Selbst 2016). Even if an engineer has good intentions, their uncritical treatment of data and their resulting systems might become instrumental in producing great social and economic harm. No matter how complex a system might be, engineers must study, evaluate and document its working principles and limitations and make users and others impacted by such a system aware of them. Technology is not ethically neutral (Dusek 2006), which motivates design procedures that are guided by clear ethical principles (Hand 2018; The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems 2017), as well as employing continuous evaluation to mitigate harmful impact. Bryson and Winfield (2017) provide a brief description of and motivations for standardizing the ethical design of AI such that its development benefits humanity. It introduces the standardization efforts of The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems (2017), and discusses in particular standardizing transparency in autonomous systems. The development and implementation of these practices are still ongoing.
When evaluating the impact of AI, engineers typically work from the narrow perspective of AI performance in limited and controlled contexts. A critical but often overlooked question is what defines a particular system as “good” or “bad” with respect to a use or user, e.g., a music listener, performer, composer or business. Defining meaningful evaluation strategies is difficult for many applications in the music domain, due to the subjectivity of music judgment and preference. This makes evaluation one of the main research challenges in the discipline of music information retrieval (Schedl et al. 2014; Sturm 2016). Engineers by and large focus on performance metrics like accuracy, precision, and reliability when developing data-driven AI techniques, and then infer from performance how useful or harmless a system is. Evaluation can instead be framed in the context of specific applications and their users, but this is rarely done (Schedl et al. 2013).
Broader perspectives are being advocated for by the developing research fields “fair, accountable and transparent machine learning” (FAT/ML),36 and “explainable AI” or “interpretable AI”. These fields argue for richer notions of evaluating AI performance and use, ones that actually reflect usefulness and real impact. Definitions of these concepts are numerous, and represents a track of research in itself. In one sense, an AI system is considered to be fair if it does not discriminate against certain individuals based on a set of protected attributes like gender, race, and age.37 Group fairness has been proposed as a category of algorithmic fairness (Kleinberg et al. 2017), which consists of more than 21 metrics to measure discrimination, some of which are contradictory. Transparency, or a clear explanation of where and how AI has been used to make some decision or product, has been also argued to be a relevant evaluation criteria for AI development and application. A related concept is interpretability (Doshi-Velez and Kim 2017) or explainability (Mittelstadt et al. 2019): the ability to explain, in an understandable way, how an AI behaves a particular way, or has arrived at a given result. Metcalf (2018) underlines the importance of distinguishing between bias and fairness. The former is a technical feature of statistical models, and the latter is related to human values that comprise social biases and prejudices about a group of people. Bias and unfairness may occur within an AI system without necessarily impacting its performance in some aspects, but can impact metrics reflecting the values of other participants.
These concepts not only promote the measurement of AI systems in ways that are more relevant to their application, but also motivate new methodologies for engineering, and for addressing the general lack of auditing tools to evaluate the social impact of AI.38 For instance, engineers can use interpretability to find and correct causes of discrimination in an AI system, to better explain its errors and improve operation, and to perform causal inference (Zeng et al. 2017). One can also see potential uses of such approaches to addressing some questions of copyright law in music AI raised in Section 2. For instance, explainability can help address questions about how specific music material was generated by a system. Explainability can also improve the deployment of AI systems in the real world, where sample bias and concept drift are common.39 The engineering of AI systems can benefit by working in transparent ways as well, e.g., clarifying the reasons why specific metrics are used and the societal values which underlie them (Glymour and Herington 2019; Kilbertus et al. 2017); describing the processes of data collection and use, which may result in discrimination (Hand 2018); even improving the composition of a research team, where a lack of team diversity can compound the impact of the problems mentioned above and reinforce blind spots (Ruiz et al. 2002). Working in these ways can help address issues related to the socio-economic impact of AI systems, define new evaluation principles centred around these issues, and prevent and mitigate negative effects, e.g., see Benthall and Haynes (2019).
Discussions about bias, fairness and transparency in the development and application of AI to music are just beginning, e.g., Holzapfel et al. (2018).40 Music creation is a part of a landscape in which various AI-driven tools are deployed and used for different purposes, e.g., composing, mixing, and streaming. For music streaming services, AI can make considerable positive and negative impact on music recommendation. AI can be a core technology of music recommendation systems designed to provide users with relevant recommendations, and thus, can influence music consumption and listening behaviour at large scales. However, such approaches can also incentivise unscrupulous behaviours that take advantage of artists and listeners (Eriksson et al. 2018).41 For instance, using AI music generation can greatly increase the revenue of a music streaming business, whereby in-house synthetic “artists” are promoted on playlists thus reducing the revenue distributed to human artists whose music is also streamed. In such cases, artists and listeners might be informed about the way AI systems work for music recommendation (Aguiar and Waldfogel 2018). It could also be mandated that users be informed about the ways in which AI was used to create the music they are listening to, or even opt-out of hearing AI-generated music altogether. Such transparency can serve to empower artists and listeners to challenge AI systems; however, these require defining or clarifying the degree to which AI is involved in the process. It is tempting to try and enumerate the involvement of AI in various stages of music creation, e.g., composition, performance, production. For each stage, one might indicate the degree of AI involvement with a numerical score. A more fine grained analysis could describe exactly how AI was involved, e.g., the design of a microphone, the creation of an impulse response, the synthesis of a particular voice, or the use of score following during performance. Consumers could then use such information to make more informed choices, in the same way that labeling food ingredients is used to identify nutritional value or potentially deadly reactions due to allergies.
This approach to transparency, however, faces problems immediately. First, decomposing music creation into stages is not so clear in many instances, and also introduces its own biases, since these stages are not always separable and do not even exist in all music. Second, designating which aspects are more and which are less central to the value consumers attach to music can be debatable. Consider, for instance, an AI application that automatically advances a music score for performing players: certainly useful AI technology, but is it material to the musical value of the recording? Do consumers themselves have enough knowledge about music and its production to evaluate the relevance of such information? Third, an accurate tally of the involvement of AI will also require extensive and sometimes intrusive documentation of all the software and hardware used in the process and its relevance to the finished product—which has the potential to become unwieldy and a hindrance to the creative process, and could even harm artists who do not have institutional backing. Requiring a young, aspiring artist to provide full and accurate documentation of the AI they have used as a condition for making their music available to audiences could be an insurmountable barrier. Investigating the provenance of each product (hardware and software) for traces of AI is perhaps not something many artists would be able or willing to do.
This is not to say that it is too difficult to inform consumers about the way AI has been involved in the creation of a particular music experience, but certainly, research must be conducted to establish exactly the harms that can result if such information is not provided to consumers and in what contexts. In the case of a person with a life-threatening allergy to a particular food ingredient, one benefit of labeling food with ingredients is clear: the avoidance of death. Such a benefit is not yet clear in the case of identifying the involvement of AI in music creation. Furthermore, inundating consumers with information but expecting them to ascertain its relevance to them may not in the end be conducive to informed choice.

5. Conclusions

Technology is a double-edged sword, with benefits and detriments that deserve to be critically analysed as it is developed, applied, improved and retired. General-purpose technologies such as AI that have the potential to affect economies and societies require study from broad and diverse perspectives, with considerations of questions that touch on law, economy, ecology, and philosophy (Bostrom and Yudkowsky 2014; Craglia 2018; Dusek 2006; Holzapfel et al. 2018). Relevant initiatives in ML research are centred on developing AI systems that are not just accurate, but also fair, accountable, transparent and interpretable. Such work moves beyond describing just the engineering of technologies to examining their use and misuse, considering their impact, and how such dimensions should be reflected in engineering praxis. This article aims to contribute to this discussion specifically for AI developed and applied to music.
We have looked at two broad perspectives, both timely and motivated by a particular music AI application. The first is from the perspective of copyright law, which arises from questions about the ownership of material generated by AI systems that may have themselves been trained on copyright protected material. Many questions have been identified, most of which cannot be answered at this time due to a lack of precedent; but what is certainly clear is the disruption AI music generation will cause to legal and societal norms as it improves. The second perspective is engineering praxis, which arises from concerns as to the impact of AI technology on music ecosystems and the ways in which such engineers work. Many questions are raised here as well that cannot be answered, but what is clear is that music AI should be developed and used in consultation with music practitioners (Holzapfel et al. 2018; Sturm et al. 2018).
The impact of music AI on career trajectories and opportunities present risks not just to job seekers, but are likely to have adverse effects on innovation in the music-ecosystem. The entry-level jobs in recording studios and small media companies that are likely to be replaced by music AI allow aspiring musicians to acquire skills and networking opportunities for embarking on their career. Adapting education curricula to include more sophisticated skills and ways of working alongside and with the support of music AI tools is only part of the answer, and temporary at that, with further technological development. Developers of the technology and the managers who decide on deploying it should consider the longer term effects of their decisions, and ways of mitigating adverse results. At the same time, the current legal framework may need adjustment to best facilitate creative activity and artistic innovation in such an evolving ecosystem.
It might be illustrative to consider a possible endpoint of music AI development: a “complete” music AI, a kind of “musical holodeck” able to generate any possible music, archiving all recorded music (scores, audio and video recordings, images, and other data about the music and its performers, etc.). It knows and understands what music is and how it works, and can not only retrieve all music fulfilling the requests of any listener, but also create any possible music in an instant. A subscriber of this system can ask to hear Led Zeppelin performing a particular ABBA song. Composer Ben-Tal can hear his piano music performed by a long-dead pianist or the Berlin Philharmonic. The system provides any subscriber with limitless access to individualised musical experiences. Will there still be a need for recording engineers, producers, and performers? Any subscriber can now be a composer of their own soundtrack. There is no need to distribute music because all music is (re)created locally. Would such a system divorce music from musical “works”? A system that responds dynamically to the wishes of a subscriber will not need to be centred on structured stretches of organised sound. Music could become an amorphous soundtrack to one’s life – delivering on Erik Satie’s vision of “Furniture Music”: music that is heard but not listened to; music that forms part of everyday life like the objects people have in their home (Templier 1969).
Such a future clearly requires a careful assessment of copyright rules. If a subscriber hears something they like, they can ask to hear something “like it” and the AI should be able to produce a new instance that meets their demands without outright copying. What should then be the approach to the ownership of AI-generated music? We would have to consider the purpose of awarding copyright. Generally speaking, copyright (and related rights) aims to provide creative people (and investors) with a reward in the form of an exclusive, quasi monopolistic, right. In droit d’auteur countries, copyright also reveals an intimate link with the personality of the author. Returning to our music AI, its musical potential is immense and it is likely that most subscribers will only ever discover a tiny sliver of that potential. It is also likely that some subscribers will have the inclination and the aptitude to be much more inventive in their interaction with the system and produce truly unique results through it. How can we enable such subscribers to dedicate their time and effort towards this and have means of sharing the fruits of their intellectual labour with others? Is there any right the designer of such an AI should retain? What would be the impact of maintaining the status quo, or of passing new rights in the market of human created and/or AI generated works? As our survey of copyright law indicates, technological developments challenge established norms. The distant future sketched above suggests that there is a need for fundamental re-thinking in this area.
Of course, the inevitability of such a music AI should be taken with a grain of salt. Human creativity can surprise in its ability to incorporate new technologies. For instance, the turntable was designed to facilitate playback at home but has become a performance tool in ways that were not intended by its developers or manufacturer. It is likely that any music AI tools will open opportunities for creative but unintended uses.

Author Contributions

Conceptualization, B.L.T.S. and E.G.; writing—original draft preparation, all; writing—review and editing, B.L.T.S. and O.B.-T. Further details: B.L.T.S. organised and supervised the work, and contributed the majority of all Sections except Section 2, which was contributed by M.I. Other contributors to Section 3 are M.M., O.B.-T. and E.G. All contributed to Section 4 and Section 5.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest. The views expressed in the article do not imply a policy position of the European Commission.

References

  1. Agres, Kat, Jamie Forth, and Geraint A. Wiggins. 2016. Evaluation of musical creativity and musical metacreation systems. Computers in Entertainment 14: 1–33. [Google Scholar] [CrossRef]
  2. Aguiar, Luis, and Joel Waldfogel. 2018. Platforms, Promotion, and Product Discovery: Evidence from Spotify Playlists. NBER Working Paper No. 24713. Cambridge: NBER. [Google Scholar]
  3. Barocas, Solon, and Andrew D. Selbst. 2016. Big data’s disparate impact. California Law Review 104: 671. [Google Scholar] [CrossRef]
  4. Benthall, Sebastian, and Bruce D. Haynes. 2019. Racial categories in machine learning. Paper presented at Conference on Fairness, Accountability, and Transparency, Atlanta, GA, USA, January 29–31; pp. 289–98. [Google Scholar]
  5. Biles, John A. 1999. Life with GenJam: Interacting with a musical IGA. Paper presented at 1999 IEEE International Conference on Systems, Man, and Cybernetics, Tokyo, Japan, October 12–15, vol. 3, pp. 652–56. [Google Scholar]
  6. Bostrom, Nick, and Eliezer Yudkowsky. 2014. Chapter The Ethics of Artificial Intelligence. In Cambridge Handbook of Artificial Intelligence. Cambridge: Cambridge University Press. [Google Scholar]
  7. Bryson, Joanna, and Alan Winfield. 2017. Standardizing ethical design for artificial intelligence and autonomous systems. Computer 50: 116–19. [Google Scholar] [CrossRef]
  8. Buning, Madeleine de Cock. 2018. Chapter Artificial Intelligence and the Creative Industry: New Challenges for the EU Paradigm for Art and Technology. In Research Handbook on the Law of Artificial Intelligence. Cheltenham: Edward Elgar Publishing. [Google Scholar]
  9. Cope, David. 1991. Computers and Musical Style. Oxford: Oxford University Press. [Google Scholar]
  10. Craglia, M., ed. 2018. Artificial Intelligence: A European Perspective. Brussels: Publications Office of the European Union. [Google Scholar]
  11. Dannenberg, Roger B., Belinda Thom, and David Watson. 1997. A machine learning approach to musical style recognition. Paper presented at International Computer Music Conference, Thessaloniki, Greece, September 25–30; pp. 344–47. [Google Scholar]
  12. Deliège, Irène, and Geraint A. Wiggins, eds. 2006. Musical Creativity: Multidisciplinary Research in Theory and Practice. Abingdon: Psychology Press. [Google Scholar]
  13. Deltorn, Jean-Marc, and Franck Macrez. 2019. Chapter Authorship in the Age of Machine learning and Artificial Intelligence. In The Oxford Handbook of Music Law and Policy. Oxford: Oxford University Press. [Google Scholar]
  14. Doshi-Velez, Finale, and Been Kim. 2017. Chapter Towards a Rigorous Science of Interpretable Machine Learning. In Springer Series on Challenges in Machine Learning. Switzerland: Springer. [Google Scholar]
  15. Drott, Eric. 2019. Music AI: Copyright, compensation, commons. Paper presented at Music & AI AIAS Conference, Aarhus, Denmark, May 28; Washington, DC: AIAS. [Google Scholar]
  16. Dusek, Val. 2006. Philosophy of Technology: An Introduction. Hoboken: Blackwell Publishing Ltd. [Google Scholar]
  17. Ebcioğlu, Kemal. 1988. An expert system for harmonizing four-part chorales. Computer Music Journal 12: 43–51. [Google Scholar] [CrossRef][Green Version]
  18. Economics, Oxford. 2019. How Robots Change the World: What Automation Really Means for Jobs and Productivity. Oxford: Oxford Economics. [Google Scholar]
  19. Eriksson, Maria, Rasmus Fleischer, Anna Johansson, Pelle Snickars, and Patrick Vonderau. 2018. Spotify Teardown: Inside the Black Box of Streaming Music. Cambridge: MIT Press. [Google Scholar]
  20. Fairbairn, Hazel. 1993. Group Playing in Traditional Irish Music: Interaction and Heterophony in the Session. Ph.D. Thesis, Cambirdge University, Cambirdge, UK. [Google Scholar]
  21. Fernández, Jose D., and Francisco Vico. 2013. AI methods in algorithmic composition: A comprehensive survey. Journal of Artificial Intelligence Research 48: 513–82. [Google Scholar] [CrossRef]
  22. Glymour, Bruce, and Jonathan Herington. 2019. Measuring the biases that matter: The ethical and casual foundations for measures of fairness in algorithms. Paper presented at Conference on Fairness, Accountability, and Transparency, Atlanta, GA, USA, January 29–31; pp. 269–78. [Google Scholar]
  23. Goodfellow, Ian, Yoshua Bengio, and Aaron Courville. 2018. Deep Learning. Cambirdge: MIT Press. [Google Scholar]
  24. Guadamuz, Andrés. 2017. Do androids dream of electric copyright? Comparative analysis of originality in artificial intelligence generated works. Intellectual Property Quarterly 2: 169–86. [Google Scholar]
  25. Hand, David J. 2018. Aspects of data ethics in a changing world: Where are we now? Big Data 6: 176–90. [Google Scholar] [CrossRef] [PubMed]
  26. Hiller, Lejaren A., and Leonard M. Isaacson. 1959. Experimental Music: Composition with an Electronic Computer. New York: McGraw-Hill Book Company. [Google Scholar]
  27. Holzapfel, André, Bob L. Sturm, and Mark Coeckelbergh. 2018. Ethical dimensions of music information retrieval technology. Transactions of the International Society for Music Information Retrieval 1: 44–55. [Google Scholar] [CrossRef]
  28. Kilbertus, Niki, Mateo Rojas Carulla, Giambattista Parascandolo, Moritz Hardt, Dominik Janzing, and Bernhard Schölkopf. 2017. Avoiding discrimination through causal reasoning. Paper presented at Advances in Neural Information Processing Systems, Long Beach, CA, USA, December 5–9; pp. 656–66. [Google Scholar]
  29. Kleinberg, Jon, Sendhil Mullainathan, and Manish Raghavan. 2017. Inherent trade-offs in the fair determination of risk scores. Paper presented at Innovations in Theoretical Computer Science, Bologna, Italy, July 3–5. [Google Scholar]
  30. Lauber-Rönsberg, Anne, and Sven Hetmank. 2019. The concept of authorship and inventorship under pressure: Does artificial intelligence shift paradigms? Journal Intellectual Property Law & Practice 14: 570–79. [Google Scholar]
  31. Lewis, George E. 2000. Too many notes: Computers, complexity and culture in Voyager. Leonardo Music Journal 10: 33–39. [Google Scholar] [CrossRef]
  32. MacMahon, Tony. 1996. The language of passion. Paper presented at Crossroads Conference, Temple Bar, Dublin, Ireland, April 19. [Google Scholar]
  33. Metcalf, Jacob. 2018. Engineering for Fairness: How a Firm Conceptual Distinction between Unfairness and Bias Makes It Easier to Address Un/fairness. Tutorial presented at ACM Conference on Fairness, Accountability, and Transparency, New York, NY, USA, February 23–24. [Google Scholar]
  34. Michaux, Benoit. 2018. Chapter Singularité technologique, singularité humaine et droit d’auteur. In Droits, Norms et Libertés Dans le Cybermonde. New York: Kluwer, pp. 401–16. [Google Scholar]
  35. Mittelstadt, Brent, Chris Russell, and Sandra Wachter. 2019. Explaining explanations in AI. Paper presented at Conference on Fairness, Accountability, and Transparency, Atlanta, GA, USA, January 29–31. [Google Scholar]
  36. Pachet, François. 2003. The continuator: Musical interaction with style. Journal of New Music Research 32: 333–41. [Google Scholar] [CrossRef]
  37. Ramalho, Anna. 2017. Will Robots Rule the (Artistic) World? A Proposed Model for the Legal Status of Creations by Artificial Intelligence Systems. Journal of Internet Law 21: 12–25. [Google Scholar] [CrossRef]
  38. Ruiz, David, Rafael Corchuelo, José A Pérez, and Miguel Toro. 2002. An algorithm for ensuring fairness and liveness in non-deterministic systems based on multiparty interactions. Paper presented at European Conference on Parallel Processing, Paderborn, Germany, August 27–30; pp. 563–72. [Google Scholar]
  39. Russell, Stuart, and Peter Norvig. 1995. Artificial Intelligence: A Modern Approach. Egnlewood Cliffs: Prentice-Hall. [Google Scholar]
  40. Sánchez-Quintana, Carlos, Francisco Moreno-Arcas, David Albarracín-Molina, José D. Fernández, and Francisco Vico. 2013. Melomics: A case-study of AI in spain. AI Magazine 34: 99–103. [Google Scholar] [CrossRef][Green Version]
  41. Schedl, Markus, Arthur Flexer, and Julián Urbano. 2013. The neglected user in music information retrieval research. Journal of Intelligent Information Systems 41: 523–39. [Google Scholar] [CrossRef]
  42. Schedl, Markus, Emilia Gomez, and Julian Urbano. 2014. Music information retrieval: Recent developments and applications. Foundations and Trends in Information Retrieval 8: 127–261. [Google Scholar] [CrossRef]
  43. Sturm, Bob L. 2016. Revisiting priorities: Improving MIR evaluation practices. Paper presented at International Symposium on Music Information Retrieval, New York, NY, USA, August 7–11. [Google Scholar]
  44. Sturm, Bob L. 2018. What do these 5,599,881 parameters mean? An analysis of a specific LSTM music transcription model, starting with the 70,281 parameters of its softmax layer. Paper presented at Music Metacreation Workshop of International Conference on Computational Creativity, Salamanca, Spain, June 25–26. [Google Scholar]
  45. Sturm, Bob L., and Oded Ben-Tal. 2017. Taking the models back to music practice: Evaluating generative transcription models built using deep learning. Journal Creative Music Systems 2. [Google Scholar] [CrossRef]
  46. Sturm, Bob L., and Oded Ben-Tal. 2018. Let’s Have Another Gan Ainm: An Experimental Album of Irish Traditional Music and Computer-Generated Tunes. Technical Report. Stockholm: KTH Royal Institute of Technology. [Google Scholar]
  47. Sturm, Bob L., Oded Ben-Tal,Úna Monaghan, Nick Collins, Dorien Herremans, Elaine Chew, Gaëtan Hadjeres, Emmanuel Deruty, and François Pachet. 2018. Machine learning research that matters for music creation: A case study. Journal New Music Research 48: 36–55. [Google Scholar] [CrossRef]
  48. Sturm, Bob L., João F. Santos, Oded Ben-Tal, and Iryna Korshunova. 2016. Music transcription modelling and composition using deep learning. Paper presented at Computer Simulation of Musical Creativity, Huddersfield, UK, June 17–19. [Google Scholar]
  49. Templier, Pierre-Daniel. 1969. Erik Satie. Translated by Elena L. French, and David S. French. Cambridge: MIT Press. [Google Scholar]
  50. The IEEE Global Initiative on Ethics of Autonomous and Intelligent Systems. 2017. Ethically Aligned Design: A Vision for Prioritizing Human Well-Being with Autonomous and Intelligent Systems, 2nd ed. Piscataway: IEEE. [Google Scholar]
  51. Zeng, Jiaming, Berk Ustun, and Cynthia Rudin. 2017. Interpretable classification models for recidivism prediction. Journal of the Royal Statistical Society: Series A (Statistics in Society) 180: 689–722. [Google Scholar] [CrossRef]
1
For instance, see the number of articles appearing like A. Moore’s “When AI Becomes an Everyday Technology”, Harvard Business Review, 7 June 2019, https://hbr.org/2019/06/when-ai-becomes-an-everyday-technology.
2
https://www.spotify.com is a music streaming service.
3
https://www.shazam.com is a music identification service.
4
https://www.amazon.com is an online retail service.
5
6
7
8
9
10
11
12
13
14
15
16
17
One source of data is https://thesession.org/.
18
19
20
See the Bottomless Tunebox YouTube channel: https://www.youtube.com/channel/UC7wzmG64y2IbTUeWji_qKhA/videos.
21
22
Art. 2 of the WIPO Performances and Phonograms Treaty (WPPT) defines phonogram as “the fixation of the sounds of a performance or of other sounds, or of a representation of sounds, other than in the form of a fixation incorporated in a cinematographic or other audiovisual work”.
23
S. 178 Copyright, Designs and Patents Act 1988.
24
Infopaq: C-5/08 Judgment of the Court (Fourth Chamber) of 16 July 2009; BSA: C-393/09, Judgment of the Court (Third Chamber) of 22 December 2010; Painer: C-145/10, Judgment of the Court (Third Chamber) of 1 December 2011; Dataco: Case 604/10, Judgment of the Court, (Third Chamber) of 1 March 2012.
25
See also presentations delivered at the European Copyright Society 2018 conference, “EU copyright, quo vadis? From the EU copyright package to the challenges of Artificial intelligence”: https://europeancopyrightsociety.org/ecs-conferences-2018-brussels. To our knowledge, there is no European case law yet dealing with AI-generated works. In China, a recent decision by Beijing Internet Court refused copyright protection for a report generated by AI. Under Chinese law, only works created by humans are eligible for protection. See Ming Chen, Beijing Internet Court denies copyright to works created solely by artificial intelligence, Journal of Intellectual Property Law & Practice, 2019, Vol. 14, No. 8.
26
In any case, humans would be needed, of course, for creating the technology. This does not mean that the generated music necessarily reflects the personality of those developers.
27
For a detailed discussion of possible authors in relation to music generated by different AI systems see Deltorn and Macrez (2019).
28
The European Commission has recently published a call for a study that will assess whether the current IPR framework is fit-for-purpose for AI-generated works/inventions: https://ec.europa.eu/digital-single-market/en/news/trends-and-developments-artificial-intelligence-challenges-intellectual-property-rights.
29
In Europe copyright lasts for 70 years from the death of the author. The duration of neighbouring rights is shorter (50 years) except for published phonograms. The use of works in the public domain is free.
30
Directive (EU) 2019/790 of the European Parliament and of the Council of 17 April 2019 on copyright and related rights in the Digital Single Market and amending Directives 96/9/EC and 2001/29/EC, OJ L 130, 17.5.2019, pp. 92–125.
31
Data mining is defined in Article 2 of the directive as “any automated analytical technique aimed at analysing text and data in digital form in order to generate information which includes but is not limited to patterns, trends and correlations”.
32
ibid., Article 3.
33
ibid., Article 4.
34
A rightholder can restrict use, e.g., explicitly using methods to block data mining is a way to opt out from the exception in case of material made publicly available online.
35
Judgment of 29 July 2019, Pelham, C-476/17, ECLI:EU:C:2019:624.
36
See the FAT/ML workshop: https://www.fatml.org.
37
“Intended” discrimination is commonly known in ML as disparate treatment, while “unintended” as disparate impact.
38
See Kate Crawford (2017), “Artificial intelligence with very real biases” https://www.wsj.com/articles/artificial-intelligencewith-very-real-biases-1508252717.
39
See, Fabien Gouyon (2018). “Overview and new challenges of music information research” https://www.slideshare.net/FabienGouyon/music-recommendation-2018-116102609.
40
Also see the tutorial, “Fairness, Accountability and Transparency in Music Information Research (FAT-MIR)” at the 2019 International Symposium on Music Information Retrieval, https://ismir2019.ewi.tudelft.nl/?q=node/41.
41
For example, see A. K. Raymond, “The Streaming Problem: How Spammers, Superstars, and Tech Giants Gamed the Music Industry”, Vulture, 5 July 2017, https://www.vulture.com/2017/07/streaming-music-cheat-codes.html.
42
The data is archived about once a week online: http://github.com/adactio/TheSession-data.
43
44
See footnote 21.
45
The first Gan Ainm on track 3, Gan Ainm, Gan Ainm, Gan Ainm.
46
47
For instance, see the discussion here: https://thesession.org/discussions/42712. See also MacMahon (1996) for a stark discussion of tradition in Irish traditional music delivered by one of its most visible advocates.
48
49
See footnote 21.
50

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.